sha
null
last_modified
null
library_name
stringclasses
154 values
text
stringlengths
1
900k
metadata
stringlengths
2
348k
pipeline_tag
stringclasses
45 values
id
stringlengths
5
122
tags
listlengths
1
1.84k
created_at
stringlengths
25
25
arxiv
listlengths
0
201
languages
listlengths
0
1.83k
tags_str
stringlengths
17
9.34k
text_str
stringlengths
0
389k
text_lists
listlengths
0
722
processed_texts
listlengths
1
723
tokens_length
listlengths
1
723
input_texts
listlengths
1
61
embeddings
listlengths
768
768
null
null
transformers
This is a multilingual NER system trained using a Frustratingly Easy Domain Adaptation architecture. It is based on CroSloEngual (https://huggingface.co/EMBEDDIA/crosloengual-bert) and supports different tagsets all using IOBES formats: 1. Wikiann (LOC, PER, ORG) 2. SlavNER 19/21 (EVT, LOC, ORG, PER, PRO) 3. SSJ500k (LOC, MISC, ORG, PER) PER: person, LOC: location, ORG: organization, EVT: event, PRO: product, MISC: Miscellaneous, MEDIA: media, ART: Artifact, TIME: time, DATE: date You can select the tagset to use in the output by configuring the model. This model manages differently uppercase words. More information about the model can be found in the paper (https://aclanthology.org/2021.bsnlp-1.12.pdf) and GitHub repository (https://github.com/EMBEDDIA/NER_FEDA).
{"language": ["hr", "sl", "en", "multilingual"], "license": "mit", "tags": ["CroSloEngual", "ner"]}
null
creat89/NER_FEDA_Sl
[ "transformers", "pytorch", "bert", "CroSloEngual", "ner", "hr", "sl", "en", "multilingual", "license:mit", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "hr", "sl", "en", "multilingual" ]
TAGS #transformers #pytorch #bert #CroSloEngual #ner #hr #sl #en #multilingual #license-mit #endpoints_compatible #region-us
This is a multilingual NER system trained using a Frustratingly Easy Domain Adaptation architecture. It is based on CroSloEngual (URL and supports different tagsets all using IOBES formats: 1. Wikiann (LOC, PER, ORG) 2. SlavNER 19/21 (EVT, LOC, ORG, PER, PRO) 3. SSJ500k (LOC, MISC, ORG, PER) PER: person, LOC: location, ORG: organization, EVT: event, PRO: product, MISC: Miscellaneous, MEDIA: media, ART: Artifact, TIME: time, DATE: date You can select the tagset to use in the output by configuring the model. This model manages differently uppercase words. More information about the model can be found in the paper (URL and GitHub repository (URL
[]
[ "TAGS\n#transformers #pytorch #bert #CroSloEngual #ner #hr #sl #en #multilingual #license-mit #endpoints_compatible #region-us \n" ]
[ 47 ]
[ "passage: TAGS\n#transformers #pytorch #bert #CroSloEngual #ner #hr #sl #en #multilingual #license-mit #endpoints_compatible #region-us \n" ]
[ -0.046101897954940796, 0.003084068186581135, -0.009912987239658833, 0.020118936896324158, 0.09581785649061203, 0.03061067871749401, 0.0955689549446106, 0.039487697184085846, 0.14023089408874512, -0.01253522839397192, 0.1302793025970459, 0.1838635951280594, -0.03774821758270264, -0.10596001148223877, -0.048574354499578476, -0.2769118547439575, 0.03592107445001602, 0.04605446010828018, 0.0001876611349871382, 0.08818645030260086, 0.07571765035390854, -0.05304563418030739, 0.07084546238183975, 0.0036447288002818823, -0.049355972558259964, 0.047309886664152145, 0.005650949198752642, -0.03486891835927963, 0.12302669882774353, 0.0599081926047802, 0.10880177468061447, 0.05217298865318298, -0.014536687172949314, -0.2148798406124115, 0.0236379262059927, -0.07252994924783707, -0.08442375808954239, -0.0039619640447199345, 0.011876353994011879, -0.10780103504657745, 0.12996073067188263, -0.001768068061210215, -0.033983200788497925, 0.05454518646001816, -0.13404406607151031, -0.15851350128650665, -0.05717615410685539, 0.09509368240833282, -0.015363861806690693, 0.035392142832279205, 0.030084362253546715, 0.1436282843351364, -0.16497276723384857, 0.06068117916584015, 0.10042638331651688, -0.32786643505096436, -0.0009008516208268702, 0.1040160208940506, 0.08549179136753082, 0.013809138908982277, -0.06858427077531815, 0.08794210106134415, 0.036245111376047134, 0.0070217144675552845, -0.10380084812641144, -0.0995572954416275, 0.03335106745362282, 0.07790547609329224, -0.0662488266825676, -0.05762188881635666, 0.10565447062253952, -0.02641178108751774, 0.025915130972862244, 0.04022420942783356, -0.022534463554620743, -0.02211112156510353, -0.004403512459248304, 0.0017171564977616072, -0.007578739896416664, 0.0795210674405098, 0.026223603636026382, -0.047605354338884354, -0.10185729712247849, 0.0671001598238945, -0.2470119446516037, 0.20362767577171326, 0.059108536690473557, 0.052988968789577484, -0.12286603450775146, 0.03229225054383278, -0.013097964227199554, -0.06506649404764175, -0.04273943230509758, -0.05362657457590103, 0.059383559972047806, 0.03192165493965149, -0.016021959483623505, 0.1518450230360031, 0.07467029243707657, 0.17995715141296387, -0.034080058336257935, 0.03794283792376518, -0.011612074449658394, 0.12994910776615143, -0.01636532135307789, 0.05997674539685249, 0.05648627504706383, -0.004642786458134651, -0.00348447123542428, -0.178485706448555, 0.04207747057080269, 0.01099359430372715, -0.15944896638393402, -0.06947366148233414, -0.087864950299263, 0.12646129727363586, -0.018535831943154335, 0.052645351737737656, -0.045039668679237366, 0.06066381186246872, -0.0175655335187912, -0.027239801362156868, 0.02016652561724186, 0.013753232546150684, 0.051943469792604446, 0.17256508767604828, -0.0384962223470211, -0.0074758464470505714, -0.022979259490966797, 0.12318085879087448, -0.046956948935985565, 0.04070562869310379, -0.03429826349020004, -0.003276701085269451, 0.08151887357234955, -0.11472442001104355, 0.06964597851037979, -0.11503278464078903, -0.10161620378494263, 0.009736756794154644, 0.021344594657421112, -0.017635449767112732, 0.022710220888257027, -0.051344357430934906, -0.003219804959371686, 0.009034756571054459, -0.06261837482452393, -0.11346893757581711, -0.07778596878051758, 0.09481405466794968, -0.038483623415231705, 0.005766252521425486, -0.19370341300964355, 0.01579875871539116, -0.07802782952785492, 0.04110882431268692, -0.03301183879375458, -0.016919059678912163, -0.04220915958285332, 0.15836010873317719, 0.01566770300269127, -0.08032304793596268, -0.14296548068523407, 0.06578236818313599, -0.0657002255320549, 0.15304335951805115, -0.11342331022024155, -0.09540978819131851, 0.23656173050403595, -0.10326364636421204, -0.12686122953891754, 0.10227407515048981, 0.006417231168597937, 0.00432013301178813, 0.08028869330883026, 0.22379741072654724, 0.04549868777394295, -0.1472753882408142, 0.06182360276579857, 0.1733553260564804, -0.08536577224731445, -0.12865838408470154, 0.05334661900997162, -0.05889151245355606, 0.0050757513381540775, 0.04860455542802811, 0.00635339692234993, 0.10538871586322784, -0.07431323081254959, -0.04350416362285614, 0.0018023509765043855, 0.014889349229633808, 0.032540369778871536, 0.053502246737480164, 0.07121594995260239, -0.10586408525705338, -0.014346176758408546, 0.05915049836039543, 0.012473384849727154, 0.05546606332063675, 0.07566056400537491, -0.07824458926916122, 0.07523985207080841, 0.08620641380548477, -0.03238515183329582, -0.11253771930932999, -0.04384974390268326, -0.06783401966094971, 0.04701196029782295, 0.06976798176765442, 0.1756046712398529, 0.06230808421969414, -0.06800560653209686, -0.05253079906105995, 0.022036606445908546, 0.15244317054748535, 0.002737026894465089, -0.00018774498312268406, -0.12074439972639084, 0.056363075971603394, -0.029165204614400864, -0.023172860965132713, -0.10322974622249603, -0.004951235838234425, 0.1568737030029297, 0.14585448801517487, -0.04912438988685608, 0.06703396886587143, -0.09541365504264832, 0.0676719918847084, -0.03582797944545746, 0.020632799714803696, 0.12695221602916718, 0.003176177153363824, -0.12083008885383606, 0.2063366174697876, -0.022861991077661514, 0.2756877541542053, 0.19385331869125366, -0.2664072811603546, 0.03504352644085884, -0.0771857425570488, -0.025529207661747932, 0.02368001639842987, 0.11642418801784515, 0.012137041427195072, 0.09690926223993301, 0.02864653244614601, 0.09156838804483414, -0.028308026492595673, -0.04136788100004196, -0.014135902747511864, -0.055842325091362, -0.08778955787420273, 0.11056928336620331, 0.11863058805465698, -0.18987371027469635, 0.18550541996955872, 0.2674977779388428, 0.06353795528411865, 0.17024104297161102, -0.07885193079710007, 0.038198839873075485, -0.03564628213644028, 0.028778357431292534, -0.050312891602516174, 0.07006462663412094, -0.24196560680866241, -0.026088887825608253, 0.03120078332722187, 0.012899274006485939, 0.0565243698656559, -0.14664432406425476, -0.09281187504529953, -0.020192502066493034, -0.003125994699075818, -0.09683786332607269, 0.09982286393642426, -0.022991131991147995, 0.05426518991589546, -0.03466043621301651, -0.10500705987215042, 0.11790910363197327, -0.002195822773501277, -0.032241325825452805, 0.12027522921562195, -0.1407175213098526, -0.16423015296459198, -0.12591402232646942, -0.14754514396190643, -0.010272295214235783, -0.012885184027254581, 0.09450490027666092, -0.04180587828159332, -0.01674993522465229, 0.09747575223445892, 0.02213745005428791, -0.15049733221530914, -0.011711008846759796, -0.08220027387142181, 0.03129400312900543, -0.12110886722803116, -0.0710226446390152, -0.08760104328393936, -0.051831867545843124, -0.04834259673953056, 0.07303786277770996, -0.11613848060369492, 0.05853838473558426, 0.08274015039205551, 0.039481695741415024, 0.06263338029384613, -0.056486163288354874, 0.16199715435504913, -0.08131183683872223, -0.062329016625881195, 0.09548679739236832, 0.004713934846222401, 0.08722091466188431, 0.1512838751077652, 0.07543521374464035, -0.05117873474955559, -0.0616474375128746, -0.02352398820221424, -0.08395214378833771, -0.2018536925315857, -0.09306910634040833, -0.14184466004371643, 0.011168778873980045, -0.06365631520748138, 0.05325627699494362, 0.02734704129397869, 0.03850555047392845, 0.014175854623317719, -0.09953257441520691, -0.02947876788675785, 0.05536040663719177, 0.35729339718818665, -0.046026941388845444, 0.06622931361198425, -0.09026601910591125, -0.08888570219278336, 0.08657511323690414, 0.04626443237066269, 0.10290469229221344, 0.13507691025733948, 0.02754874713718891, 0.1257908195257187, 0.18891699612140656, 0.11640173196792603, 0.03859338536858559, 0.011823236010968685, -0.042720671743154526, -0.00948293786495924, -0.019532080739736557, -0.03120400384068489, -0.0015897133853286505, 0.111160509288311, -0.12125701457262039, 0.0047011650167405605, -0.17910119891166687, 0.06562753021717072, 0.0389740988612175, 0.033505238592624664, -0.059628915041685104, 0.019772076979279518, 0.0787830725312233, 0.01838655397295952, -0.04606233909726143, 0.11887043714523315, 0.008213793858885765, -0.12780004739761353, 0.12549720704555511, 0.01000938843935728, 0.08338531851768494, -0.002511385828256607, 0.09371137619018555, -0.06000813469290733, -0.13501881062984467, 0.056642718613147736, 0.09899468719959259, -0.2986595928668976, 0.2900567352771759, -0.003975317347794771, -0.047995854169130325, -0.020688001066446304, -0.06404721736907959, 0.014033439569175243, 0.24491296708583832, 0.13584493100643158, 0.06061333790421486, -0.13782411813735962, -0.08429311960935593, 0.08871324360370636, 0.005843080580234528, 0.1483781933784485, 0.03840288892388344, -0.05057936906814575, -0.022661352530121803, 0.005477559752762318, -0.030877718701958656, 0.026795830577611923, -0.014663415029644966, -0.1569199115037918, 0.04796839877963066, 0.01706736907362938, 0.00284274248406291, -0.02416558563709259, -0.046898167580366135, -0.13261452317237854, 0.1027907282114029, -0.14743095636367798, -0.022756356745958328, -0.09537965804338455, -0.1526433527469635, 0.056947190314531326, -0.10013826936483383, 0.01784992404282093, -0.05335329845547676, -0.11194346100091934, -0.08915155380964279, -0.08901619166135788, 0.10981336981058121, -0.056468259543180466, 0.012069344520568848, -0.023130988702178, 0.22176940739154816, -0.052209559828042984, 0.029052745550870895, 0.00965945702046156, 0.007217562757432461, -0.0511179156601429, -0.13108868896961212, -0.013110041618347168, -0.046934884041547775, 0.012786016799509525, 0.0196582842618227, -0.08206365257501602, 0.008952477015554905, 0.036491647362709045, -0.05586567148566246, 0.19850337505340576, 0.28553906083106995, -0.017203865572810173, 0.1408652812242508, 0.1753707081079483, -0.09586455672979355, -0.2800855338573456, -0.09801515191793442, -0.20292294025421143, -0.024142390117049217, 0.01554806251078844, -0.0954023152589798, 0.06825890392065048, 0.033299192786216736, -0.0381789430975914, 0.06510796397924423, -0.26333585381507874, -0.07794348150491714, 0.13746924698352814, -0.0368921160697937, 0.43193519115448, -0.08904437720775604, -0.11173692345619202, -0.02658582665026188, -0.21687684953212738, 0.09352190047502518, 0.01961575448513031, 0.058151260018348694, -0.0193945225328207, 0.09536103904247284, 0.023044656962156296, -0.04362506791949272, 0.14327436685562134, 0.037979282438755035, 0.04158949479460716, -0.0914607122540474, -0.13056661188602448, 0.100129134953022, 0.026799924671649933, -0.05319719389081001, -0.07586289197206497, -0.02321806363761425, -0.11046075075864792, -0.04369066283106804, -0.10053863376379013, 0.09075199067592621, -0.02548946999013424, -0.041978634893894196, -0.03883076086640358, 0.018127035349607468, -0.025886287912726402, -0.017573196440935135, 0.2704150676727295, -0.05029110983014107, 0.08230139315128326, 0.004314720164984465, 0.12709830701351166, -0.14783550798892975, -0.008171911351382732, -0.09874749928712845, -0.04918679967522621, 0.013149467296898365, -0.005821107421070337, -0.0028859812300652266, 0.16494400799274445, -0.0016126487171277404, 0.04572661593556404, 0.0781601145863533, 0.00783754326403141, -0.014598065987229347, 0.12848930060863495, -0.09755370020866394, -0.1675342470407486, -0.005817699711769819, -0.08751524239778519, 0.14971491694450378, 0.08877339214086533, 0.0955345556139946, 0.055527105927467346, 0.016204895451664925, -0.0016464091604575515, -0.036455608904361725, -0.09143055230379105, -0.030053269118070602, 0.05294616520404816, -0.01795961521565914, -0.09448853880167007, 0.019243614748120308, -0.03133035823702812, -0.20829443633556366, -0.06963088363409042, 0.13842712342739105, -0.12436096370220184, -0.09639706462621689, -0.12784680724143982, 0.0919928327202797, -0.21755599975585938, -0.05964827910065651, -0.008717602118849754, -0.1604822427034378, 0.040605511516332626, 0.24471047520637512, 0.07297197729349136, 0.09982682019472122, -0.04153063893318176, -0.03091159462928772, 0.08050335198640823, -0.012501533143222332, -0.0452473945915699, -0.002638202626258135, -0.07345890253782272, 0.04825513809919357, 0.011420303955674171, 0.1301589459180832, -0.04895208403468132, -0.05325543135404587, -0.18752546608448029, 0.07700138539075851, -0.147604301571846, -0.0692610815167427, -0.12691113352775574, -0.025314273312687874, 0.027804870158433914, -0.14026212692260742, -0.05573628470301628, -0.04604241997003555, -0.1318419724702835, 0.06647725403308868, 0.0716390609741211, 0.10705170780420303, -0.06175842881202698, -0.028296874836087227, 0.14095300436019897, -0.001402225811034441, 0.09704601764678955, 0.1327422857284546, -0.06940644979476929, 0.14741447567939758, -0.17518548667430878, -0.07558301836252213, 0.09423751384019852, 0.03092215582728386, 0.038638632744550705, 0.1198599562048912, -0.019444664940238, 0.10723024606704712, 0.06588846445083618, 0.08974871784448624, -0.04277998208999634, -0.08873608708381653, 0.03429589420557022, 0.05117180943489075, -0.18947160243988037, 0.0073074460960924625, -0.046876516193151474, 0.08968362957239151, 0.007447609677910805, 0.12948505580425262, -0.021867386996746063, 0.03230380639433861, 0.005720936693251133, 0.028984609991312027, 0.019918402656912804, -0.1457792967557907, -0.009127207100391388, -0.12366056442260742, -0.013681424781680107, 0.01611066795885563, 0.27927160263061523, -0.02623768150806427, -0.0764162540435791, 0.050191931426525116, 0.07193838059902191, -0.07346288114786148, -0.019987598061561584, 0.1968526840209961, 0.06915696710348129, -0.008869391866028309, -0.10535459220409393, 0.037410784512758255, -0.05285441130399704, -0.052864495664834976, 0.11264849454164505, 0.16958968341350555, 0.08453905582427979, 0.07006988674402237, 0.04557975009083748, 0.032790061086416245, -0.04635686054825783, -0.17187519371509552, 0.06664387881755829, 0.01884235069155693, -0.068048857152462, 0.20343941450119019, 0.1265532523393631, -0.05649905651807785, 0.04243641346693039, -0.03227315470576286, -0.01831100322306156, -0.17570850253105164, -0.15195508301258087, -0.008638161234557629, -0.0747823417186737, 0.056934211403131485, -0.01381763443350792, 0.06902992725372314, 0.10654284805059433, 0.08606341481208801, -0.04615834355354309, -0.0010784080950543284, -0.06384188681840897, -0.058674849569797516, -0.030458524823188782, -0.012111633084714413, 0.08270400017499924, -0.0993049293756485, -0.018647413700819016, -0.1491028219461441, -0.10106202960014343, -0.06257490813732147, 0.019904769957065582, -0.02168901264667511, -0.05687598139047623, -0.145304873585701, -0.04359940066933632, -0.04170547425746918, 0.10188447684049606, 0.03194250911474228, 0.14200304448604584, -0.010518691502511501, 0.010889274999499321, 0.015667879953980446, 0.16553902626037598, -0.018214000388979912, -0.14261141419410706, 0.05627380684018135, 0.14736489951610565, 0.06894932687282562, 0.11401091516017914, -0.02089226059615612, 0.03653022274374962, 0.00966223981231451, 0.24988725781440735, 0.29920434951782227, -0.051639754325151443, 0.05205092206597328, 0.02470596320927143, 0.05283649265766144, 0.12386806309223175, 0.0311481524258852, 0.06846757978200912, 0.2794262170791626, -0.08743715286254883, -0.036531202495098114, -0.03738490119576454, 0.02726895734667778, -0.10126470774412155, 0.06473054736852646, -0.0243582371622324, -0.06823703646659851, -0.03488235920667648, 0.1050775870680809, -0.1941498965024948, 0.08699780702590942, 0.09645490348339081, -0.16091620922088623, -0.010715305805206299, -0.028643811121582985, 0.15429970622062683, 0.06591441482305527, 0.07206166535615921, -0.052024874836206436, -0.12876613438129425, 0.04546315595507622, 0.029952511191368103, -0.28354015946388245, -0.07063569873571396, 0.10782241821289062, 0.052259352058172226, 0.05389448627829552, -0.022733157500624657, 0.060661327093839645, 0.09383045136928558, 0.0888674333691597, 0.007678474299609661, 0.031690485775470734, 0.054214492440223694, -0.05917581915855408, -0.08982314169406891, -0.11588431149721146, 0.017428414896130562, -0.0569392554461956, 0.041838180273771286, -0.07947971671819687, 0.09357766807079315, 0.07167968153953552, -0.08815281838178635, -0.040075525641441345, 0.07987511157989502, -0.10395390540361404, 0.006217112764716148, 0.03726410120725632, 0.043377529829740524, -0.057750046253204346, -0.04184121638536453, -0.04491027817130089, 0.05065533518791199, -0.09818817675113678, -0.09175180643796921, 0.007499593310058117, -0.05019579827785492, 0.09278787672519684, -0.009661100804805756, -0.08159857988357544, -0.05562692508101463, -0.0343334898352623, 0.09438079595565796, -0.12782850861549377, 0.03492181375622749, 0.04612846300005913, 0.016658952459692955, 0.030640970915555954, -0.1980668008327484, 0.04933590069413185, 0.025749148800969124, -0.06121642142534256, -0.055691689252853394 ]
null
null
transformers
This is a multilingual NER system trained using a Frustratingly Easy Domain Adaptation architecture. It is based on LaBSE and supports different tagsets all using IOBES formats: 1. Wikiann (LOC, PER, ORG) 2. SlavNER 19/21 (EVT, LOC, ORG, PER, PRO) 3. NER-UK (LOC, MISC, ORG, PER) 4. Turku (DATE, EVT, LOC, ORG, PER, PRO, TIME) PER: person, LOC: location, ORG: organization, EVT: event, PRO: product, MISC: Miscellaneous, MEDIA: media, ART: Artifact, TIME: time, DATE: date, GEOPOLIT: Geopolitical, You can select the tagset to use in the output by configuring the model. This models manages differently uppercase words. More information about the model can be found in the paper (https://aclanthology.org/2021.bsnlp-1.12.pdf) and GitHub repository (https://github.com/EMBEDDIA/NER_FEDA).
{"language": ["multilingual", "uk"], "license": "mit", "tags": ["labse", "ner"]}
null
creat89/NER_FEDA_Uk
[ "transformers", "pytorch", "bert", "labse", "ner", "multilingual", "uk", "license:mit", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "multilingual", "uk" ]
TAGS #transformers #pytorch #bert #labse #ner #multilingual #uk #license-mit #endpoints_compatible #region-us
This is a multilingual NER system trained using a Frustratingly Easy Domain Adaptation architecture. It is based on LaBSE and supports different tagsets all using IOBES formats: 1. Wikiann (LOC, PER, ORG) 2. SlavNER 19/21 (EVT, LOC, ORG, PER, PRO) 3. NER-UK (LOC, MISC, ORG, PER) 4. Turku (DATE, EVT, LOC, ORG, PER, PRO, TIME) PER: person, LOC: location, ORG: organization, EVT: event, PRO: product, MISC: Miscellaneous, MEDIA: media, ART: Artifact, TIME: time, DATE: date, GEOPOLIT: Geopolitical, You can select the tagset to use in the output by configuring the model. This models manages differently uppercase words. More information about the model can be found in the paper (URL and GitHub repository (URL
[]
[ "TAGS\n#transformers #pytorch #bert #labse #ner #multilingual #uk #license-mit #endpoints_compatible #region-us \n" ]
[ 39 ]
[ "passage: TAGS\n#transformers #pytorch #bert #labse #ner #multilingual #uk #license-mit #endpoints_compatible #region-us \n" ]
[ -0.0443883053958416, -0.03111727349460125, -0.00857360940426588, 0.0031851688399910927, 0.08147977292537689, 0.043694525957107544, 0.12573795020580292, 0.08403627574443817, 0.1428106427192688, 0.0166592039167881, 0.13022451102733612, 0.1919403076171875, -0.05816154554486275, -0.024060022085905075, -0.001699995482340455, -0.2465478926897049, 0.07238607108592987, 0.032879434525966644, -0.06930974870920181, 0.08793062716722488, 0.11591591686010361, -0.053443543612957, 0.06847774982452393, 0.023010889068245888, -0.09131401032209396, 0.044180866330862045, 0.00856421235948801, -0.05005517229437828, 0.16446621716022491, 0.09297752380371094, 0.13968998193740845, 0.055075887590646744, -0.05436205863952637, -0.21769309043884277, 0.023585772141814232, -0.047488074749708176, -0.0782928466796875, 0.01978924870491028, 0.026010684669017792, -0.08995172381401062, 0.11271946877241135, 0.07029707729816437, -0.016544904559850693, 0.07968421280384064, -0.13348780572414398, -0.13351322710514069, -0.01779269427061081, 0.037381529808044434, -0.007756373379379511, 0.04980332404375076, 0.008175903931260109, 0.15007992088794708, -0.15026096999645233, 0.05599077418446541, 0.15754951536655426, -0.3899775743484497, 0.01447923481464386, 0.07576495409011841, 0.1133384183049202, 0.04544880613684654, -0.07763239741325378, 0.07757234573364258, 0.04644016921520233, 0.007212092634290457, -0.0939955785870552, -0.06164822354912758, -0.005739604122936726, 0.056654419749975204, -0.07544815540313721, -0.0416966937482357, 0.09262708574533463, -0.01815778762102127, -0.002387746935710311, -0.002206408651545644, -0.05910000577569008, -0.004064016509801149, -0.03944685310125351, -0.007126611191779375, -0.021920183673501015, 0.0823359489440918, 0.018751218914985657, -0.05489056557416916, -0.10967227816581726, 0.06194803863763809, -0.25595617294311523, 0.2427111715078354, 0.07479111850261688, 0.06254708766937256, -0.1056901142001152, 0.03832066431641579, -0.02702520042657852, -0.014399474486708641, -0.03633658215403557, -0.09335199743509293, 0.06964270770549774, 0.03849378600716591, -0.007195224519819021, 0.13001227378845215, 0.1022440642118454, 0.12109076976776123, -0.046540193259716034, 0.005374683998525143, -0.013542592525482178, 0.14933395385742188, 0.008197971619665623, 0.04179992526769638, 0.01826348342001438, 0.03091052547097206, -0.023825962096452713, -0.15179066359996796, 0.014598210342228413, 0.013130265288054943, -0.1310560703277588, -0.0552237369120121, -0.07794905453920364, 0.09202886372804642, -0.034504421055316925, 0.0514877513051033, -0.08526704460382462, 0.055695272982120514, 0.045859869569540024, -0.008324303664267063, -0.002742441836744547, -0.009642303921282291, 0.057389263063669205, 0.15153467655181885, -0.03604472801089287, -0.006093971896916628, -0.006737000774592161, 0.0678902342915535, -0.06822454929351807, 0.020854655653238297, -0.02491719461977482, -0.02823764830827713, 0.0763327032327652, -0.10891668498516083, 0.06998088955879211, -0.1422530859708786, -0.07560469210147858, 0.03383829444646835, 0.06753263622522354, 0.000938736426178366, 0.04568754509091377, -0.02461254596710205, 0.0008662194013595581, 0.009197166189551353, -0.06239799037575722, -0.10231123119592667, -0.06802055984735489, 0.09807050228118896, -0.01624288596212864, 0.03829942271113396, -0.23163793981075287, 0.03128678351640701, -0.09646333754062653, 0.014823595061898232, -0.02157416008412838, -0.03667515516281128, -0.06932248920202255, 0.11834071576595306, -0.03309271112084389, -0.04585357755422592, -0.057947032153606415, 0.07135599106550217, -0.08752696216106415, 0.10981620848178864, -0.06634571403265, -0.11582419276237488, 0.21301597356796265, -0.08838498592376709, -0.13130277395248413, 0.08613312244415283, -0.006711555179208517, 0.06195492669939995, 0.08480557799339294, 0.12168603390455246, 0.04166283831000328, -0.200477734208107, 0.03912419453263283, 0.14897990226745605, -0.12023475766181946, -0.12409520894289017, 0.060718514025211334, -0.04444722831249237, -0.0810980349779129, 0.06341315805912018, -0.03806931897997856, 0.13411220908164978, -0.09904427826404572, -0.05806012451648712, -0.00140188611112535, -0.00669087516143918, 0.09213199466466904, 0.06143280491232872, 0.08958674967288971, -0.13917574286460876, -0.04085119813680649, 0.07205670326948166, -0.0006608613184653223, 0.04675920680165291, 0.08120320737361908, -0.0824178010225296, 0.0847410187125206, 0.05885850638151169, -0.051772162318229675, -0.1336377114057541, -0.048520930111408234, -0.04500309005379677, 0.0072501893155276775, 0.09193234145641327, 0.2007310539484024, 0.051264237612485886, -0.04814521223306656, -0.05819845572113991, 0.019342593848705292, 0.08323124051094055, -0.0057902829721570015, -0.011116739362478256, -0.11951108276844025, 0.07347844541072845, -0.03042675368487835, -0.04162215068936348, -0.041854191571474075, -0.011623700149357319, 0.0813821330666542, 0.0760451927781105, -0.018642999231815338, 0.06348656117916107, -0.08964437991380692, 0.07173474133014679, -0.09526370465755463, 0.05800582468509674, 0.1377299726009369, -0.030057352036237717, -0.13120198249816895, 0.15623553097248077, 0.006628848612308502, 0.25780555605888367, 0.213164821267128, -0.2230830043554306, 0.03706083819270134, -0.03499838709831238, -0.0634530782699585, 0.05322718620300293, 0.07701438665390015, 0.005229269154369831, 0.049468085169792175, 0.0028846219647675753, 0.07526997476816177, -0.04768834635615349, -0.05205206945538521, 0.002414439572021365, -0.03395453840494156, -0.10390084236860275, 0.11877945065498352, 0.1288987100124359, -0.17844301462173462, 0.1703796535730362, 0.2640894055366516, 0.06314177811145782, 0.18820218741893768, -0.04602198302745819, 0.027775730937719345, 0.03641132265329361, 0.0313747301697731, -0.05783505737781525, 0.07349207252264023, -0.25524038076400757, -0.014293929561972618, 0.06156802177429199, 0.007995049469172955, 0.02929047681391239, -0.13961714506149292, -0.08819054067134857, -0.032921988517045975, -0.037935417145490646, -0.09030507504940033, 0.09002289921045303, -0.03018912300467491, 0.056917306035757065, -0.02195623330771923, -0.11476951092481613, 0.12337826937437057, -0.0021341159008443356, -0.021561184898018837, 0.12104422599077225, -0.16489234566688538, -0.1505480855703354, -0.11319107562303543, -0.14721378684043884, 0.018818669021129608, -0.008170287124812603, 0.08310037106275558, -0.045089542865753174, -0.02086292952299118, 0.0887475535273552, 0.07209749519824982, -0.17023879289627075, -0.018457025289535522, -0.12149960547685623, 0.04595242068171501, -0.11998968571424484, -0.10843853652477264, -0.05337873101234436, -0.04402138292789459, -0.02648598700761795, 0.09071813523769379, -0.1394423395395279, 0.03652497008442879, 0.06771333515644073, 0.001604184159077704, 0.0410662516951561, -0.06088254228234291, 0.26162028312683105, -0.07558297365903854, -0.06169801577925682, 0.05665111541748047, 0.043600842356681824, 0.06513789296150208, 0.1972726732492447, 0.093253955245018, -0.058436326682567596, -0.05120554566383362, -0.05329509079456329, -0.06323692947626114, -0.2556692361831665, -0.09195756167173386, -0.10270613431930542, 0.049668800085783005, -0.07282190024852753, 0.0727544054389, 0.04566404968500137, 0.032045409083366394, 0.018570654094219208, -0.14983013272285461, -0.017578447237610817, 0.05642358586192131, 0.2754708528518677, -0.05908235162496567, 0.09358623623847961, -0.0797417089343071, -0.08285894989967346, 0.09825413674116135, 0.03814253211021423, 0.12593138217926025, 0.1686519980430603, 0.02814156748354435, 0.12212977558374405, 0.16320949792861938, 0.1322070062160492, 0.08933224529027939, 0.013878470286726952, -0.04795623943209648, -0.018629005178809166, -0.006695236079394817, -0.0033843033015727997, 0.015060726553201675, 0.11655290424823761, -0.1159256100654602, 0.005144062917679548, -0.2400323450565338, 0.062336478382349014, -0.042158905416727066, 0.04178577661514282, -0.06062609702348709, 0.026472842320799828, 0.05651863291859627, -0.04605207219719887, -0.07004758715629578, 0.11602278798818588, -0.025821147486567497, -0.15054450929164886, 0.07482967525720596, -0.0034107721876353025, 0.06634882092475891, 0.05211476981639862, 0.05660397559404373, -0.0552956722676754, -0.17889098823070526, 0.06343037635087967, 0.07627712190151215, -0.3087402284145355, 0.3312089443206787, 0.005009760614484549, -0.02276959829032421, 0.008814138360321522, -0.056279901415109634, -0.031619034707546234, 0.26214733719825745, 0.173406183719635, 0.0539097785949707, -0.11520490795373917, -0.1079145222902298, 0.09868872165679932, 0.019998615607619286, 0.12783850729465485, 0.02171233855187893, -0.030857933685183525, -0.00346596771851182, -0.00821088720113039, -0.02286815643310547, 0.043450091034173965, -0.05077298730611801, -0.1576927900314331, 0.026006555184721947, 0.03418398275971413, 0.040284112095832825, -0.014240902848541737, -0.06130300089716911, -0.15386468172073364, 0.12938997149467468, -0.1923189014196396, -0.07854445278644562, -0.10799367725849152, -0.14119583368301392, 0.06029301881790161, -0.08079545944929123, 0.03737032040953636, -0.05422087758779526, -0.1104201152920723, -0.09156247228384018, -0.13628363609313965, 0.13583992421627045, -0.07919660210609436, 0.0019750664941966534, -0.058712899684906006, 0.2151688188314438, -0.007609868422150612, 0.03836429491639137, -0.02422928623855114, 0.0048594726249575615, -0.08909960836172104, -0.07721433788537979, -0.0014039709931239486, -0.001042234362103045, -0.001079237787052989, -0.012428006157279015, -0.10875118523836136, 0.05577235296368599, 0.02795126847922802, -0.06067599728703499, 0.2176428884267807, 0.23330038785934448, -0.004457545932382345, 0.13329167664051056, 0.17578239738941193, -0.0639556497335434, -0.24568775296211243, -0.11299321800470352, -0.16120189428329468, -0.027651753276586533, -0.02388039231300354, -0.10866045206785202, 0.10339666157960892, 0.04690801724791527, -0.04431585222482681, 0.05494004487991333, -0.22938655316829681, -0.08964604139328003, 0.13030868768692017, 0.003922967240214348, 0.43183434009552, -0.09493126720190048, -0.052524134516716, -0.02528732270002365, -0.2487388253211975, 0.15132242441177368, 0.05849163606762886, 0.07376794517040253, -0.029215142130851746, 0.1021343469619751, -0.01790623553097248, -0.03555960953235626, 0.1276356279850006, 0.03239491954445839, 0.012368390336632729, -0.010628709569573402, -0.131134495139122, 0.09381087869405746, 0.06916339695453644, -0.03546249493956566, -0.05168119817972183, -0.02227884717285633, -0.07325021177530289, -0.04207635298371315, -0.10937604308128357, 0.10362720489501953, -0.02157191000878811, -0.04954204335808754, -0.0905815064907074, 0.027918457984924316, -0.05791556090116501, 0.0017050192691385746, 0.27153313159942627, -0.060126569122076035, 0.13318343460559845, 0.000033295433240709826, 0.11977216601371765, -0.14591248333454132, -0.035371556878089905, -0.0493711419403553, -0.07574734836816788, 0.04204394668340683, -0.04656074196100235, -0.021009068936109543, 0.1637069135904312, -0.0250858087092638, 0.02997184358537197, 0.07641483098268509, -0.03411683812737465, -0.005784336011856794, 0.11194778978824615, -0.10258263349533081, -0.1885860115289688, -0.00017529241449665278, -0.06049462407827377, 0.14418300986289978, 0.042490921914577484, 0.06290692836046219, 0.04058540239930153, -0.006816755048930645, 0.00960373505949974, -0.05394064635038376, -0.09099522233009338, 0.004256926476955414, 0.0475616455078125, -0.02064470387995243, -0.09291303902864456, -0.02463647350668907, -0.004995990078896284, -0.18634441494941711, -0.05293947458267212, 0.09585949778556824, -0.12931832671165466, -0.098225437104702, -0.0760321170091629, 0.09722448140382767, -0.2217109352350235, -0.050831541419029236, 0.038465797901153564, -0.1431759148836136, 0.036597125232219696, 0.22441110014915466, 0.07541082054376602, 0.10419061779975891, -0.03771096095442772, -0.004507656209170818, 0.1247367113828659, 0.01017481368035078, -0.11615238338708878, 0.0541955940425396, -0.048397745937108994, -0.022875159978866577, -0.0016973656602203846, 0.11710071563720703, -0.05786103010177612, -0.043350398540496826, -0.1492023766040802, 0.05791212618350983, -0.07069352269172668, -0.0788756012916565, -0.10503106564283371, -0.056771665811538696, 0.03274638578295708, -0.14033359289169312, -0.030317282304167747, -0.08218193799257278, -0.11144521832466125, 0.02311563491821289, 0.0737898051738739, 0.07736199349164963, -0.06482022255659103, -0.005842356476932764, 0.15675047039985657, 0.018646281212568283, 0.09931337088346481, 0.11407094448804855, -0.03993496298789978, 0.156137615442276, -0.17659854888916016, -0.05895092338323593, 0.09457843005657196, 0.023577313870191574, 0.0684560164809227, 0.06655192375183105, -0.035253752022981644, 0.1077815443277359, 0.08918152004480362, 0.08538956195116043, -0.03001227416098118, -0.10133645683526993, 0.02924814075231552, 0.039211202412843704, -0.17104007303714752, -0.015256964601576328, -0.07363927364349365, 0.13485482335090637, -0.033224448561668396, 0.14608481526374817, -0.016236914321780205, 0.005142075475305319, 0.009375985711812973, 0.04650220274925232, 0.012539372779428959, -0.15241321921348572, -0.04813477024435997, -0.10774093121290207, -0.06520745158195496, -0.050592709332704544, 0.3151349723339081, -0.009899349883198738, -0.11281129717826843, 0.07626639306545258, 0.001753230462782085, -0.054487794637680054, -0.018049592152237892, 0.2364557385444641, 0.09315843135118484, 0.007324576377868652, -0.09288602322340012, 0.03554171323776245, -0.053114790469408035, -0.07899800688028336, 0.12460973858833313, 0.17410704493522644, 0.05794823542237282, 0.04466145858168602, 0.03346344083547592, 0.06102278456091881, -0.09000690281391144, -0.16606968641281128, 0.07396166026592255, -0.004293758422136307, -0.04836807772517204, 0.27261099219322205, 0.1962086260318756, -0.0743870660662651, 0.07801739126443863, 0.0036508829798549414, -0.0193620678037405, -0.18342600762844086, -0.07934283465147018, -0.02032974362373352, -0.15305669605731964, 0.06851257383823395, -0.037224430590867996, 0.02497127652168274, 0.09734928607940674, 0.08062465488910675, -0.03949001058936119, 0.029286885634064674, -0.02030632458627224, -0.050175439566373825, -0.022342640906572342, -0.032977908849716187, 0.0709572359919548, -0.06436360627412796, -0.01169878151267767, -0.13930821418762207, -0.08113247156143188, -0.06149251013994217, 0.0528176985681057, -0.09877416491508484, -0.015431505627930164, -0.16579455137252808, -0.0623597726225853, -0.06659280508756638, 0.11035335808992386, 0.002843548310920596, 0.15384504199028015, -0.008773083798587322, 0.014617523178458214, 0.007138639222830534, 0.1472489982843399, -0.008856387808918953, -0.0913223922252655, 0.044436872005462646, 0.14886407554149628, 0.06375004351139069, 0.14456689357757568, -0.005373266525566578, 0.029985705390572548, -0.0626385435461998, 0.28298231959342957, 0.30701738595962524, -0.041866958141326904, 0.032552797347307205, 0.05729331448674202, 0.0498838871717453, 0.09763623028993607, 0.04350404813885689, 0.09507111459970474, 0.2821650505065918, -0.12271733582019806, -0.0731431394815445, -0.0281531885266304, 0.042393505573272705, -0.1482304185628891, 0.06420999020338058, 0.020908333361148834, -0.11077714711427689, -0.06383457779884338, 0.07452179491519928, -0.17101038992404938, 0.1024400144815445, 0.03765791654586792, -0.10818939656019211, -0.02699446864426136, -0.03542528301477432, 0.1431645303964615, 0.060563527047634125, 0.06372226774692535, -0.048140913248062134, -0.10727736353874207, 0.025648077949881554, 0.02518484555184841, -0.2877611219882965, -0.037593159824609756, 0.13725945353507996, 0.04113301634788513, 0.04327455163002014, -0.0027579437009990215, 0.05116603150963783, 0.0815761610865593, 0.0738169252872467, -0.003670679870992899, 0.04136170819401741, 0.052207618951797485, -0.029110148549079895, -0.08380785584449768, -0.0703936517238617, 0.05859772115945816, -0.071839839220047, 0.06312081962823868, -0.03919653594493866, 0.09471087157726288, 0.060945864766836166, -0.08402280509471893, -0.07807782292366028, 0.10731973499059677, -0.11811380833387375, 0.03604375943541527, 0.0208913441747427, 0.0429772324860096, -0.061462413519620895, -0.05290345102548599, -0.037558138370513916, 0.058417078107595444, -0.06802675873041153, -0.11769861727952957, -0.018011026084423065, -0.057942334562540054, 0.0985715314745903, -0.022827453911304474, -0.13333620131015778, -0.041012488305568695, -0.051463913172483444, 0.1082782968878746, -0.13537821173667908, 0.030424373224377632, 0.11164722591638565, -0.008810766041278839, 0.055523019284009933, -0.17861251533031464, 0.04935001954436302, 0.014276864938437939, -0.07467988133430481, -0.060280296951532364 ]
null
null
transformers
# MyModel ## Model description This is the `BART-TL-all` model from the paper [BART-TL: Weakly-Supervised Topic Label Generation](https://www.aclweb.org/anthology/2021.eacl-main.121.pdf). We aim to solve the topic labeling task using generative methods, rather than selection from a pool of labels as was done in previous State of the Art works. For more details not covered here, you can read the paper or look at the open-source implementation: https://github.com/CristianViorelPopa/BART-TL-topic-label-generation. There are two models made available from the paper: * [BART-TL-all](https://huggingface.co/cristian-popa/bart-tl-all) * [BART-TL-ng](https://huggingface.co/cristian-popa/bart-tl-ng) ## Intended uses & limitations #### How to use The model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model. ```python from transformers import AutoTokenizer, AutoModelForSeq2SeqLM mname = "cristian-popa/bart-tl-all" tokenizer = AutoTokenizer.from_pretrained(mname) model = AutoModelForSeq2SeqLM.from_pretrained(mname) input = "site web google search website online internet social content user" enc = tokenizer(input, return_tensors="pt", truncation=True, padding="max_length", max_length=128) outputs = model.generate( input_ids=enc.input_ids, attention_mask=enc.attention_mask, max_length=15, min_length=1, do_sample=False, num_beams=25, length_penalty=1.0, repetition_penalty=1.5 ) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # application programming interface ``` #### Limitations and bias The model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy. ## Training data The model was fine-tuned on 5 different StackExchange corpora (see https://archive.org/download/stackexchange for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here. ## Training procedure The large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the [NETL](https://www.aclweb.org/anthology/C16-1091.pdf) method, along with other heuristic labels, such as n-grams from the topics, relevant sentences in the corpora and noun phrases. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the [paper](https://www.aclweb.org/anthology/2021.eacl-main.121.pdf) or by following [this notebook](https://github.com/CristianViorelPopa/BART-TL-topic-label-generation/blob/main/notebooks/end_to_end_workflow.ipynb). ## Eval results model | Top-1 Avg. | Top-3 Avg. | Top-5 Avg. | nDCG-1 | nDCG-3 | nDCG-5 ------------|------------|------------|------------|--------|--------|------- NETL (U) | 2.66 | 2.59 | 2.50 | 0.83 | 0.85 | 0.87 NETL (S) | 2.74 | 2.57 | 2.49 | 0.88 | 0.85 | 0.88 BART-TL-all | 2.64 | 2.52 | 2.43 | 0.83 | 0.84 | 0.87 BART-TL-ng | 2.62 | 2.50 | 2.33 | 0.82 | 0.84 | 0.85 ### BibTeX entry and citation info ```bibtex @inproceedings{popa-rebedea-2021-bart, title = "{BART}-{TL}: Weakly-Supervised Topic Label Generation", author = "Popa, Cristian and Rebedea, Traian", booktitle = "Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume", month = apr, year = "2021", address = "Online", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/2021.eacl-main.121", pages = "1418--1425", abstract = "We propose a novel solution for assigning labels to topic models by using multiple weak labelers. The method leverages generative transformers to learn accurate representations of the most important topic terms and candidate labels. This is achieved by fine-tuning pre-trained BART models on a large number of potential labels generated by state of the art non-neural models for topic labeling, enriched with different techniques. The proposed BART-TL model is able to generate valuable and novel labels in a weakly-supervised manner and can be improved by adding other weak labelers or distant supervision on similar tasks.", } ```
{"language": ["en"], "license": "apache-2.0", "tags": ["topic labeling"], "metrics": ["ndcg"], "<!-- thumbnail": "https://raw.githubusercontent.com/JetRunner/BERT-of-Theseus/master/bert-of-theseus.png -->"}
text2text-generation
cristian-popa/bart-tl-all
[ "transformers", "pytorch", "bart", "text2text-generation", "topic labeling", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bart #text2text-generation #topic labeling #en #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
MyModel ======= Model description ----------------- This is the 'BART-TL-all' model from the paper BART-TL: Weakly-Supervised Topic Label Generation. We aim to solve the topic labeling task using generative methods, rather than selection from a pool of labels as was done in previous State of the Art works. For more details not covered here, you can read the paper or look at the open-source implementation: URL There are two models made available from the paper: * BART-TL-all * BART-TL-ng Intended uses & limitations --------------------------- #### How to use The model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model. #### Limitations and bias The model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy. Training data ------------- The model was fine-tuned on 5 different StackExchange corpora (see URL for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here. Training procedure ------------------ The large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the NETL method, along with other heuristic labels, such as n-grams from the topics, relevant sentences in the corpora and noun phrases. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the paper or by following this notebook. Eval results ------------ ### BibTeX entry and citation info
[ "#### How to use\n\n\nThe model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model.", "#### Limitations and bias\n\n\nThe model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy.\n\n\nTraining data\n-------------\n\n\nThe model was fine-tuned on 5 different StackExchange corpora (see URL for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here.\n\n\nTraining procedure\n------------------\n\n\nThe large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the NETL method, along with other heuristic labels, such as n-grams from the topics, relevant sentences in the corpora and noun phrases. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the paper or by following this notebook.\n\n\nEval results\n------------", "### BibTeX entry and citation info" ]
[ "TAGS\n#transformers #pytorch #bart #text2text-generation #topic labeling #en #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "#### How to use\n\n\nThe model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model.", "#### Limitations and bias\n\n\nThe model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy.\n\n\nTraining data\n-------------\n\n\nThe model was fine-tuned on 5 different StackExchange corpora (see URL for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here.\n\n\nTraining procedure\n------------------\n\n\nThe large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the NETL method, along with other heuristic labels, such as n-grams from the topics, relevant sentences in the corpora and noun phrases. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the paper or by following this notebook.\n\n\nEval results\n------------", "### BibTeX entry and citation info" ]
[ 52, 54, 236, 11 ]
[ "passage: TAGS\n#transformers #pytorch #bart #text2text-generation #topic labeling #en #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nThe model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model.#### Limitations and bias\n\n\nThe model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy.\n\n\nTraining data\n-------------\n\n\nThe model was fine-tuned on 5 different StackExchange corpora (see URL for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here.\n\n\nTraining procedure\n------------------\n\n\nThe large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the NETL method, along with other heuristic labels, such as n-grams from the topics, relevant sentences in the corpora and noun phrases. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the paper or by following this notebook.\n\n\nEval results\n------------### BibTeX entry and citation info" ]
[ -0.025374848395586014, 0.039004284888505936, 0.0012018803972750902, -0.0021390828769654036, 0.05627782270312309, -0.012651707045733929, 0.08196865767240524, 0.07572115957736969, -0.014824253506958485, 0.05512076988816261, -0.005266357213258743, -0.028726065531373024, 0.06058561056852341, 0.09164533764123917, 0.06597661972045898, -0.23688039183616638, -0.02427762560546398, -0.019137242808938026, 0.025111205875873566, 0.07413837313652039, 0.0801633968949318, -0.07733819633722305, 0.05459797382354736, 0.03504233807325363, -0.11093968152999878, 0.014345529489219189, -0.014009845443069935, -0.04866432398557663, 0.08257357776165009, 0.0835619866847992, 0.07595868408679962, -0.0010336431441828609, 0.020149867981672287, -0.16520194709300995, 0.012523103505373001, 0.05465930700302124, -0.03233037516474724, 0.04172619804739952, -0.008979108184576035, -0.015063101425766945, 0.2015094757080078, -0.12185148149728775, 0.07069770991802216, 0.034489769488573074, -0.10205763578414917, -0.1403585970401764, -0.0698171928524971, 0.048952650278806686, 0.06267505884170532, 0.0733892172574997, -0.03587985783815384, 0.010870910249650478, -0.12984201312065125, 0.027630822733044624, 0.10064592212438583, -0.19328588247299194, -0.023060372099280357, 0.07223936915397644, 0.02025728113949299, 0.14581447839736938, -0.045333217829465866, 0.021456744521856308, 0.04448847100138664, -0.009807242080569267, 0.024089299142360687, 0.012852762825787067, 0.040141187608242035, -0.01932772621512413, -0.12328711152076721, -0.018781736493110657, 0.19605806469917297, 0.033163152635097504, -0.025639932602643967, -0.09235469996929169, -0.012363945133984089, 0.00573913287371397, -0.005389542318880558, -0.04006806015968323, -0.006444129161536694, -0.030662275850772858, 0.08311662822961807, -0.09281859546899796, -0.10850023478269577, -0.011667453683912754, 0.0045842742547392845, 0.0892564207315445, -0.013522819615900517, 0.018361195921897888, -0.037462279200553894, 0.08351389318704605, -0.027456682175397873, -0.0720236599445343, -0.005998218432068825, -0.026426732540130615, -0.06898880004882812, -0.04205002263188362, -0.02252735197544098, 0.009246132336556911, -0.05749483406543732, 0.08528144657611847, -0.004157711286097765, -0.03171727433800697, -0.05008109286427498, 0.021449534222483635, 0.042779967188835144, 0.06744495034217834, -0.04761285334825516, -0.05863207206130028, 0.036818116903305054, 0.0028639717493206263, 0.0028553286101669073, -0.018386458978056908, -0.07960375398397446, 0.06726165860891342, -0.0008613115642219782, 0.011409942060709, -0.007289382163435221, 0.013667714782059193, -0.013623302802443504, -0.0645662248134613, 0.10396476089954376, -0.096501424908638, -0.027812102809548378, 0.005944224074482918, 0.0001212509159813635, 0.0053077274933457375, 0.034477945417165756, -0.018044035881757736, -0.024628138169646263, 0.1134599968791008, -0.06307589262723923, -0.04186088591814041, -0.07820724695920944, -0.08072818070650101, -0.010292606428265572, -0.027942879125475883, -0.03270455822348595, -0.09064894914627075, -0.20244337618350983, -0.03909974545240402, -0.01929166167974472, -0.04113118350505829, -0.006094860844314098, -0.01768438331782818, 0.019536267966032028, -0.015020006336271763, -0.01866171509027481, 0.04201359301805496, -0.047658275812864304, 0.0019734289962798357, -0.030442634597420692, 0.07271069288253784, -0.001739004859700799, 0.04161064699292183, -0.09151079505681992, -0.019997790455818176, -0.08096830546855927, 0.09918474406003952, -0.09743387997150421, 0.040331318974494934, -0.0501474030315876, -0.06316972523927689, -0.05828740447759628, 0.008384767919778824, -0.032446131110191345, 0.11862676590681076, -0.2411566525697708, -0.10155875980854034, 0.22188405692577362, -0.13674941658973694, -0.04416818171739578, 0.09289900958538055, -0.058656495064496994, 0.19051308929920197, 0.13308261334896088, 0.1484132707118988, 0.11054572463035583, -0.05419793725013733, 0.012045875191688538, 0.00807588268071413, 0.024857141077518463, 0.124969482421875, 0.029876917600631714, -0.029520142823457718, -0.08425776660442352, 0.03338806703686714, -0.0417444184422493, -0.010980161838233471, -0.041108306497335434, -0.1037934347987175, 0.06212481111288071, 0.005830963607877493, 0.03267896920442581, 0.005314937327057123, 0.02663646638393402, 0.01572316326200962, -0.07575704157352448, 0.04886321723461151, 0.06014768034219742, -0.054265134036540985, 0.0027652904391288757, -0.023712513968348503, 0.08095517754554749, -0.03142314404249191, -0.005250903312116861, -0.15482088923454285, -0.05784992873668671, 0.022354478016495705, -0.057889532297849655, 0.10083726793527603, 0.12827707827091217, -0.006238583941012621, 0.036908309906721115, -0.028443872928619385, 0.04381833225488663, -0.04617217555642128, 0.017520533874630928, -0.0470028780400753, -0.14531449973583221, -0.02311975695192814, -0.062405284494161606, 0.17908786237239838, -0.1412356197834015, 0.0066263931803405285, 0.06552587449550629, 0.009014279581606388, 0.01885327138006687, -0.02184450812637806, -0.018188288435339928, 0.04846247285604477, -0.002647588262334466, -0.039111748337745667, 0.040195975452661514, 0.028115347027778625, -0.12819229066371918, -0.02621748298406601, -0.10436825454235077, -0.05807815492153168, 0.08593451976776123, 0.06871113926172256, -0.1004435196518898, -0.08555133640766144, -0.05906803160905838, -0.02791197970509529, -0.043020639568567276, -0.05431250110268593, 0.15854454040527344, 0.011350056156516075, 0.058783020824193954, -0.08793613314628601, -0.026198498904705048, 0.021638697013258934, 0.007649638224393129, 0.02080201357603073, 0.0329265370965004, 0.04627590626478195, -0.14385628700256348, 0.059971995651721954, 0.060213491320610046, 0.029332660138607025, 0.2260117083787918, 0.05588638782501221, -0.07137778401374817, 0.014338353648781776, -0.010641013272106647, 0.0023751931730657816, 0.12133446335792542, -0.15119768679141998, 0.0012428713962435722, 0.017992205917835236, -0.012768971733748913, 0.06418664008378983, -0.11056609451770782, 0.02154443971812725, 0.04523370414972305, -0.03928098827600479, 0.01849675178527832, -0.0439380407333374, -0.03175586089491844, 0.112209752202034, 0.0735037699341774, 0.021935664117336273, -0.01748288795351982, -0.05216015502810478, -0.13867239654064178, 0.21151670813560486, -0.09573441743850708, -0.2929095923900604, -0.0642627626657486, 0.08207657933235168, 0.004545880481600761, 0.004897106904536486, 0.03593382611870766, -0.12028618156909943, -0.024446604773402214, -0.09148507565259933, 0.00596793694421649, 0.06675173342227936, -0.04713333770632744, -0.04167424887418747, -0.029828494414687157, -0.041587941348552704, -0.11556001752614975, -0.024811800569295883, -0.08021716773509979, -0.09021401405334473, 0.014055823907256126, -0.022558089345693588, 0.07490658760070801, 0.15941175818443298, 0.028983348980545998, -0.011348794214427471, -0.026321861892938614, 0.1380263715982437, -0.0872647762298584, 0.05129505693912506, 0.0549904890358448, 0.04482033848762512, 0.037481676787137985, 0.08293652534484863, 0.04003308340907097, -0.056550282984972, 0.01725054532289505, 0.06478191912174225, -0.06773606687784195, -0.19421544671058655, -0.15415558218955994, -0.04637949541211128, -0.025978153571486473, 0.028528593480587006, 0.028891142457723618, 0.12461499124765396, -0.0037729176692664623, -0.027297021821141243, -0.03139037638902664, 0.009424341842532158, 0.08151570707559586, 0.04585656151175499, -0.04042676091194153, 0.07826856523752213, -0.023666907101869583, -0.03574433550238609, 0.11165247112512589, -0.010570734739303589, 0.26734036207199097, 0.038922760635614395, 0.11020205914974213, 0.08978895843029022, 0.007828119210898876, 0.034113384783267975, 0.0389973483979702, -0.01658691093325615, -0.004989570006728172, -0.04169847071170807, -0.11096638441085815, -0.012294625863432884, 0.12519261240959167, -0.0356748066842556, -0.05450219660997391, -0.0623164027929306, -0.030980275943875313, 0.060690637677907944, 0.19402699172496796, 0.016241122037172318, -0.15273137390613556, -0.07467057555913925, 0.014522177167236805, 0.01643136702477932, -0.026698289439082146, 0.01985657960176468, 0.0745992586016655, -0.1383182853460312, 0.03738125413656235, -0.047442276030778885, 0.05590251460671425, -0.016457967460155487, 0.025810606777668, 0.0573166087269783, -0.04152926802635193, -0.02078746259212494, 0.1067386120557785, -0.11537405103445053, 0.17401033639907837, 0.01359597872942686, 0.03443432226777077, -0.10598739981651306, -0.022150877863168716, -0.005321873817592859, 0.01936442404985428, 0.16003969311714172, 0.03688090667128563, -0.0187061857432127, -0.046399299055337906, -0.0534193329513073, 0.04812349006533623, 0.03411715850234032, -0.08110268414020538, 0.0553300715982914, 0.035752974450588226, 0.017497962340712547, 0.019154423847794533, -0.0239311121404171, -0.11697561293840408, -0.17434340715408325, 0.059314996004104614, -0.10086710751056671, -0.04616732895374298, -0.039796214550733566, -0.08101294189691544, 0.06058913841843605, 0.2858569920063019, -0.11056961119174957, -0.0976484939455986, -0.0931582823395729, 0.07961542904376984, 0.025420071557164192, -0.04420466721057892, 0.009872093796730042, 0.054034892469644547, 0.12028951197862625, -0.02685481496155262, -0.12529999017715454, 0.025645487010478973, -0.06017879396677017, -0.08150379359722137, -0.06256409734487534, 0.1256570667028427, 0.11306334286928177, 0.02061164192855358, 0.019414573907852173, -0.040288038551807404, -0.01881306618452072, -0.13894546031951904, -0.00013835493882652372, 0.17283155024051666, 0.03272166848182678, 0.06869496405124664, -0.10351663082838058, 0.001994180493056774, -0.08153092116117477, -0.018127717077732086, 0.13208991289138794, 0.20465973019599915, -0.09127013385295868, 0.13338486850261688, 0.1468401700258255, -0.13076111674308777, -0.155958354473114, 0.0859539583325386, 0.03101331554353237, 0.016179265454411507, 0.0023275427520275116, -0.19987569749355316, 0.13098886609077454, 0.09469186514616013, -0.006818537600338459, 0.018067440018057823, -0.21116642653942108, -0.13393408060073853, 0.07478226721286774, 0.012571830302476883, 0.0390593521296978, -0.10043828189373016, -0.02606084942817688, -0.006132635287940502, -0.06268465518951416, 0.2116919457912445, -0.17832006514072418, 0.0579785592854023, 0.003898933995515108, 0.027896972373127937, 0.010773023590445518, -0.03931251913309097, 0.08484436571598053, 0.09473834931850433, 0.06972039490938187, 0.001527162967249751, -0.019097549840807915, 0.1040133535861969, 0.0012857409892603755, 0.03482336550951004, 0.05610427260398865, 0.03022756241261959, -0.09981384873390198, -0.05043415725231171, -0.12163370102643967, 0.0856333076953888, -0.07691457122564316, -0.08862898498773575, -0.07717464119195938, 0.12471956759691238, 0.10682511329650879, -0.01807301677763462, -0.10433310270309448, -0.10616196691989899, 0.09775178134441376, 0.02757222205400467, 0.10222077369689941, -0.06222504749894142, -0.07527535408735275, 0.03059513121843338, -0.00356341153383255, 0.10012028366327286, -0.07052880525588989, 0.01892682909965515, 0.0911131352186203, 0.0434868298470974, 0.14623934030532837, 0.03134980797767639, -0.09305533021688461, -0.022113433107733727, 0.06249277666211128, -0.1434045135974884, -0.10386054217815399, -0.0667789876461029, -0.05403197184205055, -0.05982176214456558, -0.06637430191040039, 0.04598575085401535, -0.04939538612961769, -0.007862349972128868, -0.03054654225707054, 0.040938373655080795, -0.03713683784008026, 0.0963565930724144, 0.015971768647432327, 0.0425214022397995, -0.0581410713493824, 0.015348481014370918, 0.07869705557823181, -0.12279696017503738, 0.011533697135746479, 0.08363792300224304, -0.09767021238803864, -0.03269037976861, -0.09035198390483856, 0.10608736425638199, -0.1637578010559082, -0.01416875422000885, -0.018254723399877548, -0.08534263074398041, -0.0008854248444549739, 0.10918448120355606, 0.023284228518605232, 0.011308975517749786, -0.04954603686928749, 0.01783323846757412, -0.029697829857468605, 0.0575653575360775, -0.04117586836218834, 0.006940060760825872, 0.005735950544476509, 0.13801130652427673, 0.037868618965148926, 0.04139934480190277, -0.03202034533023834, -0.04953631013631821, -0.09296920895576477, -0.005610780790448189, -0.08332479745149612, 0.0232798270881176, -0.09166106581687927, 0.007772115524858236, -0.030092796310782433, 0.06940682232379913, 0.026405690237879753, -0.01912532187998295, -0.038181766867637634, 0.008061409927904606, -0.0184697937220335, 0.02613282948732376, -0.0988495871424675, 0.03648442402482033, 0.013355836272239685, -0.05714363604784012, 0.06967658549547195, -0.03552120551466942, 0.05073747783899307, -0.0042486232705414295, -0.15376003086566925, 0.04435231536626816, 0.05423755943775177, 0.06662016361951828, 0.035429757088422775, -0.11604003608226776, 0.013955638743937016, 0.03035620041191578, -0.006747977342456579, -0.013334184885025024, -0.025433465838432312, -0.09370122104883194, 0.09390179812908173, -0.020380087196826935, -0.002058945596218109, -0.07326298952102661, 0.04332165792584419, 0.03923889622092247, 0.032391343265771866, 0.04491370916366577, -0.06826385855674744, 0.06088077276945114, -0.11740238964557648, -0.05905755236744881, 0.0011629258515313268, 0.022845670580863953, -0.0019516521133482456, -0.05847083777189255, 0.037387289106845856, 0.014355680905282497, 0.25390881299972534, 0.062093090265989304, 0.069460429251194, 0.020268874242901802, -0.1339385062456131, 0.06423936039209366, -0.007607430685311556, 0.071237713098526, 0.030539395287632942, -0.019249018281698227, 0.027528604492545128, -0.002912512980401516, -0.052990030497312546, 0.07733967900276184, 0.26960861682891846, 0.09317513555288315, 0.07002261281013489, -0.01596338488161564, 0.05849657207727432, -0.09895317256450653, -0.11765959858894348, -0.03474835678935051, 0.015350895933806896, 0.09132018685340881, -0.05910055711865425, -0.020705198869109154, 0.1265740543603897, -0.1816927045583725, 0.0885203629732132, 0.023681990802288055, -0.09824202954769135, -0.10433228313922882, -0.11110048741102219, -0.022911371663212776, -0.023831717669963837, 0.005452331621199846, -0.15023021399974823, 0.009538368321955204, -0.003261393401771784, 0.0014824530808255076, -0.06816113740205765, 0.12427623569965363, -0.04228925332427025, -0.06637547165155411, 0.08012707531452179, 0.03214878961443901, 0.00829031690955162, -0.09652137756347656, 0.012890353798866272, 0.023357631638646126, 0.05309714749455452, 0.007246872875839472, 0.0351165309548378, -0.010146970860660076, 0.0496409609913826, -0.011189096607267857, -0.02576020546257496, -0.0024872974026948214, -0.021840723231434822, 0.08070813119411469, 0.1502762734889984, 0.06309176236391068, -0.07360146194696426, 0.0024960616137832403, 0.1180528923869133, -0.03477959707379341, -0.05430341511964798, -0.12513689696788788, 0.2332977056503296, -0.0047600348480045795, 0.016306515783071518, -0.0002442345430608839, -0.0810641348361969, 0.04781652241945267, 0.2072199285030365, 0.1980423480272293, -0.0047171469777822495, 0.007482382468879223, -0.020065026357769966, 0.00020159027189947665, -0.014014928601682186, 0.12610025703907013, 0.015009094029664993, 0.28122517466545105, -0.06580141931772232, 0.04875817522406578, -0.023796308785676956, -0.05780494213104248, -0.05968209728598595, 0.04789343848824501, 0.06261160224676132, 0.008001189678907394, -0.04467308893799782, 0.05927058681845665, -0.030398819595575333, -0.19415849447250366, 0.06482802331447601, -0.0885695368051529, -0.08933869004249573, -0.02497423067688942, -0.08924619108438492, -0.03111174702644348, 0.032460931688547134, -0.03607760742306709, -0.06338637322187424, 0.08439449220895767, 0.018250491470098495, -0.05676942318677902, -0.10335543751716614, 0.1096220314502716, -0.03972944989800453, 0.16455940902233124, -0.01261191163212061, 0.04919590428471565, 0.07728080451488495, 0.052128225564956665, -0.0696154311299324, 0.02687336876988411, 0.014175341464579105, -0.039005499333143234, 0.028385547921061516, 0.09246441721916199, -0.005342214368283749, 0.07834252715110779, 0.054535411298274994, -0.08795330673456192, 0.045932330191135406, 0.0032774799037724733, 0.0011050270404666662, -0.09536474943161011, 0.019846906885504723, -0.07773665338754654, 0.18985538184642792, 0.18354102969169617, -0.031886324286460876, 0.009509732015430927, -0.05681654438376427, 0.009294232353568077, 0.017961733043193817, 0.1746419370174408, -0.03700302913784981, -0.10404302924871445, 0.012524577789008617, 0.010034637525677681, 0.030655305832624435, -0.261962354183197, -0.07351183146238327, 0.0522366426885128, -0.02872975543141365, 0.013980776071548462, 0.09918464720249176, 0.044647157192230225, 0.027774231508374214, -0.046033378690481186, -0.10648221522569656, -0.013203945942223072, 0.10111916810274124, -0.12471795827150345, -0.06752505153417587 ]
null
null
transformers
# MyModel ## Model description This is the `BART-TL-ng` model from the paper [BART-TL: Weakly-Supervised Topic Label Generation](https://www.aclweb.org/anthology/2021.eacl-main.121.pdf). We aim to solve the topic labeling task using generative methods, rather than selection from a pool of labels as was done in previous State of the Art works. For more details not covered here, you can read the paper or look at the open-source implementation: https://github.com/CristianViorelPopa/BART-TL-topic-label-generation. There are two models made available from the paper: * [BART-TL-all](https://huggingface.co/cristian-popa/bart-tl-all) * [BART-TL-ng](https://huggingface.co/cristian-popa/bart-tl-ng) ## Intended uses & limitations #### How to use The model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model. ```python from transformers import AutoTokenizer, AutoModelForSeq2SeqLM mname = "cristian-popa/bart-tl-ng" tokenizer = AutoTokenizer.from_pretrained(mname) model = AutoModelForSeq2SeqLM.from_pretrained(mname) input = "site web google search website online internet social content user" enc = tokenizer(input, return_tensors="pt", truncation=True, padding="max_length", max_length=128) outputs = model.generate( input_ids=enc.input_ids, attention_mask=enc.attention_mask, max_length=15, min_length=1, do_sample=False, num_beams=25, length_penalty=1.0, repetition_penalty=1.5 ) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # windows live messenger ``` #### Limitations and bias The model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy. ## Training data The model was fine-tuned on 5 different StackExchange corpora (see https://archive.org/download/stackexchange for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here. ## Training procedure The large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the [NETL](https://www.aclweb.org/anthology/C16-1091.pdf) method, along with n-grams from the topics. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the [paper](https://www.aclweb.org/anthology/2021.eacl-main.121.pdf) or by following [this notebook](https://github.com/CristianViorelPopa/BART-TL-topic-label-generation/blob/main/notebooks/end_to_end_workflow.ipynb). ## Eval results model | Top-1 Avg. | Top-3 Avg. | Top-5 Avg. | nDCG-1 | nDCG-3 | nDCG-5 ------------|------------|------------|------------|--------|--------|------- NETL (U) | 2.66 | 2.59 | 2.50 | 0.83 | 0.85 | 0.87 NETL (S) | 2.74 | 2.57 | 2.49 | 0.88 | 0.85 | 0.88 BART-TL-all | 2.64 | 2.52 | 2.43 | 0.83 | 0.84 | 0.87 BART-TL-ng | 2.62 | 2.50 | 2.33 | 0.82 | 0.84 | 0.85 ### BibTeX entry and citation info ```bibtex @inproceedings{popa-rebedea-2021-bart, title = "{BART}-{TL}: Weakly-Supervised Topic Label Generation", author = "Popa, Cristian and Rebedea, Traian", booktitle = "Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume", month = apr, year = "2021", address = "Online", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/2021.eacl-main.121", pages = "1418--1425", abstract = "We propose a novel solution for assigning labels to topic models by using multiple weak labelers. The method leverages generative transformers to learn accurate representations of the most important topic terms and candidate labels. This is achieved by fine-tuning pre-trained BART models on a large number of potential labels generated by state of the art non-neural models for topic labeling, enriched with different techniques. The proposed BART-TL model is able to generate valuable and novel labels in a weakly-supervised manner and can be improved by adding other weak labelers or distant supervision on similar tasks.", } ```
{"language": ["en"], "license": "apache-2.0", "tags": ["topic labeling"], "metrics": ["ndcg"], "<!-- thumbnail": "https://raw.githubusercontent.com/JetRunner/BERT-of-Theseus/master/bert-of-theseus.png -->"}
text2text-generation
cristian-popa/bart-tl-ng
[ "transformers", "pytorch", "bart", "text2text-generation", "topic labeling", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bart #text2text-generation #topic labeling #en #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
MyModel ======= Model description ----------------- This is the 'BART-TL-ng' model from the paper BART-TL: Weakly-Supervised Topic Label Generation. We aim to solve the topic labeling task using generative methods, rather than selection from a pool of labels as was done in previous State of the Art works. For more details not covered here, you can read the paper or look at the open-source implementation: URL There are two models made available from the paper: * BART-TL-all * BART-TL-ng Intended uses & limitations --------------------------- #### How to use The model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model. #### Limitations and bias The model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy. Training data ------------- The model was fine-tuned on 5 different StackExchange corpora (see URL for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here. Training procedure ------------------ The large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the NETL method, along with n-grams from the topics. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the paper or by following this notebook. Eval results ------------ ### BibTeX entry and citation info
[ "#### How to use\n\n\nThe model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model.", "#### Limitations and bias\n\n\nThe model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy.\n\n\nTraining data\n-------------\n\n\nThe model was fine-tuned on 5 different StackExchange corpora (see URL for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here.\n\n\nTraining procedure\n------------------\n\n\nThe large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the NETL method, along with n-grams from the topics. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the paper or by following this notebook.\n\n\nEval results\n------------", "### BibTeX entry and citation info" ]
[ "TAGS\n#transformers #pytorch #bart #text2text-generation #topic labeling #en #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "#### How to use\n\n\nThe model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model.", "#### Limitations and bias\n\n\nThe model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy.\n\n\nTraining data\n-------------\n\n\nThe model was fine-tuned on 5 different StackExchange corpora (see URL for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here.\n\n\nTraining procedure\n------------------\n\n\nThe large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the NETL method, along with n-grams from the topics. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the paper or by following this notebook.\n\n\nEval results\n------------", "### BibTeX entry and citation info" ]
[ 52, 54, 215, 11 ]
[ "passage: TAGS\n#transformers #pytorch #bart #text2text-generation #topic labeling #en #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nThe model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model.#### Limitations and bias\n\n\nThe model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy.\n\n\nTraining data\n-------------\n\n\nThe model was fine-tuned on 5 different StackExchange corpora (see URL for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here.\n\n\nTraining procedure\n------------------\n\n\nThe large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the NETL method, along with n-grams from the topics. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the paper or by following this notebook.\n\n\nEval results\n------------### BibTeX entry and citation info" ]
[ -0.058342840522527695, 0.030573591589927673, 0.0009910708758980036, 0.029448406770825386, 0.09407609701156616, -0.006430593319237232, 0.06431687623262405, 0.09990748763084412, -0.020535092800855637, 0.04680928215384483, 0.009751266799867153, 0.00364040769636631, 0.06301063299179077, 0.19909220933914185, 0.07533171772956848, -0.24952134490013123, -0.02668389491736889, -0.031093338504433632, -0.021916141733527184, 0.09100772440433502, 0.09117823094129562, -0.0978337973356247, 0.06438127160072327, 0.0327000729739666, -0.13304725289344788, 0.006562351249158382, -0.01713912934064865, -0.07628340274095535, 0.08511122316122055, 0.06718479841947556, 0.09355984628200531, -0.023052550852298737, 0.056373778730630875, -0.15545515716075897, 0.023145727813243866, 0.07129095494747162, -0.01641656830906868, 0.06063896417617798, 0.0487065427005291, -0.020678333938121796, 0.17517170310020447, -0.125321164727211, 0.06931795924901962, 0.060781676322221756, -0.08715667575597763, -0.14160868525505066, -0.08093162626028061, 0.06786641478538513, 0.08159607648849487, 0.07530500739812851, -0.02406216226518154, -0.01739644818007946, -0.10106674581766129, 0.037807609885931015, 0.13328221440315247, -0.22689566016197205, -0.04279904440045357, 0.09441834688186646, 0.016648000106215477, 0.08622246980667114, -0.043103452771902084, 0.025955630466341972, 0.025359584018588066, 0.008822071366012096, 0.0627066045999527, 0.00790627021342516, 0.025081414729356766, -0.04929721727967262, -0.12087327986955643, -0.03505789861083031, 0.1847388595342636, 0.02671620436012745, -0.05503256618976593, -0.12186971306800842, -0.02728126011788845, 0.038787972182035446, -0.017138266935944557, -0.048493124544620514, 0.0019641402177512646, -0.0023294149432331324, 0.039050959050655365, -0.12985970079898834, -0.11857695132493973, -0.02382560819387436, -0.04059574753046036, 0.02126193605363369, -0.01126590371131897, 0.03685931861400604, -0.0706740990281105, 0.12561112642288208, -0.04233065992593765, -0.06131984665989876, 0.024106791242957115, -0.031216681003570557, -0.0751732662320137, -0.03797106444835663, -0.003976340871304274, -0.007467689923942089, -0.04018536955118179, 0.07222623378038406, -0.0018510435475036502, -0.031280551105737686, -0.07724938541650772, 0.020403580740094185, 0.08658487349748611, 0.04295308142900467, -0.08782842010259628, -0.049130238592624664, 0.035734597593545914, -0.016364876180887222, -0.020161747932434082, -0.006849395576864481, -0.055540211498737335, 0.028979478403925896, 0.002882142784073949, 0.037599075585603714, 0.011118223890662193, -0.026767058297991753, -0.033568646758794785, -0.08530446887016296, 0.1328819841146469, -0.09543154388666153, -0.012268774211406708, 0.0034273075871169567, -0.02129155769944191, 0.019330838695168495, 0.03403456509113312, -0.014767639338970184, -0.017458485439419746, 0.06998659670352936, -0.04462329298257828, -0.04089457914233208, -0.0851701870560646, -0.05928947776556015, 0.01444813422858715, -0.0024286224506795406, -0.043818362057209015, -0.0753307044506073, -0.24844898283481598, -0.05765001103281975, -0.0000597610414843075, -0.05186036601662636, -0.014315507374703884, 0.010513955727219582, -0.009551780298352242, -0.013130560517311096, -0.019312243908643723, 0.1245054081082344, -0.04913797974586487, 0.03272130712866783, -0.0880001038312912, 0.07720021158456802, 0.03168881684541702, 0.039653703570365906, -0.03528591990470886, 0.0019328133203089237, -0.1451042741537094, 0.13024908304214478, -0.07276961952447891, 0.02961386740207672, -0.07805461436510086, -0.06799259781837463, -0.06438018381595612, 0.005182104185223579, -0.01358987856656313, 0.09826919436454773, -0.23330287635326385, -0.07749452441930771, 0.23801572620868683, -0.11409841477870941, -0.03317215293645859, 0.1059221625328064, -0.08733032643795013, 0.16701965034008026, 0.11246955394744873, 0.15654504299163818, 0.11029225587844849, -0.027743080630898476, -0.013853859156370163, 0.004641373176127672, -0.012098517268896103, 0.04941832274198532, 0.0428425669670105, -0.0006375138182193041, -0.05939803645014763, 0.012979860417544842, -0.01135268248617649, 0.018503153696656227, -0.04998841881752014, -0.08357661962509155, 0.03001316450536251, -0.03557678312063217, 0.02019551768898964, -0.015422710217535496, 0.0384000800549984, 0.020128604024648666, -0.04887813329696655, 0.04966037720441818, 0.09271354228258133, -0.03092930279672146, -0.01780250109732151, -0.007291102781891823, 0.0730806514620781, -0.0596151165664196, -0.017882810905575752, -0.1714581847190857, -0.08131341636180878, 0.013912152498960495, -0.04811101406812668, 0.053962547332048416, 0.08562042564153671, -0.011072170920670033, 0.024481510743498802, -0.024122506380081177, 0.04766746237874031, -0.0510750487446785, -0.0223967544734478, -0.05786409601569176, -0.11757993698120117, -0.04671051725745201, -0.07669480890035629, 0.1559901088476181, -0.1263575255870819, 0.023070236667990685, 0.049314141273498535, 0.04606348276138306, 0.023175332695245743, -0.029774855822324753, 0.03438388183712959, 0.02312796376645565, 0.018023306503891945, -0.05012032762169838, 0.022587014362215996, 0.012333761900663376, -0.065131276845932, -0.0370250903069973, -0.0767601802945137, 0.006858821492642164, 0.08353833854198456, 0.07545866072177887, -0.09723395854234695, -0.08698149025440216, -0.08268161863088608, -0.00825247447937727, -0.10843393951654434, -0.019677050411701202, 0.16108928620815277, 0.0163166131824255, 0.07302290946245193, -0.08568139374256134, -0.0369659923017025, 0.01856652833521366, -0.001009804313071072, 0.005269660614430904, 0.034230709075927734, 0.06339166313409805, -0.15905506908893585, 0.05357871949672699, 0.06322236359119415, 0.06911996752023697, 0.2333313673734665, 0.046343844383955, -0.0878371149301529, -0.003461424959823489, -0.019974468275904655, -0.010641857981681824, 0.11016261577606201, -0.09302546083927155, -0.0017491488251835108, 0.041598811745643616, 0.008428750559687614, 0.07880750298500061, -0.10156016796827316, 0.0390683077275753, 0.04556549713015556, -0.06217455491423607, 0.013582726009190083, -0.052571557462215424, 0.000529274286236614, 0.11707046627998352, 0.06270211189985275, 0.04449890926480293, -0.013917402364313602, -0.038149669766426086, -0.11764880269765854, 0.19318106770515442, -0.08886654675006866, -0.2626831829547882, -0.09855068475008011, 0.08962839841842651, -0.0022388736251741648, 0.0025413562543690205, 0.019458815455436707, -0.0970102995634079, -0.0245984997600317, -0.07629316300153732, 0.008551460690796375, 0.03635755553841591, -0.033122628927230835, -0.021451212465763092, 0.005116978194564581, -0.02688535489141941, -0.10449887067079544, -0.03559839725494385, -0.06772641837596893, -0.10486956685781479, 0.005984237417578697, 0.005970361176878214, 0.06467245519161224, 0.13154171407222748, 0.005976049229502678, 0.0037174862809479237, -0.032312504947185516, 0.13807635009288788, -0.11079484224319458, 0.07803452014923096, 0.1077672690153122, 0.062137700617313385, 0.028223548084497452, 0.062307510524988174, 0.03146607056260109, -0.07076277583837509, 0.03591207414865494, 0.07118268311023712, -0.06969510018825531, -0.22140564024448395, -0.1311884969472885, -0.02737789787352085, -0.032960519194602966, 0.038947269320487976, 0.040127336978912354, 0.04888443276286125, 0.023387465626001358, -0.012813050299882889, -0.055598631501197815, 0.007564835250377655, 0.03651135414838791, 0.04009876027703285, -0.0466010719537735, 0.10182483494281769, -0.04065719619393349, -0.009073780849575996, 0.14718441665172577, -0.011366254650056362, 0.25477662682533264, -0.013088683597743511, 0.06522723287343979, 0.09513356536626816, 0.039059437811374664, 0.03197916969656944, 0.0722714215517044, -0.012408450245857239, 0.01142103224992752, -0.04238027334213257, -0.06244606897234917, -0.0186020415276289, 0.07575233280658722, -0.03216494247317314, -0.06200456619262695, -0.1116180568933487, -0.0196929182857275, 0.051307179033756256, 0.22882391512393951, 0.003063799813389778, -0.17597733438014984, -0.08621248602867126, 0.020625611767172813, -0.02446790598332882, -0.06691140681505203, 0.016353396698832512, 0.10100878775119781, -0.17418310046195984, 0.022296587005257607, -0.043108366429805756, 0.06633739173412323, -0.07022139430046082, 0.004406327847391367, 0.0379958301782608, -0.029929103329777718, -0.027290500700473785, 0.09984713792800903, -0.11808774620294571, 0.16601800918579102, 0.01602788455784321, 0.059623558074235916, -0.12900419533252716, 0.008938377723097801, 0.026027007028460503, -0.03340480849146843, 0.15321989357471466, 0.012870158068835735, 0.009208454750478268, -0.046813931316137314, -0.10588183999061584, 0.043973155319690704, 0.00836845114827156, -0.08309002965688705, 0.044083449989557266, 0.02996736206114292, 0.012083927169442177, 0.004998138640075922, -0.00795533787459135, -0.09857282787561417, -0.14314775168895721, 0.015880858525633812, -0.08027873933315277, -0.049369554966688156, -0.04961135610938072, -0.07697252929210663, 0.06869460642337799, 0.25060421228408813, -0.03784763440489769, -0.10615942627191544, -0.12849211692810059, 0.07380979508161545, 0.06786727905273438, -0.03141176328063011, 0.04798521474003792, 0.05887294188141823, 0.15076637268066406, -0.0087036844342947, -0.1267872303724289, 0.031166141852736473, -0.09814701229333878, -0.11663303524255753, -0.030915148556232452, 0.10102105140686035, 0.12958934903144836, 0.030847836285829544, 0.021512309089303017, -0.03598560020327568, -0.018173763528466225, -0.11852359771728516, 0.007726116105914116, 0.20454733073711395, 0.035390596836805344, 0.07373800128698349, -0.09757405519485474, -0.03905724361538887, -0.05015677958726883, -0.0028359373100101948, 0.1498568207025528, 0.16449537873268127, -0.09239063411951065, 0.17171844840049744, 0.14819326996803284, -0.11204651743173599, -0.1518813818693161, 0.0655488669872284, 0.0439140759408474, 0.05710531771183014, 0.014247988350689411, -0.23451128602027893, 0.1635744422674179, 0.12239357084035873, -0.030424607917666435, 0.012011608108878136, -0.28192970156669617, -0.13155917823314667, 0.10077385604381561, -0.00009603358193999156, 0.07728811353445053, -0.08408793061971664, -0.010091124102473259, -0.020169725641608238, -0.018051577731966972, 0.17749595642089844, -0.19531254470348358, 0.08424545079469681, -0.006354453042149544, 0.05830709636211395, 0.006056278012692928, -0.04561086744070053, 0.09640072286128998, 0.07856305688619614, 0.06964194029569626, -0.02574380300939083, 0.008902081288397312, 0.1144956573843956, -0.02633904479444027, 0.07172442972660065, 0.10801653563976288, 0.05097726732492447, -0.15742053091526031, -0.07774507999420166, -0.123231902718544, 0.08208120614290237, -0.043931037187576294, -0.09687887132167816, -0.08548503369092941, 0.11645499616861343, 0.09555312246084213, -0.01752432994544506, -0.18009839951992035, -0.06799182295799255, 0.07152710855007172, 0.047555893659591675, 0.1290031373500824, -0.07513950765132904, -0.0860796794295311, 0.014088650234043598, -0.020328694954514503, 0.062271848320961, -0.0745231956243515, 0.02116983011364937, 0.11253651976585388, 0.023655356839299202, 0.12333513051271439, 0.036072369664907455, -0.08212592452764511, 0.010394391603767872, 0.05205133929848671, -0.11706235259771347, -0.10832182317972183, -0.03595542162656784, -0.002475827233865857, -0.10464441031217575, -0.0663393959403038, 0.1027846708893776, -0.013562353327870369, -0.03259022906422615, -0.02210722491145134, 0.04942626506090164, -0.03529943898320198, 0.12882910668849945, 0.03960306569933891, 0.03345973789691925, -0.06276913732290268, 0.07921484857797623, 0.08270875364542007, -0.0966816246509552, 0.03750482201576233, 0.061507292091846466, -0.10566810518503189, -0.04214899614453316, -0.028298791497945786, 0.10918109863996506, -0.12837141752243042, -0.026688596233725548, -0.016610419377684593, -0.0786278247833252, 0.04312693700194359, 0.10079561173915863, 0.02844146080315113, 0.012124431319534779, -0.07948318123817444, -0.007384150288999081, -0.06485842913389206, 0.052358973771333694, 0.023885084316134453, -0.004932453390210867, -0.04161762073636055, 0.14737378060817719, 0.014695890247821808, 0.056870099157094955, -0.046302299946546555, -0.03702888637781143, -0.0625452846288681, 0.00039691445999778807, -0.10317511856555939, -0.00512924138456583, -0.10965092480182648, -0.009945610538125038, -0.025679145008325577, 0.04858749732375145, 0.01419808529317379, 0.0005101314163766801, -0.05660496652126312, -0.011668460443615913, -0.045271776616573334, 0.05150192603468895, -0.13089726865291595, 0.017271079123020172, 0.020816335454583168, -0.03668724000453949, 0.11176998913288116, 0.005618351045995951, 0.04666804149746895, -0.0575108602643013, -0.13081584870815277, 0.0058205099776387215, 0.044750042259693146, 0.043321214616298676, 0.028192181140184402, -0.11851637065410614, 0.036951951682567596, 0.019684063270688057, -0.039027679711580276, 0.007913587614893913, 0.01619609259068966, -0.09821455180644989, 0.07380329072475433, -0.013126696459949017, 0.014196646399796009, -0.07796042412519455, 0.04416633024811745, 0.06239810958504677, 0.04790641367435455, 0.02438671700656414, -0.06155340373516083, 0.059636130928993225, -0.14688067138195038, -0.046048544347286224, -0.03676383197307587, 0.026272805407643318, 0.057479340583086014, -0.056968722492456436, 0.057352688163518906, 0.00999210774898529, 0.24028001725673676, 0.08102734386920929, 0.028194056823849678, 0.008176000788807869, -0.14922186732292175, 0.08177367597818375, -0.011772839352488518, 0.08119329065084457, 0.019536703824996948, -0.01926739513874054, 0.08310384303331375, 0.017610007897019386, -0.011455736123025417, 0.11950110644102097, 0.2567911744117737, 0.1210012212395668, 0.03448742628097534, -0.03109383024275303, 0.046119574457407, -0.10541165620088577, -0.11118148267269135, -0.02968544140458107, -0.03591173514723778, 0.08167903125286102, -0.07698766887187958, -0.03791169077157974, 0.08438604325056076, -0.19658175110816956, 0.0854947566986084, 0.04879022389650345, -0.08471024036407471, -0.10416188836097717, -0.09216365218162537, -0.026333481073379517, -0.013496228493750095, -0.0101845758035779, -0.15818408131599426, -0.00426621874794364, 0.06318286061286926, 0.02077604830265045, -0.04341045394539833, 0.15926404297351837, -0.07252433896064758, -0.051497478038072586, 0.10409259051084518, 0.02818872220814228, 0.013267066329717636, -0.11635036021471024, -0.004668618086725473, -0.008388787508010864, 0.056736528873443604, 0.021655913442373276, 0.030244680121541023, -0.010239290073513985, 0.021173642948269844, 0.01835465431213379, -0.023499276489019394, -0.030948497354984283, -0.02695849910378456, 0.06716864556074142, 0.14122924208641052, 0.07459448277950287, -0.08119167387485504, -0.009529965929687023, 0.1708410233259201, -0.031712837517261505, -0.02768450416624546, -0.12802790105342865, 0.17781271040439606, -0.001604619319550693, -0.03558432683348656, -0.0112056415528059, -0.08406542986631393, 0.04428919777274132, 0.1900353878736496, 0.2325240671634674, -0.027795586735010147, -0.004566521849483252, 0.0027429896872490644, -0.008089184761047363, 0.005748538766056299, 0.12975434958934784, 0.05913783982396126, 0.29923075437545776, -0.09908019006252289, 0.03931696340441704, -0.0071924906224012375, -0.06713194400072098, -0.056258562952280045, 0.0674978569149971, 0.03791152313351631, 0.01761043816804886, -0.016676906496286392, 0.09435518085956573, -0.047919366508722305, -0.18710947036743164, 0.03936079517006874, -0.09078465402126312, -0.12131673097610474, -0.010458340868353844, -0.11833801120519638, -0.015906330198049545, 0.053012073040008545, -0.05400494858622551, -0.020473122596740723, 0.11533978581428528, 0.017791088670492172, -0.07696043699979782, -0.07084628194570541, 0.10521527379751205, -0.03689102083444595, 0.22013181447982788, -0.013028360903263092, -0.010242043994367123, 0.08957763016223907, 0.008665123023092747, -0.11609933525323868, -0.012756678275763988, 0.010921875014901161, -0.0012333891354501247, 0.021401172503829002, 0.11248355358839035, -0.014983473345637321, 0.03983921930193901, 0.05257648602128029, -0.0973421111702919, 0.024479549378156662, -0.04133936017751694, 0.022901717573404312, -0.11432449519634247, 0.017058394849300385, -0.08986000716686249, 0.1699211299419403, 0.17144270241260529, -0.039137180894613266, 0.022438153624534607, -0.06636509299278259, 0.0494748055934906, 0.032057806849479675, 0.1347661018371582, -0.03661283105611801, -0.12832269072532654, 0.02062162570655346, -0.018944986164569855, 0.014135508798062801, -0.2570283114910126, -0.08721671998500824, 0.06243297457695007, -0.036938123404979706, -0.010299976915121078, 0.08259129524230957, 0.018214790150523186, 0.04814095050096512, -0.026854095980525017, 0.011055383831262589, -0.03250135853886604, 0.09736742824316025, -0.1514401137828827, -0.05418457090854645 ]
null
null
null
### Preprocessing 1. Normalisation and tokenisation with moses scripts 2. truecased with model docgWP.tcmodel.[LAN] and moses scripts 3. bped with model model.caesen40k.bpe and subword-nmt - Note: no prepended tag for multilinguality ### Training Data 1. Bilingual es-ca: DOGC, Wikimatrix, OpenSubtitles, JW300, GlobalVoices * Bilingual es-ca: Translations using systems trained with 1. of Oscar and Wikipedia 2. Bilingual es-en, ca-en: United Nations, Europarl, Wikimatrix, OpenSubtitles, JW300 * Bilingual es-en, ca-en: Translations using systems trained with 1. of the missing pairs - Final training data size for the ca/es-en: 44M parallel sentences - Finetuned with 1.5M real parallel data (without backtranslations) ### Model Transformer big with guided alignments. Relevant parameters: --beam-size 6 --normalize 0.6 --enc-depth 6 --dec-depth 6 --transformer-heads 8 --transformer-preprocess n --transformer-postprocess da --transformer-dropout 0.1 --label-smoothing 0.1 --dim-emb 1024 --transformer-dim-ffn 4096 --transformer-dropout-attention 0.1 --transformer-dropout-ffn 0.1 --learn-rate 0.00015 --lr-warmup 8000 --lr-decay-inv-sqrt 8000 --optimizer-params 0.9 0.998 1e-09 --clip-norm 5 --tied-embeddings --exponential-smoothing --transformer-guided-alignment-layer 1 --guided-alignment-cost mse --guided-alignment-weight 0.1 ## Evaluation ### Test set https://github.com/PLXIV/Gebiotoolkit/tree/master/gebiocorpus_v2 ### ca2en BLEU|#:1|bs:1000|rs:12345|c:mixed|e:no|tok:13a|s:exp|v:2.0.0 = 47.8 (μ = 47.8 ± 0.9) chrF|#:1|bs:1000|rs:12345|c:mixed|e:yes|nc:6|nw:0|s:no|v:2.0.0 = 69.9 (μ = 69.9 ± 0.7) ### es2en BLEU|#:1|bs:1000|rs:12345|c:mixed|e:no|tok:13a|s:exp|v:2.0.0 = 48.9 (μ = 48.9 ± 0.9) chrF2|#:1|bs:1000|rs:12345|c:mixed|e:yes|nc:6|nw:0|s:no|v:2.0.0 = 70.5 (μ = 70.5 ± 0.7)
{"language": ["ca", "es", "en"], "tags": ["translation"]}
translation
cristinae/marian_caes2en
[ "translation", "ca", "es", "en", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "ca", "es", "en" ]
TAGS #translation #ca #es #en #region-us
### Preprocessing 1. Normalisation and tokenisation with moses scripts 2. truecased with model docgWP.tcmodel.[LAN] and moses scripts 3. bped with model URL and subword-nmt - Note: no prepended tag for multilinguality ### Training Data 1. Bilingual es-ca: DOGC, Wikimatrix, OpenSubtitles, JW300, GlobalVoices * Bilingual es-ca: Translations using systems trained with 1. of Oscar and Wikipedia 2. Bilingual es-en, ca-en: United Nations, Europarl, Wikimatrix, OpenSubtitles, JW300 * Bilingual es-en, ca-en: Translations using systems trained with 1. of the missing pairs - Final training data size for the ca/es-en: 44M parallel sentences - Finetuned with 1.5M real parallel data (without backtranslations) ### Model Transformer big with guided alignments. Relevant parameters: --beam-size 6 --normalize 0.6 --enc-depth 6 --dec-depth 6 --transformer-heads 8 --transformer-preprocess n --transformer-postprocess da --transformer-dropout 0.1 --label-smoothing 0.1 --dim-emb 1024 --transformer-dim-ffn 4096 --transformer-dropout-attention 0.1 --transformer-dropout-ffn 0.1 --learn-rate 0.00015 --lr-warmup 8000 --lr-decay-inv-sqrt 8000 --optimizer-params 0.9 0.998 1e-09 --clip-norm 5 --tied-embeddings --exponential-smoothing --transformer-guided-alignment-layer 1 --guided-alignment-cost mse --guided-alignment-weight 0.1 ## Evaluation ### Test set URL ### ca2en BLEU|#:1|bs:1000|rs:12345|c:mixed|e:no|tok:13a|s:exp|v:2.0.0 = 47.8 (μ = 47.8 ± 0.9) chrF|#:1|bs:1000|rs:12345|c:mixed|e:yes|nc:6|nw:0|s:no|v:2.0.0 = 69.9 (μ = 69.9 ± 0.7) ### es2en BLEU|#:1|bs:1000|rs:12345|c:mixed|e:no|tok:13a|s:exp|v:2.0.0 = 48.9 (μ = 48.9 ± 0.9) chrF2|#:1|bs:1000|rs:12345|c:mixed|e:yes|nc:6|nw:0|s:no|v:2.0.0 = 70.5 (μ = 70.5 ± 0.7)
[ "### Preprocessing\n1. Normalisation and tokenisation with moses scripts\n2. truecased with model docgWP.tcmodel.[LAN] and moses scripts\n3. bped with model URL and subword-nmt\n- Note: no prepended tag for multilinguality", "### Training Data\n1. Bilingual es-ca: DOGC, Wikimatrix, OpenSubtitles, JW300, GlobalVoices\n* Bilingual es-ca: Translations using systems trained with 1. of Oscar and Wikipedia\n2. Bilingual es-en, ca-en: United Nations, Europarl, Wikimatrix, OpenSubtitles, JW300\n* Bilingual es-en, ca-en: Translations using systems trained with 1. of the missing pairs\n\n- Final training data size for the ca/es-en: 44M parallel sentences\n- Finetuned with 1.5M real parallel data (without backtranslations)", "### Model\nTransformer big with guided alignments. Relevant parameters:\n\n--beam-size 6 \n\n--normalize 0.6 \n\n--enc-depth 6 --dec-depth 6 --transformer-heads 8\n\n--transformer-preprocess n --transformer-postprocess da \n\n--transformer-dropout 0.1 \n\n--label-smoothing 0.1 \n\n--dim-emb 1024 --transformer-dim-ffn 4096 \n\n--transformer-dropout-attention 0.1 \n\n--transformer-dropout-ffn 0.1 \n\n--learn-rate 0.00015 --lr-warmup 8000 --lr-decay-inv-sqrt 8000 \n\n--optimizer-params 0.9 0.998 1e-09 \n\n--clip-norm 5 \n\n--tied-embeddings \n\n--exponential-smoothing \n\n--transformer-guided-alignment-layer 1 --guided-alignment-cost mse --guided-alignment-weight 0.1", "## Evaluation", "### Test set\n\nURL", "### ca2en\n BLEU|#:1|bs:1000|rs:12345|c:mixed|e:no|tok:13a|s:exp|v:2.0.0 = 47.8 (μ = 47.8 ± 0.9)\n\n chrF|#:1|bs:1000|rs:12345|c:mixed|e:yes|nc:6|nw:0|s:no|v:2.0.0 = 69.9 (μ = 69.9 ± 0.7)", "### es2en\nBLEU|#:1|bs:1000|rs:12345|c:mixed|e:no|tok:13a|s:exp|v:2.0.0 = 48.9 (μ = 48.9 ± 0.9) \n\nchrF2|#:1|bs:1000|rs:12345|c:mixed|e:yes|nc:6|nw:0|s:no|v:2.0.0 = 70.5 (μ = 70.5 ± 0.7)" ]
[ "TAGS\n#translation #ca #es #en #region-us \n", "### Preprocessing\n1. Normalisation and tokenisation with moses scripts\n2. truecased with model docgWP.tcmodel.[LAN] and moses scripts\n3. bped with model URL and subword-nmt\n- Note: no prepended tag for multilinguality", "### Training Data\n1. Bilingual es-ca: DOGC, Wikimatrix, OpenSubtitles, JW300, GlobalVoices\n* Bilingual es-ca: Translations using systems trained with 1. of Oscar and Wikipedia\n2. Bilingual es-en, ca-en: United Nations, Europarl, Wikimatrix, OpenSubtitles, JW300\n* Bilingual es-en, ca-en: Translations using systems trained with 1. of the missing pairs\n\n- Final training data size for the ca/es-en: 44M parallel sentences\n- Finetuned with 1.5M real parallel data (without backtranslations)", "### Model\nTransformer big with guided alignments. Relevant parameters:\n\n--beam-size 6 \n\n--normalize 0.6 \n\n--enc-depth 6 --dec-depth 6 --transformer-heads 8\n\n--transformer-preprocess n --transformer-postprocess da \n\n--transformer-dropout 0.1 \n\n--label-smoothing 0.1 \n\n--dim-emb 1024 --transformer-dim-ffn 4096 \n\n--transformer-dropout-attention 0.1 \n\n--transformer-dropout-ffn 0.1 \n\n--learn-rate 0.00015 --lr-warmup 8000 --lr-decay-inv-sqrt 8000 \n\n--optimizer-params 0.9 0.998 1e-09 \n\n--clip-norm 5 \n\n--tied-embeddings \n\n--exponential-smoothing \n\n--transformer-guided-alignment-layer 1 --guided-alignment-cost mse --guided-alignment-weight 0.1", "## Evaluation", "### Test set\n\nURL", "### ca2en\n BLEU|#:1|bs:1000|rs:12345|c:mixed|e:no|tok:13a|s:exp|v:2.0.0 = 47.8 (μ = 47.8 ± 0.9)\n\n chrF|#:1|bs:1000|rs:12345|c:mixed|e:yes|nc:6|nw:0|s:no|v:2.0.0 = 69.9 (μ = 69.9 ± 0.7)", "### es2en\nBLEU|#:1|bs:1000|rs:12345|c:mixed|e:no|tok:13a|s:exp|v:2.0.0 = 48.9 (μ = 48.9 ± 0.9) \n\nchrF2|#:1|bs:1000|rs:12345|c:mixed|e:yes|nc:6|nw:0|s:no|v:2.0.0 = 70.5 (μ = 70.5 ± 0.7)" ]
[ 15, 64, 141, 210, 3, 5, 105, 106 ]
[ "passage: TAGS\n#translation #ca #es #en #region-us \n### Preprocessing\n1. Normalisation and tokenisation with moses scripts\n2. truecased with model docgWP.tcmodel.[LAN] and moses scripts\n3. bped with model URL and subword-nmt\n- Note: no prepended tag for multilinguality### Training Data\n1. Bilingual es-ca: DOGC, Wikimatrix, OpenSubtitles, JW300, GlobalVoices\n* Bilingual es-ca: Translations using systems trained with 1. of Oscar and Wikipedia\n2. Bilingual es-en, ca-en: United Nations, Europarl, Wikimatrix, OpenSubtitles, JW300\n* Bilingual es-en, ca-en: Translations using systems trained with 1. of the missing pairs\n\n- Final training data size for the ca/es-en: 44M parallel sentences\n- Finetuned with 1.5M real parallel data (without backtranslations)### Model\nTransformer big with guided alignments. Relevant parameters:\n\n--beam-size 6 \n\n--normalize 0.6 \n\n--enc-depth 6 --dec-depth 6 --transformer-heads 8\n\n--transformer-preprocess n --transformer-postprocess da \n\n--transformer-dropout 0.1 \n\n--label-smoothing 0.1 \n\n--dim-emb 1024 --transformer-dim-ffn 4096 \n\n--transformer-dropout-attention 0.1 \n\n--transformer-dropout-ffn 0.1 \n\n--learn-rate 0.00015 --lr-warmup 8000 --lr-decay-inv-sqrt 8000 \n\n--optimizer-params 0.9 0.998 1e-09 \n\n--clip-norm 5 \n\n--tied-embeddings \n\n--exponential-smoothing \n\n--transformer-guided-alignment-layer 1 --guided-alignment-cost mse --guided-alignment-weight 0.1## Evaluation### Test set\n\nURL" ]
[ -0.04639185965061188, 0.06822874397039413, -0.007070061285048723, 0.03232867270708084, 0.07837794721126556, 0.017189905047416687, 0.0353710874915123, 0.12522460520267487, 0.01575244963169098, 0.10191977769136429, 0.03591351956129074, 0.018606865778565407, 0.1128823533654213, 0.12153605371713638, 0.013359157368540764, -0.2710344195365906, 0.06310766190290451, -0.06216515600681305, -0.06520454585552216, 0.06705158203840256, 0.10712601989507675, -0.0395958386361599, 0.05069415271282196, -0.025239448994398117, -0.04079374298453331, 0.021172652021050453, -0.0400947667658329, -0.06287755072116852, 0.05354087054729462, 0.07959359884262085, 0.03172596916556358, 0.022446129471063614, 0.02131400629878044, -0.22600696980953217, 0.011241520754992962, 0.042657461017370224, 0.006942070554941893, 0.02728775143623352, 0.10687469691038132, 0.0010640494292601943, 0.04743005707859993, -0.17806628346443176, 0.019162043929100037, 0.06755266338586807, -0.06286301463842392, -0.22004581987857819, -0.1080203577876091, 0.03747222572565079, 0.06874976307153702, 0.05650030076503754, -0.03354746848344803, 0.06365031003952026, -0.06894998252391815, 0.05784019082784653, 0.11335919052362442, -0.27183791995048523, -0.03031180426478386, 0.01922469586133957, 0.040880993008613586, -0.017878007143735886, -0.0632321685552597, -0.005289653316140175, -0.01858511194586754, 0.04194530099630356, -0.01938818395137787, -0.026658762246370316, 0.060578301548957825, -0.06067555770277977, -0.09228765219449997, 0.033217672258615494, 0.09958916902542114, 0.04923150688409805, -0.07033899426460266, -0.1647229939699173, -0.06320804357528687, -0.028986506164073944, -0.04386555030941963, -0.03364621847867966, 0.06636550277471542, -0.018411889672279358, 0.11089494824409485, -0.04718326777219772, -0.05886564403772354, 0.02143622748553753, -0.12635909020900726, 0.16353711485862732, 0.05901259928941727, -0.0019053351134061813, 0.010659059509634972, 0.03818389028310776, -0.03807791322469711, -0.17463012039661407, -0.055256761610507965, -0.03176164999604225, -0.08499270677566528, -0.007302843499928713, 0.017614144831895828, -0.01849312148988247, 0.06683015078306198, 0.1659615933895111, -0.06529611349105835, 0.11101874709129333, -0.08482275903224945, 0.015052583068609238, -0.01007780060172081, 0.16671012341976166, -0.12164173275232315, -0.0951574519276619, 0.042720552533864975, 0.003163771703839302, 0.047682587057352066, 0.023063823580741882, -0.07979937642812729, -0.04974904656410217, -0.01444453839212656, 0.08952423185110092, 0.033178817480802536, 0.022948849946260452, -0.047656863927841187, -0.04308939352631569, 0.16200725734233856, -0.1644197702407837, 0.027050958946347237, 0.029321976006031036, -0.05865006893873215, 0.1821635365486145, 0.011805066838860512, -0.04164494574069977, -0.10336652398109436, 0.06723595410585403, -0.017766881734132767, 0.027719711884856224, -0.05254121869802475, -0.07045333087444305, 0.06534931808710098, 0.00043338563409633934, -0.04416409879922867, -0.07438936084508896, -0.07404346764087677, -0.05469398573040962, 0.01315934956073761, -0.08856164664030075, 0.00848545040935278, -0.029048465192317963, -0.06383585184812546, 0.05466829612851143, -0.005691868253052235, -0.013578740879893303, -0.06212864816188812, -0.004308071453124285, 0.007904532365500927, 0.08233530819416046, 0.018635479733347893, 0.011038310825824738, -0.044358670711517334, 0.042708758264780045, -0.2245938926935196, 0.1344585120677948, -0.07774863392114639, 0.008305992931127548, -0.1496967077255249, -0.04850482940673828, 0.01364655140787363, 0.04089697450399399, 0.03806710243225098, 0.14934135973453522, -0.24118946492671967, -0.02483285963535309, 0.2545061707496643, -0.05315317213535309, -0.04701533913612366, 0.13841156661510468, 0.022154534235596657, -0.028323858976364136, 0.049996696412563324, 0.11592195183038712, 0.04212046414613724, -0.0637047067284584, 0.00785769708454609, -0.03375551849603653, -0.001819354249164462, 0.0779011994600296, 0.08893938362598419, -0.07782695442438126, 0.10316482931375504, 0.026484515517950058, -0.0014285545330494642, 0.011074979789555073, -0.009979091584682465, -0.03213166818022728, 0.009121894836425781, 0.012440497986972332, 0.01280944887548685, -0.052863746881484985, 0.02090325951576233, -0.08924135565757751, -0.11441967636346817, -0.007243650499731302, 0.07310321182012558, -0.007213204633444548, 0.07064800709486008, -0.1108989343047142, -0.0362754724919796, 0.005381281487643719, 0.045624732971191406, -0.11133690178394318, -0.09177358448505402, 0.02431611530482769, -0.039049576967954636, 0.061058685183525085, -0.055606573820114136, 0.05086323991417885, 0.016896486282348633, -0.059460967779159546, -0.027949152514338493, 0.05310213193297386, -0.029746130108833313, -0.0038439983036369085, -0.13541734218597412, 0.013546884059906006, -0.025157151743769646, 0.0052035474218428135, -0.09699995070695877, 0.00039467663737013936, 0.03939451649785042, 0.15092512965202332, 0.011298760771751404, 0.005013188812881708, 0.023426935076713562, 0.01029860507696867, 0.02532286010682583, -0.06358210742473602, -0.0013090592110529542, -0.03885684534907341, -0.11817290633916855, 0.1222761869430542, -0.11185025423765182, 0.06498368829488754, 0.10352161526679993, 0.05254485458135605, -0.08397696912288666, 0.01626729592680931, 0.01150272786617279, -0.04367154836654663, 0.0035871670115739107, -0.026421988382935524, 0.11556308716535568, 0.06392046809196472, 0.11092688888311386, -0.07084035873413086, -0.030065827071666718, -0.024254705756902695, -0.03164233639836311, -0.06591867655515671, 0.12840797007083893, -0.06746010482311249, -0.19760456681251526, 0.07719383388757706, 0.1252610683441162, 0.005707701668143272, 0.15914082527160645, -0.03771594539284706, -0.05188274383544922, -0.06665240228176117, 0.036699771881103516, 0.02627362497150898, 0.016712907701730728, 0.0015298951184377074, 0.04600916802883148, 0.051718227565288544, 0.073259137570858, 0.03491372987627983, -0.06351708620786667, 0.0344332791864872, 0.03315161541104317, -0.07083142548799515, 0.03662041574716568, 0.03551597148180008, 0.028402822092175484, 0.09900988638401031, 0.06078985705971718, 0.012843489646911621, 0.007349574938416481, -0.050027865916490555, -0.07008806616067886, 0.17416246235370636, -0.17362001538276672, -0.1776224672794342, -0.1575307995080948, -0.05392633005976677, -0.0865810289978981, -0.020117968320846558, 0.0252753384411335, -0.08898710459470749, -0.10387516021728516, -0.08821197599172592, 0.09422486275434494, -0.010246383026242256, -0.02997087500989437, 0.01200594287365675, 0.01873069629073143, 0.010727182030677795, -0.10598654299974442, -0.028752947226166725, 0.010553307831287384, -0.08121258020401001, -0.00680981669574976, 0.005313746631145477, 0.01748739369213581, 0.08683431893587112, 0.03911207243800163, -0.021997496485710144, 0.0016765149775892496, 0.10791300237178802, -0.09760098159313202, 0.07878820598125458, 0.18189971148967743, -0.004283179063349962, 0.05131930857896805, 0.16738128662109375, 0.05273696780204773, -0.05756501108407974, -0.01489739678800106, 0.03871419280767441, -0.009077467955648899, -0.2729412317276001, -0.07973740249872208, -0.0494382381439209, -0.018055347725749016, 0.0058037517592310905, 0.06360701471567154, -0.06616814434528351, 0.01126775424927473, -0.07472560554742813, -0.06944098323583603, 0.046184174716472626, 0.06517188996076584, 0.1654781699180603, 0.039972756057977676, 0.0887419804930687, -0.07085038721561432, 0.0020394122693687677, 0.07242672890424728, 0.03506438434123993, 0.13311345875263214, 0.05583595857024193, 0.21598698198795319, 0.05462003871798515, -0.002020656829699874, 0.032999005168676376, 0.031679630279541016, 0.029627306386828423, -0.02601291425526142, 0.0160685908049345, -0.07261408120393753, -0.012425726279616356, 0.021038001403212547, 0.03569018095731735, 0.004229366313666105, -0.030802834779024124, 0.028268540278077126, 0.091445192694664, 0.22053954005241394, -0.0006523411138914526, -0.1646880805492401, -0.03230709955096245, 0.04858832806348801, -0.07517330348491669, -0.0581229105591774, 0.023464921861886978, 0.05428869649767876, -0.10710173845291138, 0.11756806820631027, 0.009435263462364674, 0.07514332979917526, -0.12413250654935837, -0.0010382720502093434, 0.021277492865920067, 0.12853173911571503, 0.020455345511436462, 0.035544537007808685, -0.2025926560163498, 0.12004154920578003, 0.02921088971197605, 0.05598365515470505, -0.048436086624860764, 0.06675048917531967, 0.0039522950537502766, -0.01724141836166382, 0.1374511867761612, -0.00795641727745533, -0.12056124955415726, -0.06670627743005753, -0.0548182837665081, -0.01538938470184803, 0.13587656617164612, -0.013185619376599789, 0.08724360167980194, -0.002482473384588957, -0.04468294233083725, -0.032926276326179504, -0.05323740467429161, -0.1264563500881195, -0.16243673861026764, 0.05874455347657204, -0.0011774058220908046, -0.013833187520503998, -0.06193224713206291, -0.032299213111400604, -0.13166356086730957, 0.22738789021968842, -0.16298775374889374, -0.06960590183734894, -0.1218961626291275, -0.042365871369838715, 0.17107398808002472, -0.10358306020498276, 0.06980295479297638, -0.043083555996418, 0.06386744976043701, 0.013165170326828957, -0.08150216937065125, 0.1292172521352768, -0.08257307857275009, -0.10489518940448761, -0.0069463979452848434, 0.20545533299446106, -0.003036011243239045, 0.03949975222349167, -0.003304594662040472, 0.0046327440068125725, 0.05688925087451935, -0.10593055933713913, 0.012195144779980183, 0.13090161979198456, 0.03415803611278534, 0.08402850478887558, -0.09108282625675201, -0.07666175067424774, -0.023379312828183174, -0.04646100848913193, 0.1836387813091278, 0.26713231205940247, -0.04211205244064331, 0.08587144315242767, 0.07745010405778885, -0.07747635990381241, -0.2127947360277176, 0.0023958953097462654, 0.06203586235642433, 0.043825604021549225, -0.04129290208220482, -0.19014669954776764, 0.09013767540454865, 0.11196570098400116, 0.012407327070832253, 0.09436100721359253, -0.2809896171092987, -0.14772871136665344, 0.053411439061164856, -0.0021632390562444925, 0.04283541813492775, -0.11455567181110382, -0.07282695174217224, -0.07615899294614792, -0.16871227324008942, 0.10550092160701752, -0.028728656470775604, 0.08003789186477661, 0.004764099605381489, 0.013808834366500378, 0.01590392179787159, -0.01676305942237377, 0.18276360630989075, 0.002037837402895093, 0.011778688058257103, -0.07148563861846924, 0.045439232140779495, 0.02926402911543846, -0.03296862915158272, 0.0757802352309227, 0.03017452359199524, -0.03771968185901642, -0.08863142132759094, -0.039480626583099365, -0.08237436413764954, -0.012455316260457039, -0.04206642508506775, 0.008953801356256008, -0.011939088813960552, 0.08720453083515167, 0.06787383556365967, -0.03459601104259491, 0.09635087102651596, -0.0951453223824501, -0.047637537121772766, 0.09736067056655884, 0.161001518368721, -0.023845698684453964, -0.023642482236027718, 0.0186733640730381, -0.025567540898919106, 0.019728463143110275, -0.09380586445331573, 0.07443259656429291, 0.10298959910869598, -0.03421005234122276, 0.06371613591909409, 0.01692577451467514, -0.09004848450422287, -0.032879967242479324, 0.09623879194259644, -0.028621509671211243, -0.22651731967926025, 0.018876666203141212, 0.0232190303504467, -0.06038730964064598, -0.012972825206816196, 0.15059585869312286, 0.05441631004214287, -0.03482869639992714, 0.040401216596364975, 0.06694716960191727, -0.021507572382688522, 0.14515618979930878, 0.015424409881234169, 0.03823300451040268, -0.11480613052845001, 0.16115431487560272, 0.03245433047413826, -0.1058218777179718, 0.024618178606033325, 0.22683563828468323, -0.0737217590212822, -0.060251615941524506, -0.02146177552640438, 0.05717041343450546, -0.018971050158143044, -0.045443180948495865, -0.0022556050680577755, -0.1727001816034317, 0.02049834653735161, 0.04218333587050438, 0.04519595205783844, -0.019048204645514488, 0.0184429120272398, -0.01394016481935978, -0.028069553896784782, 0.06846868246793747, 0.0605490580201149, 0.055085670202970505, -0.06599479168653488, 0.0308831874281168, -0.055984169244766235, 0.03623070567846298, -0.004360000602900982, 0.020621750503778458, -0.10623106360435486, 0.0016430328832939267, -0.140842005610466, 0.014025258831679821, -0.10032504051923752, -0.019199810922145844, -0.01856118254363537, 0.001959915040060878, -0.03562274947762489, -0.0007558366050943732, -0.082035131752491, -0.07401154190301895, -0.07235761731863022, 0.05996303632855415, -0.0929722934961319, -0.004395999014377594, 0.01919666677713394, -0.08745104819536209, 0.12022630870342255, 0.00755872530862689, 0.00615983922034502, 0.014736875891685486, -0.11242226511240005, -0.03990356996655464, 0.0055046542547643185, 0.007024066057056189, 0.042377032339572906, -0.08858808875083923, 0.038703255355358124, 0.0045117647387087345, -0.03400271013379097, -0.0003834737872239202, 0.02569626085460186, -0.11025117337703705, -0.012865918688476086, -0.07645029574632645, -0.003620545379817486, -0.0683392807841301, 0.030125059187412262, 0.12003545463085175, 0.026038048788905144, 0.10362199693918228, -0.08086249977350235, 0.06800686568021774, -0.1076691523194313, -0.0028189695440232754, -0.0005942368297837675, -0.036735616624355316, 0.12116257101297379, -0.010917754843831062, 0.08687847852706909, -0.020137762650847435, 0.09122630953788757, -0.08994853496551514, 0.05573754757642746, 0.06152684614062309, -0.07574345916509628, -0.03411426022648811, 0.07099106907844543, 0.09522873908281326, 0.018149953335523605, -0.020199626684188843, 0.01681886799633503, -0.0292719304561615, 0.0011611760128289461, 0.015608781017363071, 0.10803882777690887, 0.208360493183136, 0.08701474964618683, 0.1041194424033165, 0.017838116735219955, -0.08681899309158325, -0.05168315768241882, 0.08111163228750229, -0.049091894179582596, 0.06075522303581238, -0.014675106853246689, 0.043721552938222885, 0.09260226041078568, -0.14796395599842072, 0.08810802549123764, 0.003903869306668639, -0.05267643555998802, -0.10830242186784744, -0.15784315764904022, -0.08952310681343079, -0.03862658143043518, 0.008218671195209026, -0.09902508556842804, 0.06374038755893707, 0.07660307735204697, 0.10045521706342697, -0.016208117827773094, 0.12963950634002686, -0.08378942310810089, -0.10027429461479187, 0.13023057579994202, 0.033973731100559235, 0.023934273049235344, 0.1179545521736145, -0.030150070786476135, 0.003214652882888913, 0.04025633633136749, 0.047405850142240524, 0.09459249675273895, -0.017889853566884995, 0.012914943508803844, -0.10290756076574326, -0.08166252821683884, 0.010950771160423756, 0.04906008765101433, 0.012747756205499172, 0.15441620349884033, 0.041457291692495346, -0.05483483895659447, -0.015629692003130913, 0.15782183408737183, -0.04404256120324135, -0.11880659312009811, -0.11835665255784988, 0.06865295767784119, 0.05686162784695625, 0.06606364995241165, -0.03755431994795799, -0.09462384134531021, -0.0319533608853817, 0.13862335681915283, 0.18152275681495667, -0.06161707267165184, 0.0017546683084219694, 0.0348035991191864, 0.002964004408568144, 0.0087481290102005, 0.06650248914957047, 0.10184389352798462, 0.1602047085762024, -0.037884268909692764, 0.04073890298604965, -0.013865995220839977, -0.054295655339956284, -0.1160353422164917, 0.12561452388763428, -0.006054403726011515, 0.032812800258398056, -0.026058262214064598, 0.07730580866336823, -0.07122334837913513, -0.08888483047485352, 0.037139762192964554, -0.09725192934274673, -0.14283247292041779, -0.021186497062444687, 0.022208301350474358, 0.029586942866444588, 0.039653971791267395, 0.0439237505197525, -0.07034078985452652, 0.13607457280158997, 0.013954893685877323, -0.06788217276334763, -0.052994996309280396, 0.00626909825950861, -0.033718738704919815, 0.18290847539901733, 0.030859310179948807, 0.07858139276504517, 0.11769293248653412, 0.04740583524107933, -0.14530988037586212, 0.03829551860690117, 0.030913660302758217, -0.0866454541683197, 0.06468751281499863, 0.14090277254581451, -0.023563247174024582, 0.03498158976435661, 0.08236900717020035, -0.0776064470410347, 0.014317837543785572, 0.005403186194598675, -0.04980172961950302, -0.08022542297840118, 0.07462454587221146, -0.11007649451494217, 0.10480381548404694, 0.20232516527175903, -0.021904218941926956, 0.007218359503895044, -0.027249017730355263, 0.02488136664032936, 0.0019813315011560917, 0.0944223403930664, -0.04390599951148033, -0.15712596476078033, 0.0566747672855854, 0.022116370499134064, 0.09743818640708923, -0.1326105147600174, -0.07895626872777939, -0.0006316130748018622, 0.007773600984364748, -0.09189429879188538, 0.1145000010728836, 0.05107851326465607, 0.030498135834932327, -0.029527099803090096, -0.18762432038784027, 0.03377383574843407, 0.09188276529312134, -0.13706527650356293, -0.06224900856614113 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec-timit This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1000 - num_epochs: 30 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "wav2vec-timit", "results": []}]}
automatic-speech-recognition
cristinakuo/wav2vec-timit
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us
# wav2vec-timit This model is a fine-tuned version of facebook/wav2vec2-base on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1000 - num_epochs: 30 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
[ "# wav2vec-timit\n\nThis model is a fine-tuned version of facebook/wav2vec2-base on the None dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 30\n- mixed_precision_training: Native AMP", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.10.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n", "# wav2vec-timit\n\nThis model is a fine-tuned version of facebook/wav2vec2-base on the None dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 30\n- mixed_precision_training: Native AMP", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.10.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ 52, 32, 6, 12, 8, 3, 117, 33 ]
[ "passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n# wav2vec-timit\n\nThis model is a fine-tuned version of facebook/wav2vec2-base on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 30\n- mixed_precision_training: Native AMP### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.10.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ -0.10877498984336853, 0.09662093222141266, -0.0029397383332252502, 0.03550403565168381, 0.126646488904953, 0.015628598630428314, 0.08752505481243134, 0.10619714856147766, -0.05869685858488083, 0.06267892569303513, 0.07386357337236404, 0.005418893415480852, 0.06806017458438873, 0.1374618411064148, -0.030923208221793175, -0.21295593678951263, 0.027059225365519524, -0.015200169757008553, -0.05231095105409622, 0.08767984062433243, 0.09603627771139145, -0.08322007209062576, 0.05126743018627167, 0.033149976283311844, -0.14410389959812164, 0.023153601214289665, -0.04061105102300644, -0.084200918674469, 0.10286605358123779, 0.04474921524524689, 0.0767868384718895, 0.023776929825544357, 0.09924697875976562, -0.23043031990528107, 0.005783200729638338, 0.08554726094007492, 0.029281485825777054, 0.0674399584531784, 0.08298734575510025, 0.0031263004057109356, 0.12663237750530243, -0.14801885187625885, 0.08211545646190643, 0.056369658559560776, -0.06638497859239578, -0.1968664675951004, -0.06689172238111496, 0.10077212750911713, 0.08997321128845215, 0.10150709748268127, -0.0076980190351605415, 0.09968359023332596, -0.09635525941848755, 0.06922397017478943, 0.1962634027004242, -0.25214019417762756, -0.056458957493305206, -0.011320199817419052, 0.04585195705294609, 0.018832135945558548, -0.09840576350688934, 0.0054738763719797134, 0.04135631024837494, 0.045132748782634735, 0.08429259806871414, 0.0022272909991443157, -0.08431728184223175, -0.032898493111133575, -0.12415239959955215, -0.007053246721625328, 0.14903110265731812, 0.08631891012191772, -0.0538078248500824, -0.10097965598106384, -0.04559721052646637, -0.0594572015106678, -0.032833125442266464, -0.05044430494308472, 0.01610122062265873, -0.020094428211450577, -0.06229245290160179, -0.06762432307004929, -0.06720376014709473, -0.06341198831796646, -0.0014623187016695738, 0.13239160180091858, 0.036509159952402115, 0.0026749717071652412, -0.037988729774951935, 0.0786338523030281, 0.017936212942004204, -0.09968128055334091, -0.006951979827135801, -0.010479096323251724, -0.09649249166250229, -0.0389280691742897, -0.06104184314608574, -0.005859626457095146, 0.027684736996889114, 0.12012793868780136, -0.03434640169143677, 0.10160859674215317, 0.0005570212379097939, 0.00841433648020029, -0.026372579857707024, 0.10951117426156998, -0.04812266305088997, -0.04760728403925896, -0.007061479613184929, 0.05983440577983856, -0.011831107549369335, -0.023357165977358818, -0.06845317035913467, -0.05568648874759674, 0.07931321114301682, 0.05284101516008377, -0.04364873096346855, 0.01676325500011444, -0.04465916380286217, -0.03761421889066696, 0.005235984455794096, -0.10947742313146591, 0.036769986152648926, 0.02287694625556469, -0.04694134742021561, 0.06243867427110672, 0.03479846939444542, 0.015279448591172695, -0.05168275162577629, 0.04955633357167244, -0.044313330203294754, -0.017610013484954834, -0.038690321147441864, -0.04941916465759277, 0.021562129259109497, -0.04514089599251747, -0.009941866621375084, -0.11116272211074829, -0.14922663569450378, -0.034337420016527176, 0.041913192719221115, -0.046991243958473206, -0.04487429931759834, -0.034798379987478256, -0.03846399858593941, 0.03228400647640228, -0.031715672463178635, 0.16045024991035461, -0.05732320621609688, 0.049300629645586014, -0.021562613546848297, 0.033254507929086685, 0.007049087435007095, 0.054298434406518936, -0.037093281745910645, 0.027908051386475563, -0.07389504462480545, 0.11735209822654724, -0.10204337537288666, -0.007187202572822571, -0.12455746531486511, -0.08307620882987976, -0.012575552798807621, -0.01210657600313425, 0.0827266126871109, 0.1234971359372139, -0.1872541308403015, -0.06136789172887802, 0.14040981233119965, -0.07556524127721786, -0.03961645066738129, 0.11929165571928024, -0.025144293904304504, -0.0038792213890701532, 0.07057813555002213, 0.14464373886585236, 0.12784774601459503, -0.14638209342956543, 0.007742945570498705, 0.04272308945655823, 0.03937526419758797, 0.013993365690112114, 0.05138707160949707, -0.03210333734750748, -0.032464854419231415, 0.014230470173060894, -0.005971256643533707, 0.03017762303352356, -0.08251375705003738, -0.07797014713287354, -0.05702979490160942, -0.09979292750358582, 0.060017552226781845, -0.015603349544107914, 0.013385246507823467, -0.07732062041759491, -0.11469990760087967, 0.061918940395116806, 0.14060690999031067, -0.060776494443416595, 0.022271977737545967, -0.07892517000436783, 0.014942009933292866, 0.015281029976904392, -0.00908727291971445, -0.18470434844493866, -0.057300589978694916, 0.02633463405072689, -0.09618724882602692, 0.03881969675421715, 0.047664981335401535, 0.050085414201021194, 0.05261142924427986, -0.042295973747968674, -0.022628501057624817, -0.10675814747810364, 0.020135626196861267, -0.08070103079080582, -0.17159989476203918, -0.065264493227005, -0.03426970914006233, 0.1599881947040558, -0.2024581879377365, -0.0014569745399057865, 0.03198186308145523, 0.15644285082817078, 0.012679724022746086, -0.03507274016737938, 0.00043003089376725256, 0.06053066626191139, 0.023288344964385033, -0.0885084718465805, 0.04961921274662018, 0.014070974662899971, -0.11340519040822983, -0.04085076227784157, -0.08018512278795242, 0.04696926474571228, 0.10342492908239365, 0.04478681460022926, -0.07461195439100266, -0.017937002703547478, -0.058719102293252945, -0.04515514522790909, -0.05049068480730057, 0.0002573136880528182, 0.23708635568618774, 0.027808280661702156, 0.11380542814731598, -0.05448519438505173, -0.03775626793503761, 0.026979031041264534, 0.008490742184221745, -0.027983270585536957, 0.07300828397274017, 0.06415735930204391, -0.10357458889484406, 0.05635764077305794, 0.11486604809761047, -0.03547178953886032, 0.1411331295967102, -0.04168357327580452, -0.07935961335897446, -0.010078872554004192, -0.003901989432051778, -0.023645054548978806, 0.07583068311214447, -0.16241131722927094, 0.005374868866056204, 0.030248617753386497, 0.029923230409622192, 0.0635615810751915, -0.15572987496852875, 0.025070207193493843, 0.020888345316052437, -0.05663435533642769, -0.009417508728802204, -0.003792727831751108, 0.03545691817998886, 0.0775628462433815, 0.031660597771406174, -0.021704640239477158, 0.030119502916932106, -0.013407448306679726, -0.08602337539196014, 0.15777365863323212, -0.1318495273590088, -0.20200961828231812, -0.11423411220312119, 0.02470136433839798, -0.07521861791610718, -0.03119315207004547, 0.03761829808354378, -0.10950873047113419, -0.05098309367895126, -0.05100974813103676, 0.03972640633583069, -0.06676285713911057, 0.013505615293979645, 0.08643274754285812, 0.018459074199199677, 0.09957531094551086, -0.12274344265460968, 0.001335379434749484, -0.009818623773753643, -0.06052667275071144, -0.023107614368200302, 0.05731344223022461, 0.06021704524755478, 0.10516297817230225, -0.005062054842710495, 0.013561166822910309, -0.03601580485701561, 0.23491835594177246, -0.06469008326530457, 0.0008565475582145154, 0.14710275828838348, 0.002869020914658904, 0.04299754276871681, 0.06616554409265518, 0.025140300393104553, -0.10242624580860138, 0.024033596739172935, 0.058749597519636154, -0.010694453492760658, -0.23187269270420074, -0.06708474457263947, -0.046210214495658875, -0.07788047939538956, 0.0991303026676178, 0.04130746051669121, -0.018704378977417946, 0.04137157276272774, -0.018470797687768936, 0.036932557821273804, -0.009577760472893715, 0.06138044223189354, 0.10673097521066666, 0.03413818031549454, 0.08926046639680862, -0.029926951974630356, -0.0024278471246361732, 0.09171976894140244, -0.010211824439466, 0.22118571400642395, 0.01984224282205105, 0.05856953561306, 0.05498114973306656, 0.15964224934577942, 0.005420234054327011, 0.024226471781730652, 0.031392574310302734, 0.0009714126936160028, 0.0002608165086712688, -0.054014842957258224, -0.046515610069036484, 0.023241491988301277, 0.042402006685733795, 0.0008234973065555096, -0.11991271376609802, 0.00350527698174119, 0.0018550974782556295, 0.3238118290901184, 0.038944389671087265, -0.25084397196769714, -0.08257074654102325, 0.001798473298549652, -0.0702078565955162, -0.07576259225606918, 0.030992351472377777, 0.1152917668223381, -0.14332018792629242, 0.09683848917484283, -0.025034070014953613, 0.09625114500522614, -0.03242335468530655, 0.023484228178858757, 0.011573377065360546, 0.13598033785820007, 0.017334122210741043, 0.08177676051855087, -0.2205827534198761, 0.1954256296157837, 0.008835752494633198, 0.10631981492042542, -0.06380481272935867, 0.036702487617731094, 0.03112797997891903, 0.04275748133659363, 0.07953710108995438, -0.0011629883665591478, -0.03845692798495293, -0.11792544275522232, -0.07551047950983047, 0.04344950616359711, 0.108644999563694, -0.02668253518640995, 0.05816507339477539, -0.03430456295609474, 0.005597986746579409, 0.04350857436656952, 0.013656253926455975, -0.22417296469211578, -0.1388792246580124, -0.0016967797419056296, 0.08348690718412399, 0.03176585212349892, -0.08840300142765045, -0.09631901234388351, -0.055360421538352966, 0.1781175434589386, -0.016398994252085686, -0.028621455654501915, -0.1411036103963852, 0.04369601607322693, 0.12338346242904663, -0.04998849704861641, 0.031894803047180176, 0.02155952900648117, 0.1597699671983719, 0.00901733711361885, -0.11063000559806824, 0.05902382358908653, -0.09507922828197479, -0.1686534434556961, -0.03251172974705696, 0.154094398021698, 0.0712839663028717, 0.028560571372509003, 0.010435422882437706, -0.006148618180304766, 0.002105240011587739, -0.09214220941066742, 0.03739134222269058, 0.09805293381214142, 0.013161294162273407, 0.06296665966510773, -0.04060974344611168, -0.054933059960603714, -0.03876582533121109, -0.02395431138575077, 0.1455746740102768, 0.22120486199855804, -0.07811582833528519, 0.10536453872919083, 0.1417420655488968, -0.06267328560352325, -0.15346555411815643, 0.03201717138290405, 0.12117011845111847, 0.030585218220949173, 0.012844417244195938, -0.21831820905208588, 0.09412895888090134, 0.1079513356089592, -0.0059300372377038, 0.013644036836922169, -0.26828864216804504, -0.13218066096305847, 0.12653376162052155, 0.07545828074216843, 0.06067531183362007, -0.07508036494255066, -0.0390414334833622, -0.0754014104604721, -0.11583466827869415, 0.12027079612016678, -0.14535637199878693, 0.1147933229804039, 0.007497474551200867, 0.07585561275482178, 0.012461988255381584, -0.010392104275524616, 0.1498195379972458, 0.030621975660324097, 0.06741341948509216, -0.028260568156838417, 0.1009482890367508, 0.02284371107816696, -0.05028301104903221, 0.036595527082681656, -0.06192454695701599, 0.06161421164870262, -0.11416076123714447, -0.04131070151925087, -0.07655195146799088, 0.08888446539640427, -0.03127102553844452, -0.047379277646541595, -0.03244289010763168, 0.026667334139347076, 0.04894683137536049, -0.030313244089484215, -0.007860693149268627, -0.001455042278394103, 0.0931636393070221, 0.1144702136516571, 0.11434796452522278, -0.05298831686377525, -0.10008092224597931, -0.019951460883021355, -0.04574226215481758, 0.08912388235330582, -0.050990086048841476, 0.03147737309336662, 0.09277984499931335, 0.026356348767876625, 0.12451112270355225, 0.025713879615068436, -0.07144437730312347, 0.0027159203309565783, 0.019825303927063942, -0.10165812075138092, -0.14096058905124664, -0.029846027493476868, -0.004764875862747431, -0.08509330451488495, 0.007429995108395815, 0.14764276146888733, -0.052512120455503464, -0.007916707545518875, -0.024907764047384262, 0.0004920942592434585, -0.06068827584385872, 0.1573842316865921, 0.02472110092639923, 0.06495726108551025, -0.09986817836761475, 0.12873752415180206, 0.06284455209970474, -0.09751614183187485, 0.09520172327756882, 0.05845721811056137, -0.08624309301376343, -0.03083069808781147, 0.04437083750963211, 0.1642586588859558, 0.010196591727435589, -0.08161204308271408, -0.036447953432798386, -0.14473889768123627, 0.0798501968383789, 0.11588931828737259, 0.02593933418393135, -0.007121115457266569, -0.05358415096998215, 0.017304250970482826, -0.11217638105154037, 0.04192770645022392, 0.04482448101043701, 0.027132371440529823, -0.14577296376228333, 0.12550897896289825, 0.02909408137202263, 0.039647504687309265, -0.016568679362535477, -0.010040469467639923, -0.08777797222137451, -0.006784357596188784, -0.2095743715763092, -0.008237119764089584, -0.03066575527191162, 0.023440882563591003, -0.010183348320424557, -0.04376043751835823, -0.04483703523874283, 0.05073145404458046, -0.06613966077566147, -0.05256340280175209, 0.004711037501692772, 0.06949934363365173, -0.12189662456512451, 0.012972841039299965, 0.016446638852357864, -0.1032484769821167, 0.07911131531000137, 0.07650713622570038, 0.003618380520492792, 0.041134558618068695, -0.12442365288734436, -0.04300391301512718, 0.03507591038942337, 0.038430456072092056, 0.04966779053211212, -0.11179843544960022, -0.008124489337205887, -0.012195953167974949, 0.05058775842189789, 0.004170175641775131, 0.09512338042259216, -0.09710487723350525, -0.07369830459356308, -0.05590566247701645, -0.04561357572674751, -0.05868973582983017, 0.021639283746480942, 0.1362050622701645, 0.05006612837314606, 0.15065382421016693, -0.08839187026023865, 0.043587177991867065, -0.17376750707626343, -0.02806163765490055, -0.028535261750221252, -0.008047751151025295, -0.029429322108626366, -0.055288635194301605, 0.09381391108036041, -0.04043901711702347, 0.12295011430978775, -0.048985231667757034, 0.08328862488269806, 0.027871891856193542, -0.11258480697870255, -0.08597704023122787, 0.02016010694205761, 0.19160039722919464, 0.08632530272006989, 0.0050036064349114895, 0.06583283096551895, -0.028710229322314262, 0.07010706514120102, 0.07263333350419998, 0.1639476865530014, 0.17230698466300964, -0.016960781067609787, 0.0798993781208992, 0.07871364057064056, -0.12157124280929565, -0.1321680247783661, 0.10364412516355515, -0.060700081288814545, 0.12586013972759247, -0.06613123416900635, 0.13727733492851257, 0.057976800948381424, -0.16680939495563507, 0.0602722242474556, -0.0372600182890892, -0.0896816998720169, -0.10820595175027847, -0.04301333427429199, -0.07251179963350296, -0.1378588080406189, 0.01722487434744835, -0.09031563997268677, 0.019506892189383507, 0.038445085287094116, 0.018839789554476738, 0.007833017967641354, 0.1325487196445465, -0.04186573624610901, -0.0383242592215538, 0.10698370635509491, 0.0007293768576346338, -0.02932034805417061, -0.08569017052650452, -0.09116942435503006, 0.06656603515148163, 0.03167876973748207, 0.061824750155210495, -0.027335109189152718, -0.04196241870522499, 0.04941545054316521, 0.0000753947242628783, -0.06794491410255432, 0.03565674275159836, -0.013320892117917538, 0.043671950697898865, 0.06762328743934631, 0.05600978061556816, -0.030638400465250015, -0.029814528301358223, 0.22779348492622375, -0.08677167445421219, -0.0954098179936409, -0.14217622578144073, 0.15771393477916718, 0.006159490440040827, 0.0006626692484132946, 0.05305861309170723, -0.07521667331457138, -0.045105528086423874, 0.18441922962665558, 0.12098510563373566, -0.061361562460660934, -0.0027861965354532003, -0.01763669028878212, -0.020861808210611343, -0.052814215421676636, 0.14405715465545654, 0.12812478840351105, 0.023457789793610573, -0.03594452515244484, 0.015393108129501343, -0.011157380416989326, -0.051097970455884933, -0.08315169811248779, 0.11503266543149948, -0.013075768947601318, -0.01598230004310608, -0.027144089341163635, 0.08210363984107971, -0.01008684653788805, -0.2013113647699356, 0.015779493376612663, -0.12701018154621124, -0.1718190759420395, -0.027833426371216774, 0.04411604627966881, 0.021778492256999016, 0.04970526322722435, -0.015040060505270958, 0.006524661090224981, 0.16223646700382233, -0.01721321977674961, -0.028605088591575623, -0.10335364192724228, 0.08896081149578094, -0.05003296583890915, 0.15703432261943817, -0.0019414934795349836, 0.06220119073987007, 0.09069833159446716, 0.0630866289138794, -0.11956746131181717, 0.04276566207408905, 0.03978632390499115, -0.050480179488658905, 0.040116362273693085, 0.14336562156677246, -0.053296755999326706, 0.10124414414167404, 0.047977056354284286, -0.1424037665128708, -0.030942402780056, -0.024747811257839203, -0.014648159965872765, -0.05001937970519066, 0.016254372894763947, -0.0725824236869812, 0.14058947563171387, 0.1693616509437561, -0.04982820153236389, -0.016285572201013565, -0.07054025679826736, 0.052148401737213135, 0.028757669031620026, 0.06308937817811966, -0.03308512642979622, -0.20406474173069, 0.004705186001956463, 0.0236944742500782, 0.04122009500861168, -0.22781158983707428, -0.09892953932285309, 0.018664345145225525, -0.036652300506830215, -0.04097309708595276, 0.10311908274888992, 0.02191823348402977, 0.020190713927149773, -0.031222449615597725, -0.10497882217168808, -0.0347956120967865, 0.1415175050497055, -0.14677643775939941, -0.03699145093560219 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-latino40 This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.8795 - Wer: 1.0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 32 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 100 - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:---:| | 5.6846 | 0.83 | 100 | 2.9086 | 1.0 | | 2.8686 | 1.67 | 200 | 2.8922 | 1.0 | | 2.8805 | 2.5 | 300 | 2.9326 | 1.0 | | 2.8613 | 3.33 | 400 | 2.8698 | 1.0 | | 2.8643 | 4.17 | 500 | 2.9027 | 1.0 | | 2.8688 | 5.0 | 600 | 2.9544 | 1.0 | | 2.8689 | 5.83 | 700 | 2.8914 | 1.0 | | 2.8558 | 6.67 | 800 | 2.8762 | 1.0 | | 2.8537 | 7.5 | 900 | 2.8982 | 1.0 | | 2.8522 | 8.33 | 1000 | 2.8820 | 1.0 | | 2.8468 | 9.17 | 1100 | 2.8760 | 1.0 | | 2.8454 | 10.0 | 1200 | 2.8795 | 1.0 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.9.1 - Datasets 1.16.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "wav2vec2-latino40", "results": []}]}
automatic-speech-recognition
cristinakuo/wav2vec2-latino40
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us
wav2vec2-latino40 ================= This model is a fine-tuned version of facebook/wav2vec2-base on the None dataset. It achieves the following results on the evaluation set: * Loss: 2.8795 * Wer: 1.0 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.0001 * train\_batch\_size: 32 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 100 * num\_epochs: 10 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.12.5 * Pytorch 1.9.1 * Datasets 1.16.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 100\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.9.1\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 100\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.9.1\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ 52, 130, 4, 31 ]
[ "passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 100\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.9.1\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ -0.10980920493602753, 0.08066091686487198, -0.0021913868840783834, 0.0798632800579071, 0.14092256128787994, 0.0030789559241384268, 0.08961708098649979, 0.1398172527551651, -0.07077814638614655, 0.040380705147981644, 0.1079588532447815, 0.14933347702026367, 0.03574487194418907, 0.12165234982967377, -0.04340481013059616, -0.29026296734809875, 0.01716907136142254, 0.014433713629841805, -0.044237591326236725, 0.1325906664133072, 0.09864385426044464, -0.12339255213737488, 0.04956413060426712, 0.011601904407143593, -0.14805734157562256, 0.010426769964396954, 0.0026738583110272884, -0.08487695455551147, 0.14838279783725739, 0.00794005487114191, 0.07316295057535172, 0.015556921251118183, 0.0989600419998169, -0.2548612356185913, 0.003781266510486603, 0.041084736585617065, 0.047623373568058014, 0.061330974102020264, 0.07041196525096893, -0.0112789086997509, 0.11298255622386932, -0.10049547255039215, 0.0634230226278305, 0.035836685448884964, -0.10100731998682022, -0.29945191740989685, -0.07114958763122559, 0.022580832242965698, 0.08290360867977142, 0.11252439022064209, -0.017463045194745064, 0.10048104077577591, -0.08502072840929031, 0.10130675882101059, 0.2632434368133545, -0.28493553400039673, -0.04188449680805206, -0.0320056714117527, 0.030116042122244835, 0.05607079714536667, -0.1130281314253807, -0.0199373010545969, 0.021807949990034103, 0.051630325615406036, 0.11832799762487411, -0.02314845658838749, -0.11859384179115295, 0.012785177677869797, -0.1495315581560135, -0.03822407126426697, 0.12689848244190216, 0.044106051325798035, -0.03479421138763428, -0.08074644953012466, -0.04381341114640236, -0.14233465492725372, -0.05427202954888344, -0.014493022114038467, 0.04658723995089531, -0.03640612214803696, -0.08634860068559647, -0.017045967280864716, -0.08484596014022827, -0.08799180388450623, -0.03250779211521149, 0.163325235247612, 0.049993909895420074, -0.0054273405112326145, -0.009451371617615223, 0.08832554519176483, 0.0006423027953132987, -0.12581756711006165, -0.0027684089727699757, 0.03728927671909332, -0.030845317989587784, -0.006714310962706804, -0.0775303915143013, -0.006719156634062529, 0.025007838383316994, 0.12120231240987778, -0.07473734021186829, 0.07411094754934311, 0.013705099932849407, 0.02501939795911312, -0.11756092309951782, 0.20496274530887604, -0.05834757164120674, -0.025750350207090378, -0.013698107562959194, 0.04614108055830002, 0.005192374344915152, -0.01254340074956417, -0.09562832862138748, -0.005417653825134039, 0.1215624064207077, 0.039135344326496124, -0.06357745081186295, 0.05863622948527336, -0.03601355850696564, -0.01738746650516987, -0.04621974006295204, -0.11743055284023285, 0.03168739378452301, 0.025128355249762535, -0.08781567960977554, 0.000711713160853833, 0.023804007098078728, 0.02700847014784813, -0.04845186695456505, 0.08255426585674286, -0.05339859426021576, 0.04193750023841858, -0.06689785420894623, -0.11167975515127182, 0.013831880874931812, -0.06339839100837708, 0.015696832910180092, -0.1014920100569725, -0.15867817401885986, -0.01975177228450775, 0.04356376826763153, -0.036123570054769516, -0.03404486924409866, -0.08703483641147614, -0.07419129461050034, 0.024441922083497047, -0.04096292331814766, 0.14225170016288757, -0.06730513274669647, 0.11225902289152145, 0.029156258329749107, 0.07115056365728378, -0.018335899338126183, 0.07524584233760834, -0.0815892219543457, 0.002993879606947303, -0.1414223611354828, 0.08606751263141632, -0.07554500550031662, 0.0405312106013298, -0.12535442411899567, -0.1199832633137703, 0.02675686776638031, 0.0007289050845429301, 0.08981318771839142, 0.10596152395009995, -0.17258626222610474, -0.08019126206636429, 0.1752787083387375, -0.06593860685825348, -0.08045100420713425, 0.11808839440345764, -0.039672963321208954, 0.023581838235259056, 0.07135766744613647, 0.2478012591600418, 0.04693472757935524, -0.11165597289800644, 0.043805141001939774, -0.013276532292366028, 0.06678491830825806, -0.025329528376460075, 0.07152348011732101, -0.03138262778520584, 0.010174485854804516, 0.03288539499044418, -0.043981969356536865, 0.07483356446027756, -0.09624520689249039, -0.08675498515367508, -0.03934602811932564, -0.10191374272108078, 0.03482167422771454, 0.043031103909015656, 0.06365569680929184, -0.0985899269580841, -0.08746787160634995, 0.04025315120816231, 0.09528588503599167, -0.09891946613788605, 0.05347936600446701, -0.07100776582956314, 0.04501967132091522, -0.001897135516628623, -0.016613401472568512, -0.1871185600757599, 0.03687867894768715, 0.023578641936182976, -0.02337789349257946, 0.039804913103580475, -0.03218350186944008, 0.07619354128837585, 0.03881413862109184, -0.04651379957795143, -0.04409072548151016, -0.02065667323768139, 0.012341917492449284, -0.09351722151041031, -0.1993872970342636, -0.035728778690099716, -0.018513230606913567, 0.11530566960573196, -0.17193715274333954, 0.014598478563129902, -0.019322114065289497, 0.07519219815731049, 0.0037099465262144804, -0.019692348316311836, -0.01481874380260706, 0.08457089215517044, -0.0085816141217947, -0.05307632312178612, 0.06940632313489914, 0.0031947139650583267, -0.09319141507148743, 0.01586523838341236, -0.10217957198619843, 0.10750915110111237, 0.1332051008939743, -0.08164609968662262, -0.06618577986955643, 0.0011545104207471013, -0.049083709716796875, -0.039093658328056335, -0.02735612727701664, 0.02902553416788578, 0.20677080750465393, -0.005885649938136339, 0.13525806367397308, -0.07594040781259537, -0.02799786813557148, 0.022258900105953217, -0.023045221343636513, 0.015822691842913628, 0.13198581337928772, 0.07947170734405518, -0.041822366416454315, 0.11116132140159607, 0.12410301715135574, -0.09254572540521622, 0.11285964399576187, -0.05419359728693962, -0.08977862447500229, -0.010861373506486416, -0.007626814767718315, -0.015537419356405735, 0.08842293918132782, -0.15356121957302094, -0.014198613353073597, 0.029136179015040398, 0.0361868180334568, 0.026153240352869034, -0.225873202085495, -0.01565035991370678, 0.029006700962781906, -0.08592138439416885, -0.03842167928814888, -0.0021013568621128798, 0.012155167758464813, 0.09853563457727432, 0.003302189754322171, -0.10624036937952042, 0.009442655369639397, -0.017023036256432533, -0.084555484354496, 0.19169212877750397, -0.11786629259586334, -0.17307570576667786, -0.09892862290143967, -0.07620367407798767, -0.03259239345788956, 0.003938349429517984, 0.07807767391204834, -0.09726081788539886, -0.028173502534627914, -0.06308752298355103, 0.03210591897368431, -0.041175562888383865, 0.03680180385708809, 0.0036928232293576, -0.003402488073334098, 0.06942503899335861, -0.10922123491764069, -0.008241287432610989, -0.05346428602933884, -0.03276630863547325, 0.038681626319885254, 0.04400238022208214, 0.10152043402194977, 0.16139578819274902, 0.008259491994976997, 0.028078991919755936, -0.04554590582847595, 0.20942853391170502, -0.06462772190570831, -0.04289346933364868, 0.1388622522354126, 0.0070281364023685455, 0.04439382255077362, 0.12561090290546417, 0.04757501557469368, -0.09919650852680206, 0.0033767425920814276, 0.012305780313909054, -0.022879192605614662, -0.21807990968227386, -0.0541338287293911, -0.049014441668987274, -0.029451046139001846, 0.10054463893175125, 0.021792670711874962, 0.01725919358432293, 0.020740719512104988, 0.04053569585084915, 0.002018396742641926, -0.0015959605807438493, 0.058000557124614716, 0.14925150573253632, 0.024657532572746277, 0.12087169289588928, -0.01869051344692707, -0.04857170581817627, 0.02699047513306141, -0.0108701903373003, 0.23927874863147736, 0.012819498777389526, 0.12436876446008682, 0.050061244517564774, 0.18832843005657196, -0.0005413531907834113, 0.07564345002174377, 0.015274266712367535, -0.012130046263337135, -0.002754911780357361, -0.05093031004071236, -0.045556794852018356, 0.031116588041186333, 0.02848834916949272, 0.025044111534953117, -0.13001902401447296, -0.049802038818597794, 0.03907139599323273, 0.3274923264980316, 0.05229125916957855, -0.29431799054145813, -0.08614405244588852, -0.005382025148719549, -0.0854690670967102, -0.022076856344938278, 0.046303000301122665, 0.08580722659826279, -0.09947432577610016, 0.04448115825653076, -0.047977153211832047, 0.09024563431739807, -0.03151266276836395, 0.040820345282554626, 0.04360958933830261, 0.07487986236810684, 0.02256215550005436, 0.06914918124675751, -0.3232811391353607, 0.29180625081062317, -0.008358528837561607, 0.07779446989297867, -0.06456765532493591, -0.0008857692591845989, 0.03378576412796974, -0.0011557098478078842, 0.09022928774356842, -0.018672609701752663, -0.0801580622792244, -0.18707461655139923, -0.060589663684368134, 0.036532677710056305, 0.127633199095726, -0.011211197823286057, 0.10346045345067978, -0.03243345022201538, -0.009961643256247044, 0.07621586322784424, -0.055092498660087585, -0.10099312663078308, -0.08181507140398026, -0.00940331257879734, 0.07693351060152054, 0.03743179887533188, -0.06578808277845383, -0.10221552848815918, -0.11230611056089401, 0.1218077689409256, -0.04539094492793083, -0.027113184332847595, -0.1105206087231636, 0.061217136681079865, 0.130745530128479, -0.0792306512594223, 0.05503794550895691, 0.028080426156520844, 0.06753195822238922, 0.03414563834667206, -0.06759253889322281, 0.11237512528896332, -0.07640185952186584, -0.17026765644550323, -0.033496465533971786, 0.1469498723745346, 0.036228787153959274, 0.06863132119178772, -0.006273076869547367, 0.028511548414826393, -0.045666828751564026, -0.0804627537727356, 0.03256896138191223, 0.036437153816223145, 0.023420540615916252, 0.04869148135185242, -0.05705498158931732, -0.017071884125471115, -0.09537269175052643, -0.04085355997085571, 0.20932559669017792, 0.2342206984758377, -0.093354232609272, 0.0716480240225792, 0.06755233556032181, -0.05380915105342865, -0.18005482852458954, -0.0030158916488289833, 0.0848696231842041, 0.0050223711878061295, 0.015279422514140606, -0.19609257578849792, 0.07792839407920837, 0.06992988288402557, -0.01870499551296234, 0.07932014018297195, -0.3264188766479492, -0.14206287264823914, 0.1441699117422104, 0.12191656976938248, 0.08031284809112549, -0.13511276245117188, -0.0404108390212059, -0.01999991573393345, -0.091950923204422, 0.0915660560131073, -0.04971156641840935, 0.1435217708349228, -0.017111627385020256, 0.1045382097363472, 0.013917868956923485, -0.049031343311071396, 0.11851494759321213, 0.024936791509389877, 0.06427741050720215, -0.036918606609106064, 0.019268162548542023, -0.00964928325265646, -0.034783001989126205, 0.04311220347881317, -0.06751948595046997, 0.037642188370227814, -0.08364549279212952, -0.035955946892499924, -0.10305949300527573, 0.03376590088009834, -0.005185771267861128, -0.05732830613851547, -0.018481800332665443, 0.02000368759036064, 0.06819313764572144, -0.007115368265658617, 0.09952523559331894, -0.033060330897569656, 0.12334724515676498, 0.11199751496315002, 0.07548270374536514, -0.04602932929992676, -0.06700962036848068, -0.014362026937305927, -0.028282882645726204, 0.05370795726776123, -0.11295565962791443, 0.029776575043797493, 0.13810531795024872, 0.03917720168828964, 0.14451472461223602, 0.06972716003656387, -0.04443461820483208, 0.010397059842944145, 0.04125906527042389, -0.12701374292373657, -0.11644574999809265, -0.004148922394961119, -0.02314218319952488, -0.07639680802822113, 0.04377919062972069, 0.10687291622161865, -0.06386600434780121, -0.011042671278119087, -0.021161457523703575, 0.02330395020544529, -0.07179027050733566, 0.2252303510904312, 0.06212461739778519, 0.04856344684958458, -0.11988171190023422, 0.08925968408584595, 0.041735388338565826, -0.1158151850104332, 0.03984756022691727, 0.07972640544176102, -0.08497200906276703, -0.039511069655418396, 0.05026671290397644, 0.10819706320762634, -0.04242098703980446, -0.07139823585748672, -0.11464481800794601, -0.14505107700824738, 0.09956482797861099, 0.1452970802783966, 0.07713909447193146, 0.014507410116493702, -0.06352374702692032, 0.013524708338081837, -0.10527925938367844, 0.07716207951307297, 0.05278395861387253, 0.04045756161212921, -0.13215371966362, 0.1596628576517105, 0.01660974510014057, 0.05105000361800194, -0.01783343032002449, 0.0003049750521313399, -0.09723072499036789, 0.044797927141189575, -0.13372665643692017, -0.016621513292193413, -0.0440630204975605, 0.00647035613656044, 0.009973205626010895, -0.07510308921337128, -0.05170198529958725, 0.037644289433956146, -0.1226712167263031, -0.030972497537732124, -0.0020897414069622755, 0.05209917202591896, -0.12659986317157745, -0.04508664831519127, 0.02240194007754326, -0.07813826948404312, 0.08739304542541504, 0.09181436896324158, -0.019976526498794556, 0.08028993755578995, -0.14488914608955383, -0.028371071442961693, 0.0663955956697464, 0.008208757266402245, 0.0362272784113884, -0.13547508418560028, -0.014229604974389076, 0.009603659622371197, 0.05940083786845207, 0.014157848432660103, 0.08902829885482788, -0.1306706815958023, -0.016501907259225845, -0.042343106120824814, -0.06023785099387169, -0.06204197555780411, 0.02433887869119644, 0.07678435742855072, 0.046087849885225296, 0.1713658571243286, -0.09509948641061783, 0.04948916658759117, -0.1854909211397171, 0.005001120734959841, -0.04503310099244118, -0.09847234189510345, -0.09514760971069336, -0.052987098693847656, 0.08700865507125854, -0.04966249689459801, 0.12958776950836182, -0.015757031738758087, 0.07188122719526291, 0.018411675468087196, -0.06927963346242905, -0.010539642535150051, 0.03743515536189079, 0.25493499636650085, 0.038861505687236786, -0.029862146824598312, 0.07857978343963623, 0.04346318542957306, 0.09011953324079514, 0.15496864914894104, 0.16722559928894043, 0.19484195113182068, 0.03936972841620445, 0.10197120159864426, 0.052975527942180634, -0.0825294777750969, -0.14498275518417358, 0.06355110555887222, -0.031587496399879456, 0.11224187910556793, -0.029109589755535126, 0.2681392729282379, 0.061866383999586105, -0.17852698266506195, 0.06703627109527588, -0.03827032446861267, -0.08790795505046844, -0.1184806376695633, -0.027287298813462257, -0.08511841297149658, -0.17839950323104858, 0.004721211735159159, -0.10451129823923111, 0.05941224470734596, 0.0790998712182045, 0.03342793136835098, 0.0002967296168208122, 0.13273613154888153, 0.017104819416999817, -0.013799617066979408, 0.09511622041463852, -0.005110669881105423, -0.04267231747508049, -0.08802526444196701, -0.08089674264192581, 0.0317658968269825, -0.012727725319564342, 0.04702073708176613, -0.04121053218841553, -0.11177753657102585, 0.03308112919330597, -0.044365085661411285, -0.08819011598825455, 0.01934644766151905, 0.018091760575771332, 0.08175329864025116, 0.08086381107568741, 0.037871353328228, -0.03429818153381348, -0.0020277153234928846, 0.2526741921901703, -0.10593343526124954, -0.123429074883461, -0.10019700229167938, 0.2773582935333252, 0.04580030217766762, -0.0008489458705298603, 0.016611211001873016, -0.0669313296675682, -0.01992752216756344, 0.24284401535987854, 0.175593301653862, -0.06653659045696259, -0.0029926213901489973, -0.0014997737016528845, -0.007376679219305515, -0.033362776041030884, 0.08983216434717178, 0.16204452514648438, 0.06294775754213333, -0.07975048571825027, -0.03254455327987671, -0.0467907078564167, -0.030329039320349693, -0.0675441175699234, 0.07937409728765488, 0.0074424175545573235, -0.03071312978863716, -0.035275060683488846, 0.08556422591209412, -0.08735252916812897, -0.12293340265750885, 0.015828529372811317, -0.1928618997335434, -0.1565762311220169, -0.0158489141613245, 0.07319075614213943, 0.040658846497535706, 0.03373730182647705, -0.02140166237950325, 0.006188822444528341, 0.08847112208604813, -0.0029859363567084074, -0.0775647982954979, -0.08026209473609924, 0.09346207976341248, -0.10640282183885574, 0.1653733253479004, -0.03455295041203499, 0.07426184415817261, 0.10542125254869461, 0.09622615575790405, -0.06406783312559128, 0.102670818567276, 0.052303776144981384, -0.107117660343647, 0.05454130098223686, 0.14238789677619934, -0.042176973074674606, 0.09960916638374329, 0.03267516940832138, -0.1318810135126114, 0.026521943509578705, -0.044511228799819946, -0.0656069964170456, -0.04928657412528992, -0.044857949018478394, -0.05315857380628586, 0.11852610111236572, 0.1792403608560562, -0.05418472737073898, 0.009916502982378006, -0.06851965188980103, 0.005978732369840145, 0.03164321184158325, 0.03913820907473564, -0.06651060283184052, -0.2513844966888428, 0.010636888444423676, 0.023227639496326447, 0.0053297677077353, -0.23199740052223206, -0.09097366780042648, 0.014225414022803307, -0.06440398842096329, -0.0720076858997345, 0.1059345230460167, 0.06386196613311768, 0.031837984919548035, -0.04153471067547798, -0.09472637623548508, -0.03624151647090912, 0.1907484531402588, -0.18053385615348816, -0.07049532979726791 ]
null
null
transformers
# Cross-Encoder for MS Marco This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco) ## Usage with Transformers ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('model_name') tokenizer = AutoTokenizer.from_pretrained('model_name') features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits print(scores) ``` ## Usage with SentenceTransformers The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('model_name', max_length=512) scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')]) ``` ## Performance In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset. | Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec | | ------------- |:-------------| -----| --- | | **Version 2 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000 | cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100 | cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500 | cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800 | cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960 | **Version 1 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000 | cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900 | cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680 | cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340 | **Other models** | | | | nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900 | nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340 | nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100 | Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340 | amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330 | sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720 Note: Runtime was computed on a V100 GPU.
{"license": "apache-2.0"}
text-classification
cross-encoder/ms-marco-MiniLM-L-12-v2
[ "transformers", "pytorch", "jax", "bert", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
Cross-Encoder for MS Marco ========================== This model was trained on the MS Marco Passage Ranking task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco Usage with Transformers ----------------------- Usage with SentenceTransformers ------------------------------- The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this: Performance ----------- In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset. Note: Runtime was computed on a V100 GPU.
[]
[ "TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ 51 ]
[ "passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ -0.005313366185873747, 0.12051703780889511, -0.004693866241723299, 0.03580747917294502, 0.08237535506486893, 0.015574881806969643, 0.09972856193780899, 0.1512490212917328, -0.002342487219721079, -0.035675179213285446, 0.1512344777584076, 0.1678246557712555, -0.03263122960925102, 0.07377322763204575, -0.06331735849380493, -0.23472362756729126, 0.10480628162622452, 0.022721635177731514, -0.06391183286905289, 0.0852574035525322, 0.11916844546794891, -0.06770222634077072, 0.04979204759001732, -0.006682221312075853, -0.07624185085296631, 0.03358222916722298, 0.023558329790830612, -0.11514781415462494, 0.09597253054380417, 0.031052103266119957, 0.10091041028499603, 0.054762229323387146, -0.03489936888217926, -0.1495836228132248, 0.026294076815247536, 0.029823211953043938, -0.08911542594432831, 0.07207200676202774, 0.03851554915308952, -0.03519143536686897, 0.04294948652386665, 0.0052515072748064995, -0.0014171735383570194, 0.028817106038331985, -0.09800101071596146, -0.21412332355976105, -0.07283638417720795, 0.07232868671417236, 0.027582870796322823, 0.07712103426456451, 0.024217145517468452, 0.15384386479854584, -0.14576973021030426, 0.05114344507455826, 0.14506278932094574, -0.3469622731208801, -0.0045180246233940125, 0.13070425391197205, 0.07035423070192337, 0.04190012812614441, -0.03805508464574814, 0.04657916724681854, 0.05582089722156525, 0.0025910616386681795, 0.06670897454023361, -0.04245023429393768, -0.1239764392375946, 0.06446674466133118, -0.07143140584230423, -0.07907963544130325, 0.2631944417953491, -0.019031910225749016, 0.0633535161614418, -0.020391980186104774, -0.07111599296331406, -0.02444433979690075, 0.0007090915460139513, 0.04914804920554161, 0.029526568949222565, 0.10229268670082092, 0.07079360634088516, -0.03269055485725403, -0.14858068525791168, 0.020561350509524345, -0.18860138952732086, 0.06773007661104202, 0.0031633847393095493, 0.07213912159204483, -0.1380876749753952, 0.05910053104162216, 0.022541260346770287, -0.12431655079126358, 0.026827946305274963, -0.08040676265954971, 0.09522116184234619, 0.02466272935271263, -0.10294348746538162, 0.048924896866083145, 0.12230401486158371, 0.1840950846672058, 0.025208648294210434, 0.004822383169084787, -0.026003237813711166, 0.12904393672943115, 0.010278739035129547, 0.08728154003620148, -0.017540551722049713, -0.016535868868231773, 0.07604997605085373, -0.07513169944286346, 0.056044477969408035, -0.06161702796816826, -0.1701550930738449, -0.03133761137723923, 0.03735613077878952, 0.09264276921749115, 0.08358346670866013, 0.05583064258098602, -0.02714722789824009, 0.025454645976424217, 0.1282920092344284, -0.06134769693017006, 0.04080605506896973, -0.00510947173461318, 0.03630151227116585, 0.02500070445239544, 0.02943694032728672, 0.023852426558732986, -0.017657775431871414, 0.06957349926233292, -0.06750348955392838, -0.010275444015860558, -0.051824845373630524, -0.060393817722797394, 0.08819574117660522, -0.0934952050447464, 0.03543853387236595, -0.15935862064361572, -0.08309947699308395, 0.017079420387744904, 0.06980738043785095, 0.0021625745575875044, -0.09810265898704529, 0.07114944607019424, -0.04728621616959572, 0.0843905657529831, -0.08494804799556732, 0.002027006121352315, -0.09370028227567673, 0.05686311423778534, -0.10129949450492859, 0.055879801511764526, -0.17257225513458252, 0.06273247301578522, -0.1052277535200119, -0.0019254190847277641, -0.05104181170463562, -0.04737512394785881, -0.09146487712860107, 0.16669559478759766, -0.022341273725032806, -0.0318073108792305, -0.019266171380877495, 0.031578149646520615, -0.04990849643945694, 0.08952520042657852, -0.13634717464447021, -0.03626799210906029, 0.14703963696956635, -0.09130478650331497, -0.16326141357421875, 0.0717688575387001, 0.00046021450543776155, 0.0031928219832479954, 0.0205746628344059, 0.2063581943511963, 0.0636834129691124, -0.038732003420591354, 0.0429588183760643, 0.14875924587249756, -0.031094709411263466, -0.15359465777873993, 0.06786390393972397, -0.029294706881046295, -0.09197673946619034, 0.043804608285427094, -0.017092304304242134, 0.06857074052095413, 0.011296656914055347, -0.08280479162931442, -0.07357008755207062, -0.03364529088139534, 0.04349697753787041, 0.0174326840788126, 0.09276469796895981, -0.07873370498418808, -0.03898242488503456, -0.0033968340139836073, 0.03604090213775635, 0.05958762764930725, 0.056634191423654556, -0.03324146941304207, 0.10759710520505905, -0.00020664279873017222, 0.019652394577860832, -0.1466042846441269, -0.022884350270032883, -0.024007568135857582, 0.029566245153546333, -0.0034570449497550726, 0.1761774867773056, 0.028720151633024216, -0.09089623391628265, -0.005154056008905172, -0.005816992837935686, 0.10818760842084885, 0.07017345726490021, -0.0435544028878212, -0.122578464448452, 0.02524160034954548, -0.05991243198513985, 0.011386431753635406, -0.028770629316568375, 0.03522561118006706, 0.0614837110042572, 0.095530666410923, -0.04097152501344681, 0.10141308605670929, -0.024964042007923126, 0.0062868064269423485, -0.08510880917310715, 0.002189264167100191, 0.09858357906341553, 0.033706147223711014, -0.04600461944937706, 0.2194674164056778, -0.10151731222867966, 0.2976343631744385, 0.2267475426197052, -0.1751745343208313, 0.057022739201784134, 0.03313778340816498, -0.019422493875026703, 0.019689708948135376, 0.0130884675309062, -0.001759985345415771, -0.028527354821562767, -0.02610705979168415, 0.14932112395763397, -0.04963885620236397, -0.03732553496956825, -0.023624418303370476, -0.06671696156263351, -0.03914187103509903, 0.04179641976952553, 0.09773124009370804, -0.15121658146381378, 0.21856796741485596, 0.39029189944267273, -0.04813329502940178, 0.11297299712896347, -0.055986665189266205, 0.025421306490898132, 0.030753333121538162, -0.07714325934648514, -0.05503294616937637, 0.02333962731063366, -0.13534876704216003, -0.02139679528772831, 0.1016768291592598, 0.056578896939754486, 0.04646281898021698, -0.13725902140140533, -0.0547390952706337, 0.02008267678320408, 0.020364783704280853, -0.05890339985489845, 0.08217864483594894, 0.018010256811976433, 0.09777913242578506, -0.0105742197483778, -0.10413795709609985, 0.10512387752532959, -0.0025146319530904293, -0.046441737562417984, 0.12223948538303375, -0.1751604825258255, -0.26337674260139465, -0.10911697149276733, -0.10594077408313751, -0.0219988115131855, 0.016123976558446884, 0.14658363163471222, -0.05834493041038513, -0.04452504217624664, -0.03016248345375061, -0.08486530184745789, -0.06864798069000244, 0.03474634513258934, -0.04818902164697647, 0.05575546622276306, 0.0009074720437638462, -0.1152714416384697, -0.06752760708332062, 0.020496297627687454, -0.03415825590491295, 0.091299869120121, -0.030293690040707588, 0.07342889904975891, 0.135502427816391, 0.007198224309831858, 0.027944006025791168, -0.030184734612703323, 0.16622070968151093, -0.03921347111463547, 0.00924068596214056, 0.19809506833553314, -0.0031844095792621374, 0.0713656023144722, 0.18086189031600952, 0.03603030741214752, -0.022779395803809166, 0.01003937516361475, -0.03807242214679718, -0.08706989139318466, -0.20672155916690826, -0.11673396080732346, -0.12385633587837219, 0.03123038075864315, 0.058827221393585205, 0.10125505924224854, 0.1326202005147934, 0.05287611857056618, 0.0016830338863655925, 0.002457578666508198, -0.029018420726060867, 0.06340944021940231, 0.23763202130794525, -0.03204694390296936, 0.14382527768611908, -0.10360123962163925, -0.058313898742198944, 0.11185727268457413, 0.07886284589767456, 0.08192996680736542, 0.10665401816368103, 0.03152311220765114, 0.0695836991071701, 0.19727741181850433, 0.10917509347200394, 0.08815993368625641, 0.02420349419116974, -0.010355910286307335, -0.06285016238689423, -0.023843849077820778, -0.023379191756248474, 0.06658007949590683, 0.039418045431375504, -0.13914524018764496, -0.029837213456630707, -0.19487792253494263, 0.08332343399524689, 0.151483952999115, 0.059639498591423035, -0.13124634325504303, 0.031023966148495674, 0.10783152282238007, -0.01136285811662674, -0.04613389074802399, 0.08763978630304337, -0.02930242381989956, -0.08054038137197495, 0.07484424859285355, 0.007519237697124481, 0.12197799980640411, 0.02354556694626808, 0.07243271917104721, -0.04469209909439087, -0.16604083776474, 0.056991759687662125, 0.12605127692222595, -0.2823342978954315, 0.20038354396820068, -0.02226925455033779, -0.09456770122051239, -0.07824184000492096, 0.0034344338346272707, 0.07712201029062271, 0.22976431250572205, 0.061447788029909134, 0.03248010203242302, -0.1648344248533249, -0.04715387150645256, -0.03496819734573364, 0.012963905930519104, 0.017423907294869423, -0.02331795170903206, -0.03611692786216736, -0.06621395796537399, -0.002671534428372979, 0.020914889872074127, 0.1569015085697174, -0.022213159129023552, -0.17191611230373383, 0.06727875024080276, 0.11111561954021454, 0.02136494219303131, -0.04995092377066612, -0.043893180787563324, -0.16064417362213135, 0.16429398953914642, -0.034801822155714035, -0.03848220035433769, -0.09644030034542084, -0.11785909533500671, 0.05767476186156273, -0.043937042355537415, 0.05455517768859863, -0.08362634479999542, 0.027540860697627068, -0.06371035426855087, -0.19031928479671478, 0.11500367522239685, -0.12582644820213318, -0.04196672886610031, -0.053986817598342896, 0.08883348852396011, -0.15042348206043243, 0.04660613462328911, 0.03068644367158413, 0.06134520471096039, -0.15422704815864563, -0.11111470311880112, -0.0030049297492951155, 0.01256405096501112, 0.059495214372873306, -0.03420129418373108, -0.07899564504623413, -0.04707137122750282, 0.06431546807289124, -0.03068653494119644, 0.25191134214401245, 0.18924297392368317, -0.12640511989593506, 0.17781375348567963, 0.13471996784210205, -0.07904026657342911, -0.3402675986289978, -0.1414533108472824, -0.16139845550060272, -0.07071840018033981, 0.026064196601510048, -0.11777465045452118, 0.0958787128329277, 0.03335252031683922, -0.10947048664093018, 0.052748363465070724, -0.17017929255962372, -0.07672762125730515, 0.17991377413272858, -0.07845299690961838, 0.32294031977653503, -0.15600302815437317, -0.0581873282790184, -0.07140202820301056, -0.1397649347782135, 0.15831077098846436, -0.12095484882593155, 0.0640173852443695, -0.01085357554256916, 0.0009923696052283049, 0.000617707846686244, -0.05387040227651596, 0.15228526294231415, -0.05112951993942261, 0.02527214027941227, -0.1299176663160324, -0.04767915979027748, 0.10295214504003525, -0.06544466316699982, 0.023311404511332512, -0.1613304764032364, 0.012056032195687294, -0.1437864750623703, 0.010370835661888123, -0.08332697302103043, 0.08352860808372498, -0.009336385875940323, -0.040623150765895844, -0.05954933911561966, -0.005279048811644316, 0.04284205287694931, -0.015335856936872005, 0.29630112648010254, 0.008494893088936806, 0.13189546763896942, 0.16303826868534088, 0.022608771920204163, -0.1718173325061798, -0.01953105255961418, -0.03700250759720802, -0.0736953616142273, 0.07995044440031052, -0.2031794935464859, 0.057182274758815765, 0.07621432095766068, -0.06988967210054398, 0.06107276678085327, 0.08900803327560425, 0.01489779632538557, -0.04548985883593559, 0.16442106664180756, -0.1671867072582245, -0.0006360075785778463, -0.008117838762700558, 0.10618383437395096, 0.0521981455385685, 0.017339367419481277, 0.12554024159908295, 0.005745370872318745, -0.036925509572029114, 0.026103302836418152, 0.021035362035036087, -0.0628022700548172, 0.008186271414160728, 0.0677781030535698, 0.028965607285499573, -0.1132744550704956, 0.07926022261381149, 0.06677495688199997, -0.129411518573761, -0.010128652676939964, 0.09587135165929794, -0.1095706969499588, -0.16171106696128845, 0.00618077302351594, 0.04380159080028534, -0.08738097548484802, -0.08622825890779495, -0.024138811975717545, -0.1249731257557869, 0.04831845685839653, 0.09633947908878326, 0.13090389966964722, 0.06443681567907333, -0.0017022290267050266, -0.06870454549789429, 0.05082831531763077, 0.001563939149491489, -0.07709051668643951, 0.036532118916511536, -0.12133266031742096, -0.02846197411417961, 0.010334618389606476, 0.11554903537034988, -0.06259089708328247, -0.0042324368841946125, -0.12426569312810898, 0.001558715826831758, -0.11614777147769928, -0.029476512223482132, -0.08306972682476044, -0.033092986792325974, 0.02484716661274433, -0.08117660880088806, -0.04502910003066063, -0.01520068384706974, -0.14415021240711212, -0.024689702317118645, -0.04193141311407089, 0.09420331567525864, -0.104824498295784, -0.05173527076840401, 0.09438075870275497, -0.011454490944743156, 0.09049908816814423, 0.0417814776301384, -0.05572384595870972, 0.07632939517498016, -0.07947538793087006, -0.10600963234901428, 0.08863865584135056, 0.03787854313850403, 0.046659957617521286, -0.005956739187240601, -0.011253931559622288, 0.09342904388904572, -0.0007177229854278266, 0.042248114943504333, 0.011596323922276497, -0.12917691469192505, -0.02238038182258606, -0.010726326145231724, -0.13931554555892944, 0.015678487718105316, -0.09275448322296143, 0.14498212933540344, 0.01790434494614601, 0.17994743585586548, -0.005748189985752106, 0.031212467700242996, -0.06608235090970993, 0.025236306712031364, -0.057449471205472946, -0.17951899766921997, -0.12403986603021622, -0.06836044043302536, -0.027640171349048615, -0.028845123946666718, 0.2805347144603729, 0.08917142450809479, -0.08347166329622269, 0.07290034741163254, 0.07196228206157684, 0.01505149994045496, 0.02077554725110531, 0.22023816406726837, 0.06361716240644455, -0.02245054394006729, -0.061880506575107574, 0.019000306725502014, 0.022344766184687614, -0.03256882727146149, 0.07464652508497238, 0.1169479712843895, 0.1142251044511795, 0.07622475177049637, 0.02704872190952301, -0.016070278361439705, -0.11697600036859512, -0.1287461221218109, 0.014688584953546524, 0.119608573615551, -0.03686606511473656, 0.059426259249448776, 0.13529466092586517, -0.037811972200870514, 0.03956560790538788, -0.086575448513031, 0.012540166266262531, -0.17560049891471863, -0.09157182276248932, -0.07360459119081497, -0.13344724476337433, -0.03314107656478882, -0.052344802767038345, 0.03984220325946808, 0.1292441338300705, 0.032210953533649445, -0.03851942718029022, -0.015154266729950905, -0.015868499875068665, -0.04258604347705841, 0.03028041310608387, -0.023542780429124832, 0.0002562306181062013, -0.06029721349477768, -0.02395167201757431, -0.10368048399686813, -0.02776411361992359, -0.05679041147232056, 0.04382525384426117, -0.027991825714707375, 0.021414659917354584, -0.12156787514686584, -0.08556406199932098, -0.0529513917863369, 0.020444447174668312, -0.005759124178439379, 0.18267738819122314, 0.005337047390639782, 0.031859856098890305, 0.07273654639720917, 0.2198667675256729, -0.09987863898277283, -0.1095680221915245, -0.045107651501894, 0.17264553904533386, 0.04104243591427803, 0.06176810711622238, -0.020409779623150826, 0.0026832649018615484, -0.09016072750091553, 0.26033926010131836, 0.33503320813179016, -0.05959988385438919, 0.06000038608908653, -0.001532332506030798, 0.012618477456271648, 0.0598568320274353, 0.1106472909450531, 0.1316719800233841, 0.16129449009895325, -0.08782772719860077, -0.0036300979554653168, -0.04469006136059761, 0.003653900232166052, -0.15766099095344543, 0.07017975300550461, -0.0036263535730540752, -0.09436394274234772, -0.022905346006155014, 0.057690780609846115, -0.09149443358182907, 0.07677844911813736, 0.012266093865036964, -0.18119388818740845, -0.05720311030745506, 0.02039954625070095, 0.21919496357440948, -0.017752492800354958, 0.05047750845551491, -0.03148294985294342, -0.05256124958395958, 0.07185778021812439, -0.028453262522816658, -0.16809456050395966, -0.052983757108449936, 0.11245088279247284, -0.033219534903764725, 0.11339452862739563, -0.01019338145852089, 0.02442827820777893, 0.09866058081388474, 0.06523024290800095, -0.09986453503370285, 0.06829454749822617, 0.024860212579369545, -0.06409720331430435, -0.025689758360385895, -0.10032090544700623, -0.02310752309858799, -0.06327763199806213, 0.06990473717451096, -0.10408949106931686, 0.047875236719846725, -0.02276439219713211, -0.041091352701187134, -0.03285158798098564, 0.01771629974246025, -0.06046866253018379, 0.07364235818386078, 0.030026888474822044, -0.036611031740903854, -0.07146476209163666, -0.04183752462267876, -0.04579772427678108, 0.008587779477238655, -0.16310185194015503, -0.11530570685863495, -0.004291696939617395, -0.022359905764460564, 0.07707220315933228, 0.03829155117273331, -0.09312435239553452, -0.03533054143190384, -0.05517655238509178, 0.0333234928548336, -0.12787498533725739, 0.0459769070148468, 0.08696694672107697, -0.015718704089522362, -0.010699767619371414, -0.06356548517942429, 0.019441908225417137, 0.023609060794115067, -0.10050466656684875, -0.07246299833059311 ]
null
null
transformers
# Cross-Encoder for MS Marco This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco) ## Usage with Transformers ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('model_name') tokenizer = AutoTokenizer.from_pretrained('model_name') features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits print(scores) ``` ## Usage with SentenceTransformers The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('model_name', max_length=512) scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')]) ``` ## Performance In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset. | Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec | | ------------- |:-------------| -----| --- | | **Version 2 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000 | cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100 | cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500 | cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800 | cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960 | **Version 1 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000 | cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900 | cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680 | cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340 | **Other models** | | | | nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900 | nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340 | nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100 | Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340 | amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330 | sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720 Note: Runtime was computed on a V100 GPU.
{"license": "apache-2.0"}
text-classification
cross-encoder/ms-marco-MiniLM-L-2-v2
[ "transformers", "pytorch", "jax", "bert", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
Cross-Encoder for MS Marco ========================== This model was trained on the MS Marco Passage Ranking task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco Usage with Transformers ----------------------- Usage with SentenceTransformers ------------------------------- The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this: Performance ----------- In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset. Note: Runtime was computed on a V100 GPU.
[]
[ "TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ 51 ]
[ "passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ -0.005313366185873747, 0.12051703780889511, -0.004693866241723299, 0.03580747917294502, 0.08237535506486893, 0.015574881806969643, 0.09972856193780899, 0.1512490212917328, -0.002342487219721079, -0.035675179213285446, 0.1512344777584076, 0.1678246557712555, -0.03263122960925102, 0.07377322763204575, -0.06331735849380493, -0.23472362756729126, 0.10480628162622452, 0.022721635177731514, -0.06391183286905289, 0.0852574035525322, 0.11916844546794891, -0.06770222634077072, 0.04979204759001732, -0.006682221312075853, -0.07624185085296631, 0.03358222916722298, 0.023558329790830612, -0.11514781415462494, 0.09597253054380417, 0.031052103266119957, 0.10091041028499603, 0.054762229323387146, -0.03489936888217926, -0.1495836228132248, 0.026294076815247536, 0.029823211953043938, -0.08911542594432831, 0.07207200676202774, 0.03851554915308952, -0.03519143536686897, 0.04294948652386665, 0.0052515072748064995, -0.0014171735383570194, 0.028817106038331985, -0.09800101071596146, -0.21412332355976105, -0.07283638417720795, 0.07232868671417236, 0.027582870796322823, 0.07712103426456451, 0.024217145517468452, 0.15384386479854584, -0.14576973021030426, 0.05114344507455826, 0.14506278932094574, -0.3469622731208801, -0.0045180246233940125, 0.13070425391197205, 0.07035423070192337, 0.04190012812614441, -0.03805508464574814, 0.04657916724681854, 0.05582089722156525, 0.0025910616386681795, 0.06670897454023361, -0.04245023429393768, -0.1239764392375946, 0.06446674466133118, -0.07143140584230423, -0.07907963544130325, 0.2631944417953491, -0.019031910225749016, 0.0633535161614418, -0.020391980186104774, -0.07111599296331406, -0.02444433979690075, 0.0007090915460139513, 0.04914804920554161, 0.029526568949222565, 0.10229268670082092, 0.07079360634088516, -0.03269055485725403, -0.14858068525791168, 0.020561350509524345, -0.18860138952732086, 0.06773007661104202, 0.0031633847393095493, 0.07213912159204483, -0.1380876749753952, 0.05910053104162216, 0.022541260346770287, -0.12431655079126358, 0.026827946305274963, -0.08040676265954971, 0.09522116184234619, 0.02466272935271263, -0.10294348746538162, 0.048924896866083145, 0.12230401486158371, 0.1840950846672058, 0.025208648294210434, 0.004822383169084787, -0.026003237813711166, 0.12904393672943115, 0.010278739035129547, 0.08728154003620148, -0.017540551722049713, -0.016535868868231773, 0.07604997605085373, -0.07513169944286346, 0.056044477969408035, -0.06161702796816826, -0.1701550930738449, -0.03133761137723923, 0.03735613077878952, 0.09264276921749115, 0.08358346670866013, 0.05583064258098602, -0.02714722789824009, 0.025454645976424217, 0.1282920092344284, -0.06134769693017006, 0.04080605506896973, -0.00510947173461318, 0.03630151227116585, 0.02500070445239544, 0.02943694032728672, 0.023852426558732986, -0.017657775431871414, 0.06957349926233292, -0.06750348955392838, -0.010275444015860558, -0.051824845373630524, -0.060393817722797394, 0.08819574117660522, -0.0934952050447464, 0.03543853387236595, -0.15935862064361572, -0.08309947699308395, 0.017079420387744904, 0.06980738043785095, 0.0021625745575875044, -0.09810265898704529, 0.07114944607019424, -0.04728621616959572, 0.0843905657529831, -0.08494804799556732, 0.002027006121352315, -0.09370028227567673, 0.05686311423778534, -0.10129949450492859, 0.055879801511764526, -0.17257225513458252, 0.06273247301578522, -0.1052277535200119, -0.0019254190847277641, -0.05104181170463562, -0.04737512394785881, -0.09146487712860107, 0.16669559478759766, -0.022341273725032806, -0.0318073108792305, -0.019266171380877495, 0.031578149646520615, -0.04990849643945694, 0.08952520042657852, -0.13634717464447021, -0.03626799210906029, 0.14703963696956635, -0.09130478650331497, -0.16326141357421875, 0.0717688575387001, 0.00046021450543776155, 0.0031928219832479954, 0.0205746628344059, 0.2063581943511963, 0.0636834129691124, -0.038732003420591354, 0.0429588183760643, 0.14875924587249756, -0.031094709411263466, -0.15359465777873993, 0.06786390393972397, -0.029294706881046295, -0.09197673946619034, 0.043804608285427094, -0.017092304304242134, 0.06857074052095413, 0.011296656914055347, -0.08280479162931442, -0.07357008755207062, -0.03364529088139534, 0.04349697753787041, 0.0174326840788126, 0.09276469796895981, -0.07873370498418808, -0.03898242488503456, -0.0033968340139836073, 0.03604090213775635, 0.05958762764930725, 0.056634191423654556, -0.03324146941304207, 0.10759710520505905, -0.00020664279873017222, 0.019652394577860832, -0.1466042846441269, -0.022884350270032883, -0.024007568135857582, 0.029566245153546333, -0.0034570449497550726, 0.1761774867773056, 0.028720151633024216, -0.09089623391628265, -0.005154056008905172, -0.005816992837935686, 0.10818760842084885, 0.07017345726490021, -0.0435544028878212, -0.122578464448452, 0.02524160034954548, -0.05991243198513985, 0.011386431753635406, -0.028770629316568375, 0.03522561118006706, 0.0614837110042572, 0.095530666410923, -0.04097152501344681, 0.10141308605670929, -0.024964042007923126, 0.0062868064269423485, -0.08510880917310715, 0.002189264167100191, 0.09858357906341553, 0.033706147223711014, -0.04600461944937706, 0.2194674164056778, -0.10151731222867966, 0.2976343631744385, 0.2267475426197052, -0.1751745343208313, 0.057022739201784134, 0.03313778340816498, -0.019422493875026703, 0.019689708948135376, 0.0130884675309062, -0.001759985345415771, -0.028527354821562767, -0.02610705979168415, 0.14932112395763397, -0.04963885620236397, -0.03732553496956825, -0.023624418303370476, -0.06671696156263351, -0.03914187103509903, 0.04179641976952553, 0.09773124009370804, -0.15121658146381378, 0.21856796741485596, 0.39029189944267273, -0.04813329502940178, 0.11297299712896347, -0.055986665189266205, 0.025421306490898132, 0.030753333121538162, -0.07714325934648514, -0.05503294616937637, 0.02333962731063366, -0.13534876704216003, -0.02139679528772831, 0.1016768291592598, 0.056578896939754486, 0.04646281898021698, -0.13725902140140533, -0.0547390952706337, 0.02008267678320408, 0.020364783704280853, -0.05890339985489845, 0.08217864483594894, 0.018010256811976433, 0.09777913242578506, -0.0105742197483778, -0.10413795709609985, 0.10512387752532959, -0.0025146319530904293, -0.046441737562417984, 0.12223948538303375, -0.1751604825258255, -0.26337674260139465, -0.10911697149276733, -0.10594077408313751, -0.0219988115131855, 0.016123976558446884, 0.14658363163471222, -0.05834493041038513, -0.04452504217624664, -0.03016248345375061, -0.08486530184745789, -0.06864798069000244, 0.03474634513258934, -0.04818902164697647, 0.05575546622276306, 0.0009074720437638462, -0.1152714416384697, -0.06752760708332062, 0.020496297627687454, -0.03415825590491295, 0.091299869120121, -0.030293690040707588, 0.07342889904975891, 0.135502427816391, 0.007198224309831858, 0.027944006025791168, -0.030184734612703323, 0.16622070968151093, -0.03921347111463547, 0.00924068596214056, 0.19809506833553314, -0.0031844095792621374, 0.0713656023144722, 0.18086189031600952, 0.03603030741214752, -0.022779395803809166, 0.01003937516361475, -0.03807242214679718, -0.08706989139318466, -0.20672155916690826, -0.11673396080732346, -0.12385633587837219, 0.03123038075864315, 0.058827221393585205, 0.10125505924224854, 0.1326202005147934, 0.05287611857056618, 0.0016830338863655925, 0.002457578666508198, -0.029018420726060867, 0.06340944021940231, 0.23763202130794525, -0.03204694390296936, 0.14382527768611908, -0.10360123962163925, -0.058313898742198944, 0.11185727268457413, 0.07886284589767456, 0.08192996680736542, 0.10665401816368103, 0.03152311220765114, 0.0695836991071701, 0.19727741181850433, 0.10917509347200394, 0.08815993368625641, 0.02420349419116974, -0.010355910286307335, -0.06285016238689423, -0.023843849077820778, -0.023379191756248474, 0.06658007949590683, 0.039418045431375504, -0.13914524018764496, -0.029837213456630707, -0.19487792253494263, 0.08332343399524689, 0.151483952999115, 0.059639498591423035, -0.13124634325504303, 0.031023966148495674, 0.10783152282238007, -0.01136285811662674, -0.04613389074802399, 0.08763978630304337, -0.02930242381989956, -0.08054038137197495, 0.07484424859285355, 0.007519237697124481, 0.12197799980640411, 0.02354556694626808, 0.07243271917104721, -0.04469209909439087, -0.16604083776474, 0.056991759687662125, 0.12605127692222595, -0.2823342978954315, 0.20038354396820068, -0.02226925455033779, -0.09456770122051239, -0.07824184000492096, 0.0034344338346272707, 0.07712201029062271, 0.22976431250572205, 0.061447788029909134, 0.03248010203242302, -0.1648344248533249, -0.04715387150645256, -0.03496819734573364, 0.012963905930519104, 0.017423907294869423, -0.02331795170903206, -0.03611692786216736, -0.06621395796537399, -0.002671534428372979, 0.020914889872074127, 0.1569015085697174, -0.022213159129023552, -0.17191611230373383, 0.06727875024080276, 0.11111561954021454, 0.02136494219303131, -0.04995092377066612, -0.043893180787563324, -0.16064417362213135, 0.16429398953914642, -0.034801822155714035, -0.03848220035433769, -0.09644030034542084, -0.11785909533500671, 0.05767476186156273, -0.043937042355537415, 0.05455517768859863, -0.08362634479999542, 0.027540860697627068, -0.06371035426855087, -0.19031928479671478, 0.11500367522239685, -0.12582644820213318, -0.04196672886610031, -0.053986817598342896, 0.08883348852396011, -0.15042348206043243, 0.04660613462328911, 0.03068644367158413, 0.06134520471096039, -0.15422704815864563, -0.11111470311880112, -0.0030049297492951155, 0.01256405096501112, 0.059495214372873306, -0.03420129418373108, -0.07899564504623413, -0.04707137122750282, 0.06431546807289124, -0.03068653494119644, 0.25191134214401245, 0.18924297392368317, -0.12640511989593506, 0.17781375348567963, 0.13471996784210205, -0.07904026657342911, -0.3402675986289978, -0.1414533108472824, -0.16139845550060272, -0.07071840018033981, 0.026064196601510048, -0.11777465045452118, 0.0958787128329277, 0.03335252031683922, -0.10947048664093018, 0.052748363465070724, -0.17017929255962372, -0.07672762125730515, 0.17991377413272858, -0.07845299690961838, 0.32294031977653503, -0.15600302815437317, -0.0581873282790184, -0.07140202820301056, -0.1397649347782135, 0.15831077098846436, -0.12095484882593155, 0.0640173852443695, -0.01085357554256916, 0.0009923696052283049, 0.000617707846686244, -0.05387040227651596, 0.15228526294231415, -0.05112951993942261, 0.02527214027941227, -0.1299176663160324, -0.04767915979027748, 0.10295214504003525, -0.06544466316699982, 0.023311404511332512, -0.1613304764032364, 0.012056032195687294, -0.1437864750623703, 0.010370835661888123, -0.08332697302103043, 0.08352860808372498, -0.009336385875940323, -0.040623150765895844, -0.05954933911561966, -0.005279048811644316, 0.04284205287694931, -0.015335856936872005, 0.29630112648010254, 0.008494893088936806, 0.13189546763896942, 0.16303826868534088, 0.022608771920204163, -0.1718173325061798, -0.01953105255961418, -0.03700250759720802, -0.0736953616142273, 0.07995044440031052, -0.2031794935464859, 0.057182274758815765, 0.07621432095766068, -0.06988967210054398, 0.06107276678085327, 0.08900803327560425, 0.01489779632538557, -0.04548985883593559, 0.16442106664180756, -0.1671867072582245, -0.0006360075785778463, -0.008117838762700558, 0.10618383437395096, 0.0521981455385685, 0.017339367419481277, 0.12554024159908295, 0.005745370872318745, -0.036925509572029114, 0.026103302836418152, 0.021035362035036087, -0.0628022700548172, 0.008186271414160728, 0.0677781030535698, 0.028965607285499573, -0.1132744550704956, 0.07926022261381149, 0.06677495688199997, -0.129411518573761, -0.010128652676939964, 0.09587135165929794, -0.1095706969499588, -0.16171106696128845, 0.00618077302351594, 0.04380159080028534, -0.08738097548484802, -0.08622825890779495, -0.024138811975717545, -0.1249731257557869, 0.04831845685839653, 0.09633947908878326, 0.13090389966964722, 0.06443681567907333, -0.0017022290267050266, -0.06870454549789429, 0.05082831531763077, 0.001563939149491489, -0.07709051668643951, 0.036532118916511536, -0.12133266031742096, -0.02846197411417961, 0.010334618389606476, 0.11554903537034988, -0.06259089708328247, -0.0042324368841946125, -0.12426569312810898, 0.001558715826831758, -0.11614777147769928, -0.029476512223482132, -0.08306972682476044, -0.033092986792325974, 0.02484716661274433, -0.08117660880088806, -0.04502910003066063, -0.01520068384706974, -0.14415021240711212, -0.024689702317118645, -0.04193141311407089, 0.09420331567525864, -0.104824498295784, -0.05173527076840401, 0.09438075870275497, -0.011454490944743156, 0.09049908816814423, 0.0417814776301384, -0.05572384595870972, 0.07632939517498016, -0.07947538793087006, -0.10600963234901428, 0.08863865584135056, 0.03787854313850403, 0.046659957617521286, -0.005956739187240601, -0.011253931559622288, 0.09342904388904572, -0.0007177229854278266, 0.042248114943504333, 0.011596323922276497, -0.12917691469192505, -0.02238038182258606, -0.010726326145231724, -0.13931554555892944, 0.015678487718105316, -0.09275448322296143, 0.14498212933540344, 0.01790434494614601, 0.17994743585586548, -0.005748189985752106, 0.031212467700242996, -0.06608235090970993, 0.025236306712031364, -0.057449471205472946, -0.17951899766921997, -0.12403986603021622, -0.06836044043302536, -0.027640171349048615, -0.028845123946666718, 0.2805347144603729, 0.08917142450809479, -0.08347166329622269, 0.07290034741163254, 0.07196228206157684, 0.01505149994045496, 0.02077554725110531, 0.22023816406726837, 0.06361716240644455, -0.02245054394006729, -0.061880506575107574, 0.019000306725502014, 0.022344766184687614, -0.03256882727146149, 0.07464652508497238, 0.1169479712843895, 0.1142251044511795, 0.07622475177049637, 0.02704872190952301, -0.016070278361439705, -0.11697600036859512, -0.1287461221218109, 0.014688584953546524, 0.119608573615551, -0.03686606511473656, 0.059426259249448776, 0.13529466092586517, -0.037811972200870514, 0.03956560790538788, -0.086575448513031, 0.012540166266262531, -0.17560049891471863, -0.09157182276248932, -0.07360459119081497, -0.13344724476337433, -0.03314107656478882, -0.052344802767038345, 0.03984220325946808, 0.1292441338300705, 0.032210953533649445, -0.03851942718029022, -0.015154266729950905, -0.015868499875068665, -0.04258604347705841, 0.03028041310608387, -0.023542780429124832, 0.0002562306181062013, -0.06029721349477768, -0.02395167201757431, -0.10368048399686813, -0.02776411361992359, -0.05679041147232056, 0.04382525384426117, -0.027991825714707375, 0.021414659917354584, -0.12156787514686584, -0.08556406199932098, -0.0529513917863369, 0.020444447174668312, -0.005759124178439379, 0.18267738819122314, 0.005337047390639782, 0.031859856098890305, 0.07273654639720917, 0.2198667675256729, -0.09987863898277283, -0.1095680221915245, -0.045107651501894, 0.17264553904533386, 0.04104243591427803, 0.06176810711622238, -0.020409779623150826, 0.0026832649018615484, -0.09016072750091553, 0.26033926010131836, 0.33503320813179016, -0.05959988385438919, 0.06000038608908653, -0.001532332506030798, 0.012618477456271648, 0.0598568320274353, 0.1106472909450531, 0.1316719800233841, 0.16129449009895325, -0.08782772719860077, -0.0036300979554653168, -0.04469006136059761, 0.003653900232166052, -0.15766099095344543, 0.07017975300550461, -0.0036263535730540752, -0.09436394274234772, -0.022905346006155014, 0.057690780609846115, -0.09149443358182907, 0.07677844911813736, 0.012266093865036964, -0.18119388818740845, -0.05720311030745506, 0.02039954625070095, 0.21919496357440948, -0.017752492800354958, 0.05047750845551491, -0.03148294985294342, -0.05256124958395958, 0.07185778021812439, -0.028453262522816658, -0.16809456050395966, -0.052983757108449936, 0.11245088279247284, -0.033219534903764725, 0.11339452862739563, -0.01019338145852089, 0.02442827820777893, 0.09866058081388474, 0.06523024290800095, -0.09986453503370285, 0.06829454749822617, 0.024860212579369545, -0.06409720331430435, -0.025689758360385895, -0.10032090544700623, -0.02310752309858799, -0.06327763199806213, 0.06990473717451096, -0.10408949106931686, 0.047875236719846725, -0.02276439219713211, -0.041091352701187134, -0.03285158798098564, 0.01771629974246025, -0.06046866253018379, 0.07364235818386078, 0.030026888474822044, -0.036611031740903854, -0.07146476209163666, -0.04183752462267876, -0.04579772427678108, 0.008587779477238655, -0.16310185194015503, -0.11530570685863495, -0.004291696939617395, -0.022359905764460564, 0.07707220315933228, 0.03829155117273331, -0.09312435239553452, -0.03533054143190384, -0.05517655238509178, 0.0333234928548336, -0.12787498533725739, 0.0459769070148468, 0.08696694672107697, -0.015718704089522362, -0.010699767619371414, -0.06356548517942429, 0.019441908225417137, 0.023609060794115067, -0.10050466656684875, -0.07246299833059311 ]
null
null
transformers
# Cross-Encoder for MS Marco This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco) ## Usage with Transformers ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('model_name') tokenizer = AutoTokenizer.from_pretrained('model_name') features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits print(scores) ``` ## Usage with SentenceTransformers The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('model_name', max_length=512) scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')]) ``` ## Performance In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset. | Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec | | ------------- |:-------------| -----| --- | | **Version 2 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000 | cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100 | cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500 | cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800 | cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960 | **Version 1 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000 | cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900 | cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680 | cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340 | **Other models** | | | | nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900 | nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340 | nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100 | Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340 | amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330 | sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720 Note: Runtime was computed on a V100 GPU.
{"license": "apache-2.0"}
text-classification
cross-encoder/ms-marco-MiniLM-L-4-v2
[ "transformers", "pytorch", "jax", "bert", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
Cross-Encoder for MS Marco ========================== This model was trained on the MS Marco Passage Ranking task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco Usage with Transformers ----------------------- Usage with SentenceTransformers ------------------------------- The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this: Performance ----------- In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset. Note: Runtime was computed on a V100 GPU.
[]
[ "TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ 51 ]
[ "passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ -0.005313366185873747, 0.12051703780889511, -0.004693866241723299, 0.03580747917294502, 0.08237535506486893, 0.015574881806969643, 0.09972856193780899, 0.1512490212917328, -0.002342487219721079, -0.035675179213285446, 0.1512344777584076, 0.1678246557712555, -0.03263122960925102, 0.07377322763204575, -0.06331735849380493, -0.23472362756729126, 0.10480628162622452, 0.022721635177731514, -0.06391183286905289, 0.0852574035525322, 0.11916844546794891, -0.06770222634077072, 0.04979204759001732, -0.006682221312075853, -0.07624185085296631, 0.03358222916722298, 0.023558329790830612, -0.11514781415462494, 0.09597253054380417, 0.031052103266119957, 0.10091041028499603, 0.054762229323387146, -0.03489936888217926, -0.1495836228132248, 0.026294076815247536, 0.029823211953043938, -0.08911542594432831, 0.07207200676202774, 0.03851554915308952, -0.03519143536686897, 0.04294948652386665, 0.0052515072748064995, -0.0014171735383570194, 0.028817106038331985, -0.09800101071596146, -0.21412332355976105, -0.07283638417720795, 0.07232868671417236, 0.027582870796322823, 0.07712103426456451, 0.024217145517468452, 0.15384386479854584, -0.14576973021030426, 0.05114344507455826, 0.14506278932094574, -0.3469622731208801, -0.0045180246233940125, 0.13070425391197205, 0.07035423070192337, 0.04190012812614441, -0.03805508464574814, 0.04657916724681854, 0.05582089722156525, 0.0025910616386681795, 0.06670897454023361, -0.04245023429393768, -0.1239764392375946, 0.06446674466133118, -0.07143140584230423, -0.07907963544130325, 0.2631944417953491, -0.019031910225749016, 0.0633535161614418, -0.020391980186104774, -0.07111599296331406, -0.02444433979690075, 0.0007090915460139513, 0.04914804920554161, 0.029526568949222565, 0.10229268670082092, 0.07079360634088516, -0.03269055485725403, -0.14858068525791168, 0.020561350509524345, -0.18860138952732086, 0.06773007661104202, 0.0031633847393095493, 0.07213912159204483, -0.1380876749753952, 0.05910053104162216, 0.022541260346770287, -0.12431655079126358, 0.026827946305274963, -0.08040676265954971, 0.09522116184234619, 0.02466272935271263, -0.10294348746538162, 0.048924896866083145, 0.12230401486158371, 0.1840950846672058, 0.025208648294210434, 0.004822383169084787, -0.026003237813711166, 0.12904393672943115, 0.010278739035129547, 0.08728154003620148, -0.017540551722049713, -0.016535868868231773, 0.07604997605085373, -0.07513169944286346, 0.056044477969408035, -0.06161702796816826, -0.1701550930738449, -0.03133761137723923, 0.03735613077878952, 0.09264276921749115, 0.08358346670866013, 0.05583064258098602, -0.02714722789824009, 0.025454645976424217, 0.1282920092344284, -0.06134769693017006, 0.04080605506896973, -0.00510947173461318, 0.03630151227116585, 0.02500070445239544, 0.02943694032728672, 0.023852426558732986, -0.017657775431871414, 0.06957349926233292, -0.06750348955392838, -0.010275444015860558, -0.051824845373630524, -0.060393817722797394, 0.08819574117660522, -0.0934952050447464, 0.03543853387236595, -0.15935862064361572, -0.08309947699308395, 0.017079420387744904, 0.06980738043785095, 0.0021625745575875044, -0.09810265898704529, 0.07114944607019424, -0.04728621616959572, 0.0843905657529831, -0.08494804799556732, 0.002027006121352315, -0.09370028227567673, 0.05686311423778534, -0.10129949450492859, 0.055879801511764526, -0.17257225513458252, 0.06273247301578522, -0.1052277535200119, -0.0019254190847277641, -0.05104181170463562, -0.04737512394785881, -0.09146487712860107, 0.16669559478759766, -0.022341273725032806, -0.0318073108792305, -0.019266171380877495, 0.031578149646520615, -0.04990849643945694, 0.08952520042657852, -0.13634717464447021, -0.03626799210906029, 0.14703963696956635, -0.09130478650331497, -0.16326141357421875, 0.0717688575387001, 0.00046021450543776155, 0.0031928219832479954, 0.0205746628344059, 0.2063581943511963, 0.0636834129691124, -0.038732003420591354, 0.0429588183760643, 0.14875924587249756, -0.031094709411263466, -0.15359465777873993, 0.06786390393972397, -0.029294706881046295, -0.09197673946619034, 0.043804608285427094, -0.017092304304242134, 0.06857074052095413, 0.011296656914055347, -0.08280479162931442, -0.07357008755207062, -0.03364529088139534, 0.04349697753787041, 0.0174326840788126, 0.09276469796895981, -0.07873370498418808, -0.03898242488503456, -0.0033968340139836073, 0.03604090213775635, 0.05958762764930725, 0.056634191423654556, -0.03324146941304207, 0.10759710520505905, -0.00020664279873017222, 0.019652394577860832, -0.1466042846441269, -0.022884350270032883, -0.024007568135857582, 0.029566245153546333, -0.0034570449497550726, 0.1761774867773056, 0.028720151633024216, -0.09089623391628265, -0.005154056008905172, -0.005816992837935686, 0.10818760842084885, 0.07017345726490021, -0.0435544028878212, -0.122578464448452, 0.02524160034954548, -0.05991243198513985, 0.011386431753635406, -0.028770629316568375, 0.03522561118006706, 0.0614837110042572, 0.095530666410923, -0.04097152501344681, 0.10141308605670929, -0.024964042007923126, 0.0062868064269423485, -0.08510880917310715, 0.002189264167100191, 0.09858357906341553, 0.033706147223711014, -0.04600461944937706, 0.2194674164056778, -0.10151731222867966, 0.2976343631744385, 0.2267475426197052, -0.1751745343208313, 0.057022739201784134, 0.03313778340816498, -0.019422493875026703, 0.019689708948135376, 0.0130884675309062, -0.001759985345415771, -0.028527354821562767, -0.02610705979168415, 0.14932112395763397, -0.04963885620236397, -0.03732553496956825, -0.023624418303370476, -0.06671696156263351, -0.03914187103509903, 0.04179641976952553, 0.09773124009370804, -0.15121658146381378, 0.21856796741485596, 0.39029189944267273, -0.04813329502940178, 0.11297299712896347, -0.055986665189266205, 0.025421306490898132, 0.030753333121538162, -0.07714325934648514, -0.05503294616937637, 0.02333962731063366, -0.13534876704216003, -0.02139679528772831, 0.1016768291592598, 0.056578896939754486, 0.04646281898021698, -0.13725902140140533, -0.0547390952706337, 0.02008267678320408, 0.020364783704280853, -0.05890339985489845, 0.08217864483594894, 0.018010256811976433, 0.09777913242578506, -0.0105742197483778, -0.10413795709609985, 0.10512387752532959, -0.0025146319530904293, -0.046441737562417984, 0.12223948538303375, -0.1751604825258255, -0.26337674260139465, -0.10911697149276733, -0.10594077408313751, -0.0219988115131855, 0.016123976558446884, 0.14658363163471222, -0.05834493041038513, -0.04452504217624664, -0.03016248345375061, -0.08486530184745789, -0.06864798069000244, 0.03474634513258934, -0.04818902164697647, 0.05575546622276306, 0.0009074720437638462, -0.1152714416384697, -0.06752760708332062, 0.020496297627687454, -0.03415825590491295, 0.091299869120121, -0.030293690040707588, 0.07342889904975891, 0.135502427816391, 0.007198224309831858, 0.027944006025791168, -0.030184734612703323, 0.16622070968151093, -0.03921347111463547, 0.00924068596214056, 0.19809506833553314, -0.0031844095792621374, 0.0713656023144722, 0.18086189031600952, 0.03603030741214752, -0.022779395803809166, 0.01003937516361475, -0.03807242214679718, -0.08706989139318466, -0.20672155916690826, -0.11673396080732346, -0.12385633587837219, 0.03123038075864315, 0.058827221393585205, 0.10125505924224854, 0.1326202005147934, 0.05287611857056618, 0.0016830338863655925, 0.002457578666508198, -0.029018420726060867, 0.06340944021940231, 0.23763202130794525, -0.03204694390296936, 0.14382527768611908, -0.10360123962163925, -0.058313898742198944, 0.11185727268457413, 0.07886284589767456, 0.08192996680736542, 0.10665401816368103, 0.03152311220765114, 0.0695836991071701, 0.19727741181850433, 0.10917509347200394, 0.08815993368625641, 0.02420349419116974, -0.010355910286307335, -0.06285016238689423, -0.023843849077820778, -0.023379191756248474, 0.06658007949590683, 0.039418045431375504, -0.13914524018764496, -0.029837213456630707, -0.19487792253494263, 0.08332343399524689, 0.151483952999115, 0.059639498591423035, -0.13124634325504303, 0.031023966148495674, 0.10783152282238007, -0.01136285811662674, -0.04613389074802399, 0.08763978630304337, -0.02930242381989956, -0.08054038137197495, 0.07484424859285355, 0.007519237697124481, 0.12197799980640411, 0.02354556694626808, 0.07243271917104721, -0.04469209909439087, -0.16604083776474, 0.056991759687662125, 0.12605127692222595, -0.2823342978954315, 0.20038354396820068, -0.02226925455033779, -0.09456770122051239, -0.07824184000492096, 0.0034344338346272707, 0.07712201029062271, 0.22976431250572205, 0.061447788029909134, 0.03248010203242302, -0.1648344248533249, -0.04715387150645256, -0.03496819734573364, 0.012963905930519104, 0.017423907294869423, -0.02331795170903206, -0.03611692786216736, -0.06621395796537399, -0.002671534428372979, 0.020914889872074127, 0.1569015085697174, -0.022213159129023552, -0.17191611230373383, 0.06727875024080276, 0.11111561954021454, 0.02136494219303131, -0.04995092377066612, -0.043893180787563324, -0.16064417362213135, 0.16429398953914642, -0.034801822155714035, -0.03848220035433769, -0.09644030034542084, -0.11785909533500671, 0.05767476186156273, -0.043937042355537415, 0.05455517768859863, -0.08362634479999542, 0.027540860697627068, -0.06371035426855087, -0.19031928479671478, 0.11500367522239685, -0.12582644820213318, -0.04196672886610031, -0.053986817598342896, 0.08883348852396011, -0.15042348206043243, 0.04660613462328911, 0.03068644367158413, 0.06134520471096039, -0.15422704815864563, -0.11111470311880112, -0.0030049297492951155, 0.01256405096501112, 0.059495214372873306, -0.03420129418373108, -0.07899564504623413, -0.04707137122750282, 0.06431546807289124, -0.03068653494119644, 0.25191134214401245, 0.18924297392368317, -0.12640511989593506, 0.17781375348567963, 0.13471996784210205, -0.07904026657342911, -0.3402675986289978, -0.1414533108472824, -0.16139845550060272, -0.07071840018033981, 0.026064196601510048, -0.11777465045452118, 0.0958787128329277, 0.03335252031683922, -0.10947048664093018, 0.052748363465070724, -0.17017929255962372, -0.07672762125730515, 0.17991377413272858, -0.07845299690961838, 0.32294031977653503, -0.15600302815437317, -0.0581873282790184, -0.07140202820301056, -0.1397649347782135, 0.15831077098846436, -0.12095484882593155, 0.0640173852443695, -0.01085357554256916, 0.0009923696052283049, 0.000617707846686244, -0.05387040227651596, 0.15228526294231415, -0.05112951993942261, 0.02527214027941227, -0.1299176663160324, -0.04767915979027748, 0.10295214504003525, -0.06544466316699982, 0.023311404511332512, -0.1613304764032364, 0.012056032195687294, -0.1437864750623703, 0.010370835661888123, -0.08332697302103043, 0.08352860808372498, -0.009336385875940323, -0.040623150765895844, -0.05954933911561966, -0.005279048811644316, 0.04284205287694931, -0.015335856936872005, 0.29630112648010254, 0.008494893088936806, 0.13189546763896942, 0.16303826868534088, 0.022608771920204163, -0.1718173325061798, -0.01953105255961418, -0.03700250759720802, -0.0736953616142273, 0.07995044440031052, -0.2031794935464859, 0.057182274758815765, 0.07621432095766068, -0.06988967210054398, 0.06107276678085327, 0.08900803327560425, 0.01489779632538557, -0.04548985883593559, 0.16442106664180756, -0.1671867072582245, -0.0006360075785778463, -0.008117838762700558, 0.10618383437395096, 0.0521981455385685, 0.017339367419481277, 0.12554024159908295, 0.005745370872318745, -0.036925509572029114, 0.026103302836418152, 0.021035362035036087, -0.0628022700548172, 0.008186271414160728, 0.0677781030535698, 0.028965607285499573, -0.1132744550704956, 0.07926022261381149, 0.06677495688199997, -0.129411518573761, -0.010128652676939964, 0.09587135165929794, -0.1095706969499588, -0.16171106696128845, 0.00618077302351594, 0.04380159080028534, -0.08738097548484802, -0.08622825890779495, -0.024138811975717545, -0.1249731257557869, 0.04831845685839653, 0.09633947908878326, 0.13090389966964722, 0.06443681567907333, -0.0017022290267050266, -0.06870454549789429, 0.05082831531763077, 0.001563939149491489, -0.07709051668643951, 0.036532118916511536, -0.12133266031742096, -0.02846197411417961, 0.010334618389606476, 0.11554903537034988, -0.06259089708328247, -0.0042324368841946125, -0.12426569312810898, 0.001558715826831758, -0.11614777147769928, -0.029476512223482132, -0.08306972682476044, -0.033092986792325974, 0.02484716661274433, -0.08117660880088806, -0.04502910003066063, -0.01520068384706974, -0.14415021240711212, -0.024689702317118645, -0.04193141311407089, 0.09420331567525864, -0.104824498295784, -0.05173527076840401, 0.09438075870275497, -0.011454490944743156, 0.09049908816814423, 0.0417814776301384, -0.05572384595870972, 0.07632939517498016, -0.07947538793087006, -0.10600963234901428, 0.08863865584135056, 0.03787854313850403, 0.046659957617521286, -0.005956739187240601, -0.011253931559622288, 0.09342904388904572, -0.0007177229854278266, 0.042248114943504333, 0.011596323922276497, -0.12917691469192505, -0.02238038182258606, -0.010726326145231724, -0.13931554555892944, 0.015678487718105316, -0.09275448322296143, 0.14498212933540344, 0.01790434494614601, 0.17994743585586548, -0.005748189985752106, 0.031212467700242996, -0.06608235090970993, 0.025236306712031364, -0.057449471205472946, -0.17951899766921997, -0.12403986603021622, -0.06836044043302536, -0.027640171349048615, -0.028845123946666718, 0.2805347144603729, 0.08917142450809479, -0.08347166329622269, 0.07290034741163254, 0.07196228206157684, 0.01505149994045496, 0.02077554725110531, 0.22023816406726837, 0.06361716240644455, -0.02245054394006729, -0.061880506575107574, 0.019000306725502014, 0.022344766184687614, -0.03256882727146149, 0.07464652508497238, 0.1169479712843895, 0.1142251044511795, 0.07622475177049637, 0.02704872190952301, -0.016070278361439705, -0.11697600036859512, -0.1287461221218109, 0.014688584953546524, 0.119608573615551, -0.03686606511473656, 0.059426259249448776, 0.13529466092586517, -0.037811972200870514, 0.03956560790538788, -0.086575448513031, 0.012540166266262531, -0.17560049891471863, -0.09157182276248932, -0.07360459119081497, -0.13344724476337433, -0.03314107656478882, -0.052344802767038345, 0.03984220325946808, 0.1292441338300705, 0.032210953533649445, -0.03851942718029022, -0.015154266729950905, -0.015868499875068665, -0.04258604347705841, 0.03028041310608387, -0.023542780429124832, 0.0002562306181062013, -0.06029721349477768, -0.02395167201757431, -0.10368048399686813, -0.02776411361992359, -0.05679041147232056, 0.04382525384426117, -0.027991825714707375, 0.021414659917354584, -0.12156787514686584, -0.08556406199932098, -0.0529513917863369, 0.020444447174668312, -0.005759124178439379, 0.18267738819122314, 0.005337047390639782, 0.031859856098890305, 0.07273654639720917, 0.2198667675256729, -0.09987863898277283, -0.1095680221915245, -0.045107651501894, 0.17264553904533386, 0.04104243591427803, 0.06176810711622238, -0.020409779623150826, 0.0026832649018615484, -0.09016072750091553, 0.26033926010131836, 0.33503320813179016, -0.05959988385438919, 0.06000038608908653, -0.001532332506030798, 0.012618477456271648, 0.0598568320274353, 0.1106472909450531, 0.1316719800233841, 0.16129449009895325, -0.08782772719860077, -0.0036300979554653168, -0.04469006136059761, 0.003653900232166052, -0.15766099095344543, 0.07017975300550461, -0.0036263535730540752, -0.09436394274234772, -0.022905346006155014, 0.057690780609846115, -0.09149443358182907, 0.07677844911813736, 0.012266093865036964, -0.18119388818740845, -0.05720311030745506, 0.02039954625070095, 0.21919496357440948, -0.017752492800354958, 0.05047750845551491, -0.03148294985294342, -0.05256124958395958, 0.07185778021812439, -0.028453262522816658, -0.16809456050395966, -0.052983757108449936, 0.11245088279247284, -0.033219534903764725, 0.11339452862739563, -0.01019338145852089, 0.02442827820777893, 0.09866058081388474, 0.06523024290800095, -0.09986453503370285, 0.06829454749822617, 0.024860212579369545, -0.06409720331430435, -0.025689758360385895, -0.10032090544700623, -0.02310752309858799, -0.06327763199806213, 0.06990473717451096, -0.10408949106931686, 0.047875236719846725, -0.02276439219713211, -0.041091352701187134, -0.03285158798098564, 0.01771629974246025, -0.06046866253018379, 0.07364235818386078, 0.030026888474822044, -0.036611031740903854, -0.07146476209163666, -0.04183752462267876, -0.04579772427678108, 0.008587779477238655, -0.16310185194015503, -0.11530570685863495, -0.004291696939617395, -0.022359905764460564, 0.07707220315933228, 0.03829155117273331, -0.09312435239553452, -0.03533054143190384, -0.05517655238509178, 0.0333234928548336, -0.12787498533725739, 0.0459769070148468, 0.08696694672107697, -0.015718704089522362, -0.010699767619371414, -0.06356548517942429, 0.019441908225417137, 0.023609060794115067, -0.10050466656684875, -0.07246299833059311 ]
null
null
transformers
# Cross-Encoder for MS Marco This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco) ## Usage with Transformers ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('model_name') tokenizer = AutoTokenizer.from_pretrained('model_name') features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits print(scores) ``` ## Usage with SentenceTransformers The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('model_name', max_length=512) scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')]) ``` ## Performance In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset. | Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec | | ------------- |:-------------| -----| --- | | **Version 2 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000 | cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100 | cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500 | cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800 | cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960 | **Version 1 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000 | cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900 | cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680 | cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340 | **Other models** | | | | nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900 | nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340 | nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100 | Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340 | amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330 | sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720 Note: Runtime was computed on a V100 GPU.
{"license": "apache-2.0"}
text-classification
cross-encoder/ms-marco-MiniLM-L-6-v2
[ "transformers", "pytorch", "jax", "bert", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
Cross-Encoder for MS Marco ========================== This model was trained on the MS Marco Passage Ranking task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco Usage with Transformers ----------------------- Usage with SentenceTransformers ------------------------------- The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this: Performance ----------- In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset. Note: Runtime was computed on a V100 GPU.
[]
[ "TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ 51 ]
[ "passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ -0.005313366185873747, 0.12051703780889511, -0.004693866241723299, 0.03580747917294502, 0.08237535506486893, 0.015574881806969643, 0.09972856193780899, 0.1512490212917328, -0.002342487219721079, -0.035675179213285446, 0.1512344777584076, 0.1678246557712555, -0.03263122960925102, 0.07377322763204575, -0.06331735849380493, -0.23472362756729126, 0.10480628162622452, 0.022721635177731514, -0.06391183286905289, 0.0852574035525322, 0.11916844546794891, -0.06770222634077072, 0.04979204759001732, -0.006682221312075853, -0.07624185085296631, 0.03358222916722298, 0.023558329790830612, -0.11514781415462494, 0.09597253054380417, 0.031052103266119957, 0.10091041028499603, 0.054762229323387146, -0.03489936888217926, -0.1495836228132248, 0.026294076815247536, 0.029823211953043938, -0.08911542594432831, 0.07207200676202774, 0.03851554915308952, -0.03519143536686897, 0.04294948652386665, 0.0052515072748064995, -0.0014171735383570194, 0.028817106038331985, -0.09800101071596146, -0.21412332355976105, -0.07283638417720795, 0.07232868671417236, 0.027582870796322823, 0.07712103426456451, 0.024217145517468452, 0.15384386479854584, -0.14576973021030426, 0.05114344507455826, 0.14506278932094574, -0.3469622731208801, -0.0045180246233940125, 0.13070425391197205, 0.07035423070192337, 0.04190012812614441, -0.03805508464574814, 0.04657916724681854, 0.05582089722156525, 0.0025910616386681795, 0.06670897454023361, -0.04245023429393768, -0.1239764392375946, 0.06446674466133118, -0.07143140584230423, -0.07907963544130325, 0.2631944417953491, -0.019031910225749016, 0.0633535161614418, -0.020391980186104774, -0.07111599296331406, -0.02444433979690075, 0.0007090915460139513, 0.04914804920554161, 0.029526568949222565, 0.10229268670082092, 0.07079360634088516, -0.03269055485725403, -0.14858068525791168, 0.020561350509524345, -0.18860138952732086, 0.06773007661104202, 0.0031633847393095493, 0.07213912159204483, -0.1380876749753952, 0.05910053104162216, 0.022541260346770287, -0.12431655079126358, 0.026827946305274963, -0.08040676265954971, 0.09522116184234619, 0.02466272935271263, -0.10294348746538162, 0.048924896866083145, 0.12230401486158371, 0.1840950846672058, 0.025208648294210434, 0.004822383169084787, -0.026003237813711166, 0.12904393672943115, 0.010278739035129547, 0.08728154003620148, -0.017540551722049713, -0.016535868868231773, 0.07604997605085373, -0.07513169944286346, 0.056044477969408035, -0.06161702796816826, -0.1701550930738449, -0.03133761137723923, 0.03735613077878952, 0.09264276921749115, 0.08358346670866013, 0.05583064258098602, -0.02714722789824009, 0.025454645976424217, 0.1282920092344284, -0.06134769693017006, 0.04080605506896973, -0.00510947173461318, 0.03630151227116585, 0.02500070445239544, 0.02943694032728672, 0.023852426558732986, -0.017657775431871414, 0.06957349926233292, -0.06750348955392838, -0.010275444015860558, -0.051824845373630524, -0.060393817722797394, 0.08819574117660522, -0.0934952050447464, 0.03543853387236595, -0.15935862064361572, -0.08309947699308395, 0.017079420387744904, 0.06980738043785095, 0.0021625745575875044, -0.09810265898704529, 0.07114944607019424, -0.04728621616959572, 0.0843905657529831, -0.08494804799556732, 0.002027006121352315, -0.09370028227567673, 0.05686311423778534, -0.10129949450492859, 0.055879801511764526, -0.17257225513458252, 0.06273247301578522, -0.1052277535200119, -0.0019254190847277641, -0.05104181170463562, -0.04737512394785881, -0.09146487712860107, 0.16669559478759766, -0.022341273725032806, -0.0318073108792305, -0.019266171380877495, 0.031578149646520615, -0.04990849643945694, 0.08952520042657852, -0.13634717464447021, -0.03626799210906029, 0.14703963696956635, -0.09130478650331497, -0.16326141357421875, 0.0717688575387001, 0.00046021450543776155, 0.0031928219832479954, 0.0205746628344059, 0.2063581943511963, 0.0636834129691124, -0.038732003420591354, 0.0429588183760643, 0.14875924587249756, -0.031094709411263466, -0.15359465777873993, 0.06786390393972397, -0.029294706881046295, -0.09197673946619034, 0.043804608285427094, -0.017092304304242134, 0.06857074052095413, 0.011296656914055347, -0.08280479162931442, -0.07357008755207062, -0.03364529088139534, 0.04349697753787041, 0.0174326840788126, 0.09276469796895981, -0.07873370498418808, -0.03898242488503456, -0.0033968340139836073, 0.03604090213775635, 0.05958762764930725, 0.056634191423654556, -0.03324146941304207, 0.10759710520505905, -0.00020664279873017222, 0.019652394577860832, -0.1466042846441269, -0.022884350270032883, -0.024007568135857582, 0.029566245153546333, -0.0034570449497550726, 0.1761774867773056, 0.028720151633024216, -0.09089623391628265, -0.005154056008905172, -0.005816992837935686, 0.10818760842084885, 0.07017345726490021, -0.0435544028878212, -0.122578464448452, 0.02524160034954548, -0.05991243198513985, 0.011386431753635406, -0.028770629316568375, 0.03522561118006706, 0.0614837110042572, 0.095530666410923, -0.04097152501344681, 0.10141308605670929, -0.024964042007923126, 0.0062868064269423485, -0.08510880917310715, 0.002189264167100191, 0.09858357906341553, 0.033706147223711014, -0.04600461944937706, 0.2194674164056778, -0.10151731222867966, 0.2976343631744385, 0.2267475426197052, -0.1751745343208313, 0.057022739201784134, 0.03313778340816498, -0.019422493875026703, 0.019689708948135376, 0.0130884675309062, -0.001759985345415771, -0.028527354821562767, -0.02610705979168415, 0.14932112395763397, -0.04963885620236397, -0.03732553496956825, -0.023624418303370476, -0.06671696156263351, -0.03914187103509903, 0.04179641976952553, 0.09773124009370804, -0.15121658146381378, 0.21856796741485596, 0.39029189944267273, -0.04813329502940178, 0.11297299712896347, -0.055986665189266205, 0.025421306490898132, 0.030753333121538162, -0.07714325934648514, -0.05503294616937637, 0.02333962731063366, -0.13534876704216003, -0.02139679528772831, 0.1016768291592598, 0.056578896939754486, 0.04646281898021698, -0.13725902140140533, -0.0547390952706337, 0.02008267678320408, 0.020364783704280853, -0.05890339985489845, 0.08217864483594894, 0.018010256811976433, 0.09777913242578506, -0.0105742197483778, -0.10413795709609985, 0.10512387752532959, -0.0025146319530904293, -0.046441737562417984, 0.12223948538303375, -0.1751604825258255, -0.26337674260139465, -0.10911697149276733, -0.10594077408313751, -0.0219988115131855, 0.016123976558446884, 0.14658363163471222, -0.05834493041038513, -0.04452504217624664, -0.03016248345375061, -0.08486530184745789, -0.06864798069000244, 0.03474634513258934, -0.04818902164697647, 0.05575546622276306, 0.0009074720437638462, -0.1152714416384697, -0.06752760708332062, 0.020496297627687454, -0.03415825590491295, 0.091299869120121, -0.030293690040707588, 0.07342889904975891, 0.135502427816391, 0.007198224309831858, 0.027944006025791168, -0.030184734612703323, 0.16622070968151093, -0.03921347111463547, 0.00924068596214056, 0.19809506833553314, -0.0031844095792621374, 0.0713656023144722, 0.18086189031600952, 0.03603030741214752, -0.022779395803809166, 0.01003937516361475, -0.03807242214679718, -0.08706989139318466, -0.20672155916690826, -0.11673396080732346, -0.12385633587837219, 0.03123038075864315, 0.058827221393585205, 0.10125505924224854, 0.1326202005147934, 0.05287611857056618, 0.0016830338863655925, 0.002457578666508198, -0.029018420726060867, 0.06340944021940231, 0.23763202130794525, -0.03204694390296936, 0.14382527768611908, -0.10360123962163925, -0.058313898742198944, 0.11185727268457413, 0.07886284589767456, 0.08192996680736542, 0.10665401816368103, 0.03152311220765114, 0.0695836991071701, 0.19727741181850433, 0.10917509347200394, 0.08815993368625641, 0.02420349419116974, -0.010355910286307335, -0.06285016238689423, -0.023843849077820778, -0.023379191756248474, 0.06658007949590683, 0.039418045431375504, -0.13914524018764496, -0.029837213456630707, -0.19487792253494263, 0.08332343399524689, 0.151483952999115, 0.059639498591423035, -0.13124634325504303, 0.031023966148495674, 0.10783152282238007, -0.01136285811662674, -0.04613389074802399, 0.08763978630304337, -0.02930242381989956, -0.08054038137197495, 0.07484424859285355, 0.007519237697124481, 0.12197799980640411, 0.02354556694626808, 0.07243271917104721, -0.04469209909439087, -0.16604083776474, 0.056991759687662125, 0.12605127692222595, -0.2823342978954315, 0.20038354396820068, -0.02226925455033779, -0.09456770122051239, -0.07824184000492096, 0.0034344338346272707, 0.07712201029062271, 0.22976431250572205, 0.061447788029909134, 0.03248010203242302, -0.1648344248533249, -0.04715387150645256, -0.03496819734573364, 0.012963905930519104, 0.017423907294869423, -0.02331795170903206, -0.03611692786216736, -0.06621395796537399, -0.002671534428372979, 0.020914889872074127, 0.1569015085697174, -0.022213159129023552, -0.17191611230373383, 0.06727875024080276, 0.11111561954021454, 0.02136494219303131, -0.04995092377066612, -0.043893180787563324, -0.16064417362213135, 0.16429398953914642, -0.034801822155714035, -0.03848220035433769, -0.09644030034542084, -0.11785909533500671, 0.05767476186156273, -0.043937042355537415, 0.05455517768859863, -0.08362634479999542, 0.027540860697627068, -0.06371035426855087, -0.19031928479671478, 0.11500367522239685, -0.12582644820213318, -0.04196672886610031, -0.053986817598342896, 0.08883348852396011, -0.15042348206043243, 0.04660613462328911, 0.03068644367158413, 0.06134520471096039, -0.15422704815864563, -0.11111470311880112, -0.0030049297492951155, 0.01256405096501112, 0.059495214372873306, -0.03420129418373108, -0.07899564504623413, -0.04707137122750282, 0.06431546807289124, -0.03068653494119644, 0.25191134214401245, 0.18924297392368317, -0.12640511989593506, 0.17781375348567963, 0.13471996784210205, -0.07904026657342911, -0.3402675986289978, -0.1414533108472824, -0.16139845550060272, -0.07071840018033981, 0.026064196601510048, -0.11777465045452118, 0.0958787128329277, 0.03335252031683922, -0.10947048664093018, 0.052748363465070724, -0.17017929255962372, -0.07672762125730515, 0.17991377413272858, -0.07845299690961838, 0.32294031977653503, -0.15600302815437317, -0.0581873282790184, -0.07140202820301056, -0.1397649347782135, 0.15831077098846436, -0.12095484882593155, 0.0640173852443695, -0.01085357554256916, 0.0009923696052283049, 0.000617707846686244, -0.05387040227651596, 0.15228526294231415, -0.05112951993942261, 0.02527214027941227, -0.1299176663160324, -0.04767915979027748, 0.10295214504003525, -0.06544466316699982, 0.023311404511332512, -0.1613304764032364, 0.012056032195687294, -0.1437864750623703, 0.010370835661888123, -0.08332697302103043, 0.08352860808372498, -0.009336385875940323, -0.040623150765895844, -0.05954933911561966, -0.005279048811644316, 0.04284205287694931, -0.015335856936872005, 0.29630112648010254, 0.008494893088936806, 0.13189546763896942, 0.16303826868534088, 0.022608771920204163, -0.1718173325061798, -0.01953105255961418, -0.03700250759720802, -0.0736953616142273, 0.07995044440031052, -0.2031794935464859, 0.057182274758815765, 0.07621432095766068, -0.06988967210054398, 0.06107276678085327, 0.08900803327560425, 0.01489779632538557, -0.04548985883593559, 0.16442106664180756, -0.1671867072582245, -0.0006360075785778463, -0.008117838762700558, 0.10618383437395096, 0.0521981455385685, 0.017339367419481277, 0.12554024159908295, 0.005745370872318745, -0.036925509572029114, 0.026103302836418152, 0.021035362035036087, -0.0628022700548172, 0.008186271414160728, 0.0677781030535698, 0.028965607285499573, -0.1132744550704956, 0.07926022261381149, 0.06677495688199997, -0.129411518573761, -0.010128652676939964, 0.09587135165929794, -0.1095706969499588, -0.16171106696128845, 0.00618077302351594, 0.04380159080028534, -0.08738097548484802, -0.08622825890779495, -0.024138811975717545, -0.1249731257557869, 0.04831845685839653, 0.09633947908878326, 0.13090389966964722, 0.06443681567907333, -0.0017022290267050266, -0.06870454549789429, 0.05082831531763077, 0.001563939149491489, -0.07709051668643951, 0.036532118916511536, -0.12133266031742096, -0.02846197411417961, 0.010334618389606476, 0.11554903537034988, -0.06259089708328247, -0.0042324368841946125, -0.12426569312810898, 0.001558715826831758, -0.11614777147769928, -0.029476512223482132, -0.08306972682476044, -0.033092986792325974, 0.02484716661274433, -0.08117660880088806, -0.04502910003066063, -0.01520068384706974, -0.14415021240711212, -0.024689702317118645, -0.04193141311407089, 0.09420331567525864, -0.104824498295784, -0.05173527076840401, 0.09438075870275497, -0.011454490944743156, 0.09049908816814423, 0.0417814776301384, -0.05572384595870972, 0.07632939517498016, -0.07947538793087006, -0.10600963234901428, 0.08863865584135056, 0.03787854313850403, 0.046659957617521286, -0.005956739187240601, -0.011253931559622288, 0.09342904388904572, -0.0007177229854278266, 0.042248114943504333, 0.011596323922276497, -0.12917691469192505, -0.02238038182258606, -0.010726326145231724, -0.13931554555892944, 0.015678487718105316, -0.09275448322296143, 0.14498212933540344, 0.01790434494614601, 0.17994743585586548, -0.005748189985752106, 0.031212467700242996, -0.06608235090970993, 0.025236306712031364, -0.057449471205472946, -0.17951899766921997, -0.12403986603021622, -0.06836044043302536, -0.027640171349048615, -0.028845123946666718, 0.2805347144603729, 0.08917142450809479, -0.08347166329622269, 0.07290034741163254, 0.07196228206157684, 0.01505149994045496, 0.02077554725110531, 0.22023816406726837, 0.06361716240644455, -0.02245054394006729, -0.061880506575107574, 0.019000306725502014, 0.022344766184687614, -0.03256882727146149, 0.07464652508497238, 0.1169479712843895, 0.1142251044511795, 0.07622475177049637, 0.02704872190952301, -0.016070278361439705, -0.11697600036859512, -0.1287461221218109, 0.014688584953546524, 0.119608573615551, -0.03686606511473656, 0.059426259249448776, 0.13529466092586517, -0.037811972200870514, 0.03956560790538788, -0.086575448513031, 0.012540166266262531, -0.17560049891471863, -0.09157182276248932, -0.07360459119081497, -0.13344724476337433, -0.03314107656478882, -0.052344802767038345, 0.03984220325946808, 0.1292441338300705, 0.032210953533649445, -0.03851942718029022, -0.015154266729950905, -0.015868499875068665, -0.04258604347705841, 0.03028041310608387, -0.023542780429124832, 0.0002562306181062013, -0.06029721349477768, -0.02395167201757431, -0.10368048399686813, -0.02776411361992359, -0.05679041147232056, 0.04382525384426117, -0.027991825714707375, 0.021414659917354584, -0.12156787514686584, -0.08556406199932098, -0.0529513917863369, 0.020444447174668312, -0.005759124178439379, 0.18267738819122314, 0.005337047390639782, 0.031859856098890305, 0.07273654639720917, 0.2198667675256729, -0.09987863898277283, -0.1095680221915245, -0.045107651501894, 0.17264553904533386, 0.04104243591427803, 0.06176810711622238, -0.020409779623150826, 0.0026832649018615484, -0.09016072750091553, 0.26033926010131836, 0.33503320813179016, -0.05959988385438919, 0.06000038608908653, -0.001532332506030798, 0.012618477456271648, 0.0598568320274353, 0.1106472909450531, 0.1316719800233841, 0.16129449009895325, -0.08782772719860077, -0.0036300979554653168, -0.04469006136059761, 0.003653900232166052, -0.15766099095344543, 0.07017975300550461, -0.0036263535730540752, -0.09436394274234772, -0.022905346006155014, 0.057690780609846115, -0.09149443358182907, 0.07677844911813736, 0.012266093865036964, -0.18119388818740845, -0.05720311030745506, 0.02039954625070095, 0.21919496357440948, -0.017752492800354958, 0.05047750845551491, -0.03148294985294342, -0.05256124958395958, 0.07185778021812439, -0.028453262522816658, -0.16809456050395966, -0.052983757108449936, 0.11245088279247284, -0.033219534903764725, 0.11339452862739563, -0.01019338145852089, 0.02442827820777893, 0.09866058081388474, 0.06523024290800095, -0.09986453503370285, 0.06829454749822617, 0.024860212579369545, -0.06409720331430435, -0.025689758360385895, -0.10032090544700623, -0.02310752309858799, -0.06327763199806213, 0.06990473717451096, -0.10408949106931686, 0.047875236719846725, -0.02276439219713211, -0.041091352701187134, -0.03285158798098564, 0.01771629974246025, -0.06046866253018379, 0.07364235818386078, 0.030026888474822044, -0.036611031740903854, -0.07146476209163666, -0.04183752462267876, -0.04579772427678108, 0.008587779477238655, -0.16310185194015503, -0.11530570685863495, -0.004291696939617395, -0.022359905764460564, 0.07707220315933228, 0.03829155117273331, -0.09312435239553452, -0.03533054143190384, -0.05517655238509178, 0.0333234928548336, -0.12787498533725739, 0.0459769070148468, 0.08696694672107697, -0.015718704089522362, -0.010699767619371414, -0.06356548517942429, 0.019441908225417137, 0.023609060794115067, -0.10050466656684875, -0.07246299833059311 ]
null
null
transformers
# Cross-Encoder for MS Marco This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco) ## Usage with Transformers ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('model_name') tokenizer = AutoTokenizer.from_pretrained('model_name') features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits print(scores) ``` ## Usage with SentenceTransformers The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('model_name', max_length=512) scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')]) ``` ## Performance In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset. | Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec | | ------------- |:-------------| -----| --- | | **Version 2 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000 | cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100 | cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500 | cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800 | cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960 | **Version 1 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000 | cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900 | cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680 | cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340 | **Other models** | | | | nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900 | nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340 | nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100 | Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340 | amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330 | sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720 Note: Runtime was computed on a V100 GPU.
{"license": "apache-2.0"}
text-classification
cross-encoder/ms-marco-TinyBERT-L-2-v2
[ "transformers", "pytorch", "jax", "bert", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
Cross-Encoder for MS Marco ========================== This model was trained on the MS Marco Passage Ranking task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco Usage with Transformers ----------------------- Usage with SentenceTransformers ------------------------------- The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this: Performance ----------- In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset. Note: Runtime was computed on a V100 GPU.
[]
[ "TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ 51 ]
[ "passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ -0.005313366185873747, 0.12051703780889511, -0.004693866241723299, 0.03580747917294502, 0.08237535506486893, 0.015574881806969643, 0.09972856193780899, 0.1512490212917328, -0.002342487219721079, -0.035675179213285446, 0.1512344777584076, 0.1678246557712555, -0.03263122960925102, 0.07377322763204575, -0.06331735849380493, -0.23472362756729126, 0.10480628162622452, 0.022721635177731514, -0.06391183286905289, 0.0852574035525322, 0.11916844546794891, -0.06770222634077072, 0.04979204759001732, -0.006682221312075853, -0.07624185085296631, 0.03358222916722298, 0.023558329790830612, -0.11514781415462494, 0.09597253054380417, 0.031052103266119957, 0.10091041028499603, 0.054762229323387146, -0.03489936888217926, -0.1495836228132248, 0.026294076815247536, 0.029823211953043938, -0.08911542594432831, 0.07207200676202774, 0.03851554915308952, -0.03519143536686897, 0.04294948652386665, 0.0052515072748064995, -0.0014171735383570194, 0.028817106038331985, -0.09800101071596146, -0.21412332355976105, -0.07283638417720795, 0.07232868671417236, 0.027582870796322823, 0.07712103426456451, 0.024217145517468452, 0.15384386479854584, -0.14576973021030426, 0.05114344507455826, 0.14506278932094574, -0.3469622731208801, -0.0045180246233940125, 0.13070425391197205, 0.07035423070192337, 0.04190012812614441, -0.03805508464574814, 0.04657916724681854, 0.05582089722156525, 0.0025910616386681795, 0.06670897454023361, -0.04245023429393768, -0.1239764392375946, 0.06446674466133118, -0.07143140584230423, -0.07907963544130325, 0.2631944417953491, -0.019031910225749016, 0.0633535161614418, -0.020391980186104774, -0.07111599296331406, -0.02444433979690075, 0.0007090915460139513, 0.04914804920554161, 0.029526568949222565, 0.10229268670082092, 0.07079360634088516, -0.03269055485725403, -0.14858068525791168, 0.020561350509524345, -0.18860138952732086, 0.06773007661104202, 0.0031633847393095493, 0.07213912159204483, -0.1380876749753952, 0.05910053104162216, 0.022541260346770287, -0.12431655079126358, 0.026827946305274963, -0.08040676265954971, 0.09522116184234619, 0.02466272935271263, -0.10294348746538162, 0.048924896866083145, 0.12230401486158371, 0.1840950846672058, 0.025208648294210434, 0.004822383169084787, -0.026003237813711166, 0.12904393672943115, 0.010278739035129547, 0.08728154003620148, -0.017540551722049713, -0.016535868868231773, 0.07604997605085373, -0.07513169944286346, 0.056044477969408035, -0.06161702796816826, -0.1701550930738449, -0.03133761137723923, 0.03735613077878952, 0.09264276921749115, 0.08358346670866013, 0.05583064258098602, -0.02714722789824009, 0.025454645976424217, 0.1282920092344284, -0.06134769693017006, 0.04080605506896973, -0.00510947173461318, 0.03630151227116585, 0.02500070445239544, 0.02943694032728672, 0.023852426558732986, -0.017657775431871414, 0.06957349926233292, -0.06750348955392838, -0.010275444015860558, -0.051824845373630524, -0.060393817722797394, 0.08819574117660522, -0.0934952050447464, 0.03543853387236595, -0.15935862064361572, -0.08309947699308395, 0.017079420387744904, 0.06980738043785095, 0.0021625745575875044, -0.09810265898704529, 0.07114944607019424, -0.04728621616959572, 0.0843905657529831, -0.08494804799556732, 0.002027006121352315, -0.09370028227567673, 0.05686311423778534, -0.10129949450492859, 0.055879801511764526, -0.17257225513458252, 0.06273247301578522, -0.1052277535200119, -0.0019254190847277641, -0.05104181170463562, -0.04737512394785881, -0.09146487712860107, 0.16669559478759766, -0.022341273725032806, -0.0318073108792305, -0.019266171380877495, 0.031578149646520615, -0.04990849643945694, 0.08952520042657852, -0.13634717464447021, -0.03626799210906029, 0.14703963696956635, -0.09130478650331497, -0.16326141357421875, 0.0717688575387001, 0.00046021450543776155, 0.0031928219832479954, 0.0205746628344059, 0.2063581943511963, 0.0636834129691124, -0.038732003420591354, 0.0429588183760643, 0.14875924587249756, -0.031094709411263466, -0.15359465777873993, 0.06786390393972397, -0.029294706881046295, -0.09197673946619034, 0.043804608285427094, -0.017092304304242134, 0.06857074052095413, 0.011296656914055347, -0.08280479162931442, -0.07357008755207062, -0.03364529088139534, 0.04349697753787041, 0.0174326840788126, 0.09276469796895981, -0.07873370498418808, -0.03898242488503456, -0.0033968340139836073, 0.03604090213775635, 0.05958762764930725, 0.056634191423654556, -0.03324146941304207, 0.10759710520505905, -0.00020664279873017222, 0.019652394577860832, -0.1466042846441269, -0.022884350270032883, -0.024007568135857582, 0.029566245153546333, -0.0034570449497550726, 0.1761774867773056, 0.028720151633024216, -0.09089623391628265, -0.005154056008905172, -0.005816992837935686, 0.10818760842084885, 0.07017345726490021, -0.0435544028878212, -0.122578464448452, 0.02524160034954548, -0.05991243198513985, 0.011386431753635406, -0.028770629316568375, 0.03522561118006706, 0.0614837110042572, 0.095530666410923, -0.04097152501344681, 0.10141308605670929, -0.024964042007923126, 0.0062868064269423485, -0.08510880917310715, 0.002189264167100191, 0.09858357906341553, 0.033706147223711014, -0.04600461944937706, 0.2194674164056778, -0.10151731222867966, 0.2976343631744385, 0.2267475426197052, -0.1751745343208313, 0.057022739201784134, 0.03313778340816498, -0.019422493875026703, 0.019689708948135376, 0.0130884675309062, -0.001759985345415771, -0.028527354821562767, -0.02610705979168415, 0.14932112395763397, -0.04963885620236397, -0.03732553496956825, -0.023624418303370476, -0.06671696156263351, -0.03914187103509903, 0.04179641976952553, 0.09773124009370804, -0.15121658146381378, 0.21856796741485596, 0.39029189944267273, -0.04813329502940178, 0.11297299712896347, -0.055986665189266205, 0.025421306490898132, 0.030753333121538162, -0.07714325934648514, -0.05503294616937637, 0.02333962731063366, -0.13534876704216003, -0.02139679528772831, 0.1016768291592598, 0.056578896939754486, 0.04646281898021698, -0.13725902140140533, -0.0547390952706337, 0.02008267678320408, 0.020364783704280853, -0.05890339985489845, 0.08217864483594894, 0.018010256811976433, 0.09777913242578506, -0.0105742197483778, -0.10413795709609985, 0.10512387752532959, -0.0025146319530904293, -0.046441737562417984, 0.12223948538303375, -0.1751604825258255, -0.26337674260139465, -0.10911697149276733, -0.10594077408313751, -0.0219988115131855, 0.016123976558446884, 0.14658363163471222, -0.05834493041038513, -0.04452504217624664, -0.03016248345375061, -0.08486530184745789, -0.06864798069000244, 0.03474634513258934, -0.04818902164697647, 0.05575546622276306, 0.0009074720437638462, -0.1152714416384697, -0.06752760708332062, 0.020496297627687454, -0.03415825590491295, 0.091299869120121, -0.030293690040707588, 0.07342889904975891, 0.135502427816391, 0.007198224309831858, 0.027944006025791168, -0.030184734612703323, 0.16622070968151093, -0.03921347111463547, 0.00924068596214056, 0.19809506833553314, -0.0031844095792621374, 0.0713656023144722, 0.18086189031600952, 0.03603030741214752, -0.022779395803809166, 0.01003937516361475, -0.03807242214679718, -0.08706989139318466, -0.20672155916690826, -0.11673396080732346, -0.12385633587837219, 0.03123038075864315, 0.058827221393585205, 0.10125505924224854, 0.1326202005147934, 0.05287611857056618, 0.0016830338863655925, 0.002457578666508198, -0.029018420726060867, 0.06340944021940231, 0.23763202130794525, -0.03204694390296936, 0.14382527768611908, -0.10360123962163925, -0.058313898742198944, 0.11185727268457413, 0.07886284589767456, 0.08192996680736542, 0.10665401816368103, 0.03152311220765114, 0.0695836991071701, 0.19727741181850433, 0.10917509347200394, 0.08815993368625641, 0.02420349419116974, -0.010355910286307335, -0.06285016238689423, -0.023843849077820778, -0.023379191756248474, 0.06658007949590683, 0.039418045431375504, -0.13914524018764496, -0.029837213456630707, -0.19487792253494263, 0.08332343399524689, 0.151483952999115, 0.059639498591423035, -0.13124634325504303, 0.031023966148495674, 0.10783152282238007, -0.01136285811662674, -0.04613389074802399, 0.08763978630304337, -0.02930242381989956, -0.08054038137197495, 0.07484424859285355, 0.007519237697124481, 0.12197799980640411, 0.02354556694626808, 0.07243271917104721, -0.04469209909439087, -0.16604083776474, 0.056991759687662125, 0.12605127692222595, -0.2823342978954315, 0.20038354396820068, -0.02226925455033779, -0.09456770122051239, -0.07824184000492096, 0.0034344338346272707, 0.07712201029062271, 0.22976431250572205, 0.061447788029909134, 0.03248010203242302, -0.1648344248533249, -0.04715387150645256, -0.03496819734573364, 0.012963905930519104, 0.017423907294869423, -0.02331795170903206, -0.03611692786216736, -0.06621395796537399, -0.002671534428372979, 0.020914889872074127, 0.1569015085697174, -0.022213159129023552, -0.17191611230373383, 0.06727875024080276, 0.11111561954021454, 0.02136494219303131, -0.04995092377066612, -0.043893180787563324, -0.16064417362213135, 0.16429398953914642, -0.034801822155714035, -0.03848220035433769, -0.09644030034542084, -0.11785909533500671, 0.05767476186156273, -0.043937042355537415, 0.05455517768859863, -0.08362634479999542, 0.027540860697627068, -0.06371035426855087, -0.19031928479671478, 0.11500367522239685, -0.12582644820213318, -0.04196672886610031, -0.053986817598342896, 0.08883348852396011, -0.15042348206043243, 0.04660613462328911, 0.03068644367158413, 0.06134520471096039, -0.15422704815864563, -0.11111470311880112, -0.0030049297492951155, 0.01256405096501112, 0.059495214372873306, -0.03420129418373108, -0.07899564504623413, -0.04707137122750282, 0.06431546807289124, -0.03068653494119644, 0.25191134214401245, 0.18924297392368317, -0.12640511989593506, 0.17781375348567963, 0.13471996784210205, -0.07904026657342911, -0.3402675986289978, -0.1414533108472824, -0.16139845550060272, -0.07071840018033981, 0.026064196601510048, -0.11777465045452118, 0.0958787128329277, 0.03335252031683922, -0.10947048664093018, 0.052748363465070724, -0.17017929255962372, -0.07672762125730515, 0.17991377413272858, -0.07845299690961838, 0.32294031977653503, -0.15600302815437317, -0.0581873282790184, -0.07140202820301056, -0.1397649347782135, 0.15831077098846436, -0.12095484882593155, 0.0640173852443695, -0.01085357554256916, 0.0009923696052283049, 0.000617707846686244, -0.05387040227651596, 0.15228526294231415, -0.05112951993942261, 0.02527214027941227, -0.1299176663160324, -0.04767915979027748, 0.10295214504003525, -0.06544466316699982, 0.023311404511332512, -0.1613304764032364, 0.012056032195687294, -0.1437864750623703, 0.010370835661888123, -0.08332697302103043, 0.08352860808372498, -0.009336385875940323, -0.040623150765895844, -0.05954933911561966, -0.005279048811644316, 0.04284205287694931, -0.015335856936872005, 0.29630112648010254, 0.008494893088936806, 0.13189546763896942, 0.16303826868534088, 0.022608771920204163, -0.1718173325061798, -0.01953105255961418, -0.03700250759720802, -0.0736953616142273, 0.07995044440031052, -0.2031794935464859, 0.057182274758815765, 0.07621432095766068, -0.06988967210054398, 0.06107276678085327, 0.08900803327560425, 0.01489779632538557, -0.04548985883593559, 0.16442106664180756, -0.1671867072582245, -0.0006360075785778463, -0.008117838762700558, 0.10618383437395096, 0.0521981455385685, 0.017339367419481277, 0.12554024159908295, 0.005745370872318745, -0.036925509572029114, 0.026103302836418152, 0.021035362035036087, -0.0628022700548172, 0.008186271414160728, 0.0677781030535698, 0.028965607285499573, -0.1132744550704956, 0.07926022261381149, 0.06677495688199997, -0.129411518573761, -0.010128652676939964, 0.09587135165929794, -0.1095706969499588, -0.16171106696128845, 0.00618077302351594, 0.04380159080028534, -0.08738097548484802, -0.08622825890779495, -0.024138811975717545, -0.1249731257557869, 0.04831845685839653, 0.09633947908878326, 0.13090389966964722, 0.06443681567907333, -0.0017022290267050266, -0.06870454549789429, 0.05082831531763077, 0.001563939149491489, -0.07709051668643951, 0.036532118916511536, -0.12133266031742096, -0.02846197411417961, 0.010334618389606476, 0.11554903537034988, -0.06259089708328247, -0.0042324368841946125, -0.12426569312810898, 0.001558715826831758, -0.11614777147769928, -0.029476512223482132, -0.08306972682476044, -0.033092986792325974, 0.02484716661274433, -0.08117660880088806, -0.04502910003066063, -0.01520068384706974, -0.14415021240711212, -0.024689702317118645, -0.04193141311407089, 0.09420331567525864, -0.104824498295784, -0.05173527076840401, 0.09438075870275497, -0.011454490944743156, 0.09049908816814423, 0.0417814776301384, -0.05572384595870972, 0.07632939517498016, -0.07947538793087006, -0.10600963234901428, 0.08863865584135056, 0.03787854313850403, 0.046659957617521286, -0.005956739187240601, -0.011253931559622288, 0.09342904388904572, -0.0007177229854278266, 0.042248114943504333, 0.011596323922276497, -0.12917691469192505, -0.02238038182258606, -0.010726326145231724, -0.13931554555892944, 0.015678487718105316, -0.09275448322296143, 0.14498212933540344, 0.01790434494614601, 0.17994743585586548, -0.005748189985752106, 0.031212467700242996, -0.06608235090970993, 0.025236306712031364, -0.057449471205472946, -0.17951899766921997, -0.12403986603021622, -0.06836044043302536, -0.027640171349048615, -0.028845123946666718, 0.2805347144603729, 0.08917142450809479, -0.08347166329622269, 0.07290034741163254, 0.07196228206157684, 0.01505149994045496, 0.02077554725110531, 0.22023816406726837, 0.06361716240644455, -0.02245054394006729, -0.061880506575107574, 0.019000306725502014, 0.022344766184687614, -0.03256882727146149, 0.07464652508497238, 0.1169479712843895, 0.1142251044511795, 0.07622475177049637, 0.02704872190952301, -0.016070278361439705, -0.11697600036859512, -0.1287461221218109, 0.014688584953546524, 0.119608573615551, -0.03686606511473656, 0.059426259249448776, 0.13529466092586517, -0.037811972200870514, 0.03956560790538788, -0.086575448513031, 0.012540166266262531, -0.17560049891471863, -0.09157182276248932, -0.07360459119081497, -0.13344724476337433, -0.03314107656478882, -0.052344802767038345, 0.03984220325946808, 0.1292441338300705, 0.032210953533649445, -0.03851942718029022, -0.015154266729950905, -0.015868499875068665, -0.04258604347705841, 0.03028041310608387, -0.023542780429124832, 0.0002562306181062013, -0.06029721349477768, -0.02395167201757431, -0.10368048399686813, -0.02776411361992359, -0.05679041147232056, 0.04382525384426117, -0.027991825714707375, 0.021414659917354584, -0.12156787514686584, -0.08556406199932098, -0.0529513917863369, 0.020444447174668312, -0.005759124178439379, 0.18267738819122314, 0.005337047390639782, 0.031859856098890305, 0.07273654639720917, 0.2198667675256729, -0.09987863898277283, -0.1095680221915245, -0.045107651501894, 0.17264553904533386, 0.04104243591427803, 0.06176810711622238, -0.020409779623150826, 0.0026832649018615484, -0.09016072750091553, 0.26033926010131836, 0.33503320813179016, -0.05959988385438919, 0.06000038608908653, -0.001532332506030798, 0.012618477456271648, 0.0598568320274353, 0.1106472909450531, 0.1316719800233841, 0.16129449009895325, -0.08782772719860077, -0.0036300979554653168, -0.04469006136059761, 0.003653900232166052, -0.15766099095344543, 0.07017975300550461, -0.0036263535730540752, -0.09436394274234772, -0.022905346006155014, 0.057690780609846115, -0.09149443358182907, 0.07677844911813736, 0.012266093865036964, -0.18119388818740845, -0.05720311030745506, 0.02039954625070095, 0.21919496357440948, -0.017752492800354958, 0.05047750845551491, -0.03148294985294342, -0.05256124958395958, 0.07185778021812439, -0.028453262522816658, -0.16809456050395966, -0.052983757108449936, 0.11245088279247284, -0.033219534903764725, 0.11339452862739563, -0.01019338145852089, 0.02442827820777893, 0.09866058081388474, 0.06523024290800095, -0.09986453503370285, 0.06829454749822617, 0.024860212579369545, -0.06409720331430435, -0.025689758360385895, -0.10032090544700623, -0.02310752309858799, -0.06327763199806213, 0.06990473717451096, -0.10408949106931686, 0.047875236719846725, -0.02276439219713211, -0.041091352701187134, -0.03285158798098564, 0.01771629974246025, -0.06046866253018379, 0.07364235818386078, 0.030026888474822044, -0.036611031740903854, -0.07146476209163666, -0.04183752462267876, -0.04579772427678108, 0.008587779477238655, -0.16310185194015503, -0.11530570685863495, -0.004291696939617395, -0.022359905764460564, 0.07707220315933228, 0.03829155117273331, -0.09312435239553452, -0.03533054143190384, -0.05517655238509178, 0.0333234928548336, -0.12787498533725739, 0.0459769070148468, 0.08696694672107697, -0.015718704089522362, -0.010699767619371414, -0.06356548517942429, 0.019441908225417137, 0.023609060794115067, -0.10050466656684875, -0.07246299833059311 ]
null
null
transformers
# Cross-Encoder for MS Marco This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco) ## Usage with Transformers ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('model_name') tokenizer = AutoTokenizer.from_pretrained('model_name') features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits print(scores) ``` ## Usage with SentenceTransformers The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('model_name', max_length=512) scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')]) ``` ## Performance In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset. | Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec | | ------------- |:-------------| -----| --- | | **Version 2 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000 | cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100 | cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500 | cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800 | cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960 | **Version 1 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000 | cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900 | cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680 | cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340 | **Other models** | | | | nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900 | nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340 | nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100 | Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340 | amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330 | sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720 Note: Runtime was computed on a V100 GPU.
{"license": "apache-2.0"}
text-classification
cross-encoder/ms-marco-TinyBERT-L-2
[ "transformers", "pytorch", "jax", "bert", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
Cross-Encoder for MS Marco ========================== This model was trained on the MS Marco Passage Ranking task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco Usage with Transformers ----------------------- Usage with SentenceTransformers ------------------------------- The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this: Performance ----------- In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset. Note: Runtime was computed on a V100 GPU.
[]
[ "TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ 51 ]
[ "passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ -0.005313366185873747, 0.12051703780889511, -0.004693866241723299, 0.03580747917294502, 0.08237535506486893, 0.015574881806969643, 0.09972856193780899, 0.1512490212917328, -0.002342487219721079, -0.035675179213285446, 0.1512344777584076, 0.1678246557712555, -0.03263122960925102, 0.07377322763204575, -0.06331735849380493, -0.23472362756729126, 0.10480628162622452, 0.022721635177731514, -0.06391183286905289, 0.0852574035525322, 0.11916844546794891, -0.06770222634077072, 0.04979204759001732, -0.006682221312075853, -0.07624185085296631, 0.03358222916722298, 0.023558329790830612, -0.11514781415462494, 0.09597253054380417, 0.031052103266119957, 0.10091041028499603, 0.054762229323387146, -0.03489936888217926, -0.1495836228132248, 0.026294076815247536, 0.029823211953043938, -0.08911542594432831, 0.07207200676202774, 0.03851554915308952, -0.03519143536686897, 0.04294948652386665, 0.0052515072748064995, -0.0014171735383570194, 0.028817106038331985, -0.09800101071596146, -0.21412332355976105, -0.07283638417720795, 0.07232868671417236, 0.027582870796322823, 0.07712103426456451, 0.024217145517468452, 0.15384386479854584, -0.14576973021030426, 0.05114344507455826, 0.14506278932094574, -0.3469622731208801, -0.0045180246233940125, 0.13070425391197205, 0.07035423070192337, 0.04190012812614441, -0.03805508464574814, 0.04657916724681854, 0.05582089722156525, 0.0025910616386681795, 0.06670897454023361, -0.04245023429393768, -0.1239764392375946, 0.06446674466133118, -0.07143140584230423, -0.07907963544130325, 0.2631944417953491, -0.019031910225749016, 0.0633535161614418, -0.020391980186104774, -0.07111599296331406, -0.02444433979690075, 0.0007090915460139513, 0.04914804920554161, 0.029526568949222565, 0.10229268670082092, 0.07079360634088516, -0.03269055485725403, -0.14858068525791168, 0.020561350509524345, -0.18860138952732086, 0.06773007661104202, 0.0031633847393095493, 0.07213912159204483, -0.1380876749753952, 0.05910053104162216, 0.022541260346770287, -0.12431655079126358, 0.026827946305274963, -0.08040676265954971, 0.09522116184234619, 0.02466272935271263, -0.10294348746538162, 0.048924896866083145, 0.12230401486158371, 0.1840950846672058, 0.025208648294210434, 0.004822383169084787, -0.026003237813711166, 0.12904393672943115, 0.010278739035129547, 0.08728154003620148, -0.017540551722049713, -0.016535868868231773, 0.07604997605085373, -0.07513169944286346, 0.056044477969408035, -0.06161702796816826, -0.1701550930738449, -0.03133761137723923, 0.03735613077878952, 0.09264276921749115, 0.08358346670866013, 0.05583064258098602, -0.02714722789824009, 0.025454645976424217, 0.1282920092344284, -0.06134769693017006, 0.04080605506896973, -0.00510947173461318, 0.03630151227116585, 0.02500070445239544, 0.02943694032728672, 0.023852426558732986, -0.017657775431871414, 0.06957349926233292, -0.06750348955392838, -0.010275444015860558, -0.051824845373630524, -0.060393817722797394, 0.08819574117660522, -0.0934952050447464, 0.03543853387236595, -0.15935862064361572, -0.08309947699308395, 0.017079420387744904, 0.06980738043785095, 0.0021625745575875044, -0.09810265898704529, 0.07114944607019424, -0.04728621616959572, 0.0843905657529831, -0.08494804799556732, 0.002027006121352315, -0.09370028227567673, 0.05686311423778534, -0.10129949450492859, 0.055879801511764526, -0.17257225513458252, 0.06273247301578522, -0.1052277535200119, -0.0019254190847277641, -0.05104181170463562, -0.04737512394785881, -0.09146487712860107, 0.16669559478759766, -0.022341273725032806, -0.0318073108792305, -0.019266171380877495, 0.031578149646520615, -0.04990849643945694, 0.08952520042657852, -0.13634717464447021, -0.03626799210906029, 0.14703963696956635, -0.09130478650331497, -0.16326141357421875, 0.0717688575387001, 0.00046021450543776155, 0.0031928219832479954, 0.0205746628344059, 0.2063581943511963, 0.0636834129691124, -0.038732003420591354, 0.0429588183760643, 0.14875924587249756, -0.031094709411263466, -0.15359465777873993, 0.06786390393972397, -0.029294706881046295, -0.09197673946619034, 0.043804608285427094, -0.017092304304242134, 0.06857074052095413, 0.011296656914055347, -0.08280479162931442, -0.07357008755207062, -0.03364529088139534, 0.04349697753787041, 0.0174326840788126, 0.09276469796895981, -0.07873370498418808, -0.03898242488503456, -0.0033968340139836073, 0.03604090213775635, 0.05958762764930725, 0.056634191423654556, -0.03324146941304207, 0.10759710520505905, -0.00020664279873017222, 0.019652394577860832, -0.1466042846441269, -0.022884350270032883, -0.024007568135857582, 0.029566245153546333, -0.0034570449497550726, 0.1761774867773056, 0.028720151633024216, -0.09089623391628265, -0.005154056008905172, -0.005816992837935686, 0.10818760842084885, 0.07017345726490021, -0.0435544028878212, -0.122578464448452, 0.02524160034954548, -0.05991243198513985, 0.011386431753635406, -0.028770629316568375, 0.03522561118006706, 0.0614837110042572, 0.095530666410923, -0.04097152501344681, 0.10141308605670929, -0.024964042007923126, 0.0062868064269423485, -0.08510880917310715, 0.002189264167100191, 0.09858357906341553, 0.033706147223711014, -0.04600461944937706, 0.2194674164056778, -0.10151731222867966, 0.2976343631744385, 0.2267475426197052, -0.1751745343208313, 0.057022739201784134, 0.03313778340816498, -0.019422493875026703, 0.019689708948135376, 0.0130884675309062, -0.001759985345415771, -0.028527354821562767, -0.02610705979168415, 0.14932112395763397, -0.04963885620236397, -0.03732553496956825, -0.023624418303370476, -0.06671696156263351, -0.03914187103509903, 0.04179641976952553, 0.09773124009370804, -0.15121658146381378, 0.21856796741485596, 0.39029189944267273, -0.04813329502940178, 0.11297299712896347, -0.055986665189266205, 0.025421306490898132, 0.030753333121538162, -0.07714325934648514, -0.05503294616937637, 0.02333962731063366, -0.13534876704216003, -0.02139679528772831, 0.1016768291592598, 0.056578896939754486, 0.04646281898021698, -0.13725902140140533, -0.0547390952706337, 0.02008267678320408, 0.020364783704280853, -0.05890339985489845, 0.08217864483594894, 0.018010256811976433, 0.09777913242578506, -0.0105742197483778, -0.10413795709609985, 0.10512387752532959, -0.0025146319530904293, -0.046441737562417984, 0.12223948538303375, -0.1751604825258255, -0.26337674260139465, -0.10911697149276733, -0.10594077408313751, -0.0219988115131855, 0.016123976558446884, 0.14658363163471222, -0.05834493041038513, -0.04452504217624664, -0.03016248345375061, -0.08486530184745789, -0.06864798069000244, 0.03474634513258934, -0.04818902164697647, 0.05575546622276306, 0.0009074720437638462, -0.1152714416384697, -0.06752760708332062, 0.020496297627687454, -0.03415825590491295, 0.091299869120121, -0.030293690040707588, 0.07342889904975891, 0.135502427816391, 0.007198224309831858, 0.027944006025791168, -0.030184734612703323, 0.16622070968151093, -0.03921347111463547, 0.00924068596214056, 0.19809506833553314, -0.0031844095792621374, 0.0713656023144722, 0.18086189031600952, 0.03603030741214752, -0.022779395803809166, 0.01003937516361475, -0.03807242214679718, -0.08706989139318466, -0.20672155916690826, -0.11673396080732346, -0.12385633587837219, 0.03123038075864315, 0.058827221393585205, 0.10125505924224854, 0.1326202005147934, 0.05287611857056618, 0.0016830338863655925, 0.002457578666508198, -0.029018420726060867, 0.06340944021940231, 0.23763202130794525, -0.03204694390296936, 0.14382527768611908, -0.10360123962163925, -0.058313898742198944, 0.11185727268457413, 0.07886284589767456, 0.08192996680736542, 0.10665401816368103, 0.03152311220765114, 0.0695836991071701, 0.19727741181850433, 0.10917509347200394, 0.08815993368625641, 0.02420349419116974, -0.010355910286307335, -0.06285016238689423, -0.023843849077820778, -0.023379191756248474, 0.06658007949590683, 0.039418045431375504, -0.13914524018764496, -0.029837213456630707, -0.19487792253494263, 0.08332343399524689, 0.151483952999115, 0.059639498591423035, -0.13124634325504303, 0.031023966148495674, 0.10783152282238007, -0.01136285811662674, -0.04613389074802399, 0.08763978630304337, -0.02930242381989956, -0.08054038137197495, 0.07484424859285355, 0.007519237697124481, 0.12197799980640411, 0.02354556694626808, 0.07243271917104721, -0.04469209909439087, -0.16604083776474, 0.056991759687662125, 0.12605127692222595, -0.2823342978954315, 0.20038354396820068, -0.02226925455033779, -0.09456770122051239, -0.07824184000492096, 0.0034344338346272707, 0.07712201029062271, 0.22976431250572205, 0.061447788029909134, 0.03248010203242302, -0.1648344248533249, -0.04715387150645256, -0.03496819734573364, 0.012963905930519104, 0.017423907294869423, -0.02331795170903206, -0.03611692786216736, -0.06621395796537399, -0.002671534428372979, 0.020914889872074127, 0.1569015085697174, -0.022213159129023552, -0.17191611230373383, 0.06727875024080276, 0.11111561954021454, 0.02136494219303131, -0.04995092377066612, -0.043893180787563324, -0.16064417362213135, 0.16429398953914642, -0.034801822155714035, -0.03848220035433769, -0.09644030034542084, -0.11785909533500671, 0.05767476186156273, -0.043937042355537415, 0.05455517768859863, -0.08362634479999542, 0.027540860697627068, -0.06371035426855087, -0.19031928479671478, 0.11500367522239685, -0.12582644820213318, -0.04196672886610031, -0.053986817598342896, 0.08883348852396011, -0.15042348206043243, 0.04660613462328911, 0.03068644367158413, 0.06134520471096039, -0.15422704815864563, -0.11111470311880112, -0.0030049297492951155, 0.01256405096501112, 0.059495214372873306, -0.03420129418373108, -0.07899564504623413, -0.04707137122750282, 0.06431546807289124, -0.03068653494119644, 0.25191134214401245, 0.18924297392368317, -0.12640511989593506, 0.17781375348567963, 0.13471996784210205, -0.07904026657342911, -0.3402675986289978, -0.1414533108472824, -0.16139845550060272, -0.07071840018033981, 0.026064196601510048, -0.11777465045452118, 0.0958787128329277, 0.03335252031683922, -0.10947048664093018, 0.052748363465070724, -0.17017929255962372, -0.07672762125730515, 0.17991377413272858, -0.07845299690961838, 0.32294031977653503, -0.15600302815437317, -0.0581873282790184, -0.07140202820301056, -0.1397649347782135, 0.15831077098846436, -0.12095484882593155, 0.0640173852443695, -0.01085357554256916, 0.0009923696052283049, 0.000617707846686244, -0.05387040227651596, 0.15228526294231415, -0.05112951993942261, 0.02527214027941227, -0.1299176663160324, -0.04767915979027748, 0.10295214504003525, -0.06544466316699982, 0.023311404511332512, -0.1613304764032364, 0.012056032195687294, -0.1437864750623703, 0.010370835661888123, -0.08332697302103043, 0.08352860808372498, -0.009336385875940323, -0.040623150765895844, -0.05954933911561966, -0.005279048811644316, 0.04284205287694931, -0.015335856936872005, 0.29630112648010254, 0.008494893088936806, 0.13189546763896942, 0.16303826868534088, 0.022608771920204163, -0.1718173325061798, -0.01953105255961418, -0.03700250759720802, -0.0736953616142273, 0.07995044440031052, -0.2031794935464859, 0.057182274758815765, 0.07621432095766068, -0.06988967210054398, 0.06107276678085327, 0.08900803327560425, 0.01489779632538557, -0.04548985883593559, 0.16442106664180756, -0.1671867072582245, -0.0006360075785778463, -0.008117838762700558, 0.10618383437395096, 0.0521981455385685, 0.017339367419481277, 0.12554024159908295, 0.005745370872318745, -0.036925509572029114, 0.026103302836418152, 0.021035362035036087, -0.0628022700548172, 0.008186271414160728, 0.0677781030535698, 0.028965607285499573, -0.1132744550704956, 0.07926022261381149, 0.06677495688199997, -0.129411518573761, -0.010128652676939964, 0.09587135165929794, -0.1095706969499588, -0.16171106696128845, 0.00618077302351594, 0.04380159080028534, -0.08738097548484802, -0.08622825890779495, -0.024138811975717545, -0.1249731257557869, 0.04831845685839653, 0.09633947908878326, 0.13090389966964722, 0.06443681567907333, -0.0017022290267050266, -0.06870454549789429, 0.05082831531763077, 0.001563939149491489, -0.07709051668643951, 0.036532118916511536, -0.12133266031742096, -0.02846197411417961, 0.010334618389606476, 0.11554903537034988, -0.06259089708328247, -0.0042324368841946125, -0.12426569312810898, 0.001558715826831758, -0.11614777147769928, -0.029476512223482132, -0.08306972682476044, -0.033092986792325974, 0.02484716661274433, -0.08117660880088806, -0.04502910003066063, -0.01520068384706974, -0.14415021240711212, -0.024689702317118645, -0.04193141311407089, 0.09420331567525864, -0.104824498295784, -0.05173527076840401, 0.09438075870275497, -0.011454490944743156, 0.09049908816814423, 0.0417814776301384, -0.05572384595870972, 0.07632939517498016, -0.07947538793087006, -0.10600963234901428, 0.08863865584135056, 0.03787854313850403, 0.046659957617521286, -0.005956739187240601, -0.011253931559622288, 0.09342904388904572, -0.0007177229854278266, 0.042248114943504333, 0.011596323922276497, -0.12917691469192505, -0.02238038182258606, -0.010726326145231724, -0.13931554555892944, 0.015678487718105316, -0.09275448322296143, 0.14498212933540344, 0.01790434494614601, 0.17994743585586548, -0.005748189985752106, 0.031212467700242996, -0.06608235090970993, 0.025236306712031364, -0.057449471205472946, -0.17951899766921997, -0.12403986603021622, -0.06836044043302536, -0.027640171349048615, -0.028845123946666718, 0.2805347144603729, 0.08917142450809479, -0.08347166329622269, 0.07290034741163254, 0.07196228206157684, 0.01505149994045496, 0.02077554725110531, 0.22023816406726837, 0.06361716240644455, -0.02245054394006729, -0.061880506575107574, 0.019000306725502014, 0.022344766184687614, -0.03256882727146149, 0.07464652508497238, 0.1169479712843895, 0.1142251044511795, 0.07622475177049637, 0.02704872190952301, -0.016070278361439705, -0.11697600036859512, -0.1287461221218109, 0.014688584953546524, 0.119608573615551, -0.03686606511473656, 0.059426259249448776, 0.13529466092586517, -0.037811972200870514, 0.03956560790538788, -0.086575448513031, 0.012540166266262531, -0.17560049891471863, -0.09157182276248932, -0.07360459119081497, -0.13344724476337433, -0.03314107656478882, -0.052344802767038345, 0.03984220325946808, 0.1292441338300705, 0.032210953533649445, -0.03851942718029022, -0.015154266729950905, -0.015868499875068665, -0.04258604347705841, 0.03028041310608387, -0.023542780429124832, 0.0002562306181062013, -0.06029721349477768, -0.02395167201757431, -0.10368048399686813, -0.02776411361992359, -0.05679041147232056, 0.04382525384426117, -0.027991825714707375, 0.021414659917354584, -0.12156787514686584, -0.08556406199932098, -0.0529513917863369, 0.020444447174668312, -0.005759124178439379, 0.18267738819122314, 0.005337047390639782, 0.031859856098890305, 0.07273654639720917, 0.2198667675256729, -0.09987863898277283, -0.1095680221915245, -0.045107651501894, 0.17264553904533386, 0.04104243591427803, 0.06176810711622238, -0.020409779623150826, 0.0026832649018615484, -0.09016072750091553, 0.26033926010131836, 0.33503320813179016, -0.05959988385438919, 0.06000038608908653, -0.001532332506030798, 0.012618477456271648, 0.0598568320274353, 0.1106472909450531, 0.1316719800233841, 0.16129449009895325, -0.08782772719860077, -0.0036300979554653168, -0.04469006136059761, 0.003653900232166052, -0.15766099095344543, 0.07017975300550461, -0.0036263535730540752, -0.09436394274234772, -0.022905346006155014, 0.057690780609846115, -0.09149443358182907, 0.07677844911813736, 0.012266093865036964, -0.18119388818740845, -0.05720311030745506, 0.02039954625070095, 0.21919496357440948, -0.017752492800354958, 0.05047750845551491, -0.03148294985294342, -0.05256124958395958, 0.07185778021812439, -0.028453262522816658, -0.16809456050395966, -0.052983757108449936, 0.11245088279247284, -0.033219534903764725, 0.11339452862739563, -0.01019338145852089, 0.02442827820777893, 0.09866058081388474, 0.06523024290800095, -0.09986453503370285, 0.06829454749822617, 0.024860212579369545, -0.06409720331430435, -0.025689758360385895, -0.10032090544700623, -0.02310752309858799, -0.06327763199806213, 0.06990473717451096, -0.10408949106931686, 0.047875236719846725, -0.02276439219713211, -0.041091352701187134, -0.03285158798098564, 0.01771629974246025, -0.06046866253018379, 0.07364235818386078, 0.030026888474822044, -0.036611031740903854, -0.07146476209163666, -0.04183752462267876, -0.04579772427678108, 0.008587779477238655, -0.16310185194015503, -0.11530570685863495, -0.004291696939617395, -0.022359905764460564, 0.07707220315933228, 0.03829155117273331, -0.09312435239553452, -0.03533054143190384, -0.05517655238509178, 0.0333234928548336, -0.12787498533725739, 0.0459769070148468, 0.08696694672107697, -0.015718704089522362, -0.010699767619371414, -0.06356548517942429, 0.019441908225417137, 0.023609060794115067, -0.10050466656684875, -0.07246299833059311 ]
null
null
transformers
# Cross-Encoder for MS Marco This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco) ## Usage with Transformers ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('model_name') tokenizer = AutoTokenizer.from_pretrained('model_name') features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits print(scores) ``` ## Usage with SentenceTransformers The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('model_name', max_length=512) scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')]) ``` ## Performance In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset. | Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec | | ------------- |:-------------| -----| --- | | **Version 2 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000 | cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100 | cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500 | cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800 | cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960 | **Version 1 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000 | cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900 | cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680 | cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340 | **Other models** | | | | nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900 | nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340 | nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100 | Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340 | amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330 | sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720 Note: Runtime was computed on a V100 GPU.
{"license": "apache-2.0"}
text-classification
cross-encoder/ms-marco-TinyBERT-L-4
[ "transformers", "pytorch", "jax", "bert", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
Cross-Encoder for MS Marco ========================== This model was trained on the MS Marco Passage Ranking task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco Usage with Transformers ----------------------- Usage with SentenceTransformers ------------------------------- The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this: Performance ----------- In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset. Note: Runtime was computed on a V100 GPU.
[]
[ "TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 47 ]
[ "passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.031222937628626823, 0.13048580288887024, -0.006592764053493738, 0.035769447684288025, 0.11023839563131332, 0.023622313514351845, 0.1118343323469162, 0.14141945540905, 0.011887344531714916, -0.07865098118782043, 0.14571447670459747, 0.2256961315870285, -0.011237370781600475, 0.06962031126022339, -0.07936780899763107, -0.23767422139644623, 0.08903764188289642, 0.04316740483045578, -0.02329845540225506, 0.09714217483997345, 0.11549460887908936, -0.0445655882358551, 0.0482446625828743, -0.005831177346408367, -0.07852476090192795, 0.02374298870563507, 0.05234311893582344, -0.11925916373729706, 0.07824034988880157, 0.038800518959760666, 0.08641701191663742, 0.047947775572538376, -0.02588331513106823, -0.18091587722301483, 0.020593609660863876, 0.014527560211718082, -0.08456390351057053, 0.06201760843396187, 0.06459520757198334, -0.05026666447520256, 0.028283588588237762, 0.04363081604242325, -0.00040016864659264684, 0.05442293360829353, -0.08365041762590408, -0.18777050077915192, -0.07788987457752228, 0.09138200432062149, 0.0822538360953331, 0.07590723037719727, 0.041403573006391525, 0.14720378816127777, -0.12174759060144424, 0.06750258803367615, 0.11365879327058792, -0.3379494249820709, -0.00012072643585270271, 0.0726609155535698, 0.019400611519813538, 0.02270316518843174, -0.02585512027144432, 0.032551705837249756, 0.04672111198306084, 0.0034406250342726707, 0.024993298575282097, -0.06176041439175606, -0.11698492616415024, 0.03534247726202011, -0.04940676689147949, -0.06621868163347244, 0.24049511551856995, -0.013352715410292149, 0.029132870957255363, -0.0047011785209178925, -0.06106462702155113, 0.0031755072996020317, -0.027250586077570915, 0.04743704944849014, 0.022055545821785927, 0.11091086268424988, 0.07261817902326584, -0.016392147168517113, -0.1279716044664383, 0.0008985644089989364, -0.1909598708152771, 0.08758813887834549, 0.020127490162849426, 0.08088716119527817, -0.15236561000347137, 0.055559054017066956, 0.028175966814160347, -0.11545844376087189, 0.0005364827229641378, -0.08256713300943375, 0.10449586063623428, 0.011525196954607964, -0.06949394196271896, 0.04138573631644249, 0.13127945363521576, 0.2090783268213272, 0.053110383450984955, 0.030188433825969696, -0.050601422786712646, 0.11591508984565735, -0.023805562406778336, 0.08434237539768219, 0.03620612993836403, -0.007659158669412136, 0.11160806566476822, -0.11625667661428452, 0.05364016443490982, -0.03825801610946655, -0.1572587639093399, -0.018107563257217407, 0.027624348178505898, 0.11380995064973831, 0.04983453080058098, 0.0565413273870945, -0.05016535893082619, -0.0004966393462382257, 0.14144660532474518, -0.06383424252271652, 0.01579524390399456, 0.00790349394083023, 0.028509506955742836, 0.06099670007824898, 0.026157883927226067, 0.02817358262836933, -0.0572095662355423, 0.09920138865709305, -0.04989606887102127, -0.011875666677951813, -0.037143491208553314, -0.008297450840473175, 0.0870564803481102, -0.08851300179958344, 0.05059066042304039, -0.15062427520751953, -0.14653214812278748, 0.030236298218369484, 0.0720018520951271, 0.012049839831888676, -0.09377799183130264, 0.034003641456365585, -0.010921157896518707, 0.047230709344148636, -0.08446047455072403, -0.016658423468470573, -0.08630052953958511, 0.07090666890144348, -0.09199928492307663, 0.024058466777205467, -0.15290489792823792, 0.06158958747982979, -0.11669103801250458, -0.017135046422481537, -0.051425229758024216, -0.03311094269156456, -0.09284032881259918, 0.1942492127418518, -0.043372999876737595, -0.03203170374035835, 0.013681466691195965, 0.005522655323147774, -0.06157875433564186, 0.12065905332565308, -0.09521565586328506, -0.06215933337807655, 0.19024963676929474, -0.1084100529551506, -0.1800156831741333, 0.07037068903446198, 0.001334881060756743, 0.00610776524990797, 0.0773395523428917, 0.1616002321243286, 0.09672778099775314, -0.02310440130531788, 0.09237653762102127, 0.14067962765693665, -0.06062621623277664, -0.19009415805339813, 0.05171007290482521, -0.04433976858854294, -0.14356684684753418, 0.06429912894964218, -0.017190508544445038, 0.0829833447933197, 0.002651303308084607, -0.07413939386606216, -0.052518080919981, -0.043487370014190674, 0.021368304267525673, 0.025271417573094368, 0.08122741430997849, -0.07590355724096298, -0.008831245824694633, -0.031189359724521637, 0.034308433532714844, 0.05369502305984497, 0.04838934913277626, -0.06191778555512428, 0.061439886689186096, 0.02486778050661087, 0.033942271023988724, -0.14579053223133087, -0.007896673865616322, -0.0017176179680973291, 0.03740164265036583, -0.0015597393503412604, 0.07824065536260605, 0.036536093801259995, -0.08376944065093994, -0.008764523081481457, -0.015120083466172218, 0.14143557846546173, 0.0484895296394825, -0.026612315326929092, -0.12957508862018585, 0.049588754773139954, -0.0399571917951107, 0.02532694675028324, -0.01934116706252098, 0.024379892274737358, 0.0456324964761734, 0.08728967607021332, -0.029270634055137634, 0.09425725787878036, -0.034797847270965576, -0.0007632380584254861, -0.06166345998644829, 0.0052313413470983505, 0.12030249834060669, 0.04904041439294815, -0.06734607368707657, 0.19378910958766937, -0.05520053952932358, 0.28220629692077637, 0.20493178069591522, -0.18299666047096252, 0.06226447969675064, 0.0013990707229822874, -0.010337709449231625, -0.000578522973228246, 0.03439730405807495, 0.0335940420627594, 0.03681229054927826, 0.010316891595721245, 0.17409954965114594, -0.052756331861019135, -0.04641488194465637, -0.01663317158818245, -0.05680250748991966, -0.013033241964876652, 0.051225002855062485, 0.14472876489162445, -0.1932520568370819, 0.19265632331371307, 0.32094016671180725, -0.01836542785167694, 0.08962716907262802, -0.07507248222827911, 0.018457842990756035, 0.055129729211330414, -0.05118462070822716, -0.022528575733304024, -0.02265028841793537, -0.12742535769939423, -0.000777439447119832, 0.09433021396398544, 0.040656160563230515, 0.04232742637395859, -0.13055551052093506, -0.05610394477844238, 0.00009042652527568862, -0.00696546072140336, -0.050355494022369385, 0.057761259377002716, 0.004052021540701389, 0.09031425416469574, -0.02385767363011837, -0.10479959100484848, 0.1368882954120636, -0.0031609961297363043, -0.07879859209060669, 0.1476660817861557, -0.1670645922422409, -0.2675674557685852, -0.13479484617710114, -0.1487317681312561, -0.027203327044844627, 0.01467166654765606, 0.13942782580852509, -0.060368604958057404, -0.0525057278573513, 0.012911345809698105, -0.09812077134847641, -0.037766944617033005, 0.028387591242790222, -0.0297428946942091, 0.06439930200576782, 0.004743938334286213, -0.10729271918535233, -0.0666976347565651, 0.01350174006074667, -0.031026283279061317, 0.0922672301530838, -0.08812612295150757, 0.06524661928415298, 0.12273956090211868, 0.017020998522639275, 0.037275176495313644, -0.03228471800684929, 0.13733430206775665, -0.03686793893575668, -0.008670826442539692, 0.2096070498228073, -0.03775579482316971, 0.07707484066486359, 0.150436133146286, 0.04063309356570244, -0.04427816718816757, 0.015577026642858982, -0.049652185291051865, -0.07288471609354019, -0.26840829849243164, -0.10533773899078369, -0.10727188736200333, 0.053441211581230164, 0.06932765990495682, 0.09347819536924362, 0.14388203620910645, 0.07078276574611664, -0.010041584260761738, 0.018074555322527885, 0.006245976779609919, 0.08051405102014542, 0.23954537510871887, -0.003942674491554499, 0.12337736785411835, -0.10948646068572998, -0.053607091307640076, 0.1174393892288208, 0.055854130536317825, 0.10559877753257751, 0.12827354669570923, 0.0634327083826065, 0.056444261223077774, 0.16488425433635712, 0.10907312482595444, 0.11669280380010605, 0.02187339775264263, -0.006276066415011883, -0.05460451915860176, -0.015361282974481583, -0.06251870095729828, 0.03777467831969261, -0.013609152287244797, -0.12285177409648895, -0.04584629461169243, -0.1615571826696396, 0.06845147162675858, 0.18546631932258606, 0.020312797278165817, -0.14464733004570007, 0.025239532813429832, 0.09377814084291458, -0.01945749670267105, -0.053647808730602264, 0.10041272640228271, -0.11533371359109879, -0.11224431544542313, 0.11980185657739639, -0.028893087059259415, 0.14660626649856567, -0.016873618587851524, 0.04937293007969856, -0.013568560592830181, -0.13501383364200592, 0.06583409011363983, 0.14391310513019562, -0.30061182379722595, 0.19650478661060333, -0.007461524568498135, -0.043495722115039825, -0.08261296898126602, 0.003955775871872902, 0.06695891916751862, 0.24799129366874695, 0.0889478474855423, 0.01616300828754902, -0.12551507353782654, -0.0395657904446125, -0.062256503850221634, 0.028937330469489098, 0.02037951722741127, -0.0179972555488348, -0.057268936187028885, -0.06969304382801056, -0.022713175043463707, -0.00025611647288315, 0.03760894015431404, -0.026397885754704475, -0.15958911180496216, 0.05402941629290581, 0.09700854867696762, 0.052172884345054626, -0.05016152188181877, -0.03343906253576279, -0.10161250084638596, 0.19424182176589966, -0.08696603775024414, -0.07252633571624756, -0.09862473607063293, -0.13247954845428467, 0.025107014924287796, -0.06244604289531708, 0.061074692755937576, -0.08559167385101318, 0.011953501962125301, -0.047901470214128494, -0.21219149231910706, 0.10887160897254944, -0.13795332610607147, -0.04415391758084297, -0.05766300484538078, 0.12810908257961273, -0.11148732900619507, 0.022969752550125122, 0.034964412450790405, 0.016335586085915565, -0.10780297219753265, -0.10891950130462646, -0.013977473601698875, 0.030052734538912773, 0.05653700977563858, -0.04045039415359497, -0.09685365110635757, -0.04719218611717224, 0.013199465349316597, -0.03896326571702957, 0.24409520626068115, 0.17588753998279572, -0.08934667706489563, 0.19285185635089874, 0.18291422724723816, -0.09928718209266663, -0.3110765814781189, -0.15143781900405884, -0.13517117500305176, -0.10138122737407684, -0.023945894092321396, -0.15794737637043, 0.12962974607944489, 0.04799811169505119, -0.09087679535150528, 0.08863683044910431, -0.15966655313968658, -0.0805826410651207, 0.19894154369831085, -0.033684100955724716, 0.3326871395111084, -0.14180131256580353, -0.08361893892288208, -0.09042175114154816, -0.17759105563163757, 0.15467439591884613, -0.0708114504814148, 0.052896592766046524, -0.01295430026948452, 0.002900085411965847, -0.007546346168965101, -0.04754116013646126, 0.11410257965326309, -0.025622457265853882, 0.01940036565065384, -0.1225917786359787, -0.012375768274068832, 0.06971857696771622, -0.029210612177848816, 0.029483215883374214, -0.1414121687412262, 0.02156968228518963, -0.11909553408622742, -0.021438248455524445, -0.06554215401411057, 0.0683499425649643, 0.0028498645406216383, -0.04397595301270485, -0.019701682031154633, -0.024387499317526817, 0.0310470312833786, -0.016381951048970222, 0.2860105335712433, 0.023737331852316856, 0.09919518977403641, 0.11846044659614563, 0.08302575349807739, -0.18312297761440277, -0.011074194684624672, -0.10109596699476242, -0.08606860041618347, 0.06998859345912933, -0.1615401655435562, 0.06356693059206009, 0.09814772009849548, -0.061531227082014084, 0.061257727444171906, 0.08415394276380539, 0.019474968314170837, -0.0573870912194252, 0.13776203989982605, -0.17544475197792053, 0.031624019145965576, -0.017194677144289017, 0.11635028570890427, 0.06414506584405899, 0.05926978960633278, 0.11567357182502747, 0.013719623908400536, -0.05012055113911629, 0.025427423417568207, 0.02964070253074169, -0.05026624724268913, 0.027212506160140038, 0.05781693384051323, 0.008329174481332302, -0.1355985403060913, 0.10132720321416855, 0.03726547211408615, -0.1443147212266922, -0.025202171877026558, 0.10883472114801407, -0.1498357653617859, -0.13316985964775085, 0.03644883260130882, 0.08920352905988693, -0.11446324735879898, -0.10948441922664642, -0.049708377569913864, -0.13059930503368378, 0.07466436177492142, 0.111910380423069, 0.12298257648944855, 0.06841865926980972, -0.013371058739721775, -0.06788983941078186, 0.0383983813226223, 0.0029060887172818184, -0.07893457263708115, 0.02773320861160755, -0.1128721833229065, -0.06500279903411865, 0.017772741615772247, 0.10809708386659622, -0.053233202546834946, -0.017664581537246704, -0.11155769228935242, 0.03594955801963806, -0.1270546019077301, 0.004668314475566149, -0.08084987848997116, -0.004947167355567217, 0.017740819603204727, -0.06953023374080658, -0.03667932003736496, -0.005178960505872965, -0.12781259417533875, -0.015513245016336441, -0.02598072960972786, 0.07093318551778793, -0.10114233940839767, -0.05152695253491402, 0.09541739523410797, -0.012447931803762913, 0.0973988026380539, 0.0750475525856018, -0.06654786318540573, 0.10046802461147308, -0.1393914669752121, -0.09427568316459656, 0.09227922558784485, 0.04359714314341545, 0.03399856761097908, -0.008078759536147118, 0.016459334641695023, 0.1202658861875534, -0.033905576914548874, 0.043104592710733414, 0.024574751034379005, -0.147372767329216, -0.03224669396877289, 0.0014853047905489802, -0.14264920353889465, -0.009076599031686783, -0.09593779593706131, 0.12641161680221558, 0.019937938079237938, 0.19958613812923431, -0.019179530441761017, 0.04851064458489418, -0.04842802882194519, 0.01850293017923832, -0.03834648057818413, -0.17755559086799622, -0.14522022008895874, -0.07037320733070374, -0.038313861936330795, -0.018554124981164932, 0.251798540353775, 0.04093918949365616, -0.042603056877851486, 0.07759085297584534, 0.06571243703365326, 0.016727644950151443, 0.0212676003575325, 0.24752330780029297, 0.06033453717827797, -0.010926000773906708, -0.06129152700304985, 0.01934751868247986, 0.017677469179034233, -0.056927431374788284, 0.09427402168512344, 0.09811029583215714, 0.068744957447052, 0.06465623527765274, 0.016684463247656822, 0.026975518092513084, -0.10663208365440369, -0.14640776813030243, 0.015742281451821327, 0.09490249305963516, 0.01810981146991253, 0.11160247772932053, 0.12422402948141098, -0.03371858596801758, 0.0225676279515028, -0.06432820111513138, -0.0060334219597280025, -0.18783824145793915, -0.10917981714010239, -0.08849360793828964, -0.12464868277311325, 0.0013375792186707258, -0.048415862023830414, 0.0207132026553154, 0.08100679516792297, 0.04537936672568321, -0.06281905621290207, -0.02731252834200859, -0.021290594711899757, -0.044625598937273026, 0.04943249374628067, -0.03116297721862793, -0.023418989032506943, -0.029619375243782997, -0.031456053256988525, -0.10911190509796143, -0.04594817012548447, -0.05295524001121521, 0.040317900478839874, -0.013821911998093128, 0.040155891329050064, -0.10918360203504562, -0.08355793356895447, -0.03505280986428261, 0.029276607558131218, -0.006034370511770248, 0.17343491315841675, 0.011322897858917713, 0.03913239762187004, 0.096920907497406, 0.17550869286060333, -0.07044066488742828, -0.12284383922815323, -0.05195518955588341, 0.19905413687229156, 0.05575999990105629, 0.06016545370221138, 0.015420405194163322, 0.01774226874113083, -0.05175405368208885, 0.3056987226009369, 0.28822413086891174, -0.04234142228960991, 0.03812962397933006, -0.012957140803337097, 0.017551559954881668, 0.08646208047866821, 0.14314526319503784, 0.1225508525967598, 0.17820580303668976, -0.07520273327827454, -0.020763490349054337, -0.033969681710004807, 0.0011272527044638991, -0.16123616695404053, 0.0776604637503624, -0.01038258709013462, -0.08717965334653854, -0.01985633186995983, 0.09582911431789398, -0.111983522772789, 0.1065826490521431, 0.0007611355395056307, -0.12746141850948334, -0.04855658859014511, -0.01268674898892641, 0.18558096885681152, 0.010363814420998096, 0.029743928462266922, -0.018424078822135925, -0.06368906795978546, 0.10191062837839127, -0.01379233319312334, -0.20874279737472534, -0.07239016890525818, 0.09477964043617249, -0.033585019409656525, 0.12148012965917587, 0.005590854678303003, 0.05478139966726303, 0.07904579490423203, 0.06767632812261581, -0.09296553581953049, 0.06549388915300369, 0.012761259451508522, -0.05543714389204979, 0.0046954466961324215, -0.09866335242986679, -0.024628590792417526, -0.06140054389834404, 0.04757868871092796, -0.07632169872522354, 0.04376089945435524, -0.03289888799190521, -0.06448434293270111, -0.04339258000254631, 0.06421805173158646, -0.07498079538345337, 0.05848152935504913, 0.030363943427801132, -0.026018934324383736, -0.07160912454128265, -0.061227958649396896, -0.02492934837937355, 0.023381372913718224, -0.1795109510421753, -0.09561535716056824, -0.007375465705990791, -0.031752556562423706, 0.09167668968439102, 0.045114073902368546, -0.10676556080579758, -0.019531425088644028, -0.09544464945793152, 0.016059521585702896, -0.16992658376693726, 0.04376090317964554, 0.06283412128686905, -0.003013976849615574, -0.006386113353073597, -0.06914757937192917, 0.025922639295458794, 0.02013215981423855, -0.09332647174596786, -0.08685147762298584 ]
null
null
transformers
# Cross-Encoder for MS Marco This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco) ## Usage with Transformers ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('model_name') tokenizer = AutoTokenizer.from_pretrained('model_name') features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits print(scores) ``` ## Usage with SentenceTransformers The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('model_name', max_length=512) scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')]) ``` ## Performance In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset. | Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec | | ------------- |:-------------| -----| --- | | **Version 2 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000 | cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100 | cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500 | cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800 | cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960 | **Version 1 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000 | cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900 | cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680 | cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340 | **Other models** | | | | nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900 | nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340 | nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100 | Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340 | amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330 | sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720 Note: Runtime was computed on a V100 GPU.
{"license": "apache-2.0"}
text-classification
cross-encoder/ms-marco-TinyBERT-L-6
[ "transformers", "pytorch", "jax", "bert", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
Cross-Encoder for MS Marco ========================== This model was trained on the MS Marco Passage Ranking task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco Usage with Transformers ----------------------- Usage with SentenceTransformers ------------------------------- The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this: Performance ----------- In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset. Note: Runtime was computed on a V100 GPU.
[]
[ "TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ 51 ]
[ "passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ -0.005313366185873747, 0.12051703780889511, -0.004693866241723299, 0.03580747917294502, 0.08237535506486893, 0.015574881806969643, 0.09972856193780899, 0.1512490212917328, -0.002342487219721079, -0.035675179213285446, 0.1512344777584076, 0.1678246557712555, -0.03263122960925102, 0.07377322763204575, -0.06331735849380493, -0.23472362756729126, 0.10480628162622452, 0.022721635177731514, -0.06391183286905289, 0.0852574035525322, 0.11916844546794891, -0.06770222634077072, 0.04979204759001732, -0.006682221312075853, -0.07624185085296631, 0.03358222916722298, 0.023558329790830612, -0.11514781415462494, 0.09597253054380417, 0.031052103266119957, 0.10091041028499603, 0.054762229323387146, -0.03489936888217926, -0.1495836228132248, 0.026294076815247536, 0.029823211953043938, -0.08911542594432831, 0.07207200676202774, 0.03851554915308952, -0.03519143536686897, 0.04294948652386665, 0.0052515072748064995, -0.0014171735383570194, 0.028817106038331985, -0.09800101071596146, -0.21412332355976105, -0.07283638417720795, 0.07232868671417236, 0.027582870796322823, 0.07712103426456451, 0.024217145517468452, 0.15384386479854584, -0.14576973021030426, 0.05114344507455826, 0.14506278932094574, -0.3469622731208801, -0.0045180246233940125, 0.13070425391197205, 0.07035423070192337, 0.04190012812614441, -0.03805508464574814, 0.04657916724681854, 0.05582089722156525, 0.0025910616386681795, 0.06670897454023361, -0.04245023429393768, -0.1239764392375946, 0.06446674466133118, -0.07143140584230423, -0.07907963544130325, 0.2631944417953491, -0.019031910225749016, 0.0633535161614418, -0.020391980186104774, -0.07111599296331406, -0.02444433979690075, 0.0007090915460139513, 0.04914804920554161, 0.029526568949222565, 0.10229268670082092, 0.07079360634088516, -0.03269055485725403, -0.14858068525791168, 0.020561350509524345, -0.18860138952732086, 0.06773007661104202, 0.0031633847393095493, 0.07213912159204483, -0.1380876749753952, 0.05910053104162216, 0.022541260346770287, -0.12431655079126358, 0.026827946305274963, -0.08040676265954971, 0.09522116184234619, 0.02466272935271263, -0.10294348746538162, 0.048924896866083145, 0.12230401486158371, 0.1840950846672058, 0.025208648294210434, 0.004822383169084787, -0.026003237813711166, 0.12904393672943115, 0.010278739035129547, 0.08728154003620148, -0.017540551722049713, -0.016535868868231773, 0.07604997605085373, -0.07513169944286346, 0.056044477969408035, -0.06161702796816826, -0.1701550930738449, -0.03133761137723923, 0.03735613077878952, 0.09264276921749115, 0.08358346670866013, 0.05583064258098602, -0.02714722789824009, 0.025454645976424217, 0.1282920092344284, -0.06134769693017006, 0.04080605506896973, -0.00510947173461318, 0.03630151227116585, 0.02500070445239544, 0.02943694032728672, 0.023852426558732986, -0.017657775431871414, 0.06957349926233292, -0.06750348955392838, -0.010275444015860558, -0.051824845373630524, -0.060393817722797394, 0.08819574117660522, -0.0934952050447464, 0.03543853387236595, -0.15935862064361572, -0.08309947699308395, 0.017079420387744904, 0.06980738043785095, 0.0021625745575875044, -0.09810265898704529, 0.07114944607019424, -0.04728621616959572, 0.0843905657529831, -0.08494804799556732, 0.002027006121352315, -0.09370028227567673, 0.05686311423778534, -0.10129949450492859, 0.055879801511764526, -0.17257225513458252, 0.06273247301578522, -0.1052277535200119, -0.0019254190847277641, -0.05104181170463562, -0.04737512394785881, -0.09146487712860107, 0.16669559478759766, -0.022341273725032806, -0.0318073108792305, -0.019266171380877495, 0.031578149646520615, -0.04990849643945694, 0.08952520042657852, -0.13634717464447021, -0.03626799210906029, 0.14703963696956635, -0.09130478650331497, -0.16326141357421875, 0.0717688575387001, 0.00046021450543776155, 0.0031928219832479954, 0.0205746628344059, 0.2063581943511963, 0.0636834129691124, -0.038732003420591354, 0.0429588183760643, 0.14875924587249756, -0.031094709411263466, -0.15359465777873993, 0.06786390393972397, -0.029294706881046295, -0.09197673946619034, 0.043804608285427094, -0.017092304304242134, 0.06857074052095413, 0.011296656914055347, -0.08280479162931442, -0.07357008755207062, -0.03364529088139534, 0.04349697753787041, 0.0174326840788126, 0.09276469796895981, -0.07873370498418808, -0.03898242488503456, -0.0033968340139836073, 0.03604090213775635, 0.05958762764930725, 0.056634191423654556, -0.03324146941304207, 0.10759710520505905, -0.00020664279873017222, 0.019652394577860832, -0.1466042846441269, -0.022884350270032883, -0.024007568135857582, 0.029566245153546333, -0.0034570449497550726, 0.1761774867773056, 0.028720151633024216, -0.09089623391628265, -0.005154056008905172, -0.005816992837935686, 0.10818760842084885, 0.07017345726490021, -0.0435544028878212, -0.122578464448452, 0.02524160034954548, -0.05991243198513985, 0.011386431753635406, -0.028770629316568375, 0.03522561118006706, 0.0614837110042572, 0.095530666410923, -0.04097152501344681, 0.10141308605670929, -0.024964042007923126, 0.0062868064269423485, -0.08510880917310715, 0.002189264167100191, 0.09858357906341553, 0.033706147223711014, -0.04600461944937706, 0.2194674164056778, -0.10151731222867966, 0.2976343631744385, 0.2267475426197052, -0.1751745343208313, 0.057022739201784134, 0.03313778340816498, -0.019422493875026703, 0.019689708948135376, 0.0130884675309062, -0.001759985345415771, -0.028527354821562767, -0.02610705979168415, 0.14932112395763397, -0.04963885620236397, -0.03732553496956825, -0.023624418303370476, -0.06671696156263351, -0.03914187103509903, 0.04179641976952553, 0.09773124009370804, -0.15121658146381378, 0.21856796741485596, 0.39029189944267273, -0.04813329502940178, 0.11297299712896347, -0.055986665189266205, 0.025421306490898132, 0.030753333121538162, -0.07714325934648514, -0.05503294616937637, 0.02333962731063366, -0.13534876704216003, -0.02139679528772831, 0.1016768291592598, 0.056578896939754486, 0.04646281898021698, -0.13725902140140533, -0.0547390952706337, 0.02008267678320408, 0.020364783704280853, -0.05890339985489845, 0.08217864483594894, 0.018010256811976433, 0.09777913242578506, -0.0105742197483778, -0.10413795709609985, 0.10512387752532959, -0.0025146319530904293, -0.046441737562417984, 0.12223948538303375, -0.1751604825258255, -0.26337674260139465, -0.10911697149276733, -0.10594077408313751, -0.0219988115131855, 0.016123976558446884, 0.14658363163471222, -0.05834493041038513, -0.04452504217624664, -0.03016248345375061, -0.08486530184745789, -0.06864798069000244, 0.03474634513258934, -0.04818902164697647, 0.05575546622276306, 0.0009074720437638462, -0.1152714416384697, -0.06752760708332062, 0.020496297627687454, -0.03415825590491295, 0.091299869120121, -0.030293690040707588, 0.07342889904975891, 0.135502427816391, 0.007198224309831858, 0.027944006025791168, -0.030184734612703323, 0.16622070968151093, -0.03921347111463547, 0.00924068596214056, 0.19809506833553314, -0.0031844095792621374, 0.0713656023144722, 0.18086189031600952, 0.03603030741214752, -0.022779395803809166, 0.01003937516361475, -0.03807242214679718, -0.08706989139318466, -0.20672155916690826, -0.11673396080732346, -0.12385633587837219, 0.03123038075864315, 0.058827221393585205, 0.10125505924224854, 0.1326202005147934, 0.05287611857056618, 0.0016830338863655925, 0.002457578666508198, -0.029018420726060867, 0.06340944021940231, 0.23763202130794525, -0.03204694390296936, 0.14382527768611908, -0.10360123962163925, -0.058313898742198944, 0.11185727268457413, 0.07886284589767456, 0.08192996680736542, 0.10665401816368103, 0.03152311220765114, 0.0695836991071701, 0.19727741181850433, 0.10917509347200394, 0.08815993368625641, 0.02420349419116974, -0.010355910286307335, -0.06285016238689423, -0.023843849077820778, -0.023379191756248474, 0.06658007949590683, 0.039418045431375504, -0.13914524018764496, -0.029837213456630707, -0.19487792253494263, 0.08332343399524689, 0.151483952999115, 0.059639498591423035, -0.13124634325504303, 0.031023966148495674, 0.10783152282238007, -0.01136285811662674, -0.04613389074802399, 0.08763978630304337, -0.02930242381989956, -0.08054038137197495, 0.07484424859285355, 0.007519237697124481, 0.12197799980640411, 0.02354556694626808, 0.07243271917104721, -0.04469209909439087, -0.16604083776474, 0.056991759687662125, 0.12605127692222595, -0.2823342978954315, 0.20038354396820068, -0.02226925455033779, -0.09456770122051239, -0.07824184000492096, 0.0034344338346272707, 0.07712201029062271, 0.22976431250572205, 0.061447788029909134, 0.03248010203242302, -0.1648344248533249, -0.04715387150645256, -0.03496819734573364, 0.012963905930519104, 0.017423907294869423, -0.02331795170903206, -0.03611692786216736, -0.06621395796537399, -0.002671534428372979, 0.020914889872074127, 0.1569015085697174, -0.022213159129023552, -0.17191611230373383, 0.06727875024080276, 0.11111561954021454, 0.02136494219303131, -0.04995092377066612, -0.043893180787563324, -0.16064417362213135, 0.16429398953914642, -0.034801822155714035, -0.03848220035433769, -0.09644030034542084, -0.11785909533500671, 0.05767476186156273, -0.043937042355537415, 0.05455517768859863, -0.08362634479999542, 0.027540860697627068, -0.06371035426855087, -0.19031928479671478, 0.11500367522239685, -0.12582644820213318, -0.04196672886610031, -0.053986817598342896, 0.08883348852396011, -0.15042348206043243, 0.04660613462328911, 0.03068644367158413, 0.06134520471096039, -0.15422704815864563, -0.11111470311880112, -0.0030049297492951155, 0.01256405096501112, 0.059495214372873306, -0.03420129418373108, -0.07899564504623413, -0.04707137122750282, 0.06431546807289124, -0.03068653494119644, 0.25191134214401245, 0.18924297392368317, -0.12640511989593506, 0.17781375348567963, 0.13471996784210205, -0.07904026657342911, -0.3402675986289978, -0.1414533108472824, -0.16139845550060272, -0.07071840018033981, 0.026064196601510048, -0.11777465045452118, 0.0958787128329277, 0.03335252031683922, -0.10947048664093018, 0.052748363465070724, -0.17017929255962372, -0.07672762125730515, 0.17991377413272858, -0.07845299690961838, 0.32294031977653503, -0.15600302815437317, -0.0581873282790184, -0.07140202820301056, -0.1397649347782135, 0.15831077098846436, -0.12095484882593155, 0.0640173852443695, -0.01085357554256916, 0.0009923696052283049, 0.000617707846686244, -0.05387040227651596, 0.15228526294231415, -0.05112951993942261, 0.02527214027941227, -0.1299176663160324, -0.04767915979027748, 0.10295214504003525, -0.06544466316699982, 0.023311404511332512, -0.1613304764032364, 0.012056032195687294, -0.1437864750623703, 0.010370835661888123, -0.08332697302103043, 0.08352860808372498, -0.009336385875940323, -0.040623150765895844, -0.05954933911561966, -0.005279048811644316, 0.04284205287694931, -0.015335856936872005, 0.29630112648010254, 0.008494893088936806, 0.13189546763896942, 0.16303826868534088, 0.022608771920204163, -0.1718173325061798, -0.01953105255961418, -0.03700250759720802, -0.0736953616142273, 0.07995044440031052, -0.2031794935464859, 0.057182274758815765, 0.07621432095766068, -0.06988967210054398, 0.06107276678085327, 0.08900803327560425, 0.01489779632538557, -0.04548985883593559, 0.16442106664180756, -0.1671867072582245, -0.0006360075785778463, -0.008117838762700558, 0.10618383437395096, 0.0521981455385685, 0.017339367419481277, 0.12554024159908295, 0.005745370872318745, -0.036925509572029114, 0.026103302836418152, 0.021035362035036087, -0.0628022700548172, 0.008186271414160728, 0.0677781030535698, 0.028965607285499573, -0.1132744550704956, 0.07926022261381149, 0.06677495688199997, -0.129411518573761, -0.010128652676939964, 0.09587135165929794, -0.1095706969499588, -0.16171106696128845, 0.00618077302351594, 0.04380159080028534, -0.08738097548484802, -0.08622825890779495, -0.024138811975717545, -0.1249731257557869, 0.04831845685839653, 0.09633947908878326, 0.13090389966964722, 0.06443681567907333, -0.0017022290267050266, -0.06870454549789429, 0.05082831531763077, 0.001563939149491489, -0.07709051668643951, 0.036532118916511536, -0.12133266031742096, -0.02846197411417961, 0.010334618389606476, 0.11554903537034988, -0.06259089708328247, -0.0042324368841946125, -0.12426569312810898, 0.001558715826831758, -0.11614777147769928, -0.029476512223482132, -0.08306972682476044, -0.033092986792325974, 0.02484716661274433, -0.08117660880088806, -0.04502910003066063, -0.01520068384706974, -0.14415021240711212, -0.024689702317118645, -0.04193141311407089, 0.09420331567525864, -0.104824498295784, -0.05173527076840401, 0.09438075870275497, -0.011454490944743156, 0.09049908816814423, 0.0417814776301384, -0.05572384595870972, 0.07632939517498016, -0.07947538793087006, -0.10600963234901428, 0.08863865584135056, 0.03787854313850403, 0.046659957617521286, -0.005956739187240601, -0.011253931559622288, 0.09342904388904572, -0.0007177229854278266, 0.042248114943504333, 0.011596323922276497, -0.12917691469192505, -0.02238038182258606, -0.010726326145231724, -0.13931554555892944, 0.015678487718105316, -0.09275448322296143, 0.14498212933540344, 0.01790434494614601, 0.17994743585586548, -0.005748189985752106, 0.031212467700242996, -0.06608235090970993, 0.025236306712031364, -0.057449471205472946, -0.17951899766921997, -0.12403986603021622, -0.06836044043302536, -0.027640171349048615, -0.028845123946666718, 0.2805347144603729, 0.08917142450809479, -0.08347166329622269, 0.07290034741163254, 0.07196228206157684, 0.01505149994045496, 0.02077554725110531, 0.22023816406726837, 0.06361716240644455, -0.02245054394006729, -0.061880506575107574, 0.019000306725502014, 0.022344766184687614, -0.03256882727146149, 0.07464652508497238, 0.1169479712843895, 0.1142251044511795, 0.07622475177049637, 0.02704872190952301, -0.016070278361439705, -0.11697600036859512, -0.1287461221218109, 0.014688584953546524, 0.119608573615551, -0.03686606511473656, 0.059426259249448776, 0.13529466092586517, -0.037811972200870514, 0.03956560790538788, -0.086575448513031, 0.012540166266262531, -0.17560049891471863, -0.09157182276248932, -0.07360459119081497, -0.13344724476337433, -0.03314107656478882, -0.052344802767038345, 0.03984220325946808, 0.1292441338300705, 0.032210953533649445, -0.03851942718029022, -0.015154266729950905, -0.015868499875068665, -0.04258604347705841, 0.03028041310608387, -0.023542780429124832, 0.0002562306181062013, -0.06029721349477768, -0.02395167201757431, -0.10368048399686813, -0.02776411361992359, -0.05679041147232056, 0.04382525384426117, -0.027991825714707375, 0.021414659917354584, -0.12156787514686584, -0.08556406199932098, -0.0529513917863369, 0.020444447174668312, -0.005759124178439379, 0.18267738819122314, 0.005337047390639782, 0.031859856098890305, 0.07273654639720917, 0.2198667675256729, -0.09987863898277283, -0.1095680221915245, -0.045107651501894, 0.17264553904533386, 0.04104243591427803, 0.06176810711622238, -0.020409779623150826, 0.0026832649018615484, -0.09016072750091553, 0.26033926010131836, 0.33503320813179016, -0.05959988385438919, 0.06000038608908653, -0.001532332506030798, 0.012618477456271648, 0.0598568320274353, 0.1106472909450531, 0.1316719800233841, 0.16129449009895325, -0.08782772719860077, -0.0036300979554653168, -0.04469006136059761, 0.003653900232166052, -0.15766099095344543, 0.07017975300550461, -0.0036263535730540752, -0.09436394274234772, -0.022905346006155014, 0.057690780609846115, -0.09149443358182907, 0.07677844911813736, 0.012266093865036964, -0.18119388818740845, -0.05720311030745506, 0.02039954625070095, 0.21919496357440948, -0.017752492800354958, 0.05047750845551491, -0.03148294985294342, -0.05256124958395958, 0.07185778021812439, -0.028453262522816658, -0.16809456050395966, -0.052983757108449936, 0.11245088279247284, -0.033219534903764725, 0.11339452862739563, -0.01019338145852089, 0.02442827820777893, 0.09866058081388474, 0.06523024290800095, -0.09986453503370285, 0.06829454749822617, 0.024860212579369545, -0.06409720331430435, -0.025689758360385895, -0.10032090544700623, -0.02310752309858799, -0.06327763199806213, 0.06990473717451096, -0.10408949106931686, 0.047875236719846725, -0.02276439219713211, -0.041091352701187134, -0.03285158798098564, 0.01771629974246025, -0.06046866253018379, 0.07364235818386078, 0.030026888474822044, -0.036611031740903854, -0.07146476209163666, -0.04183752462267876, -0.04579772427678108, 0.008587779477238655, -0.16310185194015503, -0.11530570685863495, -0.004291696939617395, -0.022359905764460564, 0.07707220315933228, 0.03829155117273331, -0.09312435239553452, -0.03533054143190384, -0.05517655238509178, 0.0333234928548336, -0.12787498533725739, 0.0459769070148468, 0.08696694672107697, -0.015718704089522362, -0.010699767619371414, -0.06356548517942429, 0.019441908225417137, 0.023609060794115067, -0.10050466656684875, -0.07246299833059311 ]
null
null
transformers
# Cross-Encoder for MS Marco This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco) ## Usage with Transformers ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('model_name') tokenizer = AutoTokenizer.from_pretrained('model_name') features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits print(scores) ``` ## Usage with SentenceTransformers The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('model_name', max_length=512) scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')]) ``` ## Performance In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset. | Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec | | ------------- |:-------------| -----| --- | | **Version 2 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000 | cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100 | cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500 | cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800 | cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960 | **Version 1 models** | | | | cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000 | cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900 | cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680 | cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340 | **Other models** | | | | nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900 | nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340 | nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100 | Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340 | amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330 | sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720 Note: Runtime was computed on a V100 GPU.
{"license": "apache-2.0"}
text-classification
cross-encoder/ms-marco-electra-base
[ "transformers", "pytorch", "electra", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #electra #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
Cross-Encoder for MS Marco ========================== This model was trained on the MS Marco Passage Ranking task. The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco Usage with Transformers ----------------------- Usage with SentenceTransformers ------------------------------- The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this: Performance ----------- In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset. Note: Runtime was computed on a V100 GPU.
[]
[ "TAGS\n#transformers #pytorch #electra #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 45 ]
[ "passage: TAGS\n#transformers #pytorch #electra #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.04024451971054077, 0.1394282877445221, -0.006416637450456619, 0.028075577691197395, 0.15309758484363556, 0.01857059821486473, 0.09045905619859695, 0.12177535146474838, 0.011697805486619473, -0.049882590770721436, 0.13924641907215118, 0.23919491469860077, -0.006401306949555874, 0.08411108702421188, -0.09589783847332001, -0.2561156153678894, 0.09550803154706955, 0.040639109909534454, -0.024907702580094337, 0.09694972634315491, 0.1262826919555664, -0.0372859351336956, 0.0483354777097702, 0.008995389565825462, -0.04211917147040367, 0.025731168687343597, 0.028592536225914955, -0.14406731724739075, 0.09406385570764542, 0.0621630996465683, 0.09353066235780716, 0.03694679215550423, -0.024086112156510353, -0.1928902119398117, 0.026310071349143982, 0.002302405424416065, -0.09785549342632294, 0.052400074899196625, 0.03380037099123001, -0.07051724940538406, 0.08566132187843323, 0.058593396097421646, 0.008490554988384247, 0.053670577704906464, -0.08374246954917908, -0.14458279311656952, -0.07852615416049957, 0.07859311252832413, 0.08631940931081772, 0.08421549201011658, 0.03726954758167267, 0.15745727717876434, -0.11553414165973663, 0.05882599949836731, 0.09277986735105515, -0.3072034418582916, 0.0111320149153471, 0.03397609293460846, 0.041928134858608246, 0.015585123561322689, 0.001007251557894051, 0.03173469379544258, 0.05382874235510826, 0.019103143364191055, 0.01713571324944496, -0.038105905055999756, -0.11122983694076538, 0.036092307418584824, -0.05960645526647568, -0.0862732008099556, 0.2517440617084503, -0.010789081454277039, 0.03434799611568451, -0.013364816084504128, -0.06594661623239517, -0.007940991781651974, -0.00317173614166677, 0.0677703320980072, 0.0001087129203369841, 0.09997144341468811, 0.10118997097015381, -0.008642055094242096, -0.13583175837993622, 0.013400900177657604, -0.21188800036907196, 0.13454893231391907, 0.025790706276893616, 0.07544925063848495, -0.12801332771778107, 0.0582682304084301, 0.0406012162566185, -0.10812640935182571, -0.007128621451556683, -0.0808103159070015, 0.10145684331655502, 0.00537505280226469, -0.08779856562614441, 0.025743994861841202, 0.14844615757465363, 0.2594009041786194, 0.0794718861579895, 0.015034779906272888, -0.016524020582437515, 0.11119507998228073, 0.0022961304057389498, 0.07593061774969101, 0.028438640758395195, 0.01706576719880104, 0.10973414778709412, -0.1238437294960022, 0.07520311325788498, -0.04095808416604996, -0.16450434923171997, -0.02486010454595089, -0.0017762617208063602, 0.12833058834075928, 0.04751913249492645, 0.04408520087599754, -0.0488656722009182, 0.004286855924874544, 0.15226060152053833, -0.07314354181289673, 0.0011187534546479583, -0.0037508595269173384, 0.000644550658762455, 0.07853707671165466, 0.04579019919037819, 0.03356843441724777, -0.0823741927742958, 0.09596745669841766, -0.04118073731660843, -0.020657889544963837, -0.01995593123137951, -0.01896272599697113, 0.09006624668836594, -0.13740576803684235, 0.05944223329424858, -0.17719943821430206, -0.15579520165920258, 0.02343272976577282, 0.0567639134824276, 0.013866236433386803, -0.11190372705459595, 0.028464799746870995, 0.002686518244445324, 0.04053020849823952, -0.0864289179444313, -0.06876781582832336, -0.09028354287147522, 0.07457564771175385, -0.07199487090110779, 0.022805485874414444, -0.15917567908763885, 0.07116898894309998, -0.1058075800538063, -0.027042923495173454, -0.050866615027189255, -0.017219403758645058, -0.06962284445762634, 0.17894667387008667, -0.051072124391794205, -0.04391330108046532, 0.027105361223220825, 0.021444199606776237, -0.0567220002412796, 0.1178547590970993, -0.06839759647846222, -0.06930509209632874, 0.17668576538562775, -0.14144107699394226, -0.17492343485355377, 0.06966202706098557, 0.008232333697378635, -0.020347634330391884, 0.08187971264123917, 0.10969138890504837, 0.0805254802107811, -0.030976276844739914, 0.05714128538966179, 0.09338152408599854, -0.0957266092300415, -0.19246309995651245, 0.04815148934721947, -0.044398706406354904, -0.09391974657773972, 0.048111721873283386, -0.006508640479296446, 0.09925005584955215, 0.004879864864051342, -0.06286478787660599, -0.05159604176878929, -0.040863148868083954, 0.04632221907377243, 0.012806168757379055, 0.04861249774694443, -0.07744750380516052, -0.021156389266252518, -0.012659648433327675, 0.043738968670368195, 0.038132574409246445, 0.043341927230358124, -0.08769343048334122, 0.07317201793193817, 0.03563517704606056, 0.029601987451314926, -0.15311481058597565, -0.010886318050324917, -0.02645612694323063, 0.021045319736003876, 0.014990728348493576, 0.06638722121715546, 0.017479626461863518, -0.06817279011011124, 0.007545228116214275, -0.018256766721606255, 0.115231953561306, 0.03910911828279495, -0.005711002741008997, -0.10419747978448868, 0.045774057507514954, -0.040243010967969894, 0.04841990768909454, -0.04744097590446472, 0.03727738931775093, 0.029863713309168816, 0.09556417912244797, -0.04455886408686638, 0.10975157469511032, -0.024643883109092712, 0.013350252993404865, -0.08191020786762238, 0.005608132109045982, 0.0900387167930603, 0.04622197523713112, -0.08618860691785812, 0.16641731560230255, -0.09750746935606003, 0.25317510962486267, 0.190445214509964, -0.17965662479400635, 0.05708780884742737, -0.005851298104971647, -0.007946527563035488, 0.0018073483370244503, -0.00042592413956299424, 0.025624345988035202, 0.050502557307481766, 0.02388840541243553, 0.17720672488212585, -0.041322074830532074, -0.025940822437405586, -0.015320357866585255, -0.06915087252855301, 0.0011532638454809785, 0.05418197810649872, 0.18481361865997314, -0.18892161548137665, 0.17879219353199005, 0.25811824202537537, -0.029792683199048042, 0.07442477345466614, -0.06929922848939896, 0.02268984541296959, 0.07036911696195602, -0.052214909344911575, -0.011649871245026588, -0.0621630996465683, -0.09995310008525848, 0.014498937875032425, 0.09725575894117355, 0.02947993576526642, 0.04098421335220337, -0.12630021572113037, -0.05124010890722275, 0.016034454107284546, -0.013077307492494583, -0.05003990978002548, 0.06871742010116577, 0.015374417416751385, 0.08863312751054764, -0.03990410268306732, -0.10956624150276184, 0.13317373394966125, 0.016460346058011055, -0.07863547652959824, 0.15531882643699646, -0.15623947978019714, -0.265079140663147, -0.1546567678451538, -0.12503203749656677, -0.009020118042826653, 0.026540031656622887, 0.1575431078672409, -0.04157167673110962, -0.06726272404193878, 0.009546620771288872, -0.07972870767116547, 0.0012169755063951015, 0.014581684954464436, -0.02148028463125229, 0.05833166465163231, -0.01464900467544794, -0.10707701742649078, -0.05607196316123009, 0.021814193576574326, -0.014264367520809174, 0.11674091964960098, -0.08108065277338028, 0.07857286930084229, 0.13430099189281464, 0.012132087722420692, 0.02663179486989975, -0.026369832456111908, 0.12187351286411285, -0.04903124272823334, -0.0022207251749932766, 0.250207781791687, -0.04527536779642105, 0.08283510059118271, 0.1511315256357193, 0.02071426808834076, -0.04144661873579025, 0.02659960463643074, -0.05726110190153122, -0.08931757509708405, -0.2876962423324585, -0.13703086972236633, -0.09963475167751312, 0.0399465374648571, 0.06808318942785263, 0.09846090525388718, 0.15532617270946503, 0.08374204486608505, -0.02614392712712288, -0.016173049807548523, 0.005293949972838163, 0.07990442216396332, 0.2997828423976898, 0.01686898246407509, 0.12262818962335587, -0.1313197761774063, -0.0659727081656456, 0.11678404361009598, 0.04156997427344322, 0.14259211719036102, 0.13028112053871155, 0.07489220798015594, 0.058195099234580994, 0.11204276978969574, 0.11276265233755112, 0.13001494109630585, 0.05837985873222351, 0.0038573709316551685, -0.045596618205308914, 0.004708520136773586, -0.060464538633823395, 0.04057109355926514, -0.011596299707889557, -0.1525481790304184, -0.056350432336330414, -0.1299498826265335, 0.06227819249033928, 0.18490688502788544, 0.011361589655280113, -0.17781929671764374, 0.01701386459171772, 0.07427063584327698, -0.04004724696278572, -0.06352955847978592, 0.07894083857536316, -0.12588126957416534, -0.1232147216796875, 0.153847336769104, -0.027396807447075844, 0.1386658102273941, -0.036836251616477966, 0.03411301225423813, -0.039069753140211105, -0.1176995038986206, 0.05875086411833763, 0.14249558746814728, -0.253315269947052, 0.194421648979187, -0.019304737448692322, -0.03533247485756874, -0.08516760915517807, 0.005249192006886005, 0.06463167071342468, 0.245378315448761, 0.05297144502401352, 0.004224834498018026, -0.16488181054592133, -0.05632767453789711, -0.04478565603494644, 0.038439832627773285, 0.002593998098745942, -0.021312732249498367, -0.04078371822834015, -0.0646892637014389, -0.01242074929177761, -0.016343146562576294, 0.015083102509379387, -0.004813728388398886, -0.16914843022823334, 0.04277850314974785, 0.11934252083301544, 0.09082981199026108, -0.03510992228984833, -0.0336446650326252, -0.09059744328260422, 0.1370229870080948, -0.139541894197464, -0.09630466997623444, -0.10171578079462051, -0.11446097493171692, 0.0366479754447937, -0.07600505650043488, 0.10345278680324554, -0.07974762469530106, 0.01806996949017048, -0.05596334487199783, -0.17790526151657104, 0.09679289162158966, -0.13203047215938568, -0.04479781910777092, -0.03123793564736843, 0.10218585282564163, -0.08275012671947479, 0.007968029007315636, 0.042733967304229736, 0.020291447639465332, -0.11023463308811188, -0.1100921556353569, -0.020460499450564384, 0.03185494244098663, 0.06743910908699036, 0.01607770286500454, -0.06943298131227493, -0.07042031735181808, 0.012650764547288418, -0.06748540699481964, 0.2516140937805176, 0.17416873574256897, -0.07574023306369781, 0.1617659032344818, 0.17042982578277588, -0.0878354087471962, -0.27645131945610046, -0.19003352522850037, -0.1284037083387375, -0.08563404530286789, 0.0072863115929067135, -0.18424472212791443, 0.11343926191329956, 0.06691261380910873, -0.08302415907382965, 0.10592896491289139, -0.16273947060108185, -0.06736753135919571, 0.2054462432861328, -0.06705047935247421, 0.2871759235858917, -0.12948204576969147, -0.07449353486299515, -0.10361569374799728, -0.16432106494903564, 0.13637487590312958, -0.038800567388534546, 0.0602928102016449, -0.014459117315709591, 0.0265655517578125, -0.007882952690124512, -0.036956220865249634, 0.12710867822170258, -0.014967162162065506, 0.010111716575920582, -0.14166724681854248, 0.007903330028057098, 0.019002703949809074, -0.024332717061042786, 0.059651266783475876, -0.17806826531887054, 0.020020918920636177, -0.132245272397995, -0.03736512362957001, -0.048099830746650696, 0.09129953384399414, 0.03351367637515068, -0.021978409960865974, -0.034593887627124786, -0.05407708138227463, 0.030377250164747238, -0.0008416110649704933, 0.2564811408519745, 0.015424934215843678, 0.10035118460655212, 0.11503051221370697, 0.09328581392765045, -0.1466190218925476, -0.008832348510622978, -0.12192293256521225, -0.07277058809995651, 0.05477006733417511, -0.1393607258796692, 0.05706823989748955, 0.11153620481491089, -0.07808275520801544, 0.06865844875574112, 0.07176221162080765, 0.025690356269478798, -0.02029947191476822, 0.14105558395385742, -0.1430954933166504, 0.04574832320213318, -0.004595305770635605, 0.14741545915603638, 0.0640120580792427, 0.06773573160171509, 0.12599653005599976, 0.024393822997808456, -0.05459459498524666, 0.012501217424869537, 0.024452969431877136, -0.028911540284752846, 0.038295239210128784, 0.05621488764882088, 0.017675278708338737, -0.13390614092350006, 0.0964098572731018, 0.049044277518987656, -0.1492808759212494, -0.003441132139414549, 0.10814365744590759, -0.14798347651958466, -0.1383768618106842, 0.02242901362478733, 0.06571763753890991, -0.2122051864862442, -0.1294255256652832, -0.06758717447519302, -0.1314913034439087, 0.08356214314699173, 0.15618059039115906, 0.11304860562086105, 0.051549751311540604, -0.008750321343541145, -0.06876935809850693, 0.027284245938062668, -0.01640613190829754, -0.04039619863033295, 0.015225742943584919, -0.11308631300926208, -0.06689400225877762, 0.028903523460030556, 0.10405153036117554, -0.045516327023506165, -0.03453921899199486, -0.06854874640703201, 0.04426014795899391, -0.1456802487373352, 0.0278960969299078, -0.07803834974765778, -0.005698648747056723, 0.022678500041365623, -0.06405041366815567, -0.027794290333986282, -0.011602615006268024, -0.12157195061445236, -0.004956862423568964, -0.03411935642361641, 0.0728229209780693, -0.10907217860221863, -0.05473417788743973, 0.09611295163631439, -0.015284935012459755, 0.11465848982334137, 0.07213246822357178, -0.08271103352308273, 0.07898499816656113, -0.14159905910491943, -0.09054803103208542, 0.1161518394947052, 0.037348922342061996, 0.010520939715206623, -0.024960726499557495, 0.03030102141201496, 0.11350668966770172, -0.045109786093235016, 0.029737936332821846, -0.01893732324242592, -0.1524437814950943, -0.024923453107476234, -0.014129013754427433, -0.13219018280506134, -0.025927329435944557, -0.10580310225486755, 0.1152222603559494, 0.019027678295969963, 0.2183641642332077, -0.021270370110869408, 0.046936046332120895, -0.02855025976896286, 0.023033645004034042, -0.04410076141357422, -0.15587663650512695, -0.14079056680202484, -0.06641631573438644, -0.04875990003347397, -0.009176901541650295, 0.25361019372940063, 0.019943062216043472, -0.03752049803733826, 0.07169941812753677, 0.11796827614307404, 0.0026142774149775505, 0.01600082591176033, 0.24229003489017487, 0.06534610688686371, -0.009016556665301323, -0.08455488830804825, -0.007550650741904974, 0.012588622979819775, -0.10933476686477661, 0.0960666686296463, 0.06629017740488052, 0.05052799731492996, 0.07054765522480011, 0.014023473486304283, 0.031026961281895638, -0.10648991167545319, -0.17085735499858856, 0.03367779403924942, 0.08086472004652023, 0.05332938954234123, 0.1047055795788765, 0.1367921382188797, -0.0185170229524374, 0.005783441010862589, -0.04609041288495064, 0.0018499634461477399, -0.18199138343334198, -0.14808903634548187, -0.0893586128950119, -0.12070519477128983, 0.010083816014230251, -0.04544419050216675, 0.025865929201245308, 0.11981025338172913, 0.03441726788878441, -0.07407727837562561, -0.04958098381757736, -0.013592684641480446, -0.05714087933301926, 0.018829094246029854, -0.043657951056957245, -0.004283376969397068, 0.0077462331391870975, -0.045971550047397614, -0.09671482443809509, -0.026886535808444023, -0.04157080501317978, 0.04619024321436882, -0.01343106385320425, 0.06469076126813889, -0.13383154571056366, -0.07984981685876846, -0.03289703279733658, 0.04114929959177971, -0.0024660674389451742, 0.1623561531305313, 0.012266039848327637, 0.031103331595659256, 0.10458961874246597, 0.17347536981105804, -0.05008632689714432, -0.15108516812324524, -0.05923185124993324, 0.14117980003356934, 0.0859256163239479, 0.06251697987318039, 0.015807492658495903, 0.0033862146083265543, -0.06002199277281761, 0.2585621774196625, 0.23214176297187805, -0.01547662541270256, 0.03307679295539856, -0.04826577007770538, 0.021224405616521835, 0.11025270074605942, 0.10972068458795547, 0.12476218491792679, 0.17299842834472656, -0.07445333153009415, -0.023825248703360558, -0.05336571857333183, 0.020602496340870857, -0.17988669872283936, 0.04538416862487793, -0.02621188387274742, -0.09329324215650558, -0.0012173529248684645, 0.12378253042697906, -0.1319345235824585, 0.11799726635217667, -0.005860974546521902, -0.09502888470888138, -0.044742804020643234, -0.01523672230541706, 0.18205033242702484, 0.04049639031291008, 0.024224212393164635, -0.02306114137172699, -0.09810671210289001, 0.1000199019908905, -0.003367006778717041, -0.22396411001682281, -0.05375676229596138, 0.08661927282810211, -0.012011992745101452, 0.07649670541286469, 0.0002248782111564651, 0.07127409428358078, 0.08224761486053467, 0.07496672123670578, -0.07651036977767944, 0.09888625890016556, 0.01087881438434124, -0.017818013206124306, 0.05433560535311699, -0.0963621437549591, -0.023679740726947784, -0.08142072707414627, 0.04777707904577255, -0.08670085668563843, 0.04945719614624977, -0.006629862356930971, -0.04589192196726799, -0.02703871764242649, 0.06799701601266861, -0.0724881961941719, 0.05630885809659958, 0.003155894111841917, -0.03590865060687065, -0.06303263455629349, -0.06387297809123993, -0.024654222652316093, 0.038610395044088364, -0.19419945776462555, -0.08653511106967926, -0.01225605420768261, -0.03622356057167053, 0.06964404881000519, 0.05298729985952377, -0.09691957384347916, -0.010787577368319035, -0.13345758616924286, 0.007239778526127338, -0.16423611342906952, 0.043623339384794235, 0.07419291883707047, -0.006328477989882231, -0.014007709920406342, -0.015221481211483479, 0.020509619265794754, 0.02966417372226715, -0.1079070121049881, -0.08729416131973267 ]
null
null
transformers
# Cross-Encoder for MS MARCO - EN-DE This is a cross-lingual Cross-Encoder model for EN-DE that can be used for passage re-ranking. It was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task. The model can be used for Information Retrieval: See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html). The training code is available in this repository, see `train_script.py`. ## Usage with SentenceTransformers When you have [SentenceTransformers](https://www.sbert.net/) installed, you can use the model like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('model_name', max_length=512) query = 'How many people live in Berlin?' docs = ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'] pairs = [(query, doc) for doc in docs] scores = model.predict(pairs) ``` ## Usage with Transformers With the transformers library, you can use the model like this: ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('model_name') tokenizer = AutoTokenizer.from_pretrained('model_name') features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits print(scores) ``` ## Performance The performance was evaluated on three datasets: - **TREC-DL19 EN-EN**: The original [TREC 2019 Deep Learning Track](https://microsoft.github.io/msmarco/TREC-Deep-Learning-2019.html): Given an English query and 1000 documents (retrieved by BM25 lexical search), rank documents with according to their relevance. We compute NDCG@10. BM25 achieves a score of 45.46, a perfect re-ranker can achieve a score of 95.47. - **TREC-DL19 DE-EN**: The English queries of TREC-DL19 have been translated by a German native speaker to German. We rank the German queries versus the English passages from the original TREC-DL19 setup. We compute NDCG@10. - **GermanDPR DE-DE**: The [GermanDPR](https://www.deepset.ai/germanquad) dataset provides German queries and German passages from Wikipedia. We indexed the 2.8 Million paragraphs from German Wikipedia and retrieved for each query the top 100 most relevant passages using BM25 lexical search with Elasticsearch. We compute MRR@10. BM25 achieves a score of 35.85, a perfect re-ranker can achieve a score of 76.27. We also check the performance of bi-encoders using the same evaluation: The retrieved documents from BM25 lexical search are re-ranked using query & passage embeddings with cosine-similarity. Bi-Encoders can also be used for end-to-end semantic search. | Model-Name | TREC-DL19 EN-EN | TREC-DL19 DE-EN | GermanDPR DE-DE | Docs / Sec | | ------------- |:-------------:| :-----: | :---: | :----: | | BM25 | 45.46 | - | 35.85 | -| | **Cross-Encoder Re-Rankers** | | | | | [cross-encoder/msmarco-MiniLM-L6-en-de-v1](https://huggingface.co/cross-encoder/msmarco-MiniLM-L6-en-de-v1) | 72.43 | 65.53 | 46.77 | 1600 | | [cross-encoder/msmarco-MiniLM-L12-en-de-v1](https://huggingface.co/cross-encoder/msmarco-MiniLM-L12-en-de-v1) | 72.94 | 66.07 | 49.91 | 900 | | [svalabs/cross-electra-ms-marco-german-uncased](https://huggingface.co/svalabs/cross-electra-ms-marco-german-uncased) (DE only) | - | - | 53.67 | 260 | | [deepset/gbert-base-germandpr-reranking](https://huggingface.co/deepset/gbert-base-germandpr-reranking) (DE only) | - | - | 53.59 | 260 | | **Bi-Encoders (re-ranking)** | | | | | [sentence-transformers/msmarco-distilbert-multilingual-en-de-v2-tmp-lng-aligned](https://huggingface.co/sentence-transformers/msmarco-distilbert-multilingual-en-de-v2-tmp-lng-aligned) | 63.38 | 58.28 | 37.88 | 940 | | [sentence-transformers/msmarco-distilbert-multilingual-en-de-v2-tmp-trained-scratch](https://huggingface.co/sentence-transformers/msmarco-distilbert-multilingual-en-de-v2-tmp-trained-scratch) | 65.51 | 58.69 | 38.32 | 940 | | [svalabs/bi-electra-ms-marco-german-uncased](https://huggingface.co/svalabs/bi-electra-ms-marco-german-uncased) (DE only) | - | - | 34.31 | 450 | | [deepset/gbert-base-germandpr-question_encoder](https://huggingface.co/deepset/gbert-base-germandpr-question_encoder) (DE only) | - | - | 42.55 | 450 | Note: Docs / Sec gives the number of (query, document) pairs we can re-rank within a second on a V100 GPU.
{"license": "apache-2.0"}
text-classification
cross-encoder/msmarco-MiniLM-L12-en-de-v1
[ "transformers", "pytorch", "bert", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
Cross-Encoder for MS MARCO - EN-DE ================================== This is a cross-lingual Cross-Encoder model for EN-DE that can be used for passage re-ranking. It was trained on the MS Marco Passage Ranking task. The model can be used for Information Retrieval: See URL Retrieve & Re-rank. The training code is available in this repository, see 'train\_script.py'. Usage with SentenceTransformers ------------------------------- When you have SentenceTransformers installed, you can use the model like this: Usage with Transformers ----------------------- With the transformers library, you can use the model like this: Performance ----------- The performance was evaluated on three datasets: * TREC-DL19 EN-EN: The original TREC 2019 Deep Learning Track: Given an English query and 1000 documents (retrieved by BM25 lexical search), rank documents with according to their relevance. We compute NDCG@10. BM25 achieves a score of 45.46, a perfect re-ranker can achieve a score of 95.47. * TREC-DL19 DE-EN: The English queries of TREC-DL19 have been translated by a German native speaker to German. We rank the German queries versus the English passages from the original TREC-DL19 setup. We compute NDCG@10. * GermanDPR DE-DE: The GermanDPR dataset provides German queries and German passages from Wikipedia. We indexed the 2.8 Million paragraphs from German Wikipedia and retrieved for each query the top 100 most relevant passages using BM25 lexical search with Elasticsearch. We compute MRR@10. BM25 achieves a score of 35.85, a perfect re-ranker can achieve a score of 76.27. We also check the performance of bi-encoders using the same evaluation: The retrieved documents from BM25 lexical search are re-ranked using query & passage embeddings with cosine-similarity. Bi-Encoders can also be used for end-to-end semantic search. Note: Docs / Sec gives the number of (query, document) pairs we can re-rank within a second on a V100 GPU.
[]
[ "TAGS\n#transformers #pytorch #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ 48 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ -0.011929498054087162, 0.1109679788351059, -0.004657326731830835, 0.036296773701906204, 0.08949180692434311, 0.01468193531036377, 0.1052534207701683, 0.14302724599838257, 0.00024416513042524457, -0.031693242490291595, 0.1409340649843216, 0.17007292807102203, -0.03366948291659355, 0.0719553679227829, -0.07565809041261673, -0.24093829095363617, 0.10701024532318115, 0.03007899411022663, -0.0497613325715065, 0.0865328311920166, 0.11767031252384186, -0.06318635493516922, 0.0450325682759285, -0.002566360170021653, -0.062239885330200195, 0.03159003704786301, 0.01839093491435051, -0.12103993445634842, 0.1024293452501297, 0.03353764861822128, 0.10411366820335388, 0.050487738102674484, -0.03610473498702049, -0.15597668290138245, 0.02397605963051319, 0.02407376654446125, -0.09014851599931717, 0.06582070142030716, 0.042276762425899506, -0.03844092786312103, 0.04500403627753258, 0.005638193804770708, -0.001494656433351338, 0.03218536451458931, -0.09311587363481522, -0.19191624224185944, -0.07325631380081177, 0.07541981339454651, 0.023964963853359222, 0.08019165694713593, 0.023596404120326042, 0.15867139399051666, -0.14464761316776276, 0.05357449874281883, 0.13281987607479095, -0.34267017245292664, 0.0014128641923889518, 0.12683846056461334, 0.06185078248381615, 0.03614772856235504, -0.03367999568581581, 0.05115452781319618, 0.05234549194574356, 0.01213220227509737, 0.053060855716466904, -0.038291096687316895, -0.12301186472177505, 0.06110190600156784, -0.07589741796255112, -0.08090674132108688, 0.2470371127128601, -0.020999861881136894, 0.06525891274213791, -0.02700621820986271, -0.07864753156900406, -0.015083619393408298, 0.0020947037264704704, 0.05717647820711136, 0.023739924654364586, 0.10291964560747147, 0.07124892622232437, -0.034733641892671585, -0.15194028615951538, 0.028371602296829224, -0.2115340530872345, 0.07593068480491638, 0.003470568684861064, 0.0676390752196312, -0.12714898586273193, 0.06812293827533722, 0.01710754819214344, -0.11910293996334076, 0.02460039034485817, -0.07910972088575363, 0.10053464025259018, 0.02612384781241417, -0.10773281753063202, 0.0519438311457634, 0.12186801433563232, 0.198510080575943, 0.02410656027495861, -0.0036114167887717485, -0.03310275822877884, 0.1262211948633194, 0.0009386635501869023, 0.09358032047748566, -0.0214508268982172, -0.0191908348351717, 0.07524799555540085, -0.08410874754190445, 0.06491826474666595, -0.06015123426914215, -0.1755080223083496, -0.04224881902337074, 0.02791069634258747, 0.08877983689308167, 0.0867546871304512, 0.06018778309226036, -0.023768959566950798, 0.027124647051095963, 0.14338789880275726, -0.06075986847281456, 0.035603079944849014, -0.0025226115249097347, 0.03203103691339493, 0.02896995283663273, 0.04888737201690674, 0.024378491565585136, -0.028933117166161537, 0.0786980390548706, -0.0584336593747139, -0.012409459799528122, -0.042271099984645844, -0.05858355015516281, 0.08736852556467056, -0.1067584678530693, 0.026230912655591965, -0.16472218930721283, -0.0831773653626442, 0.01852068118751049, 0.06086645647883415, 0.0007838314631953835, -0.09153826534748077, 0.07393558323383331, -0.040153417736291885, 0.07722245901823044, -0.08859124779701233, 0.0032021517399698496, -0.09344978630542755, 0.057036347687244415, -0.1038811206817627, 0.04983983933925629, -0.18570169806480408, 0.06897924840450287, -0.10177050530910492, -0.0035829299595206976, -0.048835549503564835, -0.03398766368627548, -0.0877050906419754, 0.1558571755886078, -0.02481936104595661, -0.03799108788371086, -0.02024412341415882, 0.031119640916585922, -0.046811964362859726, 0.08756688982248306, -0.1204364225268364, -0.03759213536977768, 0.1446077972650528, -0.09290962666273117, -0.1673811823129654, 0.06678478419780731, 0.004953649360686541, 0.0004414678260218352, 0.023817099630832672, 0.2068229466676712, 0.06968206912279129, -0.031353238970041275, 0.03667166829109192, 0.15588927268981934, -0.036718402057886124, -0.1597534567117691, 0.05581526458263397, -0.036597561091184616, -0.07861929386854172, 0.04289300739765167, -0.020597059279680252, 0.07218319177627563, 0.013152756728231907, -0.0813322514295578, -0.07919875532388687, -0.03760673850774765, 0.056981250643730164, 0.010442853905260563, 0.09089142829179764, -0.0736541673541069, -0.03648189455270767, 0.012337546795606613, 0.029745422303676605, 0.05869894474744797, 0.06688299775123596, -0.03138571232557297, 0.10668440163135529, 0.013989963568747044, 0.018267758190631866, -0.15533185005187988, -0.016854720190167427, -0.027512626722455025, 0.033168140798807144, 0.0027375284116715193, 0.16843344271183014, 0.028516249731183052, -0.08179625123739243, -0.005133424885571003, -0.0027407007291913033, 0.1116393432021141, 0.07179897278547287, -0.04622254520654678, -0.1110997349023819, 0.02547508105635643, -0.05257298797369003, 0.013891373760998249, -0.02169971913099289, 0.0376596562564373, 0.06746970117092133, 0.10060187429189682, -0.054126087576150894, 0.11025400459766388, -0.02443268522620201, 0.01063956692814827, -0.08723245561122894, 0.0012760007521137595, 0.09992439299821854, 0.03174593299627304, -0.06503839790821075, 0.2191265970468521, -0.11319835484027863, 0.2770872712135315, 0.2274661362171173, -0.1816493421792984, 0.06211845576763153, 0.010664228349924088, -0.0253766942769289, 0.016376934945583344, 0.00973495189100504, -0.005615867208689451, -0.017049739137291908, -0.014084290713071823, 0.1513899713754654, -0.04333732649683952, -0.03485063835978508, -0.028493531048297882, -0.06856098771095276, -0.03320147842168808, 0.043046679347753525, 0.09611020982265472, -0.14282885193824768, 0.21591053903102875, 0.3771267533302307, -0.03934421390295029, 0.1023554727435112, -0.05790102481842041, 0.029091954231262207, 0.03956243023276329, -0.07205614447593689, -0.05272269248962402, 0.0077779870480299, -0.14636331796646118, -0.011522377841174603, 0.11102425307035446, 0.05511642247438431, 0.054583270102739334, -0.14314037561416626, -0.047039732336997986, 0.022951988503336906, 0.011035983450710773, -0.05264776572585106, 0.08396845310926437, 0.028842180967330933, 0.10083223134279251, -0.018218370154500008, -0.09925662726163864, 0.10900299996137619, 0.0034230707678943872, -0.04919014126062393, 0.1237892210483551, -0.17483505606651306, -0.2666981816291809, -0.12263266742229462, -0.11647240817546844, -0.03185277059674263, 0.02454465627670288, 0.1553153693675995, -0.05448857694864273, -0.04608921706676483, -0.0322968028485775, -0.07930576056241989, -0.03999684005975723, 0.03336773440241814, -0.055547602474689484, 0.051703520119190216, 0.004006838425993919, -0.12592794001102448, -0.0660366490483284, 0.019448257982730865, -0.0385357066988945, 0.10267893224954605, -0.03189206123352051, 0.0755273625254631, 0.1416429579257965, 0.008576683700084686, 0.024744180962443352, -0.028986504301428795, 0.1633431613445282, -0.04187033325433731, 0.017040017992258072, 0.2164611518383026, -0.008377695456147194, 0.07641872763633728, 0.18111683428287506, 0.03418927267193794, -0.021802285686135292, 0.013025326654314995, -0.036622799932956696, -0.09112220257520676, -0.21751870214939117, -0.13478365540504456, -0.11826911568641663, 0.022339027374982834, 0.04935091361403465, 0.10200902074575424, 0.12801212072372437, 0.05823846161365509, -0.0066392021253705025, -0.0018668529810383916, -0.03226018324494362, 0.05748061463236809, 0.2714395523071289, -0.03477463126182556, 0.14819587767124176, -0.10991998016834259, -0.06027120351791382, 0.12198864668607712, 0.0809679850935936, 0.10323648899793625, 0.10058606415987015, 0.02176450751721859, 0.07683505862951279, 0.17835702002048492, 0.10593564063310623, 0.09460754692554474, 0.03239644691348076, -0.0022547899279743433, -0.053684331476688385, -0.020651666447520256, -0.021006306633353233, 0.06357185542583466, 0.03231940418481827, -0.15682291984558105, -0.02531340718269348, -0.19327817857265472, 0.09269165247678757, 0.15072937309741974, 0.054205313324928284, -0.1248549297451973, 0.032165903598070145, 0.09894848614931107, -0.013840306550264359, -0.04561185464262962, 0.0807991772890091, -0.029446594417095184, -0.08341233432292938, 0.08656627684831619, 0.009248846210539341, 0.12374022603034973, 0.0148047786206007, 0.07288803160190582, -0.04254555702209473, -0.17002898454666138, 0.059988267719745636, 0.129883274435997, -0.27484795451164246, 0.20877805352210999, -0.026468385010957718, -0.09377090632915497, -0.0814322680234909, 0.0007435722509399056, 0.07990670949220657, 0.2424503117799759, 0.04600871354341507, 0.034760482609272, -0.14131808280944824, -0.0642140805721283, -0.04361937195062637, 0.0123197752982378, 0.023223701864480972, -0.01777510903775692, -0.03755996376276016, -0.06066042184829712, -0.0020570766646414995, 0.021542828530073166, 0.14491324126720428, -0.019954849034547806, -0.17723368108272552, 0.0679810494184494, 0.1139739453792572, 0.020291222259402275, -0.05237312614917755, -0.04039272293448448, -0.16760264337062836, 0.15746107697486877, -0.04734361916780472, -0.05293496698141098, -0.09933813661336899, -0.11485697329044342, 0.05627376213669777, -0.04504356533288956, 0.06802637875080109, -0.08537643402814865, 0.018626293167471886, -0.07503805309534073, -0.1825282722711563, 0.10788697749376297, -0.1172252967953682, -0.04584076628088951, -0.04347958415746689, 0.0893806666135788, -0.14524905383586884, 0.04412192851305008, 0.036863263696432114, 0.05305350199341774, -0.1468811184167862, -0.11997350305318832, -0.012302463874220848, 0.013916125521063805, 0.06559175997972488, -0.026150045916438103, -0.08259673416614532, -0.05353347212076187, 0.05847072973847389, -0.032503772526979446, 0.256398469209671, 0.1776508241891861, -0.1236819177865982, 0.1738201528787613, 0.13180740177631378, -0.07809863984584808, -0.3454630970954895, -0.14732776582241058, -0.15849560499191284, -0.07533957064151764, 0.0235506072640419, -0.11653820425271988, 0.10535410046577454, 0.03611930459737778, -0.10646899044513702, 0.05344188213348389, -0.17473436892032623, -0.08326824009418488, 0.19634291529655457, -0.08626771718263626, 0.30544474720954895, -0.15087944269180298, -0.06414706259965897, -0.07380376756191254, -0.1340590864419937, 0.14970876276493073, -0.12145139276981354, 0.07197150588035583, -0.012666989117860794, 0.022509323433041573, -0.0005715735605917871, -0.05197785422205925, 0.1541745513677597, -0.03773857280611992, 0.025772521272301674, -0.1358940750360489, -0.03740015998482704, 0.0981537252664566, -0.06408584862947464, 0.032355599105358124, -0.1558215320110321, 0.017573487013578415, -0.13796880841255188, 0.006053187418729067, -0.07947039604187012, 0.09438689798116684, -0.003972133621573448, -0.041665222495794296, -0.06218144670128822, -0.012985005043447018, 0.038268182426691055, -0.013135287910699844, 0.26377928256988525, 0.0018693876918405294, 0.11646835505962372, 0.1741446703672409, 0.02918052300810814, -0.16934606432914734, -0.019878719002008438, -0.0496620275080204, -0.07310948520898819, 0.06996731460094452, -0.18846379220485687, 0.05133286863565445, 0.08059290796518326, -0.07819657772779465, 0.057422809302806854, 0.08675773441791534, 0.022930843755602837, -0.04511556401848793, 0.16467398405075073, -0.16118508577346802, -0.0038538638036698103, -0.005934640299528837, 0.10133349150419235, 0.05118110030889511, 0.021290142089128494, 0.13166667520999908, 0.017662744969129562, -0.031207405030727386, 0.02763739600777626, 0.017550131306052208, -0.06095219403505325, 0.019465308636426926, 0.06818114966154099, 0.027803484350442886, -0.1139194443821907, 0.0774909257888794, 0.06724732369184494, -0.12166350334882736, -0.01452089287340641, 0.08962465077638626, -0.1174732893705368, -0.16081809997558594, -0.000930252717807889, 0.031174184754490852, -0.11175857484340668, -0.0874028131365776, -0.01669522002339363, -0.1344173103570938, 0.05237565562129021, 0.10343516618013382, 0.13125064969062805, 0.07057765126228333, -0.011259262450039387, -0.069544717669487, 0.05991494655609131, -0.003658764064311981, -0.06904593855142593, 0.04123999923467636, -0.11242011934518814, -0.01196740660816431, 0.009768360294401646, 0.11199247092008591, -0.062344446778297424, -0.005306920036673546, -0.11247684061527252, 0.0028180480003356934, -0.14773620665073395, -0.02894614078104496, -0.07614428550004959, -0.031011177226901054, 0.02214771695435047, -0.08436717092990875, -0.039100389927625656, -0.014332986436784267, -0.13986822962760925, -0.023398039862513542, -0.04736753925681114, 0.09517386555671692, -0.11655298620462418, -0.055166248232126236, 0.09626336395740509, -0.014291220344603062, 0.09398853778839111, 0.044903021305799484, -0.07022914290428162, 0.06434562802314758, -0.09496074169874191, -0.11316927522420883, 0.0959542915225029, 0.04407193884253502, 0.04776572808623314, 0.002675051800906658, -0.009801552630960941, 0.0964970737695694, -0.0004672814975492656, 0.04560285806655884, 0.029566984623670578, -0.13350655138492584, -0.0154465576633811, -0.023506714031100273, -0.1332763433456421, 0.013271442614495754, -0.0963786169886589, 0.14655865728855133, 0.02278325706720352, 0.1790017932653427, -0.006577861960977316, 0.03352291136980057, -0.06583039462566376, 0.02249913476407528, -0.054776906967163086, -0.1789943426847458, -0.11596626788377762, -0.06996689736843109, -0.029697518795728683, -0.02017328143119812, 0.2981296181678772, 0.08490652590990067, -0.08169146627187729, 0.07198161631822586, 0.09890054166316986, 0.014289051294326782, 0.010698960162699223, 0.22402669489383698, 0.06494433432817459, -0.015415017493069172, -0.06803465634584427, 0.013257022015750408, 0.018359694629907608, -0.0557638444006443, 0.07606630772352219, 0.11115698516368866, 0.0904100313782692, 0.07412039488554001, 0.03014315851032734, -0.010024036280810833, -0.14307071268558502, -0.14855130016803741, 0.009788756258785725, 0.11858081817626953, -0.04020537808537483, 0.05838916823267937, 0.11923600733280182, -0.046544451266527176, 0.04348647594451904, -0.0696643590927124, 0.006750994361937046, -0.17443858087062836, -0.09916237741708755, -0.07066202163696289, -0.1291390359401703, -0.028648626059293747, -0.05008450523018837, 0.03779691457748413, 0.13888272643089294, 0.03173473849892616, -0.03977445513010025, -0.0031719948165118694, -0.03568726405501366, -0.04079877957701683, 0.028132043778896332, -0.027076909318566322, 0.008340715430676937, -0.060388870537281036, -0.03699566051363945, -0.10331396758556366, -0.012213758192956448, -0.05525130033493042, 0.046983782202005386, -0.02489228919148445, 0.022161008790135384, -0.12678012251853943, -0.07989493757486343, -0.058136142790317535, 0.021603651344776154, -0.005209204275161028, 0.1960669904947281, 0.0022967387922108173, 0.03324297443032265, 0.073024682700634, 0.21338893473148346, -0.09978115558624268, -0.13632486760616302, -0.03914077952504158, 0.16939033567905426, 0.040507927536964417, 0.05812970548868179, -0.020286252722144127, -0.0033802634570747614, -0.09350753575563431, 0.2765567898750305, 0.3296546936035156, -0.06144540011882782, 0.059222228825092316, -0.002428764943033457, 0.015853505581617355, 0.07856421172618866, 0.11569953709840775, 0.13662907481193542, 0.16383591294288635, -0.08296691626310349, -0.01556343026459217, -0.047031085938215256, 0.0007883654325269163, -0.16509802639484406, 0.0851517915725708, -0.007957077585160732, -0.09604888409376144, -0.027654562145471573, 0.06391361355781555, -0.09459078311920166, 0.08622651547193527, 0.007731959223747253, -0.1700391173362732, -0.0547446571290493, 0.02429783344268799, 0.21584919095039368, -0.01918049342930317, 0.048848070204257965, -0.027314655482769012, -0.052800264209508896, 0.07536984980106354, -0.027961991727352142, -0.17292775213718414, -0.030522989109158516, 0.10590537637472153, -0.03492295742034912, 0.1096210852265358, -0.005980927962809801, 0.02427011914551258, 0.09679204970598221, 0.07734619826078415, -0.09253719449043274, 0.08694414794445038, 0.023542344570159912, -0.0641736388206482, -0.014475682750344276, -0.10951966792345047, -0.02482822723686695, -0.08250825107097626, 0.06349749118089676, -0.11941738426685333, 0.054807331413030624, 0.002499351045116782, -0.0450555719435215, -0.03437528386712074, 0.008221501484513283, -0.06355006247758865, 0.06977298855781555, 0.028691314160823822, -0.038544271141290665, -0.07007227838039398, -0.04650425910949707, -0.052271898835897446, -0.0004606132279150188, -0.1830073893070221, -0.10946376621723175, -0.007405308540910482, -0.023830890655517578, 0.08056151121854782, 0.04008973389863968, -0.08659695833921432, -0.039661966264247894, -0.059462614357471466, 0.037761665880680084, -0.13403542339801788, 0.05026970058679581, 0.08200293034315109, -0.01854803040623665, -0.008469896391034126, -0.052380140870809555, 0.018549039959907532, 0.021080557256937027, -0.09995073825120926, -0.06921897828578949 ]
null
null
transformers
# Cross-Encoder for MS MARCO - EN-DE This is a cross-lingual Cross-Encoder model for EN-DE that can be used for passage re-ranking. It was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task. The model can be used for Information Retrieval: See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html). The training code is available in this repository, see `train_script.py`. ## Usage with SentenceTransformers When you have [SentenceTransformers](https://www.sbert.net/) installed, you can use the model like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('model_name', max_length=512) query = 'How many people live in Berlin?' docs = ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'] pairs = [(query, doc) for doc in docs] scores = model.predict(pairs) ``` ## Usage with Transformers With the transformers library, you can use the model like this: ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('model_name') tokenizer = AutoTokenizer.from_pretrained('model_name') features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits print(scores) ``` ## Performance The performance was evaluated on three datasets: - **TREC-DL19 EN-EN**: The original [TREC 2019 Deep Learning Track](https://microsoft.github.io/msmarco/TREC-Deep-Learning-2019.html): Given an English query and 1000 documents (retrieved by BM25 lexical search), rank documents with according to their relevance. We compute NDCG@10. BM25 achieves a score of 45.46, a perfect re-ranker can achieve a score of 95.47. - **TREC-DL19 DE-EN**: The English queries of TREC-DL19 have been translated by a German native speaker to German. We rank the German queries versus the English passages from the original TREC-DL19 setup. We compute NDCG@10. - **GermanDPR DE-DE**: The [GermanDPR](https://www.deepset.ai/germanquad) dataset provides German queries and German passages from Wikipedia. We indexed the 2.8 Million paragraphs from German Wikipedia and retrieved for each query the top 100 most relevant passages using BM25 lexical search with Elasticsearch. We compute MRR@10. BM25 achieves a score of 35.85, a perfect re-ranker can achieve a score of 76.27. We also check the performance of bi-encoders using the same evaluation: The retrieved documents from BM25 lexical search are re-ranked using query & passage embeddings with cosine-similarity. Bi-Encoders can also be used for end-to-end semantic search. | Model-Name | TREC-DL19 EN-EN | TREC-DL19 DE-EN | GermanDPR DE-DE | Docs / Sec | | ------------- |:-------------:| :-----: | :---: | :----: | | BM25 | 45.46 | - | 35.85 | -| | **Cross-Encoder Re-Rankers** | | | | | [cross-encoder/msmarco-MiniLM-L6-en-de-v1](https://huggingface.co/cross-encoder/msmarco-MiniLM-L6-en-de-v1) | 72.43 | 65.53 | 46.77 | 1600 | | [cross-encoder/msmarco-MiniLM-L12-en-de-v1](https://huggingface.co/cross-encoder/msmarco-MiniLM-L12-en-de-v1) | 72.94 | 66.07 | 49.91 | 900 | | [svalabs/cross-electra-ms-marco-german-uncased](https://huggingface.co/svalabs/cross-electra-ms-marco-german-uncased) (DE only) | - | - | 53.67 | 260 | | [deepset/gbert-base-germandpr-reranking](https://huggingface.co/deepset/gbert-base-germandpr-reranking) (DE only) | - | - | 53.59 | 260 | | **Bi-Encoders (re-ranking)** | | | | | [sentence-transformers/msmarco-distilbert-multilingual-en-de-v2-tmp-lng-aligned](https://huggingface.co/sentence-transformers/msmarco-distilbert-multilingual-en-de-v2-tmp-lng-aligned) | 63.38 | 58.28 | 37.88 | 940 | | [sentence-transformers/msmarco-distilbert-multilingual-en-de-v2-tmp-trained-scratch](https://huggingface.co/sentence-transformers/msmarco-distilbert-multilingual-en-de-v2-tmp-trained-scratch) | 65.51 | 58.69 | 38.32 | 940 | | [svalabs/bi-electra-ms-marco-german-uncased](https://huggingface.co/svalabs/bi-electra-ms-marco-german-uncased) (DE only) | - | - | 34.31 | 450 | | [deepset/gbert-base-germandpr-question_encoder](https://huggingface.co/deepset/gbert-base-germandpr-question_encoder) (DE only) | - | - | 42.55 | 450 | Note: Docs / Sec gives the number of (query, document) pairs we can re-rank within a second on a V100 GPU.
{"license": "apache-2.0"}
text-classification
cross-encoder/msmarco-MiniLM-L6-en-de-v1
[ "transformers", "pytorch", "bert", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
Cross-Encoder for MS MARCO - EN-DE ================================== This is a cross-lingual Cross-Encoder model for EN-DE that can be used for passage re-ranking. It was trained on the MS Marco Passage Ranking task. The model can be used for Information Retrieval: See URL Retrieve & Re-rank. The training code is available in this repository, see 'train\_script.py'. Usage with SentenceTransformers ------------------------------- When you have SentenceTransformers installed, you can use the model like this: Usage with Transformers ----------------------- With the transformers library, you can use the model like this: Performance ----------- The performance was evaluated on three datasets: * TREC-DL19 EN-EN: The original TREC 2019 Deep Learning Track: Given an English query and 1000 documents (retrieved by BM25 lexical search), rank documents with according to their relevance. We compute NDCG@10. BM25 achieves a score of 45.46, a perfect re-ranker can achieve a score of 95.47. * TREC-DL19 DE-EN: The English queries of TREC-DL19 have been translated by a German native speaker to German. We rank the German queries versus the English passages from the original TREC-DL19 setup. We compute NDCG@10. * GermanDPR DE-DE: The GermanDPR dataset provides German queries and German passages from Wikipedia. We indexed the 2.8 Million paragraphs from German Wikipedia and retrieved for each query the top 100 most relevant passages using BM25 lexical search with Elasticsearch. We compute MRR@10. BM25 achieves a score of 35.85, a perfect re-ranker can achieve a score of 76.27. We also check the performance of bi-encoders using the same evaluation: The retrieved documents from BM25 lexical search are re-ranked using query & passage embeddings with cosine-similarity. Bi-Encoders can also be used for end-to-end semantic search. Note: Docs / Sec gives the number of (query, document) pairs we can re-rank within a second on a V100 GPU.
[]
[ "TAGS\n#transformers #pytorch #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 44 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.033624548465013504, 0.11583641916513443, -0.006466314196586609, 0.03693987429141998, 0.11971456557512283, 0.02166047692298889, 0.12048175185918808, 0.1321706920862198, 0.01429376658052206, -0.07648304849863052, 0.13479697704315186, 0.22184574604034424, -0.012524329125881195, 0.06779264658689499, -0.08648756891489029, -0.24625052511692047, 0.09145351499319077, 0.047294795513153076, -0.014880131930112839, 0.09696296602487564, 0.11783695966005325, -0.039152491837739944, 0.04760254546999931, -0.001044782460667193, -0.06468931585550308, 0.021656470373272896, 0.04621468111872673, -0.12399723380804062, 0.08577025681734085, 0.04309383034706116, 0.08981439471244812, 0.043990809470415115, -0.02796093188226223, -0.18899650871753693, 0.017944956198334694, 0.004928646609187126, -0.08463240414857864, 0.05281449854373932, 0.06530068069696426, -0.05573006346821785, 0.03398431837558746, 0.04542854055762291, -0.0016487948596477509, 0.05704890564084053, -0.07702083140611649, -0.16284798085689545, -0.07866179198026657, 0.09533879160881042, 0.07475922256708145, 0.07963944226503372, 0.04054330289363861, 0.14942073822021484, -0.12101521342992783, 0.07066715508699417, 0.10073971748352051, -0.34290656447410583, 0.006926678586751223, 0.07116013020277023, 0.014098972082138062, 0.017672913148999214, -0.019676387310028076, 0.03557991981506348, 0.04636206105351448, 0.012471802532672882, 0.013373504392802715, -0.05803173407912254, -0.11233073472976685, 0.033623434603214264, -0.05398819223046303, -0.06736930459737778, 0.23002012073993683, -0.017524652183055878, 0.029936393722891808, -0.010111764073371887, -0.06318812817335129, 0.013274782337248325, -0.021011902019381523, 0.05844603851437569, 0.015864230692386627, 0.10881900787353516, 0.07598614692687988, -0.022992447018623352, -0.134002223610878, 0.009127716533839703, -0.21311087906360626, 0.09606034308671951, 0.021664461120963097, 0.07777068018913269, -0.1399146467447281, 0.06432689726352692, 0.020169859752058983, -0.11113383620977402, 0.0031303863506764174, -0.08473936468362808, 0.11171188205480576, 0.014474593102931976, -0.07460460811853409, 0.043622326105833054, 0.13258297741413116, 0.22116319835186005, 0.05043065547943115, 0.01951022446155548, -0.058310430496931076, 0.11443714052438736, -0.029090533033013344, 0.09178415685892105, 0.02817927859723568, -0.011186737567186356, 0.10911470651626587, -0.12157052755355835, 0.06358799338340759, -0.035286471247673035, -0.1657385379076004, -0.02933056652545929, 0.015140064992010593, 0.11050593107938766, 0.05076056346297264, 0.06345071643590927, -0.04509174823760986, 0.00138133205473423, 0.15608423948287964, -0.06623833626508713, 0.012619871646165848, 0.00755408825352788, 0.022991495206952095, 0.06565660983324051, 0.04396168887615204, 0.02844872511923313, -0.0714418962597847, 0.10865788906812668, -0.03904176875948906, -0.012435059994459152, -0.031150242313742638, -0.011157393455505371, 0.08781024068593979, -0.09729834645986557, 0.04240095987915993, -0.15245147049427032, -0.14401228725910187, 0.03192957490682602, 0.0642581358551979, 0.011705230921506882, -0.08818046003580093, 0.036137018352746964, -0.005530418362468481, 0.03979998826980591, -0.09016831964254379, -0.02225935459136963, -0.08854800462722778, 0.0716094896197319, -0.09036583453416824, 0.01928316429257393, -0.17267097532749176, 0.06887461990118027, -0.11407924443483353, -0.017700498923659325, -0.05338679254055023, -0.0243575070053339, -0.09160128980875015, 0.18350261449813843, -0.04107033088803291, -0.04106784239411354, 0.012812543660402298, 0.008747126907110214, -0.05867214873433113, 0.11271152645349503, -0.07905924320220947, -0.06208209693431854, 0.193105086684227, -0.11312826722860336, -0.18387918174266815, 0.06954911351203918, 0.005549502093344927, -0.0009717301581986248, 0.07701529562473297, 0.162315234541893, 0.09539500623941422, -0.01430345606058836, 0.08831996470689774, 0.14527755975723267, -0.06741485744714737, -0.1899579018354416, 0.039836782962083817, -0.0518929660320282, -0.1316661685705185, 0.06082112714648247, -0.023314109072089195, 0.08668104559183121, 0.005240552127361298, -0.07265958935022354, -0.054311603307724, -0.04594247043132782, 0.03554585576057434, 0.015470363199710846, 0.07424456626176834, -0.0780770555138588, -0.005386367905884981, -0.02062210813164711, 0.026860764250159264, 0.04755967855453491, 0.05645165964961052, -0.0616346038877964, 0.06374672800302505, 0.038931526243686676, 0.03387635946273804, -0.1492423564195633, -0.0026758909225463867, -0.006619501393288374, 0.0394667387008667, 0.008541746996343136, 0.07236725091934204, 0.034056663513183594, -0.07543224841356277, -0.007346620317548513, -0.00990211684256792, 0.13999104499816895, 0.04745945334434509, -0.027293691411614418, -0.11992690712213516, 0.051172077655792236, -0.03158222511410713, 0.032270461320877075, -0.012440373189747334, 0.0255413930863142, 0.05202852562069893, 0.0896928682923317, -0.04332510009407997, 0.10814680904150009, -0.0314299538731575, 0.008204556070268154, -0.06421206146478653, 0.005224584136158228, 0.11843527108430862, 0.04398849606513977, -0.08439838886260986, 0.19714058935642242, -0.07091148942708969, 0.26326096057891846, 0.21244864165782928, -0.19821707904338837, 0.06781923025846481, -0.020836390554904938, -0.01435175072401762, -0.0025727238971740007, 0.030823349952697754, 0.03073660098016262, 0.04172210767865181, 0.01771463453769684, 0.17754846811294556, -0.04812063276767731, -0.04125499352812767, -0.02247837372124195, -0.06345567107200623, -0.007959081791341305, 0.05230538174510002, 0.1404448002576828, -0.18668782711029053, 0.1913488507270813, 0.31166794896125793, -0.01400954183191061, 0.08600834012031555, -0.07593902945518494, 0.024557704105973244, 0.06549926102161407, -0.04413771256804466, -0.022823473438620567, -0.03249483183026314, -0.1398867517709732, 0.004700822290033102, 0.10399795323610306, 0.039237093180418015, 0.05066521093249321, -0.13973669707775116, -0.048357754945755005, 0.003572310321033001, -0.01914861798286438, -0.04409148171544075, 0.05693091079592705, 0.01529010757803917, 0.09553027153015137, -0.029352953657507896, -0.1036546602845192, 0.14165401458740234, 0.0015804258873686194, -0.08185664564371109, 0.14734046161174774, -0.1686059981584549, -0.2672240436077118, -0.15006466209888458, -0.15754730999469757, -0.040469471365213394, 0.021722793579101562, 0.14752502739429474, -0.05383254215121269, -0.05828969553112984, 0.010263088159263134, -0.0916106328368187, -0.010198148898780346, 0.02441413700580597, -0.03552008420228958, 0.058104515075683594, 0.005071738734841347, -0.11582627147436142, -0.06767059117555618, 0.013409256935119629, -0.03647346422076225, 0.10347112268209457, -0.09030836820602417, 0.07127439975738525, 0.12871627509593964, 0.01243792474269867, 0.03326582908630371, -0.03339415043592453, 0.13468076288700104, -0.037399765104055405, -0.0030397437512874603, 0.22738684713840485, -0.04533824697136879, 0.08281181007623672, 0.1512300968170166, 0.036025270819664, -0.04110533371567726, 0.020001502707600594, -0.05199878290295601, -0.07664331048727036, -0.2774033844470978, -0.12388425320386887, -0.10269982367753983, 0.04752686247229576, 0.058578431606292725, 0.09034254401922226, 0.14106637239456177, 0.07592342048883438, -0.019996946677565575, 0.015969308093190193, 0.007639035116881132, 0.07632344216108322, 0.27253058552742004, -0.004745368380099535, 0.12732580304145813, -0.11762017011642456, -0.057013943791389465, 0.1265517622232437, 0.062044817954301834, 0.1236846074461937, 0.1234099343419075, 0.051585108041763306, 0.06550563126802444, 0.14724965393543243, 0.10624245554208755, 0.12493196874856949, 0.028105320408940315, 0.004841153975576162, -0.04727954789996147, -0.011976541019976139, -0.059970151633024216, 0.033797647804021835, -0.022963635623455048, -0.14457081258296967, -0.042962756007909775, -0.1554267853498459, 0.07811342179775238, 0.18483006954193115, 0.015438108704984188, -0.13616302609443665, 0.026686720550060272, 0.08427965641021729, -0.020005637779831886, -0.05305152013897896, 0.09713099151849747, -0.10846424102783203, -0.11656232923269272, 0.13418035209178925, -0.025590287521481514, 0.14492498338222504, -0.0248755794018507, 0.053225770592689514, -0.010453574359416962, -0.13999760150909424, 0.06725247949361801, 0.15039514005184174, -0.29247668385505676, 0.20505018532276154, -0.010331127792596817, -0.04482366144657135, -0.08720868080854416, 0.0013150222366675735, 0.07004573196172714, 0.25857123732566833, 0.07095246762037277, 0.01650865562260151, -0.10063847154378891, -0.0516783781349659, -0.06302125006914139, 0.02917596511542797, 0.029199955984950066, -0.011154877953231335, -0.05557602643966675, -0.0649041160941124, -0.021162277087569237, -0.0042198156006634235, 0.03514545410871506, -0.020032672211527824, -0.16620703041553497, 0.05486385524272919, 0.10046247392892838, 0.049888525158166885, -0.051651325076818466, -0.027695588767528534, -0.11299943178892136, 0.1882910132408142, -0.10712026804685593, -0.08979795128107071, -0.09791862219572067, -0.12910650670528412, 0.024837469682097435, -0.06235082820057869, 0.0702935978770256, -0.08618972450494766, 0.0010052000870928168, -0.05893877521157265, -0.20419736206531525, 0.10277000069618225, -0.12839920818805695, -0.0499797947704792, -0.043673690408468246, 0.12797610461711884, -0.1115511879324913, 0.02084740810096264, 0.040034614503383636, 0.003994327504187822, -0.1022302433848381, -0.12035572528839111, -0.022925307974219322, 0.03245508298277855, 0.059485625475645065, -0.026264069601893425, -0.101990707218647, -0.0571921281516552, 0.004007372539490461, -0.041260719299316406, 0.2471790313720703, 0.1690734475851059, -0.08446425944566727, 0.18276453018188477, 0.18056119978427887, -0.09941285848617554, -0.31816133856773376, -0.1598256230354309, -0.13836894929409027, -0.10192114859819412, -0.027664819732308388, -0.15457311272621155, 0.13434267044067383, 0.04874061420559883, -0.08777574449777603, 0.08117468655109406, -0.15966583788394928, -0.08619657903909683, 0.21040642261505127, -0.04534730687737465, 0.3174908459186554, -0.13340862095355988, -0.08690694719552994, -0.09293768554925919, -0.1719014197587967, 0.14943470060825348, -0.07211921364068985, 0.06140616536140442, -0.012918728403747082, 0.027143200859427452, -0.008217151276767254, -0.04619397595524788, 0.1133522018790245, -0.014066650532186031, 0.024132272228598595, -0.1293950229883194, -0.0024696949403733015, 0.06233421340584755, -0.02529490925371647, 0.042755890637636185, -0.13924568891525269, 0.023840680718421936, -0.11120016127824783, -0.026454707607626915, -0.061154644936323166, 0.07823684066534042, 0.00931283738464117, -0.0443255715072155, -0.026689087972044945, -0.03375519812107086, 0.0254792720079422, -0.009869445115327835, 0.25119978189468384, 0.015602859668433666, 0.07951269298791885, 0.12697459757328033, 0.09628882259130478, -0.18091686069965363, -0.004001361317932606, -0.11060860753059387, -0.08483202010393143, 0.059636712074279785, -0.14534153044223785, 0.05873894691467285, 0.10457886010408401, -0.07313844561576843, 0.05897313356399536, 0.08260530233383179, 0.025774210691452026, -0.05598508194088936, 0.13984863460063934, -0.17353367805480957, 0.025570666417479515, -0.01700710318982601, 0.10868968814611435, 0.06578559428453445, 0.06253332644701004, 0.12149979919195175, 0.02378910779953003, -0.045509472489356995, 0.02385631762444973, 0.026733553037047386, -0.05165635421872139, 0.03943267837166786, 0.05652942880988121, 0.007127806544303894, -0.135700523853302, 0.10076140612363815, 0.03949320688843727, -0.139295756816864, -0.02660050429403782, 0.1049129068851471, -0.15733122825622559, -0.13474465906620026, 0.02149812877178192, 0.07954824715852737, -0.14437325298786163, -0.11576265841722488, -0.04053666815161705, -0.14157921075820923, 0.0704789087176323, 0.1180836483836174, 0.12505967915058136, 0.07465346157550812, -0.01606745831668377, -0.07044725865125656, 0.04563111439347267, -0.0016263002762570977, -0.06975575536489487, 0.033102910965681076, -0.10268332809209824, -0.04989679530262947, 0.014775268733501434, 0.10792622715234756, -0.052679866552352905, -0.017179174348711967, -0.09913178533315659, 0.03468547388911247, -0.1543911248445511, 0.004061874467879534, -0.073458231985569, -0.006152780260890722, 0.019225608557462692, -0.07365509122610092, -0.02818438597023487, -0.0066046081483364105, -0.12461689859628677, -0.01529979333281517, -0.03191344812512398, 0.07186026871204376, -0.11214398592710495, -0.0533626489341259, 0.09820953011512756, -0.01407189667224884, 0.10279780626296997, 0.07704830169677734, -0.07897642254829407, 0.08706441521644592, -0.15200023353099823, -0.10375725477933884, 0.09744801372289658, 0.05007246136665344, 0.03702830895781517, 0.005307680461555719, 0.015036889351904392, 0.12421702593564987, -0.033559706062078476, 0.045524436980485916, 0.04527132213115692, -0.15067242085933685, -0.025562534108757973, -0.01080042403191328, -0.13799738883972168, -0.008309535682201385, -0.09745097905397415, 0.13043686747550964, 0.023049110546708107, 0.19978578388690948, -0.025248127058148384, 0.04954942688345909, -0.04377560690045357, 0.017911331728100777, -0.03635789826512337, -0.1728820949792862, -0.13917817175388336, -0.07161751389503479, -0.04154269024729729, -0.012728624045848846, 0.2695949673652649, 0.041973695158958435, -0.03874577581882477, 0.0755695179104805, 0.08605275303125381, 0.017459535971283913, 0.013623190112411976, 0.24184226989746094, 0.05891833081841469, -0.006889847572892904, -0.07207736372947693, 0.0095368018373847, 0.01267328392714262, -0.07895752787590027, 0.09543026238679886, 0.09202968329191208, 0.04860799387097359, 0.062353868037462234, 0.02349235676229, 0.03302542492747307, -0.13147249817848206, -0.16768448054790497, 0.017875514924526215, 0.09444275498390198, 0.01440428476780653, 0.11145492643117905, 0.11040886491537094, -0.039926040917634964, 0.02716808021068573, -0.049468379467725754, -0.012980833649635315, -0.18488742411136627, -0.12038124352693558, -0.08426732569932938, -0.12213902920484543, 0.003872788278385997, -0.04934149980545044, 0.01755918748676777, 0.09066841751337051, 0.04828392341732979, -0.067583829164505, -0.012107289396226406, -0.04697847366333008, -0.04348131641745567, 0.04423016682267189, -0.03289409726858139, -0.011101330630481243, -0.024587200954556465, -0.04467953369021416, -0.10911267250776291, -0.0322733148932457, -0.04883580282330513, 0.04706401005387306, -0.008699473924934864, 0.03918442130088806, -0.11204695701599121, -0.0767931342124939, -0.04038625955581665, 0.029535427689552307, -0.0040898569859564304, 0.1869017481803894, 0.009583257138729095, 0.038710206747055054, 0.09553047269582748, 0.16753792762756348, -0.07017162442207336, -0.15256629884243011, -0.047502581030130386, 0.20098082721233368, 0.05657508969306946, 0.06080913171172142, 0.012499228119850159, 0.011208959855139256, -0.056554656475782394, 0.3228624165058136, 0.28597143292427063, -0.04547877982258797, 0.036143917590379715, -0.018416233360767365, 0.021603884175419807, 0.10241707414388657, 0.14358292520046234, 0.127496138215065, 0.1839708834886551, -0.06740538030862808, -0.026008620858192444, -0.03907419368624687, 0.0018298098584637046, -0.17193500697612762, 0.08618982881307602, -0.01088035199791193, -0.09144339710474014, -0.022995738312602043, 0.10249924659729004, -0.11244610697031021, 0.115500807762146, -0.0033514192327857018, -0.11603543907403946, -0.043959882110357285, -0.0058623687364161015, 0.1857547014951706, 0.007513022515922785, 0.02933778427541256, -0.017118019983172417, -0.06501377373933792, 0.10304711014032364, -0.011885383166372776, -0.21393372118473053, -0.04481937363743782, 0.08612168580293655, -0.037672173231840134, 0.11618560552597046, 0.011411667801439762, 0.057896751910448074, 0.07590916007757187, 0.08023592829704285, -0.08374308794736862, 0.08137734979391098, 0.011783070862293243, -0.049928218126297, 0.014140583574771881, -0.10634700208902359, -0.023432763293385506, -0.08135902136564255, 0.04243294522166252, -0.08825377374887466, 0.047646839171648026, -0.0020005095284432173, -0.0684451311826706, -0.04261155426502228, 0.0507601760327816, -0.08038213849067688, 0.056206125766038895, 0.03303709998726845, -0.028430424630641937, -0.06526947766542435, -0.06521723419427872, -0.031366121023893356, 0.01531433779746294, -0.19917620718479156, -0.09116730839014053, -0.0053079086355865, -0.03789437934756279, 0.09222039580345154, 0.04602926969528198, -0.09527811408042908, -0.016809822991490364, -0.09963963180780411, 0.025631384924054146, -0.17426390945911407, 0.047836631536483765, 0.06116890534758568, -0.003921943251043558, -0.006005494389683008, -0.05374446138739586, 0.02516728825867176, 0.016484463587403297, -0.0931040346622467, -0.08561339229345322 ]
null
null
transformers
# Cross-Encoder for Natural Language Inference This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. ## Training Data The model was trained on the [SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral. ## Performance For evaluation results, see [SBERT.net - Pretrained Cross-Encoder](https://www.sbert.net/docs/pretrained_cross-encoders.html#nli). ## Usage Pre-trained models can be used like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('cross-encoder/nli-MiniLM2-L6-H768') scores = model.predict([('A man is eating pizza', 'A man eats something'), ('A black race car starts up in front of a crowd of people.', 'A man is driving down a lonely road.')]) #Convert scores to labels label_mapping = ['contradiction', 'entailment', 'neutral'] labels = [label_mapping[score_max] for score_max in scores.argmax(axis=1)] ``` ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('cross-encoder/nli-MiniLM2-L6-H768') tokenizer = AutoTokenizer.from_pretrained('cross-encoder/nli-MiniLM2-L6-H768') features = tokenizer(['A man is eating pizza', 'A black race car starts up in front of a crowd of people.'], ['A man eats something', 'A man is driving down a lonely road.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits label_mapping = ['contradiction', 'entailment', 'neutral'] labels = [label_mapping[score_max] for score_max in scores.argmax(dim=1)] print(labels) ``` ## Zero-Shot Classification This model can also be used for zero-shot-classification: ```python from transformers import pipeline classifier = pipeline("zero-shot-classification", model='cross-encoder/nli-MiniLM2-L6-H768') sent = "Apple just announced the newest iPhone X" candidate_labels = ["technology", "sports", "politics"] res = classifier(sent, candidate_labels) print(res) ```
{"language": "en", "license": "apache-2.0", "tags": ["MiniLMv2"], "datasets": ["multi_nli", "snli"], "metrics": ["accuracy"], "pipeline_tag": "zero-shot-classification"}
zero-shot-classification
cross-encoder/nli-MiniLM2-L6-H768
[ "transformers", "pytorch", "roberta", "text-classification", "MiniLMv2", "zero-shot-classification", "en", "dataset:multi_nli", "dataset:snli", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #roberta #text-classification #MiniLMv2 #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# Cross-Encoder for Natural Language Inference This model was trained using SentenceTransformers Cross-Encoder class. ## Training Data The model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral. ## Performance For evaluation results, see URL - Pretrained Cross-Encoder. ## Usage Pre-trained models can be used like this: ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ## Zero-Shot Classification This model can also be used for zero-shot-classification:
[ "# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.", "## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):", "## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ "TAGS\n#transformers #pytorch #roberta #text-classification #MiniLMv2 #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.", "## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):", "## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ 74, 30, 49, 18, 14, 33, 21 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #text-classification #MiniLMv2 #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ -0.02044294774532318, 0.15575215220451355, -0.005729633383452892, 0.10251614451408386, 0.12102481722831726, 0.0196793582290411, 0.24352902173995972, 0.08777108788490295, -0.061177175492048264, -0.0498560331761837, 0.027248933911323547, 0.21822427213191986, 0.0010040567722171545, 0.04355757310986519, -0.03316637873649597, -0.21052862703800201, 0.0374772734940052, 0.0011222735047340393, 0.035875651985406876, 0.10555332899093628, 0.1079602912068367, -0.05825333669781685, 0.02936483733355999, -0.0074266353622078896, -0.1471531242132187, 0.01978910155594349, 0.014657000079751015, -0.05597163364291191, 0.09088216722011566, 0.03162437304854393, 0.07640507817268372, 0.044130727648735046, 0.0359363928437233, -0.11705342680215836, 0.008199307136237621, 0.04898254573345184, -0.004913500510156155, 0.0370866060256958, -0.00022700687986798584, -0.0944342240691185, 0.04685598239302635, -0.05163112282752991, 0.06335128843784332, 0.06336657702922821, -0.06337660551071167, -0.08779341727495193, 0.01903083175420761, 0.006823698990046978, 0.0888151302933693, 0.0916251465678215, -0.027888722717761993, 0.176200270652771, -0.14434005320072174, 0.09549557417631149, 0.06898857653141022, -0.22526176273822784, -0.029416201636195183, 0.13454563915729523, -0.017796028405427933, 0.07568223029375076, -0.04063544049859047, 0.001407303148880601, -0.005642368923872709, 0.026956800371408463, 0.06969509273767471, -0.050709933042526245, -0.08801715075969696, -0.019106287509202957, -0.10946095734834671, 0.026862667873501778, 0.18969419598579407, -0.06000250205397606, -0.04209117963910103, -0.11292178183794022, -0.028622079640626907, -0.04354577511548996, -0.038097579032182693, -0.013464096933603287, 0.017812706530094147, 0.04009467735886574, -0.04261704161763191, -0.007979943417012691, -0.08131571859121323, -0.0506574884057045, -0.08506361395120621, 0.142503559589386, 0.0439772866666317, 0.02919222228229046, -0.14018838107585907, 0.06360571086406708, 0.014296100474894047, -0.10294913500547409, -0.026166347786784172, -0.0562506727874279, -0.13996000587940216, 0.013341892510652542, -0.13175155222415924, -0.11357545852661133, 0.08036861568689346, 0.14524029195308685, 0.033483244478702545, 0.057364143431186676, 0.022740498185157776, 0.06059505417943001, 0.07478686422109604, 0.07375437766313553, -0.01647154428064823, -0.005091042723506689, 0.011495700106024742, -0.018717771396040916, 0.027981381863355637, -0.05632516369223595, -0.12219145894050598, 0.012404533103108406, 0.05071580037474632, 0.044075433164834976, 0.004659691359847784, 0.06904308497905731, -0.05017746612429619, -0.07942458987236023, -0.017594285309314728, -0.08466748893260956, 0.012220519594848156, 0.0021516575943678617, 0.008066737093031406, 0.20753417909145355, -0.0032892751041799784, -0.001783298677764833, -0.0777028277516365, -0.02729785069823265, -0.08844969421625137, 0.029972929507493973, -0.09422492980957031, -0.09456250071525574, -0.005738576874136925, 0.017432570457458496, -0.03443697839975357, -0.10961238294839859, -0.10451719909906387, 0.006420790683478117, 0.04920955002307892, -0.047186944633722305, -0.03329421207308769, -0.10712062567472458, -0.025519581511616707, 0.012389824725687504, 0.0024639535695314407, -0.0523810014128685, -0.010154986754059792, 0.016318874433636665, 0.009967450983822346, 0.06681060791015625, -0.07963141053915024, -0.0034459151793271303, -0.1686994731426239, -0.03197946771979332, 0.0888838991522789, 0.14580626785755157, -0.026265665888786316, -0.0024822617415338755, -0.0786450132727623, -0.046546462923288345, 0.029771046712994576, 0.022189827635884285, -0.022078340873122215, 0.1154167577624321, -0.2562636435031891, -0.06268885731697083, 0.18092875182628632, -0.16840341687202454, -0.04971366748213768, 0.11542101204395294, -0.040555015206336975, -0.0159752257168293, 0.11004742234945297, 0.05568857863545418, 0.1427459865808487, -0.013338529504835606, 0.020268959924578667, 0.06892896443605423, -0.05401294305920601, 0.020948922261595726, 0.07549042999744415, 0.07451295107603073, -0.07352538406848907, 0.06212400272488594, -0.042508386075496674, -0.0092052286490798, -0.028451910242438316, -0.07829519361257553, -0.041404034942388535, -0.02517317421734333, 0.1021743044257164, 0.051869772374629974, 0.05929679423570633, -0.04298311099410057, -0.09496255218982697, 0.1357431411743164, 0.0936148390173912, -0.10954777151346207, 0.036424580961465836, -0.10238581895828247, -0.04425984248518944, -0.026203801855444908, -0.00053686264436692, -0.1986764818429947, -0.04732391983270645, 0.019514815881848335, -0.020605089142918587, 0.11984080821275711, 0.07286829501390457, 0.004813093226402998, 0.05050431936979294, -0.013205836527049541, 0.0251864455640316, 0.04538772255182266, 0.019697532057762146, -0.03238143026828766, -0.15861572325229645, -0.05194970220327377, -0.022638393566012383, 0.10143778473138809, -0.11849323660135269, 0.033251602202653885, 0.003522925078868866, 0.022640027105808258, -0.016404464840888977, -0.004675028845667839, -0.027870770543813705, 0.08268404006958008, -0.060549378395080566, -0.011601543985307217, 0.10490240901708603, 0.03367161750793457, -0.09695181250572205, 0.015321983024477959, -0.1289159059524536, -0.13752931356430054, 0.08735805749893188, -0.10627978295087814, -0.04266037419438362, -0.02050606906414032, -0.004458249546587467, 0.0010316065745428205, -0.04309717193245888, 0.0142204063013196, 0.12857533991336823, 0.044687509536743164, 0.08479636907577515, -0.0655670315027237, -0.03472926840186119, -0.016005247831344604, -0.1163235604763031, 0.027767043560743332, 0.07834739983081818, 0.007977491244673729, -0.2155103236436844, 0.12728577852249146, 0.016468778252601624, -0.145059734582901, 0.12234603613615036, 0.0013666314771398902, 0.019802941009402275, -0.05205942317843437, -0.06094863638281822, 0.02666199766099453, -0.015601716935634613, -0.06755883246660233, 0.024576008319854736, 0.05809354409575462, 0.03223351016640663, 0.010222329758107662, -0.10010291635990143, 0.021383119747042656, 0.02337442710995674, -0.003969837911427021, 0.0544491671025753, 0.021903283894062042, 0.004492407664656639, 0.07696536183357239, -0.02368658222258091, 0.02025931514799595, 0.06709286570549011, -0.03731979802250862, -0.1081533432006836, 0.21347522735595703, -0.11439304053783417, -0.2484542280435562, -0.09996219724416733, 0.007433710619807243, -0.11364053934812546, 0.013527089729905128, 0.03378510847687721, -0.09448453783988953, -0.008910873904824257, -0.07073260098695755, -0.05484704673290253, -0.0688294991850853, -0.03356403857469559, -0.09563419967889786, 0.05820997431874275, 0.014633937738835812, -0.1319122165441513, 0.003746438305824995, 0.01911775954067707, -0.14036278426647186, 0.06801030784845352, -0.025866447016596794, -0.02374478429555893, 0.20698107779026031, -0.006399189122021198, 0.023344149813055992, -0.007683551870286465, 0.16975949704647064, 0.0035672851372510195, -0.018430767580866814, 0.19002118706703186, 0.016030965372920036, 0.047893352806568146, 0.08078493177890778, 0.050072427839040756, -0.037740107625722885, 0.025159437209367752, 0.02966536581516266, -0.08309731632471085, -0.12931333482265472, -0.1112164705991745, -0.07011828571557999, -0.038261134177446365, 0.06872250139713287, -0.0005302509525790811, 0.08435732871294022, 0.08520771563053131, 0.011725612916052341, 0.012095702812075615, 0.05306835472583771, 0.10781484842300415, 0.12893790006637573, 0.07121352851390839, 0.15692542493343353, -0.047855451703071594, -0.0684313252568245, 0.03605961427092552, 0.00816289521753788, 0.14798010885715485, 0.024254849180579185, 0.08398832380771637, 0.13952401280403137, -0.02514726296067238, 0.06087009981274605, 0.025777742266654968, -0.006576240062713623, 0.014605501666665077, -0.0035993834026157856, -0.07545624673366547, -0.03985719382762909, 0.008608873002231121, -0.04186517372727394, 0.01744694449007511, -0.028046291321516037, -0.011955139227211475, 0.10657468438148499, 0.1183527484536171, 0.08778827637434006, -0.2335188090801239, -0.061750371009111404, 0.05188007280230522, -0.00852731429040432, 0.011962666176259518, 0.03651585057377815, 0.03392406553030014, -0.10650632530450821, 0.05428740754723549, -0.024489751085639, 0.08219022303819656, -0.09053569287061691, 0.02439223602414131, -0.03778831660747528, 0.07468104362487793, 0.06307021528482437, 0.12548880279064178, -0.3033726215362549, 0.06063929572701454, -0.020806297659873962, 0.05622200667858124, -0.08107896149158478, 0.0028053312562406063, 0.015252925455570221, 0.07426614314317703, 0.08526664972305298, 0.01456899382174015, -0.0823117196559906, -0.045259974896907806, -0.022170212119817734, 0.026017745956778526, 0.05752839148044586, -0.0417073555290699, 0.08663494884967804, -0.044031962752342224, 0.012591813690960407, -0.021412985399365425, 0.10096929967403412, -0.13842929899692535, -0.18496187031269073, 0.05729715898633003, 0.07190593332052231, -0.06036533787846565, -0.00896046869456768, -0.06729908287525177, -0.0046204193495213985, 0.18266166746616364, -0.16754329204559326, -0.10630062222480774, -0.11823683977127075, 0.043637026101350784, 0.03401126340031624, -0.06585201621055603, -0.027233652770519257, -0.024582572281360626, 0.15820975601673126, -0.025506220757961273, -0.17854087054729462, 0.005970880389213562, -0.05736926943063736, -0.10791557282209396, -0.02273452840745449, 0.05576005578041077, 0.09391968697309494, 0.04176970571279526, 0.04560280218720436, 0.03514290973544121, -0.005637523718178272, -0.13116592168807983, -0.09910948574542999, 0.05187121406197548, 0.06251195818185806, 0.048435330390930176, -0.04185543209314346, -0.13902944326400757, -0.03716715797781944, 0.03413102403283119, 0.20758028328418732, 0.15466174483299255, -0.06234185770153999, 0.09755624085664749, 0.10137150436639786, -0.13250549137592316, -0.20437651872634888, -0.10882659256458282, 0.05001349374651909, -0.0022695756051689386, 0.05168755725026131, -0.1302715539932251, 0.10319656878709793, 0.083262600004673, -0.01484473142772913, -0.11818551272153854, -0.36828625202178955, -0.09672967344522476, 0.06553153693675995, 0.11521992832422256, 0.1520194113254547, -0.10205123573541641, -0.005296474788337946, -0.07552354037761688, -0.044183917343616486, 0.35867756605148315, -0.09099519997835159, 0.10859952867031097, -0.010641813278198242, -0.030311493203043938, 0.02501179277896881, -0.02564670890569687, 0.0802052766084671, 0.03555964305996895, 0.0896834284067154, -0.010414876975119114, -0.07110468298196793, 0.11722871661186218, -0.015368213877081871, 0.05958518385887146, -0.10305779427289963, 0.05555308610200882, -0.021252896636724472, -0.04688797891139984, -0.06845952570438385, -0.0042302715592086315, -0.050310663878917694, -0.06645714491605759, -0.06168019026517868, 0.08734352886676788, 0.062900610268116, -0.017363497987389565, 0.17581401765346527, -0.10376202315092087, 0.07722096145153046, 0.18952016532421112, 0.16043440997600555, 0.06037820130586624, -0.10005547106266022, -0.05394621193408966, -0.01869809627532959, 0.09961840510368347, -0.11019805818796158, 0.09213057905435562, 0.05014769732952118, 0.0008330723503604531, 0.10419861227273941, 0.10926486551761627, -0.00010683351138141006, 0.012845735065639019, 0.06380632519721985, -0.07594796270132065, -0.017352456226944923, 0.0023655621334910393, 0.06880790740251541, 0.00939999707043171, 0.06447158753871918, 0.15127430856227875, -0.05777039751410484, -0.039313800632953644, 0.011953848414123058, 0.018955323845148087, -0.06653337180614471, 0.09266457706689835, 0.014808623120188713, 0.06856025755405426, -0.09951932728290558, 0.08257970213890076, 0.08482392877340317, 0.05402697995305061, 0.044281087815761566, 0.09554260969161987, -0.0932822898030281, -0.10082618147134781, 0.039596740156412125, 0.061615992337465286, -0.04861472174525261, -0.0343339629471302, -0.07771088927984238, -0.10094257444143295, -0.021274495869874954, 0.04030368849635124, 0.1061166450381279, 0.07835422456264496, -0.11980900168418884, -0.05509219691157341, -0.10747860372066498, 0.046549104154109955, -0.018050633370876312, 0.014276295900344849, -0.06947217136621475, 0.14072272181510925, 0.023404918611049652, -0.001454650075174868, -0.06352926790714264, -0.08115922659635544, -0.15791098773479462, 0.026657739654183388, -0.04205377772450447, 0.023753900080919266, -0.10338426381349564, 0.0004739674914162606, 0.023971648886799812, 0.022347131744027138, -0.00845251139253378, 0.022289467975497246, -0.02622738853096962, -0.007509141229093075, -0.006241797469556332, 0.08058474957942963, -0.05678468570113182, -0.005870653782039881, 0.008782824501395226, -0.061331361532211304, 0.037134040147066116, 0.011782398447394371, -0.08965231478214264, 0.06462793797254562, -0.14839473366737366, -0.008881854824721813, -0.003317304654046893, 0.07068034261465073, 0.010659690015017986, -0.08165115863084793, 0.028718244284391403, 0.06622877717018127, 0.05330555886030197, 0.040457505732774734, 0.09237898886203766, -0.09132218360900879, -0.011976861394941807, -0.03474588692188263, -0.059702664613723755, -0.06024169921875, 0.0002460430550854653, 0.06025345250964165, 0.04950081557035446, 0.19933414459228516, -0.08655760437250137, 0.06564944982528687, -0.11729869991540909, -0.03846811130642891, 0.03819186985492706, -0.05719383805990219, -0.09395851194858551, -0.09235701709985733, 0.023392481729388237, -0.02268671616911888, 0.1527499556541443, 0.002475373214110732, 0.05926771089434624, 0.010905974544584751, 0.028507882729172707, 0.06112102046608925, 0.04705872759222984, 0.18926985561847687, 0.04899565875530243, -0.0006252169259823859, 0.03444909304380417, 0.0069234031252563, -0.00851255189627409, -0.005979177076369524, 0.11397765576839447, 0.04995276778936386, -0.09392622858285904, 0.06481017917394638, 0.004640630912035704, 0.07064500451087952, -0.04637962952256203, 0.060036927461624146, 0.014667286537587643, 0.03148027881979942, -0.04418458044528961, 0.047379978001117706, 0.14499598741531372, -0.09075529873371124, 0.10548415780067444, 0.02095014415681362, -0.10149512439966202, -0.14682315289974213, -0.2836573123931885, -0.07666631042957306, -0.07261471450328827, -0.029271677136421204, -0.1452789604663849, -0.009410695172846317, 0.10793043673038483, 0.051152389496564865, -0.034224238246679306, 0.01657513529062271, 0.014843174256384373, -0.028330761939287186, -0.0035902399104088545, -0.05603792145848274, 0.03995335474610329, 0.011219297535717487, 0.05928570777177811, 0.04003645479679108, 0.04911723732948303, 0.06697992980480194, 0.036973632872104645, 0.0662851557135582, -0.019685739651322365, -0.07325179129838943, -0.10154934972524643, 0.02310566045343876, -0.015174109488725662, 0.018684856593608856, 0.10841603577136993, 0.060141220688819885, -0.08344266563653946, -0.010309474542737007, 0.17818959057331085, -0.06457918137311935, -0.19236153364181519, -0.13457652926445007, 0.3051201105117798, 0.018925899639725685, 0.053329791873693466, 0.020012937486171722, -0.04247530922293663, -0.05543932318687439, 0.21223171055316925, 0.13816148042678833, -0.04326262325048447, 0.003548911539837718, 0.04428713023662567, -0.025400912389159203, 0.0012141154147684574, 0.12842924892902374, -0.049275193363428116, 0.2147340476512909, -0.04683703929185867, 0.11148975044488907, -0.05481105297803879, -0.025474337860941887, -0.09800951182842255, 0.07831015437841415, -0.023340299725532532, -0.026296326890587807, -0.024348052218556404, 0.07491261512041092, -0.12954653799533844, 0.11372479051351547, 0.10086604952812195, -0.016529874876141548, -0.11656635254621506, -0.0038510828744620085, 0.05810235068202019, -0.04559818655252457, 0.06644906103610992, -0.02578437328338623, 0.020790429785847664, 0.0480940118432045, -0.012924017384648323, -0.02156239189207554, -0.04710719361901283, 0.02637610211968422, -0.031752072274684906, 0.09451976418495178, -0.014266686514019966, 0.149828240275383, 0.10387542098760605, 0.008918853476643562, -0.06441456079483032, 0.11729678511619568, -0.05231069400906563, -0.03512819483876228, 0.09054436534643173, 0.0698077604174614, -0.03000200167298317, 0.09521470963954926, 0.0763562023639679, -0.13672411441802979, 0.02393573336303234, -0.006911805365234613, -0.05196363106369972, -0.03412194177508354, 0.009850500151515007, -0.05475502088665962, 0.0817679911851883, 0.11484018713235855, -0.049366556107997894, -0.031070103868842125, -0.048949237912893295, 0.08188489079475403, -0.02629435621201992, -0.030747873708605766, -0.07633061707019806, -0.1555677205324173, -0.010053691454231739, 0.07328963279724121, 0.040582116693258286, -0.17272479832172394, -0.013014145195484161, 0.013647903688251972, -0.03834797814488411, -0.06634438782930374, 0.04233178123831749, 0.05373592674732208, 0.018995555117726326, -0.06601491570472717, -0.22447992861270905, 0.00313325016759336, 0.07953866571187973, -0.12006951123476028, -0.1079668253660202 ]
null
null
transformers
# Cross-Encoder for Natural Language Inference This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. ## Training Data The model was trained on the [SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral. ## Performance For evaluation results, see [SBERT.net - Pretrained Cross-Encoder](https://www.sbert.net/docs/pretrained_cross-encoders.html#nli). ## Usage Pre-trained models can be used like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('cross-encoder/nli-deberta-base') scores = model.predict([('A man is eating pizza', 'A man eats something'), ('A black race car starts up in front of a crowd of people.', 'A man is driving down a lonely road.')]) #Convert scores to labels label_mapping = ['contradiction', 'entailment', 'neutral'] labels = [label_mapping[score_max] for score_max in scores.argmax(axis=1)] ``` ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('cross-encoder/nli-deberta-base') tokenizer = AutoTokenizer.from_pretrained('cross-encoder/nli-deberta-base') features = tokenizer(['A man is eating pizza', 'A black race car starts up in front of a crowd of people.'], ['A man eats something', 'A man is driving down a lonely road.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits label_mapping = ['contradiction', 'entailment', 'neutral'] labels = [label_mapping[score_max] for score_max in scores.argmax(dim=1)] print(labels) ``` ## Zero-Shot Classification This model can also be used for zero-shot-classification: ```python from transformers import pipeline classifier = pipeline("zero-shot-classification", model='cross-encoder/nli-deberta-base') sent = "Apple just announced the newest iPhone X" candidate_labels = ["technology", "sports", "politics"] res = classifier(sent, candidate_labels) print(res) ```
{"language": "en", "license": "apache-2.0", "tags": ["deberta-base-base"], "datasets": ["multi_nli", "snli"], "metrics": ["accuracy"], "pipeline_tag": "zero-shot-classification"}
zero-shot-classification
cross-encoder/nli-deberta-base
[ "transformers", "pytorch", "deberta", "text-classification", "deberta-base-base", "zero-shot-classification", "en", "dataset:multi_nli", "dataset:snli", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #deberta #text-classification #deberta-base-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
# Cross-Encoder for Natural Language Inference This model was trained using SentenceTransformers Cross-Encoder class. ## Training Data The model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral. ## Performance For evaluation results, see URL - Pretrained Cross-Encoder. ## Usage Pre-trained models can be used like this: ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ## Zero-Shot Classification This model can also be used for zero-shot-classification:
[ "# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.", "## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):", "## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ "TAGS\n#transformers #pytorch #deberta #text-classification #deberta-base-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.", "## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):", "## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ 82, 30, 49, 18, 14, 33, 21 ]
[ "passage: TAGS\n#transformers #pytorch #deberta #text-classification #deberta-base-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ -0.017919236794114113, 0.17484505474567413, -0.006778021343052387, 0.09020459651947021, 0.1217234879732132, 0.022064127027988434, 0.2579192519187927, 0.0706629753112793, -0.02560870721936226, -0.026418372988700867, 0.03061394952237606, 0.2159518152475357, -0.0013462663628160954, 0.05589500442147255, -0.03857579827308655, -0.2438502162694931, 0.061350636184215546, 0.027842139825224876, 0.022447824478149414, 0.07621462643146515, 0.1191072091460228, -0.0333460196852684, 0.026534980162978172, -0.026621505618095398, -0.11767103523015976, 0.06971481442451477, 0.006729434244334698, -0.046108637005090714, 0.11075051873922348, 0.043982140719890594, 0.12971849739551544, 0.07993969321250916, 0.007482636719942093, -0.16267713904380798, 0.0033050712663680315, 0.04660211503505707, -0.04120033606886864, 0.018665317445993423, -0.010308874770998955, -0.10247897356748581, -0.0069606006145477295, -0.03359629586338997, 0.09808439016342163, 0.042958714067935944, -0.09473854303359985, -0.13059258460998535, 0.004744038451462984, -0.009968029335141182, 0.08492931723594666, 0.09815505146980286, -0.044237349182367325, 0.09669717401266098, -0.12101228535175323, 0.09154734760522842, 0.06905952095985413, -0.22493839263916016, -0.008864817209541798, 0.0754816010594368, 0.007986187934875488, 0.0940021201968193, -0.049159515649080276, -0.008431563153862953, -0.0072252945974469185, 0.04261929169297218, 0.07692787796258926, -0.05560019612312317, -0.029140371829271317, -0.01132947951555252, -0.11610542982816696, 0.03487422317266464, 0.16286525130271912, -0.053832922130823135, -0.05808834731578827, -0.08846239000558853, -0.05754591524600983, -0.023027431219816208, -0.012039309367537498, -0.021956484764814377, 0.03269496187567711, 0.04693632572889328, -0.03188761696219444, -0.023934174329042435, -0.09819990396499634, -0.030339164659380913, -0.09693356603384018, 0.1425454169511795, 0.052155353128910065, 0.01640072651207447, -0.1095743328332901, 0.08313541114330292, -0.006854149512946606, -0.12270649522542953, -0.03743729740381241, -0.05163949355483055, -0.15208719670772552, 0.010412071831524372, -0.14689771831035614, -0.1657213568687439, 0.0636163204908371, 0.22127321362495422, -0.006838707253336906, 0.061513639986515045, -0.03846382722258568, 0.04549495503306389, 0.07374730706214905, 0.1022772416472435, -0.022257618606090546, -0.01428849808871746, 0.009882316924631596, 0.00492192292585969, 0.05656255781650543, -0.05912761762738228, -0.13208015263080597, -0.00998496450483799, 0.02265138551592827, 0.054991092532873154, -0.0008808944839984179, 0.06360004097223282, -0.03538055717945099, -0.04901706799864769, 0.0227778609842062, -0.0906081423163414, 0.0026485577691346407, 0.014134993776679039, 0.029431795701384544, 0.17693224549293518, 0.020862584933638573, -0.004859711974859238, -0.05569184198975563, -0.009419133886694908, -0.0977562889456749, 0.015554077923297882, -0.08444737643003464, -0.11969055980443954, 0.011663319543004036, 0.003652011277154088, -0.021681562066078186, -0.1281861960887909, -0.10168514400720596, 0.004108457826077938, 0.03208392485976219, -0.03909071907401085, 0.015188808552920818, -0.1010080948472023, 0.0065541681833565235, 0.03226438909769058, -0.025443706661462784, -0.12368056178092957, 0.0002779632341116667, 0.02452290616929531, 0.04530460014939308, 0.07497736811637878, -0.09423666447401047, 0.024494370445609093, -0.15544532239437103, -0.028563059866428375, -0.01667734421789646, 0.14176183938980103, -0.02910003252327442, -0.019413335248827934, -0.06585237383842468, -0.007243653293699026, 0.058851540088653564, 0.045494481921195984, -0.0012533236294984818, 0.09859539568424225, -0.23281124234199524, -0.07907988131046295, 0.17719946801662445, -0.16429319977760315, -0.0486748069524765, 0.12165110558271408, -0.041907504200935364, -0.010634295642375946, 0.1148764044046402, 0.0759708508849144, 0.11857113242149353, -0.05272052064538002, 0.01327597163617611, 0.059210531413555145, -0.013807910494506359, 0.08182888478040695, 0.05671772360801697, 0.06395997107028961, -0.048123303800821304, 0.049450382590293884, -0.07918567210435867, 0.025966141372919083, 0.003685257863253355, -0.0503033809363842, -0.010366802103817463, -0.026216426864266396, 0.11877400428056717, 0.009352277033030987, 0.05625995621085167, -0.016213785856962204, -0.12251374870538712, 0.21402542293071747, 0.08897387236356735, -0.1363316923379898, 0.058858033269643784, -0.06846489757299423, -0.05604909360408783, -0.026184428483247757, 0.0015726321144029498, -0.18267063796520233, -0.09330670535564423, 0.03376764804124832, -0.054645828902721405, 0.11028531938791275, 0.07214705646038055, 0.006064467132091522, 0.07033234089612961, -0.05192290246486664, -0.00424180319532752, 0.01720810867846012, 0.038721177726984024, -0.036984633654356, -0.1573333591222763, -0.043651033192873, -0.03378076106309891, 0.10459751635789871, -0.083980031311512, 0.05748550593852997, 0.020577292889356613, 0.05778523534536362, -0.018668560311198235, 0.006940379738807678, -0.04644136503338814, 0.08093221485614777, -0.07215810567140579, 0.0030081323347985744, 0.0480438768863678, 0.03287649527192116, -0.12552571296691895, 0.05625665932893753, -0.16400094330310822, -0.08858004957437515, 0.08336731791496277, -0.10597740858793259, -0.05624821409583092, 0.005034140311181545, 0.013031069189310074, -0.018484337255358696, -0.05029313638806343, -0.061814866960048676, 0.1293589025735855, 0.03966318815946579, 0.10980992019176483, -0.05287345498800278, -0.01964675262570381, -0.0482243075966835, -0.09251663833856583, 0.011578804813325405, 0.05432790145277977, -0.040208250284194946, -0.18161581456661224, 0.14689861238002777, 0.04261213168501854, -0.18580032885074615, 0.1192583292722702, 0.011556273326277733, -0.013267622329294682, -0.07316488772630692, -0.042044542729854584, 0.012686844915151596, 0.002280525164678693, -0.07316605001688004, 0.023419227451086044, 0.0275421142578125, 0.04828672111034393, 0.017010854557156563, -0.08768188953399658, 0.018734991550445557, 0.018892882391810417, -0.010525209829211235, -0.013075338676571846, 0.011025841347873211, -0.000451444648206234, 0.06450878083705902, -0.007773975841701031, -0.011740980669856071, 0.08586942404508591, -0.03331875056028366, -0.11060875654220581, 0.21197614073753357, -0.13663940131664276, -0.29283758997917175, -0.11560633778572083, -0.01821517013013363, -0.11588021367788315, 0.05138146132230759, 0.08542856574058533, -0.0784367173910141, -0.017021579667925835, -0.055768195539712906, 0.03742534667253494, -0.07697836309671402, -0.02671816572546959, -0.13119091093540192, 0.03968813270330429, 0.002618422033265233, -0.13116998970508575, 0.011808295734226704, 0.02306325174868107, -0.10027164965867996, 0.05156473070383072, -0.03342854231595993, -0.011255438439548016, 0.17475736141204834, 0.01947406306862831, -0.007189058233052492, -0.010381417348980904, 0.15660187602043152, -0.012113670818507671, 0.013772371225059032, 0.1531538963317871, -0.01804499886929989, 0.0691995620727539, 0.10988116264343262, 0.05142567306756973, -0.03948983550071716, 0.032541390508413315, 0.0425097830593586, -0.05227792263031006, -0.18176481127738953, -0.04530734568834305, -0.06120879203081131, -0.03588156774640083, 0.045826781541109085, -0.002554042264819145, 0.1386830061674118, 0.0584985613822937, -0.02608049474656582, 0.028696592897176743, 0.07444869726896286, 0.11012503504753113, 0.20868901908397675, 0.02215469628572464, 0.14364799857139587, -0.05212293565273285, -0.08216919004917145, 0.025713404640555382, 0.08192639797925949, 0.10736413300037384, 0.02556353434920311, 0.13540080189704895, 0.1181240975856781, -0.0464794747531414, 0.047419823706150055, 0.029400385916233063, 0.0036782200913876295, 0.028799811378121376, -0.011339631862938404, -0.07278546690940857, -0.0010116202756762505, 0.03521711379289627, -0.02378656528890133, -0.025403685867786407, -0.038471147418022156, -0.0721053034067154, 0.09753517806529999, 0.09115509688854218, 0.11036071181297302, -0.2238849252462387, -0.042538877576589584, 0.06578190624713898, -0.024620208889245987, -0.018519602715969086, 0.025245273485779762, 0.016439203172922134, -0.07241172343492508, 0.08101996034383774, -0.011768458411097527, 0.06556490063667297, -0.1087847426533699, 0.022372644394636154, -0.012932389974594116, 0.061735186725854874, 0.040821243077516556, 0.10025244206190109, -0.3246529996395111, 0.03476742282509804, -0.005999711807817221, 0.029654929414391518, -0.03695162013173103, 0.03486808016896248, -0.008343764580786228, 0.10219413787126541, 0.03985029458999634, -0.008391914889216423, -0.08303898572921753, -0.06871290504932404, 0.009689846076071262, -0.007260762155056, 0.07025502622127533, -0.0629740059375763, 0.08817509561777115, -0.010449778288602829, 0.0016933290753513575, 0.006975602358579636, 0.09686606377363205, -0.14434391260147095, -0.2131495624780655, 0.08673153072595596, 0.05021044984459877, 0.010274029336869717, -0.019899999722838402, -0.063311867415905, 0.02985469065606594, 0.1368086189031601, -0.2048020213842392, -0.11194822192192078, -0.1205817312002182, 0.013885971158742905, 0.06674546748399734, -0.0719052106142044, -0.009029691107571125, -0.01671578735113144, 0.14161917567253113, -0.047132838517427444, -0.15904663503170013, 0.048942841589450836, -0.04190973564982414, -0.08424278348684311, -0.06921082735061646, 0.05643795430660248, 0.09686369448900223, 0.030242890119552612, 0.029346643015742302, 0.02473868802189827, -0.0022532748989760876, -0.10308665782213211, -0.0937262773513794, 0.03814380615949631, 0.046387914568185806, 0.041833728551864624, -0.054832275956869125, -0.07265356183052063, -0.05505887418985367, 0.034558236598968506, 0.1707107424736023, 0.15727956593036652, -0.06044219061732292, 0.07659310102462769, 0.1564161479473114, -0.12999629974365234, -0.17598319053649902, -0.0958087146282196, 0.052657872438430786, -0.0022943171206861734, 0.010701284743845463, -0.1882411539554596, 0.07577592134475708, 0.10238347202539444, -0.014260304160416126, -0.09820341318845749, -0.31268560886383057, -0.10006320476531982, 0.08948606997728348, 0.07572510838508606, 0.14116118848323822, -0.14300470054149628, -0.022387029603123665, -0.09737753868103027, -0.02373063750565052, 0.32563692331314087, -0.07908619195222855, 0.11513449996709824, 0.006708592176437378, -0.005339004099369049, 0.017618998885154724, -0.007111316546797752, 0.10118609666824341, 0.06391410529613495, 0.07238401472568512, -0.0010573635809123516, -0.03157266229391098, 0.12429160624742508, -0.043994881212711334, 0.0791054517030716, -0.0807863101363182, 0.03458205983042717, -0.0036144908517599106, -0.05679265782237053, -0.06492902338504791, -0.0016901958733797073, -0.05981303006410599, -0.07778781652450562, -0.03170030191540718, 0.02783580683171749, 0.049517083913087845, -0.03456871584057808, 0.19143566489219666, -0.09633027017116547, 0.04450080916285515, 0.16865786910057068, 0.11210394650697708, 0.09611877799034119, -0.1323700100183487, -0.023861965164542198, -0.05238739401102066, 0.09035442769527435, -0.14058387279510498, 0.09721042215824127, 0.04075280949473381, -0.019153106957674026, 0.13013513386249542, 0.10257936269044876, -0.04201320558786392, 0.005299472715705633, 0.05364769324660301, -0.069093719124794, -0.013505956158041954, -0.006708675064146519, 0.02270290069282055, -0.013517465442419052, 0.0885765329003334, 0.18381044268608093, -0.03942682594060898, -0.039468929171562195, 0.01676282100379467, 0.0241100937128067, -0.05941447243094444, 0.09033694118261337, 0.031980983912944794, 0.05016849935054779, -0.0845252051949501, 0.1065291091799736, 0.09383998066186905, -0.018360288813710213, 0.03850986808538437, 0.09511259198188782, -0.07044090330600739, -0.09541173279285431, -0.07573738694190979, 0.06390374153852463, -0.13005295395851135, -0.03210396692156792, -0.09693922102451324, -0.10981699824333191, -0.020465118810534477, 0.12053918093442917, 0.07337996363639832, 0.13589875400066376, -0.09528075158596039, -0.0058266520500183105, -0.09304293245077133, 0.04597464203834534, -0.042746372520923615, 0.07824336737394333, -0.07284919917583466, 0.10182245820760727, 0.01442404929548502, -0.009569516405463219, -0.06037160009145737, -0.057536300271749496, -0.16596199572086334, 0.01353226788341999, -0.11248540133237839, 0.013369976542890072, -0.07564810663461685, -0.003428440308198333, 0.02889503911137581, -0.013047188520431519, 0.02233140915632248, 0.015427744947373867, -0.014243711717426777, -0.001076067448593676, -0.007512488402426243, 0.07452798634767532, -0.09254355728626251, -0.0224663857370615, -0.0007979341316968203, -0.10577289015054703, 0.07481153309345245, 0.022001992911100388, -0.05968869850039482, 0.06870705634355545, -0.12014054507017136, 0.018218807876110077, -0.013477991335093975, 0.049837224185466766, 0.008990928530693054, -0.07807828485965729, 0.010413790121674538, 0.02911846712231636, 0.03793303668498993, 0.02012135088443756, 0.0702735185623169, -0.10435860604047775, -0.005537795834243298, 0.0025856862775981426, -0.0888323113322258, -0.06285945326089859, -0.005251996219158173, 0.08319589495658875, 0.06082065775990486, 0.16617128252983093, -0.0849226638674736, 0.07734136283397675, -0.08354036509990692, -0.024099072441458702, 0.038741666823625565, -0.06119823828339577, -0.026489082723855972, -0.09731624275445938, 0.024220485240221024, -0.0221981443464756, 0.15727193653583527, -0.05304912477731705, 0.10377173870801926, 0.025893371552228928, 0.007442222908139229, 0.10159438848495483, 0.07511613517999649, 0.16644248366355896, 0.03194189444184303, -0.0273452065885067, -0.044500235468149185, 0.0070147085934877396, -0.00298611749894917, 0.012824890203773975, 0.0338033027946949, 0.04943566024303436, -0.044641975313425064, 0.09158478677272797, 0.04284714534878731, 0.0834619477391243, -0.09450089186429977, 0.0035546980798244476, 0.03256475180387497, 0.003104743082076311, -0.04372749105095863, 0.05473312735557556, 0.14064237475395203, -0.07580794394016266, 0.09932668507099152, 0.03387482091784477, -0.10877352207899094, -0.1351385861635208, -0.1863681524991989, -0.05050908029079437, -0.07281213998794556, -0.04108026623725891, -0.15000101923942566, -0.005527449306100607, 0.08127135038375854, 0.02964053489267826, -0.03025227040052414, 0.03608755022287369, 0.11143360286951065, -0.04370906949043274, -0.00046194856986403465, -0.03292416036128998, 0.04451778531074524, 0.08353973925113678, 0.03823191672563553, 0.03840858116745949, 0.06304603070020676, 0.0344526544213295, 0.028357230126857758, 0.07804963737726212, -0.0008654832490719855, -0.08324922621250153, -0.09817887842655182, 0.020858142524957657, 0.012600918300449848, 0.002442109165713191, 0.1435920149087906, 0.047566432505846024, -0.09237156063318253, 0.0035901819355785847, 0.18544302880764008, -0.07793515175580978, -0.1289753019809723, -0.15431572496891022, 0.30636870861053467, 0.04648807272315025, 0.0860166996717453, 0.0031372213270515203, -0.08658036589622498, -0.07398116588592529, 0.1415596306324005, 0.11029724031686783, -0.01313875149935484, -0.0018154880963265896, 0.07163608819246292, -0.011961987242102623, 0.027820246294140816, 0.09145978093147278, -0.01949695497751236, 0.2702353596687317, -0.03983939811587334, 0.09685558080673218, -0.009753075428307056, -0.018274527043104172, -0.08570487052202225, 0.09368117898702621, -0.05075543373823166, -0.011814014986157417, -0.04103449732065201, 0.06783640384674072, -0.10198454558849335, 0.027102762833237648, 0.07177355140447617, -0.009137939661741257, -0.09852461516857147, -0.01163004245609045, 0.09070277959108353, -0.06330209225416183, 0.07219487428665161, 0.00593530060723424, -0.00828663632273674, 0.13653971254825592, -0.027425991371273994, -0.018974507227540016, -0.04480079934000969, 0.029946543276309967, -0.004020014777779579, 0.05136985331773758, -0.023417271673679352, 0.1450430303812027, 0.11358147859573364, 0.04622943699359894, -0.04121633619070053, 0.13102847337722778, -0.030566634610295296, -0.04576524347066879, 0.079828180372715, 0.0841294452548027, -0.03201373666524887, 0.1340978443622589, 0.05817287042737007, -0.1661582887172699, 0.050168827176094055, -0.0027838374953716993, -0.09595660120248795, -0.048232778906822205, 0.009206850081682205, -0.05545344203710556, 0.04720744490623474, 0.121429443359375, -0.056501343846321106, -0.027821805328130722, -0.03406064212322235, 0.01209392212331295, 0.003256601979956031, -0.020942598581314087, -0.05718604102730751, -0.15206564962863922, 0.01133764162659645, 0.1001632884144783, 0.04775266721844673, -0.2003321647644043, -0.020460201427340508, 0.023056350648403168, -0.00819737371057272, -0.09109323471784592, 0.05224178358912468, 0.020707691088318825, 0.025865281000733376, -0.05955204740166664, -0.24627716839313507, -0.008364363573491573, 0.07661200314760208, -0.13099971413612366, -0.12398263067007065 ]
null
null
transformers
# Cross-Encoder for Natural Language Inference This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. This model is based on [microsoft/deberta-v3-base](https://huggingface.co/microsoft/deberta-v3-base) ## Training Data The model was trained on the [SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral. ## Performance - Accuracy on SNLI-test dataset: 92.38 - Accuracy on MNLI mismatched set: 90.04 For futher evaluation results, see [SBERT.net - Pretrained Cross-Encoder](https://www.sbert.net/docs/pretrained_cross-encoders.html#nli). ## Usage Pre-trained models can be used like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('cross-encoder/nli-deberta-v3-base') scores = model.predict([('A man is eating pizza', 'A man eats something'), ('A black race car starts up in front of a crowd of people.', 'A man is driving down a lonely road.')]) #Convert scores to labels label_mapping = ['contradiction', 'entailment', 'neutral'] labels = [label_mapping[score_max] for score_max in scores.argmax(axis=1)] ``` ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('cross-encoder/nli-deberta-v3-base') tokenizer = AutoTokenizer.from_pretrained('cross-encoder/nli-deberta-v3-base') features = tokenizer(['A man is eating pizza', 'A black race car starts up in front of a crowd of people.'], ['A man eats something', 'A man is driving down a lonely road.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits label_mapping = ['contradiction', 'entailment', 'neutral'] labels = [label_mapping[score_max] for score_max in scores.argmax(dim=1)] print(labels) ``` ## Zero-Shot Classification This model can also be used for zero-shot-classification: ```python from transformers import pipeline classifier = pipeline("zero-shot-classification", model='cross-encoder/nli-deberta-v3-base') sent = "Apple just announced the newest iPhone X" candidate_labels = ["technology", "sports", "politics"] res = classifier(sent, candidate_labels) print(res) ```
{"language": "en", "license": "apache-2.0", "tags": ["microsoft/deberta-v3-base"], "datasets": ["multi_nli", "snli"], "metrics": ["accuracy"], "pipeline_tag": "zero-shot-classification"}
zero-shot-classification
cross-encoder/nli-deberta-v3-base
[ "transformers", "pytorch", "deberta-v2", "text-classification", "microsoft/deberta-v3-base", "zero-shot-classification", "en", "dataset:multi_nli", "dataset:snli", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# Cross-Encoder for Natural Language Inference This model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-base ## Training Data The model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral. ## Performance - Accuracy on SNLI-test dataset: 92.38 - Accuracy on MNLI mismatched set: 90.04 For futher evaluation results, see URL - Pretrained Cross-Encoder. ## Usage Pre-trained models can be used like this: ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ## Zero-Shot Classification This model can also be used for zero-shot-classification:
[ "# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-base", "## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.", "## Performance\r\n- Accuracy on SNLI-test dataset: 92.38\r\n- Accuracy on MNLI mismatched set: 90.04\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):", "## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ "TAGS\n#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-base", "## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.", "## Performance\r\n- Accuracy on SNLI-test dataset: 92.38\r\n- Accuracy on MNLI mismatched set: 90.04\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):", "## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ 83, 45, 49, 51, 14, 33, 21 ]
[ "passage: TAGS\n#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-base## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.## Performance\r\n- Accuracy on SNLI-test dataset: 92.38\r\n- Accuracy on MNLI mismatched set: 90.04\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ -0.05504194274544716, 0.18498900532722473, -0.00550383236259222, 0.10590871423482895, 0.11327389627695084, 0.027110308408737183, 0.22715409100055695, 0.06776661425828934, 0.047567833214998245, 0.009399577043950558, -0.022405998781323433, 0.11900611966848373, 0.04960653930902481, 0.087869793176651, -0.012860161252319813, -0.17446796596050262, 0.022154854610562325, -0.060964759439229965, 0.04712884873151779, 0.08188575506210327, 0.0980304479598999, -0.0625922754406929, 0.0420386902987957, -0.013875761069357395, -0.09353657811880112, 0.029924092814326286, -0.022783271968364716, 0.00029585283482447267, 0.056830354034900665, 0.08876778930425644, 0.11350555717945099, 0.03970613703131676, 0.054754313081502914, -0.20077818632125854, 0.00981262605637312, 0.07366976886987686, -0.018998906016349792, 0.020158149302005768, 0.05771129950881004, -0.09196603298187256, 0.02219942770898342, 0.010708280839025974, 0.05780442804098129, 0.07142206281423569, -0.06680011004209518, -0.09239906817674637, -0.04663665592670441, 0.020442496985197067, 0.08217284083366394, 0.09513447433710098, -0.031120484694838524, 0.09341774880886078, -0.047955457121133804, 0.08692276477813721, 0.033932022750377655, -0.20560957491397858, -0.01999054104089737, 0.07665058225393295, -0.00525390449911356, 0.10038639605045319, -0.07866141945123672, -0.04238595440983772, 0.0256351325660944, 0.0037330882623791695, -0.02014920301735401, -0.016897566616535187, 0.03941161185503006, -0.008122952654957771, -0.13860152661800385, -0.009864633902907372, 0.12632808089256287, -0.01790344901382923, -0.07836967706680298, -0.11451000720262527, -0.05198528617620468, -0.01190938614308834, 0.027615707367658615, -0.012339326553046703, 0.046395160257816315, 0.044713109731674194, 0.03577519208192825, -0.034368909895420074, -0.09168870747089386, -0.04361942037940025, -0.06804998964071274, 0.11794951558113098, 0.0334017276763916, 0.037278905510902405, -0.0888497531414032, 0.10708896070718765, 0.04664485529065132, -0.1272583156824112, -0.07123062759637833, -0.06702927500009537, -0.1781214475631714, 0.0028527032118290663, -0.11030040681362152, -0.15808755159378052, 0.043982647359371185, 0.20869985222816467, 0.02005097270011902, 0.049231406301259995, -0.01257365196943283, 0.030337277799844742, 0.11087530106306076, 0.10425414144992828, -0.014277372509241104, -0.02953529916703701, -0.018479187041521072, 0.03434520214796066, 0.00937954243272543, -0.0347900390625, -0.07278183102607727, 0.03311194106936455, -0.037407368421554565, 0.052623506635427475, 0.02084149606525898, 0.020093923434615135, -0.06911692768335342, -0.07502651959657669, 0.06647256016731262, -0.13196279108524323, 0.013282383792102337, 0.03074638545513153, -0.021044738590717316, 0.1381591260433197, 0.04739317670464516, -0.01108813751488924, -0.05282898619771004, -0.006589648779481649, -0.07417338341474533, 0.025174275040626526, -0.11721331626176834, -0.12351434677839279, 0.01577894203364849, 0.01619052328169346, -0.04012269526720047, -0.1093621551990509, -0.10973720997571945, -0.06980929523706436, 0.04586409404873848, -0.032204337418079376, 0.04230799898505211, -0.1013774424791336, 0.0031543162185698748, 0.018516983836889267, 0.008407749235630035, -0.07309696823358536, -0.013904559426009655, 0.006470143795013428, 0.006821855902671814, 0.05714721232652664, 0.0030589206144213676, -0.00819788221269846, -0.11710900813341141, -0.03597593680024147, 0.03594997897744179, 0.16159865260124207, -0.07749796658754349, 0.01715139113366604, -0.06503398716449738, -0.006782847456634045, 0.035939574241638184, 0.029043620452284813, 0.006536621134728193, 0.11206056922674179, -0.2584216892719269, -0.030810268595814705, 0.19104233384132385, -0.18144071102142334, -0.060482531785964966, 0.09351000189781189, -0.04705575853586197, -0.01418253593146801, 0.08187954127788544, 0.035303402692079544, 0.19866596162319183, -0.08584849536418915, -0.023834193125367165, 0.03971109166741371, -0.025841763243079185, 0.09634270519018173, 0.07864458858966827, 0.026394125074148178, -0.025367368012666702, 0.024069787934422493, -0.07283791899681091, 0.00022469974646810442, -0.027742238715291023, -0.07841481268405914, -0.012613152153789997, -0.056705292314291, 0.016994521021842957, -0.006013673730194569, -0.0005946015007793903, -0.0015652847941964865, -0.10310592502355576, 0.12788431346416473, 0.10505099594593048, -0.0985238254070282, 0.020611582323908806, -0.11138486117124557, -0.07536660879850388, -0.009428991936147213, 0.00015262819943018258, -0.18932463228702545, -0.09250278770923615, 0.02911352552473545, -0.11723294109106064, 0.050749171525239944, 0.07531702518463135, 0.010720492340624332, 0.058598339557647705, 0.0015055234543979168, -0.008644422516226768, -0.0001814083952922374, -0.006822328083217144, -0.0433112196624279, -0.15571190416812897, -0.04337691515684128, -0.03842359781265259, 0.07450328767299652, -0.1083345040678978, 0.030728261917829514, 0.014297927729785442, 0.07298529148101807, 0.0011833281023427844, -0.050213128328323364, -0.0036800713278353214, 0.04567810520529747, -0.05741357058286667, -0.023806100711226463, 0.02381780743598938, 0.03125442937016487, -0.06499721854925156, 0.0068349954672157764, -0.10206270217895508, -0.09298504889011383, 0.0743955448269844, 0.030032623559236526, -0.045575957745313644, 0.0054344660602509975, -0.026768343523144722, -0.004357573576271534, -0.05072944983839989, -0.029448527842760086, 0.15033066272735596, 0.03960495442152023, 0.08960851281881332, -0.05824699252843857, -0.04624389484524727, -0.022326815873384476, -0.04489469528198242, -0.0016865329816937447, 0.07372523099184036, -0.044137462973594666, -0.17496323585510254, 0.08889777213335037, -0.010667950846254826, -0.07524282485246658, 0.11211281269788742, 0.015653088688850403, -0.05002203583717346, -0.08680889755487442, 0.000930127571336925, 0.02013438194990158, 0.024658076465129852, -0.04671543836593628, 0.041982270777225494, 0.034399766474962234, 0.03163986653089523, 0.020479237660765648, -0.10359732061624527, 0.029420889914035797, 0.04479675367474556, -0.0278167724609375, -0.00014570071652997285, 0.0036901403218507767, -0.012195444665849209, 0.05434782803058624, -0.029096171259880066, 0.053911492228507996, 0.010792218148708344, -0.08213572949171066, -0.14955881237983704, 0.1994115114212036, -0.12241100519895554, -0.2808363735675812, -0.13221433758735657, 0.04644712060689926, -0.06551565229892731, 0.009064863435924053, 0.05141991004347801, -0.072455033659935, -0.04749874025583267, -0.09251312166452408, -0.037831857800483704, -0.06092586740851402, -0.033548105508089066, -0.0441695861518383, 0.055724505335092545, 0.02894425019621849, -0.1518130749464035, 0.02382926642894745, 0.01567000523209572, -0.08721353858709335, 0.023121468722820282, 0.05228535458445549, 0.03364504128694534, 0.12510274350643158, 0.026103613898158073, 0.017344849184155464, 0.014317595399916172, 0.2099318504333496, -0.010842018760740757, -0.009409492835402489, 0.16040030121803284, -0.04278382658958435, 0.06758035719394684, 0.0841597244143486, 0.035817213356494904, -0.07283150404691696, 0.04025954380631447, 0.043102845549583435, -0.040682584047317505, -0.2191053032875061, -0.0374254435300827, -0.0453595407307148, -0.053051259368658066, 0.04747210815548897, 0.017003025859594345, 0.00003742859189515002, 0.026984430849552155, -0.036188360303640366, -0.033118754625320435, 0.04294206202030182, 0.0830555334687233, 0.08507087826728821, 0.015191234648227692, 0.10150206834077835, -0.06411205232143402, 0.0007889331318438053, 0.06862052530050278, 0.02737266756594181, 0.13403761386871338, 0.018215110525488853, 0.062092628329992294, 0.14115402102470398, 0.0007196529186330736, 0.03844102472066879, 0.01684563048183918, -0.024710746482014656, 0.042225368320941925, 0.016566842794418335, -0.08079472184181213, -0.03737872838973999, 0.06574597954750061, -0.04520781338214874, -0.00975245051085949, -0.029636720195412636, -0.04465973377227783, 0.07290121912956238, 0.08444289863109589, 0.06730324029922485, -0.19123469293117523, -0.07690318673849106, 0.04482538625597954, -0.03502175211906433, -0.019144471734762192, 0.00846159178763628, 0.07304035872220993, -0.08820124715566635, 0.07147428393363953, -0.019599756225943565, 0.08889251202344894, -0.13368314504623413, -0.02138609066605568, 0.031508851796388626, 0.07431261986494064, 0.035433635115623474, 0.08450385183095932, -0.2430252879858017, 0.043314170092344284, -0.005113723687827587, 0.04611924663186073, -0.017762305215001106, 0.0583326481282711, 0.016640249639749527, 0.0056463307701051235, 0.09610150009393692, 0.016486236825585365, -0.03809536620974541, -0.06276077777147293, -0.04973018169403076, 0.01815425045788288, 0.09234899282455444, -0.0842365026473999, 0.1107988953590393, 0.0009532049880363047, -0.011686963029205799, -0.030774207785725594, 0.04680319130420685, -0.20535685122013092, -0.20215071737766266, 0.06971310824155807, -0.005529276095330715, -0.010821208357810974, -0.03295360133051872, -0.06344439089298248, -0.021343041211366653, 0.18232578039169312, -0.2676972448825836, -0.09968461096286774, -0.1292494386434555, 0.06743105500936508, 0.08672717958688736, -0.06368785351514816, -0.004118633456528187, 0.00032067415304481983, 0.13790972530841827, -0.03186510130763054, -0.13643941283226013, 0.0022550728172063828, -0.07091901451349258, -0.09933602064847946, -0.0700153186917305, 0.0577910952270031, 0.1268482357263565, 0.016119161620736122, 0.012887459248304367, 0.014917830005288124, -0.021414864808321, -0.12657129764556885, -0.07375578582286835, 0.1955774426460266, 0.005497160833328962, 0.050536543130874634, -0.052707739174366, -0.11646358668804169, -0.053731419146060944, 0.00653095031157136, 0.10237399488687515, 0.14385227859020233, -0.06381053477525711, 0.09386065602302551, 0.10036513209342957, -0.14982110261917114, -0.14412900805473328, -0.06292212009429932, 0.10731718689203262, 0.03679124638438225, 0.026879504323005676, -0.1850062757730484, 0.05401420593261719, 0.07810583710670471, 0.0007905822130851448, -0.05665091797709465, -0.32659003138542175, -0.10491985827684402, 0.030602186918258667, 0.08680696785449982, 0.027519717812538147, -0.10047727078199387, -0.05276799947023392, -0.06236030533909798, -0.07109225541353226, 0.24590544402599335, 0.013673312962055206, 0.08993037790060043, 0.019197948276996613, -0.06662249565124512, 0.0380304679274559, -0.033092107623815536, 0.10368650406599045, 0.06475149840116501, 0.08532814681529999, -0.029724158346652985, 0.028986984863877296, 0.048918794840574265, -0.03748374804854393, 0.10038186609745026, -0.03363662213087082, 0.08263049274682999, -0.08053994923830032, -0.0643891766667366, -0.012862388975918293, -0.030838094651699066, -0.04473068192601204, -0.033143624663352966, -0.04451719671487808, 0.056636396795511246, 0.09044865518808365, -0.022968539968132973, 0.042835816740989685, -0.08179818093776703, 0.0030684301164001226, 0.17638689279556274, 0.13431280851364136, 0.08338948339223862, -0.1430440992116928, -0.03945200517773628, -0.005747866816818714, 0.0784253180027008, -0.15915971994400024, 0.06068247929215431, 0.09306389838457108, 0.00307631678879261, 0.11957995593547821, 0.06650403887033463, -0.061954572796821594, 0.0008464406710118055, 0.07332736253738403, -0.08233359456062317, -0.08449359983205795, -0.026530278846621513, -0.014443869702517986, -0.05248107761144638, 0.006567076314240694, 0.1620473712682724, -0.032912857830524445, -0.01715313270688057, -0.004230120684951544, 0.03040088340640068, -0.02396770939230919, 0.12579575181007385, 0.007509858347475529, 0.05384809151291847, -0.08447986841201782, 0.10498879849910736, 0.07405369728803635, -0.055420421063899994, 0.025117255747318268, 0.08802884072065353, -0.09397705644369125, -0.07642081379890442, -0.07873471081256866, 0.07749859988689423, -0.10678935796022415, -0.015011227689683437, -0.07683262974023819, -0.06368281692266464, 0.008625335060060024, 0.11357968300580978, 0.07616094499826431, 0.12236165255308151, -0.03562367334961891, -0.009489047341048717, -0.044686999171972275, 0.08360257744789124, 0.01730869710445404, 0.013971706852316856, -0.0729127898812294, 0.09206555783748627, 0.027562865987420082, -0.022792065516114235, -0.022118747234344482, -0.05011564493179321, -0.11403799802064896, 0.0007474387530237436, -0.07230600714683533, 0.03707944229245186, -0.044822100549936295, 0.009820621460676193, 0.02886822819709778, -0.01765349693596363, 0.013824311085045338, 0.044652070850133896, -0.01898236945271492, -0.041413240134716034, -0.026295199990272522, 0.08962386846542358, -0.11560934036970139, -0.012414404191076756, 0.0022809773217886686, -0.08495457470417023, 0.07077190279960632, 0.02228407934308052, -0.045053694397211075, 0.009669739753007889, -0.1592082679271698, 0.013810086995363235, -0.009634409099817276, 0.09126240015029907, 0.0067637208849191666, -0.08613287657499313, 0.05600525811314583, 0.0405578538775444, -0.01700415089726448, 0.012183732353150845, 0.11135251075029373, -0.10732860863208771, -0.04200209677219391, 0.01903790980577469, -0.010048873722553253, -0.0683654174208641, -0.02090873382985592, 0.1011577770113945, 0.07878413051366806, 0.1601315140724182, -0.05765629932284355, 0.09059855341911316, -0.16924983263015747, -0.04424883425235748, 0.06704110652208328, 0.014047033153474331, -0.0403677336871624, -0.08153467625379562, 0.04822923615574837, -0.014816482551395893, 0.156684011220932, -0.010644150897860527, 0.09035526216030121, 0.028607718646526337, -0.005412376951426268, 0.05004942789673805, 0.056467726826667786, 0.1252620816230774, 0.032749395817518234, 0.018169699236750603, -0.001147633302025497, -0.019088109955191612, -0.02833385206758976, -0.08665604144334793, 0.05277206003665924, 0.07217884808778763, 0.0028929272666573524, 0.05970098823308945, 0.0585295744240284, 0.022957148030400276, -0.05473305657505989, 0.04166874289512634, -0.04177645593881607, 0.031668052077293396, -0.05330437049269676, 0.030936282128095627, 0.11086554825305939, -0.09851198643445969, 0.10417468845844269, 0.04388988018035889, -0.08361519873142242, -0.11755797266960144, -0.1959228366613388, -0.058607012033462524, -0.06164124235510826, -0.00793041754513979, -0.14971981942653656, 0.03239493817090988, 0.02468680776655674, 0.030236756429076195, -0.01650729961693287, 0.02879367209970951, 0.03597574308514595, -0.05494645982980728, 0.009652252309024334, -0.013384037651121616, 0.04743177816271782, 0.06195243448019028, 0.029018133878707886, 0.08591712266206741, 0.05816486477851868, 0.034173645079135895, 0.05000169947743416, 0.09236400574445724, 0.01564439758658409, -0.02632337436079979, -0.07658162713050842, 0.01973372884094715, -0.010552431456744671, 0.015115015208721161, 0.130109965801239, 0.08645438402891159, -0.036448486149311066, -0.004318690858781338, 0.17335094511508942, -0.06728760898113251, -0.11852625757455826, -0.16631029546260834, 0.23736071586608887, 0.03406469523906708, 0.07335697114467621, 0.01075048092752695, -0.0871499553322792, -0.04426463693380356, 0.16756469011306763, 0.11053695529699326, -0.0447482094168663, 0.00004068798807566054, 0.05820777267217636, -0.011789142154157162, 0.047675687819719315, 0.07236865907907486, -0.010301457718014717, 0.285607248544693, -0.015102813020348549, 0.08420159667730331, -0.027701037004590034, -0.025477083399891853, -0.010551589541137218, 0.09621260315179825, -0.06445971131324768, 0.0005277849268168211, -0.04907277226448059, 0.10025610029697418, -0.05446722358465195, -0.027208935469388962, 0.1188337504863739, -0.030321670696139336, -0.09363868832588196, 0.003559492528438568, 0.10779121518135071, -0.02672160416841507, 0.04499704763293266, -0.012052743695676327, -0.008704240433871746, 0.20586338639259338, -0.0336524173617363, -0.05189778655767441, -0.08592905104160309, -0.02672574482858181, -0.08246854692697525, 0.09190065413713455, 0.006385164801031351, 0.14535409212112427, 0.09763486683368683, 0.04096727445721626, -0.09245656430721283, 0.08173714578151703, -0.026385648176074028, -0.09223491698503494, 0.0661483034491539, 0.10716433823108673, -0.025122305378317833, 0.16293202340602875, 0.06686834245920181, -0.10190567374229431, 0.03913125395774841, 0.006544200703501701, -0.04689357057213783, -0.073078453540802, 0.0356152281165123, -0.07678929716348648, 0.09748276323080063, 0.14592063426971436, -0.03842494636774063, 0.0007231294875964522, -0.006602349691092968, 0.0662732943892479, -0.027565699070692062, 0.05327728018164635, -0.030983202159404755, -0.16287203133106232, 0.022207386791706085, 0.028720468282699585, 0.04467839002609253, -0.14724937081336975, -0.050950564444065094, 0.02882760763168335, -0.024051189422607422, -0.025614513084292412, 0.08672451972961426, 0.03650067001581192, 0.040117595344781876, -0.04832733795046806, -0.2077750712633133, 0.03914104029536247, 0.07718996703624725, -0.09839004278182983, -0.10816533118486404 ]
null
null
transformers
# Cross-Encoder for Natural Language Inference This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. This model is based on [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) ## Training Data The model was trained on the [SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral. ## Performance - Accuracy on SNLI-test dataset: 92.20 - Accuracy on MNLI mismatched set: 90.49 For futher evaluation results, see [SBERT.net - Pretrained Cross-Encoder](https://www.sbert.net/docs/pretrained_cross-encoders.html#nli). ## Usage Pre-trained models can be used like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('cross-encoder/nli-deberta-v3-large') scores = model.predict([('A man is eating pizza', 'A man eats something'), ('A black race car starts up in front of a crowd of people.', 'A man is driving down a lonely road.')]) #Convert scores to labels label_mapping = ['contradiction', 'entailment', 'neutral'] labels = [label_mapping[score_max] for score_max in scores.argmax(axis=1)] ``` ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('cross-encoder/nli-deberta-v3-large') tokenizer = AutoTokenizer.from_pretrained('cross-encoder/nli-deberta-v3-large') features = tokenizer(['A man is eating pizza', 'A black race car starts up in front of a crowd of people.'], ['A man eats something', 'A man is driving down a lonely road.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits label_mapping = ['contradiction', 'entailment', 'neutral'] labels = [label_mapping[score_max] for score_max in scores.argmax(dim=1)] print(labels) ``` ## Zero-Shot Classification This model can also be used for zero-shot-classification: ```python from transformers import pipeline classifier = pipeline("zero-shot-classification", model='cross-encoder/nli-deberta-v3-large') sent = "Apple just announced the newest iPhone X" candidate_labels = ["technology", "sports", "politics"] res = classifier(sent, candidate_labels) print(res) ```
{"language": "en", "license": "apache-2.0", "tags": ["microsoft/deberta-v3-large"], "datasets": ["multi_nli", "snli"], "metrics": ["accuracy"], "pipeline_tag": "zero-shot-classification"}
zero-shot-classification
cross-encoder/nli-deberta-v3-large
[ "transformers", "pytorch", "deberta-v2", "text-classification", "microsoft/deberta-v3-large", "zero-shot-classification", "en", "dataset:multi_nli", "dataset:snli", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-large #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# Cross-Encoder for Natural Language Inference This model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-large ## Training Data The model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral. ## Performance - Accuracy on SNLI-test dataset: 92.20 - Accuracy on MNLI mismatched set: 90.49 For futher evaluation results, see URL - Pretrained Cross-Encoder. ## Usage Pre-trained models can be used like this: ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ## Zero-Shot Classification This model can also be used for zero-shot-classification:
[ "# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-large", "## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.", "## Performance\r\n- Accuracy on SNLI-test dataset: 92.20\r\n- Accuracy on MNLI mismatched set: 90.49\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):", "## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ "TAGS\n#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-large #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-large", "## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.", "## Performance\r\n- Accuracy on SNLI-test dataset: 92.20\r\n- Accuracy on MNLI mismatched set: 90.49\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):", "## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ 84, 46, 49, 51, 14, 33, 21 ]
[ "passage: TAGS\n#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-large #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-large## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.## Performance\r\n- Accuracy on SNLI-test dataset: 92.20\r\n- Accuracy on MNLI mismatched set: 90.49\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ -0.057665105909109116, 0.18749360740184784, -0.005360424518585205, 0.11260697990655899, 0.124428391456604, 0.03597496077418327, 0.21414512395858765, 0.0684213638305664, 0.05462667718529701, 0.013975394889712334, -0.021411990746855736, 0.12714605033397675, 0.05493280291557312, 0.08521971851587296, -0.03299973905086517, -0.17481763660907745, 0.030826937407255173, -0.06417763978242874, 0.033448584377765656, 0.07868829369544983, 0.10413143038749695, -0.05694631114602089, 0.04402574896812439, -0.02547227218747139, -0.11947084218263626, 0.0278790220618248, -0.009689955040812492, 0.00522234570235014, 0.051854293793439865, 0.09441712498664856, 0.12076503783464432, 0.03237863630056381, 0.06096956133842468, -0.19429568946361542, 0.008188631385564804, 0.06436579674482346, -0.029312923550605774, 0.02400778792798519, 0.06048731133341789, -0.07863526046276093, 0.013365150429308414, -0.0184487272053957, 0.05757369101047516, 0.0826721042394638, -0.0772474929690361, -0.07403121143579483, -0.0579853430390358, 0.022782428190112114, 0.07859579473733902, 0.09012099355459213, -0.03662097081542015, 0.10435523092746735, -0.059308234602212906, 0.09157734364271164, 0.03107433207333088, -0.21792584657669067, -0.027574265375733376, 0.10919419676065445, -0.010593520477414131, 0.06392355263233185, -0.09540204703807831, -0.03552809730172157, 0.03054763562977314, 0.003580765100196004, -0.021235376596450806, -0.032983433455228806, 0.037567514926195145, -0.025145642459392548, -0.13379289209842682, 0.004098758567124605, 0.15568944811820984, -0.005732879042625427, -0.07971493899822235, -0.10764528810977936, -0.04420115798711777, 0.011245384812355042, 0.0212345402687788, -0.013802890665829182, 0.049024973064661026, 0.03860615938901901, 0.042909663170576096, -0.027878105640411377, -0.08155112713575363, -0.041080545634031296, -0.06712549179792404, 0.11067065596580505, 0.023882165551185608, 0.019579702988266945, -0.07150612026453018, 0.08721789717674255, 0.013381178490817547, -0.12713360786437988, -0.07692398875951767, -0.06587696075439453, -0.20380225777626038, -0.01908659003674984, -0.11178566515445709, -0.1266910284757614, 0.04743029549717903, 0.22238828241825104, 0.016788771376013756, 0.05589614436030388, -0.01910209096968174, 0.02586151286959648, 0.11089242994785309, 0.11586755514144897, -0.01817133091390133, -0.03449416905641556, -0.013699338771402836, 0.03577441722154617, 0.0001439151237718761, -0.03508862853050232, -0.07775447517633438, 0.021693643182516098, -0.023777026683092117, 0.05819425731897354, 0.012987148016691208, 0.025885554030537605, -0.07576637715101242, -0.0728289857506752, 0.04582131281495094, -0.1463281810283661, 0.02260354720056057, 0.030153706669807434, -0.021976225078105927, 0.13877995312213898, 0.03682136535644531, -0.010269691236317158, -0.0543697252869606, -0.006753763649612665, -0.08732923120260239, 0.0291579682379961, -0.10947098582983017, -0.12894666194915771, 0.017647719010710716, 0.01578080840408802, -0.03457295149564743, -0.10102622210979462, -0.10997718572616577, -0.0799730196595192, 0.04159274324774742, -0.042273376137018204, 0.03651728481054306, -0.10499851405620575, -0.010253130458295345, 0.018152723088860512, 0.018452053889632225, -0.07126293331384659, -0.0051747336983680725, 0.003612604457885027, 0.013888021931052208, 0.04588911309838295, -0.01944245956838131, -0.00347140827216208, -0.11644599586725235, -0.03008832223713398, 0.018221072852611542, 0.17437005043029785, -0.08743807673454285, 0.022183049470186234, -0.06788342446088791, -0.012129595503211021, 0.04933780059218407, 0.031438734382390976, 0.006341559812426567, 0.11363165825605392, -0.27857905626296997, -0.02079509198665619, 0.1852819323539734, -0.17586204409599304, -0.058162715286016464, 0.08968862146139145, -0.04986872151494026, 0.003432365832850337, 0.09094945341348648, 0.04772567376494408, 0.1907980591058731, -0.0633150264620781, -0.04464982450008392, 0.03979748114943504, -0.028503358364105225, 0.09688862413167953, 0.07944066822528839, 0.029896317049860954, -0.03609561547636986, 0.022328805178403854, -0.08807741850614548, -0.0077741811983287334, -0.02598041668534279, -0.07691919803619385, -0.019325541332364082, -0.05756010860204697, 0.016163425520062447, -0.0025257959496229887, -0.0017187452176585793, -0.002616995247080922, -0.09822903573513031, 0.15441814064979553, 0.10775012522935867, -0.08834459632635117, 0.013333420269191265, -0.11338664591312408, -0.0627516582608223, 0.0017005131812766194, 0.007857901975512505, -0.1924450397491455, -0.10430307686328888, 0.023818427696824074, -0.11876226216554642, 0.06154824048280716, 0.0645974799990654, -0.00046903264592401683, 0.050695259124040604, -0.008530653081834316, -0.005592285189777613, -0.011149031110107899, -0.002284345217049122, -0.04571639373898506, -0.1634732484817505, -0.046855852007865906, -0.033279091119766235, 0.07354672253131866, -0.12968425452709198, 0.038164619356393814, 0.020890727639198303, 0.07344438880681992, -0.008698543533682823, -0.0475444421172142, -0.005303744226694107, 0.04697900637984276, -0.05418785661458969, -0.031500641256570816, 0.02776838093996048, 0.022325146943330765, -0.07177898287773132, -0.01577751711010933, -0.09356905519962311, -0.08638206869363785, 0.0749434307217598, 0.024311508983373642, -0.06713926792144775, 0.00669783353805542, -0.013919364660978317, -0.0008909713942557573, -0.052165985107421875, -0.037982694804668427, 0.1494857668876648, 0.03555867448449135, 0.08326481282711029, -0.05364582687616348, -0.04386705160140991, -0.021648945286870003, -0.03292682021856308, -0.0075813839212059975, 0.08220168948173523, -0.039504263550043106, -0.21701526641845703, 0.09629640728235245, 0.001731791766360402, -0.07103391736745834, 0.10204152762889862, 0.007804511114954948, -0.03915666416287422, -0.09736821055412292, 0.003968504257500172, 0.019769337028265, 0.03015170246362686, -0.052374277263879776, 0.042239103466272354, 0.02941642515361309, 0.024998938664793968, 0.012696952559053898, -0.08994367718696594, 0.024478083476424217, 0.03306983411312103, -0.021085407584905624, -0.018499024212360382, -0.0017380345379933715, -0.005166277289390564, 0.055426955223083496, -0.028111081570386887, 0.05027680844068527, 0.0182187519967556, -0.07955874502658844, -0.14706432819366455, 0.21186062693595886, -0.1267666071653366, -0.29290419816970825, -0.13089533150196075, 0.046468522399663925, -0.0774347186088562, 0.0051574790850281715, 0.03937140479683876, -0.05986529216170311, -0.05219099670648575, -0.07977934181690216, -0.03320649638772011, -0.05274564027786255, -0.027074161916971207, -0.036772120743989944, 0.06054796278476715, 0.03860192373394966, -0.15082554519176483, 0.02644861675798893, 0.016771407797932625, -0.0918116644024849, 0.025218816474080086, 0.04909722879528999, 0.034891873598098755, 0.12289142608642578, 0.023938274011015892, 0.015967542305588722, 0.010422807186841965, 0.22915177047252655, -0.01910499483346939, -0.007999210618436337, 0.13789662718772888, -0.042449429631233215, 0.056360840797424316, 0.08899103105068207, 0.04169352725148201, -0.09107819199562073, 0.034382421523332596, 0.05126716569066048, -0.03418070077896118, -0.22638624906539917, -0.0401419922709465, -0.052265699952840805, -0.05004623159766197, 0.05254204198718071, 0.012379410676658154, -0.011889515444636345, 0.02845088578760624, -0.026638058945536613, -0.017692891880869865, 0.040743667632341385, 0.08135311305522919, 0.1022188737988472, 0.01662721112370491, 0.10964755713939667, -0.05655030161142349, 0.012636630795896053, 0.0697571188211441, 0.012473252601921558, 0.12328147143125534, 0.03019139915704727, 0.05531429126858711, 0.13001295924186707, -0.00199308292940259, 0.0321027971804142, 0.025434719398617744, -0.022264394909143448, 0.03257638216018677, 0.019793175160884857, -0.0775173008441925, -0.038814958184957504, 0.06517823785543442, -0.029327796772122383, -0.0013128501595929265, -0.041973937302827835, -0.04573383927345276, 0.08372483402490616, 0.06903909891843796, 0.07332003116607666, -0.19415488839149475, -0.08053550869226456, 0.04872570186853409, -0.025881577283143997, -0.02876596711575985, 0.005629075225442648, 0.06838619709014893, -0.08527381718158722, 0.08008579909801483, -0.01814793050289154, 0.09516307711601257, -0.12457230687141418, -0.016709117218852043, 0.015517441555857658, 0.07965568453073502, 0.02490548975765705, 0.09043360501527786, -0.2550288140773773, 0.04072898253798485, 0.00717624556273222, 0.047693051397800446, -0.008393471129238605, 0.05171489343047142, 0.02945207804441452, 0.02867080457508564, 0.10762064903974533, 0.02067841775715351, -0.03271696716547012, -0.09061522781848907, -0.05023936927318573, 0.018165867775678635, 0.09380891174077988, -0.07836149632930756, 0.10808750987052917, -0.007234653457999229, -0.0056988815777003765, -0.01972959190607071, 0.041364751756191254, -0.21893079578876495, -0.19074194133281708, 0.07411524653434753, -0.011108919978141785, 0.00011937936506001279, -0.02815171703696251, -0.06802035868167877, -0.027272144332528114, 0.20286639034748077, -0.2734784185886383, -0.08458000421524048, -0.13907702267169952, 0.0876479223370552, 0.0947323814034462, -0.061831433326005936, -0.008452769368886948, 0.004895468708127737, 0.13755019009113312, -0.03773697838187218, -0.13174059987068176, 0.011345510371029377, -0.06858566403388977, -0.10028401762247086, -0.0618094727396965, 0.06277436017990112, 0.12373785674571991, 0.02303234487771988, 0.0250016450881958, 0.020958080887794495, 0.0052667343989014626, -0.1396549791097641, -0.06933476775884628, 0.1692645400762558, 0.011821463704109192, 0.07224257290363312, -0.06350373476743698, -0.1261928379535675, -0.060596950352191925, 0.02175183594226837, 0.10319159924983978, 0.16373462975025177, -0.06260588020086288, 0.06989205628633499, 0.08469809591770172, -0.14647924900054932, -0.15073134005069733, -0.03442138060927391, 0.09728588908910751, 0.03737456351518631, 0.03593964874744415, -0.18439321219921112, 0.06819945573806763, 0.08834658563137054, 0.0033898521214723587, -0.04533231258392334, -0.3093365728855133, -0.10703980922698975, 0.04386529698967934, 0.07531942427158356, 0.03564247861504555, -0.10586807876825333, -0.05846872180700302, -0.06543514877557755, -0.04251181706786156, 0.25640934705734253, 0.0031355072278529406, 0.09846021980047226, 0.020927825942635536, -0.07630974799394608, 0.03578653931617737, -0.02363801933825016, 0.11309444159269333, 0.08084964752197266, 0.09372159093618393, -0.02151012234389782, 0.007526206783950329, 0.0659996047616005, -0.03831157833337784, 0.09619487076997757, -0.017913738265633583, 0.08665180951356888, -0.08355838060379028, -0.057030245661735535, -0.012653544545173645, -0.00773481372743845, -0.05056479945778847, -0.03332303836941719, -0.0456065833568573, 0.07295481115579605, 0.08371299505233765, -0.018338747322559357, 0.05189519375562668, -0.07958797365427017, -0.0008753683650866151, 0.1818644106388092, 0.13421708345413208, 0.08429650217294693, -0.13534115254878998, -0.02595195733010769, -0.007799254730343819, 0.07737915962934494, -0.13849106431007385, 0.0631863996386528, 0.09942299872636795, 0.009675749577581882, 0.134225994348526, 0.07184412330389023, -0.05271662026643753, -0.009729314595460892, 0.06637807935476303, -0.07509785890579224, -0.08206158131361008, -0.028724312782287598, -0.03602799028158188, -0.059088852256536484, 0.0010326291667297482, 0.16228578984737396, -0.03392717242240906, -0.01251340750604868, 0.004647531546652317, 0.03306873142719269, -0.02317374013364315, 0.1279098093509674, 0.011217324994504452, 0.059141725301742554, -0.09118495881557465, 0.10162673890590668, 0.07630746811628342, -0.04885835200548172, 0.029999015852808952, 0.09039470553398132, -0.09347172826528549, -0.08237362653017044, -0.07664462178945541, 0.08065587282180786, -0.11686337739229202, -0.012881292961537838, -0.07209533452987671, -0.07654780894517899, 0.013541803695261478, 0.13123466074466705, 0.08874864131212234, 0.12391982972621918, -0.03176678717136383, -0.01613692007958889, -0.05146809294819832, 0.0788135975599289, 0.006876029074192047, 0.014342489652335644, -0.08744140714406967, 0.09091003984212875, 0.031344179064035416, -0.027464235201478004, -0.025579849258065224, -0.04624859616160393, -0.13640174269676208, -0.009153662249445915, -0.057192154228687286, 0.04208645969629288, -0.04431177303195, 0.018927663564682007, 0.02675640769302845, -0.019677074626088142, 0.009271005168557167, 0.03907862678170204, -0.016868602484464645, -0.04957958310842514, -0.01973636820912361, 0.09832814335823059, -0.11968693137168884, -0.016389477998018265, 0.006816430017352104, -0.08819790184497833, 0.06730537116527557, 0.009467188268899918, -0.03805196285247803, 0.01764019951224327, -0.16711702942848206, 0.011368012987077236, -0.0017038426594808698, 0.08446966111660004, 0.01649138517677784, -0.11123771220445633, 0.05102556571364403, 0.030675696209073067, 0.0002710152475629002, 0.02728324383497238, 0.12810318171977997, -0.11076090484857559, -0.04538813605904579, 0.004113869741559029, -0.011708800680935383, -0.06045517325401306, -0.017462842166423798, 0.11436296254396439, 0.07116929441690445, 0.17041409015655518, -0.05567152425646782, 0.08613038063049316, -0.1765887290239334, -0.039721257984638214, 0.05763218551874161, 0.001283330493606627, -0.03540359064936638, -0.0763227716088295, 0.051812853664159775, -0.020768260583281517, 0.15189769864082336, -0.01785372383892536, 0.09527633339166641, 0.038822516798973083, 0.00433873338624835, 0.033376555889844894, 0.060359273105859756, 0.13976849615573883, 0.04243898391723633, 0.017001481726765633, 0.0028528545517474413, -0.006474299356341362, -0.02161298505961895, -0.0928155854344368, 0.07097221165895462, 0.06980274617671967, -0.03997572883963585, 0.05717054009437561, 0.06018413230776787, 0.006740938872098923, -0.04455997422337532, 0.0706065371632576, -0.039364349097013474, 0.02398986741900444, -0.051083970814943314, 0.019606031477451324, 0.10718338936567307, -0.10580746084451675, 0.10925860702991486, 0.03248471021652222, -0.0947871059179306, -0.11583490669727325, -0.18595130741596222, -0.06794553995132446, -0.04599566012620926, -0.011483184061944485, -0.16359439492225647, 0.033263932913541794, 0.03364426642656326, 0.022364115342497826, -0.010881992056965828, 0.031191887333989143, -0.0006552398554049432, -0.04614849016070366, 0.020301055163145065, -0.008696205914020538, 0.027063438668847084, 0.04934412240982056, 0.02049436792731285, 0.08580917865037918, 0.06293024867773056, 0.04137958213686943, 0.04419529810547829, 0.0972491055727005, 0.009964424185454845, -0.0365554504096508, -0.07838965952396393, 0.022900940850377083, 0.0012353123165667057, 0.008685890585184097, 0.10989752411842346, 0.08343306183815002, -0.04763570800423622, -0.003344073425978422, 0.17858655750751495, -0.06960365921258926, -0.109405018389225, -0.1677924394607544, 0.23614457249641418, 0.03791625425219536, 0.07592961937189102, 0.0038883064407855272, -0.09718609601259232, -0.03092035837471485, 0.16161982715129852, 0.11521202325820923, -0.04700302705168724, -0.004755087662488222, 0.04495500773191452, -0.009700769558548927, 0.05009165778756142, 0.0812082290649414, -0.011078368872404099, 0.24896250665187836, -0.015392433851957321, 0.07938766479492188, -0.03243725746870041, -0.03668249398469925, 0.0007912452565506101, 0.09442874044179916, -0.04320375993847847, 0.0045418692752718925, -0.04999171569943428, 0.11236628144979477, -0.054178789258003235, -0.03703644871711731, 0.12046990543603897, -0.03526090085506439, -0.10279686003923416, -0.008671681396663189, 0.07478763163089752, -0.024839574471116066, 0.04309184104204178, -0.016671204939484596, -0.009069737046957016, 0.18809691071510315, -0.033953920006752014, -0.06159922853112221, -0.09707266837358475, -0.023511532694101334, -0.06726019829511642, 0.08833751082420349, 0.004294530022889376, 0.1456068754196167, 0.10218770056962967, 0.034757621586322784, -0.08349652588367462, 0.07914342731237411, -0.02416873723268509, -0.0700109675526619, 0.07606734335422516, 0.12370195984840393, -0.020728524774312973, 0.17198072373867035, 0.06389904022216797, -0.1258181631565094, 0.03290615603327751, 0.002535546664148569, -0.0544767901301384, -0.07193843275308609, 0.041357822716236115, -0.08097375184297562, 0.09928510338068008, 0.16572722792625427, -0.03551780804991722, 0.0014993146760389209, -0.013682534918189049, 0.06789954006671906, -0.023382846266031265, 0.05415410175919533, -0.04138858616352081, -0.16572260856628418, 0.030495984479784966, 0.014189633540809155, 0.03612922504544258, -0.14817483723163605, -0.047600679099559784, 0.01902780868113041, -0.01818658970296383, -0.027880078181624413, 0.09874292463064194, 0.028974974527955055, 0.042395126074552536, -0.052695002406835556, -0.22470088303089142, 0.0348617359995842, 0.08085587620735168, -0.08967912197113037, -0.10403793305158615 ]
null
null
transformers
# Cross-Encoder for Natural Language Inference This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. This model is based on [microsoft/deberta-v3-small](https://huggingface.co/microsoft/deberta-v3-small) ## Training Data The model was trained on the [SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral. ## Performance - Accuracy on SNLI-test dataset: 91.65 - Accuracy on MNLI mismatched set: 87.55 For futher evaluation results, see [SBERT.net - Pretrained Cross-Encoder](https://www.sbert.net/docs/pretrained_cross-encoders.html#nli). ## Usage Pre-trained models can be used like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('cross-encoder/nli-deberta-v3-small') scores = model.predict([('A man is eating pizza', 'A man eats something'), ('A black race car starts up in front of a crowd of people.', 'A man is driving down a lonely road.')]) #Convert scores to labels label_mapping = ['contradiction', 'entailment', 'neutral'] labels = [label_mapping[score_max] for score_max in scores.argmax(axis=1)] ``` ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('cross-encoder/nli-deberta-v3-small') tokenizer = AutoTokenizer.from_pretrained('cross-encoder/nli-deberta-v3-small') features = tokenizer(['A man is eating pizza', 'A black race car starts up in front of a crowd of people.'], ['A man eats something', 'A man is driving down a lonely road.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits label_mapping = ['contradiction', 'entailment', 'neutral'] labels = [label_mapping[score_max] for score_max in scores.argmax(dim=1)] print(labels) ``` ## Zero-Shot Classification This model can also be used for zero-shot-classification: ```python from transformers import pipeline classifier = pipeline("zero-shot-classification", model='cross-encoder/nli-deberta-v3-small') sent = "Apple just announced the newest iPhone X" candidate_labels = ["technology", "sports", "politics"] res = classifier(sent, candidate_labels) print(res) ```
{"language": "en", "license": "apache-2.0", "tags": ["microsoft/deberta-v3-small"], "datasets": ["multi_nli", "snli"], "metrics": ["accuracy"], "pipeline_tag": "zero-shot-classification"}
zero-shot-classification
cross-encoder/nli-deberta-v3-small
[ "transformers", "pytorch", "deberta-v2", "text-classification", "microsoft/deberta-v3-small", "zero-shot-classification", "en", "dataset:multi_nli", "dataset:snli", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-small #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
# Cross-Encoder for Natural Language Inference This model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-small ## Training Data The model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral. ## Performance - Accuracy on SNLI-test dataset: 91.65 - Accuracy on MNLI mismatched set: 87.55 For futher evaluation results, see URL - Pretrained Cross-Encoder. ## Usage Pre-trained models can be used like this: ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ## Zero-Shot Classification This model can also be used for zero-shot-classification:
[ "# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-small", "## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.", "## Performance\r\n- Accuracy on SNLI-test dataset: 91.65\r\n- Accuracy on MNLI mismatched set: 87.55\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):", "## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ "TAGS\n#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-small #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-small", "## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.", "## Performance\r\n- Accuracy on SNLI-test dataset: 91.65\r\n- Accuracy on MNLI mismatched set: 87.55\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):", "## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ 88, 46, 49, 51, 14, 33, 21 ]
[ "passage: TAGS\n#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-small #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-small## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.## Performance\r\n- Accuracy on SNLI-test dataset: 91.65\r\n- Accuracy on MNLI mismatched set: 87.55\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ -0.056961070746183395, 0.1961870789527893, -0.005332413595169783, 0.09397640079259872, 0.12969684600830078, 0.030759280547499657, 0.20943328738212585, 0.0907621681690216, 0.017969153821468353, 0.015267309732735157, -0.017602231353521347, 0.10468098521232605, 0.06298787891864777, 0.08545204997062683, -0.03651503100991249, -0.17780393362045288, 0.04796379432082176, -0.03407283499836922, 0.03159286826848984, 0.09035217016935349, 0.10395898669958115, -0.04964560642838478, 0.03876478970050812, -0.011030765250325203, -0.1415451467037201, 0.03361905366182327, 0.0051561384461820126, -0.0020623458549380302, 0.053033627569675446, 0.09050386399030685, 0.13333773612976074, 0.02632060833275318, 0.05859701335430145, -0.17763419449329376, 0.009237413294613361, 0.07520344108343124, -0.013635381124913692, 0.03653927147388458, 0.0638813003897667, -0.052451565861701965, 0.004688677377998829, -0.02969725802540779, 0.05727693811058998, 0.0856747031211853, -0.07996667921543121, -0.12344317138195038, -0.06564246118068695, 0.02343943901360035, 0.07945317775011063, 0.09610721468925476, -0.03643164783716202, 0.10329966992139816, -0.07307448983192444, 0.09137709438800812, 0.07578229159116745, -0.22034852206707, -0.02919360250234604, 0.12271309643983841, -0.01669551245868206, 0.0544452928006649, -0.09719231724739075, -0.03607233986258507, 0.019358912482857704, 0.0038723223842680454, 0.01262687798589468, -0.025430399924516678, -0.006691235117614269, -0.026973122730851173, -0.11965502053499222, -0.00014310423284769058, 0.13069888949394226, -0.0028886974323540926, -0.07842600345611572, -0.10936353355646133, -0.05831489711999893, -0.022318540140986443, 0.016667451709508896, -0.016764741390943527, 0.04643956199288368, 0.03739266097545624, 0.002302109729498625, -0.032362136989831924, -0.07539698481559753, -0.04761741682887077, -0.06762035936117172, 0.09118301421403885, 0.021787207573652267, 0.028133036568760872, -0.07599619776010513, 0.09970089793205261, 0.005289139691740274, -0.1182696744799614, -0.0789690762758255, -0.054308827966451645, -0.21623556315898895, -0.022545037791132927, -0.10969150811433792, -0.11233719438314438, 0.032436005771160126, 0.20078672468662262, 0.00407967297360301, 0.063546322286129, -0.0010119822109118104, 0.023612026125192642, 0.06472960859537125, 0.1400364637374878, -0.033656924962997437, -0.010959330014884472, -0.024638643488287926, 0.030695592984557152, -0.0041329036466777325, -0.024394407868385315, -0.06526073068380356, -0.024274786934256554, 0.007041584700345993, 0.05273788794875145, -0.000617657438851893, 0.023635249584913254, -0.06414564698934555, -0.04882162809371948, 0.0041771549731493, -0.14638164639472961, 0.034152802079916, 0.027090927585959435, -0.04794420674443245, 0.13472896814346313, 0.051494523882865906, -0.021505294367671013, -0.07091271877288818, 0.0069885896518826485, -0.08569799363613129, 0.04107192903757095, -0.11503090709447861, -0.13300921022891998, 0.027024706825613976, 0.00036001738044433296, -0.04776831343770027, -0.0877830907702446, -0.11901956051588058, -0.068480484187603, 0.036347195506095886, -0.041209787130355835, 0.04326841980218887, -0.08692765235900879, -0.021107451990246773, 0.02897862158715725, 0.02264774963259697, -0.06347545981407166, -0.005577965173870325, 0.0141081428155303, -0.0005748371477238834, 0.05079640820622444, -0.017347725108265877, 0.014431588351726532, -0.12141559273004532, -0.023558735847473145, -0.014693436212837696, 0.17323268949985504, -0.08072654157876968, -0.00003345928780618124, -0.08439966291189194, -0.03274482861161232, 0.0396089106798172, 0.01695961505174637, 0.023262349888682365, 0.11895835399627686, -0.23950646817684174, -0.021843325346708298, 0.18407565355300903, -0.15927079319953918, -0.03525390475988388, 0.0889754593372345, -0.05120646953582764, -0.009058183059096336, 0.09082705527544022, 0.060031957924366, 0.19853036105632782, -0.0685625895857811, -0.046013642102479935, 0.05430838465690613, -0.025606682524085045, 0.06836490333080292, 0.07663080096244812, 0.03238728269934654, -0.03465449437499046, 0.01956859417259693, -0.08739784359931946, 0.0023523413110524416, -0.03580824285745621, -0.06834175437688828, -0.03907546401023865, -0.0403694212436676, 0.018120774999260902, 0.008914916776120663, 0.011326510459184647, -0.009339330717921257, -0.10252520442008972, 0.1758558452129364, 0.10552716255187988, -0.0841434970498085, 0.019238483160734177, -0.11075163632631302, -0.06033647805452347, -0.0031039307359606028, 0.004072209820151329, -0.1965838223695755, -0.11076483875513077, 0.024568570777773857, -0.1334027647972107, 0.04687068611383438, 0.051188234239816666, 0.012907381169497967, 0.04356514289975166, -0.02970122918486595, -0.004830125253647566, -0.024140704423189163, -0.010315688326954842, -0.049343518912792206, -0.13780340552330017, -0.06606730818748474, -0.023271184414625168, 0.09432083368301392, -0.1415734440088272, 0.04146553575992584, 0.026056276634335518, 0.0861760675907135, -0.004940844606608152, -0.06198602914810181, 0.0006747650913894176, 0.03693617880344391, -0.043233055621385574, -0.044995155185461044, 0.02960134483873844, 0.04101593792438507, -0.08540904521942139, 0.0017956276424229145, -0.11070018261671066, -0.05321377515792847, 0.07073202729225159, 0.028861263766884804, -0.08048144727945328, -0.0030741558875888586, -0.012171059846878052, -0.004922389984130859, -0.07879284769296646, -0.0442563071846962, 0.10113686323165894, 0.029502592980861664, 0.10207447409629822, -0.0669952854514122, -0.043319184333086014, -0.031170563772320747, -0.03509499877691269, 0.0033279391936957836, 0.09502648562192917, -0.048851050436496735, -0.21788381040096283, 0.10462065786123276, 0.000007496769285353366, -0.08834182471036911, 0.11559999734163284, -0.0033718696795403957, -0.03926501050591469, -0.09929623454809189, 0.0036536515690386295, 0.015375284478068352, 0.03322527930140495, -0.03470364585518837, 0.055137235671281815, 0.033362049609422684, 0.026268688961863518, -0.00538782449439168, -0.11924227327108383, 0.022927885875105858, 0.04354250803589821, -0.0400046780705452, -0.016311755403876305, 0.0017426759004592896, 0.01467554084956646, 0.07017359882593155, -0.02757679671049118, 0.04390476271510124, 0.022832009941339493, -0.0770452469587326, -0.14205944538116455, 0.22206856310367584, -0.1199897825717926, -0.2605627179145813, -0.1535775065422058, 0.0587504543364048, -0.06399208307266235, -0.0035633051302284002, 0.03243519738316536, -0.06811096519231796, -0.044749531894922256, -0.08056598156690598, -0.04170610383152962, -0.06297550350427628, -0.026499466970562935, -0.04852226376533508, 0.0514201857149601, 0.060543373227119446, -0.15224848687648773, 0.020836113020777702, 0.014202283695340157, -0.09774819761514664, 0.022519497200846672, 0.025742901489138603, 0.020679831504821777, 0.13922634720802307, -0.0005475696525536478, 0.020450416952371597, -0.0017292104894295335, 0.2370370626449585, -0.021612582728266716, 0.003287105355411768, 0.1335442066192627, -0.019259387627243996, 0.06566687673330307, 0.08157089352607727, 0.043811701238155365, -0.06757771968841553, 0.03084348328411579, 0.05638407915830612, -0.04038724675774574, -0.23690764605998993, -0.02760658785700798, -0.04851305112242699, -0.05192003399133682, 0.052688419818878174, 0.014444472268223763, 0.006107872352004051, 0.04240432754158974, -0.038990601897239685, -0.010389072820544243, 0.06150193512439728, 0.08314985036849976, 0.11515584588050842, 0.036032963544130325, 0.11449538171291351, -0.06641249358654022, 0.02337203361093998, 0.05747813358902931, -0.0010139275109395385, 0.14964117109775543, 0.013843987137079239, 0.06882432848215103, 0.10457455366849899, 0.019567999988794327, 0.037688400596380234, 0.023961113765835762, 0.0010024991352111101, 0.028486931696534157, 0.02710648439824581, -0.08003382384777069, -0.023219913244247437, 0.03872571885585785, -0.039589010179042816, 0.02129344455897808, -0.054115258157253265, -0.03512758016586304, 0.06833846867084503, 0.12211459130048752, 0.08409421890974045, -0.2301958054304123, -0.0886494368314743, 0.030209293588995934, -0.036933816969394684, -0.030981486663222313, -0.0060872784815728664, 0.08728189021348953, -0.09017068147659302, 0.07668162137269974, -0.03618237376213074, 0.10074330121278763, -0.10408969223499298, -0.014148456044495106, 0.03471336513757706, 0.07646972686052322, 0.014898359775543213, 0.08425787836313248, -0.2505313456058502, 0.07498232275247574, 0.008051376789808273, 0.06272706389427185, 0.0032530322205275297, 0.033703722059726715, 0.03904540091753006, 0.04015763849020004, 0.09559804201126099, 0.02999647706747055, -0.03591477498412132, -0.10999628156423569, -0.0336206778883934, 0.002025442197918892, 0.09012547880411148, -0.0427987277507782, 0.10987574607133865, -0.022830365225672722, -0.0014936849474906921, -0.0023441307712346315, 0.04861946776509285, -0.2045290768146515, -0.1693422794342041, 0.056700605899095535, 0.029933782294392586, 0.006762049160897732, -0.035537052899599075, -0.0720948651432991, 0.005997186526656151, 0.1846836507320404, -0.23794373869895935, -0.08160409331321716, -0.14408522844314575, 0.09477245807647705, 0.09690804034471512, -0.06397052854299545, 0.00017520775145385414, -0.013649235479533672, 0.1489323377609253, -0.02142726257443428, -0.1405046433210373, 0.013837188482284546, -0.06044485419988632, -0.10533513873815536, -0.054327525198459625, 0.0592300184071064, 0.11367049813270569, 0.033552054315805435, 0.02613644115626812, 0.014130808413028717, 0.014584802091121674, -0.1273113191127777, -0.05459503456950188, 0.15321394801139832, 0.002254319842904806, 0.10390106588602066, -0.05416392907500267, -0.11052535474300385, -0.04752272367477417, 0.025736380368471146, 0.10757150501012802, 0.16233095526695251, -0.0800897628068924, 0.04041724652051926, 0.07141215354204178, -0.1516539603471756, -0.15825630724430084, -0.00425605708733201, 0.09616869688034058, 0.037083376199007034, 0.03072933293879032, -0.18486237525939941, 0.07317405939102173, 0.10363899916410446, -0.004574956838041544, -0.07986216247081757, -0.34693461656570435, -0.10520916432142258, 0.04198843240737915, 0.08313769847154617, 0.060070645064115524, -0.10958565771579742, -0.05167347937822342, -0.06506418436765671, -0.08158134669065475, 0.2552858293056488, -0.011383550241589546, 0.1138865202665329, 0.004456536844372749, -0.06698188185691833, 0.03861719369888306, -0.028944261372089386, 0.11940091103315353, 0.06652441620826721, 0.10273183882236481, -0.02294248528778553, 0.022330382838845253, 0.07942374795675278, -0.05146781727671623, 0.10527573525905609, -0.004938979633152485, 0.07758183032274246, -0.10695798695087433, -0.03981383144855499, -0.033583950251340866, 0.0010570435551926494, -0.04799271374940872, -0.023285062983632088, -0.06632781028747559, 0.07830964773893356, 0.09005264937877655, -0.02042083442211151, 0.05805366858839989, -0.0709240511059761, 0.04380493611097336, 0.1671580672264099, 0.11670037358999252, 0.09546897560358047, -0.15661010146141052, -0.014498712494969368, -0.0010534921893849969, 0.07967856526374817, -0.132485032081604, 0.06265733391046524, 0.11045993119478226, 0.027785461395978928, 0.1395520716905594, 0.07954127341508865, -0.044443730264902115, -0.007938376627862453, 0.052884820848703384, -0.060271210968494415, -0.11546752601861954, -0.029542604461312294, -0.019196825101971626, -0.061455484479665756, -0.00043220765655860305, 0.14659729599952698, -0.05325944349169731, -0.001288223429583013, 0.01698659174144268, 0.026382988318800926, -0.02045580744743347, 0.13848403096199036, 0.0190631952136755, 0.06719701737165451, -0.08813715726137161, 0.12404534220695496, 0.08111557364463806, -0.05381942167878151, 0.03241436555981636, 0.08617100864648819, -0.09342825412750244, -0.061714768409729004, -0.05896897614002228, 0.07398626953363419, -0.12855491042137146, -0.028941912576556206, -0.08085085451602936, -0.09971784800291061, 0.028459830209612846, 0.13796161115169525, 0.08680640161037445, 0.12122590094804764, -0.03262415528297424, -0.014813854359090328, -0.06274110078811646, 0.0715794563293457, 0.030078357085585594, 0.02865665592253208, -0.10288037359714508, 0.10995189100503922, 0.030523695051670074, -0.01837310567498207, -0.024026336148381233, -0.049234695732593536, -0.14208920300006866, 0.0006098106387071311, -0.0974268987774849, 0.03894205391407013, -0.03128518909215927, 0.004905003122985363, 0.016475951299071312, -0.023368461057543755, 0.006408669985830784, 0.04357844591140747, -0.02703693136572838, -0.047953732311725616, -0.021202903240919113, 0.08244408667087555, -0.12416809797286987, -0.01840076968073845, 0.011747940443456173, -0.08701442182064056, 0.05822761356830597, 0.027990076690912247, -0.04509776458144188, 0.004477526526898146, -0.16352279484272003, -0.00909185316413641, -0.003301572985947132, 0.059549976140260696, 0.012296210043132305, -0.1166386753320694, 0.053857821971178055, 0.02099844254553318, 0.014055428095161915, 0.03278478980064392, 0.12304199486970901, -0.11628343909978867, -0.056314315646886826, -0.016660217195749283, 0.00044560778769664466, -0.07505793869495392, 0.015817973762750626, 0.13673697412014008, 0.08051448315382004, 0.16762056946754456, -0.06189186871051788, 0.07298401743173599, -0.19466251134872437, -0.03820746764540672, 0.050433285534381866, -0.01597246713936329, -0.03462592139840126, -0.062134359031915665, 0.05127105861902237, -0.014583575539290905, 0.11440566927194595, -0.021732686087489128, 0.08002149313688278, 0.05129130929708481, 0.000621872313786298, 0.011416231282055378, 0.04960817098617554, 0.14353249967098236, 0.04175664857029915, -0.0038538945373147726, 0.053394611924886703, -0.017675647512078285, -0.004553397186100483, -0.05275626853108406, 0.09570778906345367, 0.08556406199932098, -0.04584071412682533, 0.07026151567697525, 0.04271816462278366, 0.011462559923529625, -0.059924717992544174, 0.06624244153499603, -0.030110515654087067, 0.049340542405843735, -0.054501358419656754, 0.04801312834024429, 0.12171947211027145, -0.11846218258142471, 0.12039345502853394, 0.015215769410133362, -0.09585969150066376, -0.12975963950157166, -0.17491841316223145, -0.0667714923620224, -0.06470268219709396, -0.012252843007445335, -0.15192881226539612, 0.05179985985159874, 0.0718541368842125, 0.03516662120819092, -0.010590044781565666, 0.04866155609488487, -0.008129975758492947, -0.05039401724934578, 0.04115501046180725, -0.0190975870937109, 0.02007700875401497, 0.038339391350746155, 0.009153026156127453, 0.11093249917030334, 0.055144380778074265, 0.05472366511821747, 0.05135929211974144, 0.0981246754527092, 0.008300126530230045, -0.034893184900283813, -0.0797475129365921, 0.02014467678964138, 0.008098161779344082, 0.015479384921491146, 0.09278547763824463, 0.07891716808080673, -0.03576907142996788, -0.016950344666838646, 0.2205762267112732, -0.08238531649112701, -0.09898889064788818, -0.16472327709197998, 0.24246849119663239, 0.034275513142347336, 0.056518424302339554, 0.015396383590996265, -0.09635913372039795, -0.032805293798446655, 0.14195358753204346, 0.09264257550239563, -0.03852815926074982, -0.011131549254059792, 0.028302785009145737, -0.019546370953321457, 0.04360827058553696, 0.09876316040754318, 0.0009901679586619139, 0.21663962304592133, -0.032520487904548645, 0.10163549333810806, -0.026748700067400932, -0.039204709231853485, -0.01982319913804531, 0.09718060493469238, -0.039680689573287964, 0.011060677468776703, -0.06099942699074745, 0.10252456367015839, -0.0558469295501709, -0.05870893970131874, 0.11199069023132324, -0.03100733831524849, -0.10865721106529236, -0.022357655689120293, 0.05840349197387695, -0.028117772191762924, 0.0430804044008255, -0.019487028941512108, 0.016798168420791626, 0.1654404103755951, -0.02948765829205513, -0.06529358774423599, -0.12318524718284607, 0.001640239148400724, -0.07495937496423721, 0.10043440759181976, 0.013007666915655136, 0.12557460367679596, 0.10131639987230301, 0.017857175320386887, -0.09460242092609406, 0.07473164796829224, -0.016450976952910423, -0.05922538787126541, 0.05306198075413704, 0.13161322474479675, -0.014576971530914307, 0.1667606383562088, 0.04486963897943497, -0.12680228054523468, 0.02564382553100586, 0.0007472023135051131, -0.053585659712553024, -0.08150466531515121, 0.02220173180103302, -0.061192091554403305, 0.10534800589084625, 0.19582420587539673, -0.03958282992243767, 0.011014478281140327, -0.02779238112270832, 0.0411464124917984, -0.009685077704489231, 0.03688124567270279, -0.017235610634088516, -0.16092664003372192, 0.04031168669462204, 0.012002391740679741, 0.04712655395269394, -0.15618866682052612, -0.051108330488204956, 0.012334921397268772, -0.03568067401647568, -0.02219984494149685, 0.11052946746349335, 0.03266215696930885, 0.04590963199734688, -0.04853643476963043, -0.2415795475244522, 0.02950061485171318, 0.10446979850530624, -0.09330163151025772, -0.09403061866760254 ]
null
null
transformers
# Cross-Encoder for Natural Language Inference This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. This model is based on [microsoft/deberta-v3-xsmall](https://huggingface.co/microsoft/deberta-v3-xsmall) ## Training Data The model was trained on the [SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral. ## Performance - Accuracy on SNLI-test dataset: 91.64 - Accuracy on MNLI mismatched set: 87.77 For futher evaluation results, see [SBERT.net - Pretrained Cross-Encoder](https://www.sbert.net/docs/pretrained_cross-encoders.html#nli). ## Usage Pre-trained models can be used like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('cross-encoder/nli-deberta-v3-xsmall') scores = model.predict([('A man is eating pizza', 'A man eats something'), ('A black race car starts up in front of a crowd of people.', 'A man is driving down a lonely road.')]) #Convert scores to labels label_mapping = ['contradiction', 'entailment', 'neutral'] labels = [label_mapping[score_max] for score_max in scores.argmax(axis=1)] ``` ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('cross-encoder/nli-deberta-v3-xsmall') tokenizer = AutoTokenizer.from_pretrained('cross-encoder/nli-deberta-v3-xsmall') features = tokenizer(['A man is eating pizza', 'A black race car starts up in front of a crowd of people.'], ['A man eats something', 'A man is driving down a lonely road.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits label_mapping = ['contradiction', 'entailment', 'neutral'] labels = [label_mapping[score_max] for score_max in scores.argmax(dim=1)] print(labels) ``` ## Zero-Shot Classification This model can also be used for zero-shot-classification: ```python from transformers import pipeline classifier = pipeline("zero-shot-classification", model='cross-encoder/nli-deberta-v3-xsmall') sent = "Apple just announced the newest iPhone X" candidate_labels = ["technology", "sports", "politics"] res = classifier(sent, candidate_labels) print(res) ```
{"language": "en", "license": "apache-2.0", "tags": ["microsoft/deberta-v3-xsmall"], "datasets": ["multi_nli", "snli"], "metrics": ["accuracy"], "pipeline_tag": "zero-shot-classification"}
zero-shot-classification
cross-encoder/nli-deberta-v3-xsmall
[ "transformers", "pytorch", "deberta-v2", "text-classification", "microsoft/deberta-v3-xsmall", "zero-shot-classification", "en", "dataset:multi_nli", "dataset:snli", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-xsmall #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
# Cross-Encoder for Natural Language Inference This model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-xsmall ## Training Data The model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral. ## Performance - Accuracy on SNLI-test dataset: 91.64 - Accuracy on MNLI mismatched set: 87.77 For futher evaluation results, see URL - Pretrained Cross-Encoder. ## Usage Pre-trained models can be used like this: ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ## Zero-Shot Classification This model can also be used for zero-shot-classification:
[ "# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-xsmall", "## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.", "## Performance\r\n- Accuracy on SNLI-test dataset: 91.64\r\n- Accuracy on MNLI mismatched set: 87.77\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):", "## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ "TAGS\n#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-xsmall #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-xsmall", "## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.", "## Performance\r\n- Accuracy on SNLI-test dataset: 91.64\r\n- Accuracy on MNLI mismatched set: 87.77\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):", "## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ 89, 47, 49, 51, 14, 33, 21 ]
[ "passage: TAGS\n#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-xsmall #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-xsmall## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.## Performance\r\n- Accuracy on SNLI-test dataset: 91.64\r\n- Accuracy on MNLI mismatched set: 87.77\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ -0.05526004731655121, 0.20002509653568268, -0.004549776203930378, 0.09243170917034149, 0.1255296766757965, 0.04860931262373924, 0.2057001292705536, 0.10026445984840393, 0.023124130442738533, 0.010743226855993271, 0.0033932949882000685, 0.10624440014362335, 0.0735110193490982, 0.06944567710161209, -0.03670136258006096, -0.15651534497737885, 0.06273180991411209, -0.031661514192819595, 0.03551706671714783, 0.07941190898418427, 0.09823884814977646, -0.03796064481139183, 0.027161527425050735, -0.011348406784236431, -0.16056472063064575, 0.03263018652796745, 0.013877912424504757, 0.008690397255122662, 0.04654226079583168, 0.07941348105669022, 0.134487122297287, 0.022073298692703247, 0.054972339421510696, -0.15428295731544495, 0.0071214670315384865, 0.08016258478164673, -0.008101080544292927, 0.040968116372823715, 0.06426200270652771, -0.042856525629758835, -0.013149209320545197, -0.021201087161898613, 0.049097828567028046, 0.09118204563856125, -0.08311184495687485, -0.12524905800819397, -0.07852311432361603, 0.041662704199552536, 0.0712871104478836, 0.10549604147672653, -0.034251656383275986, 0.11924458295106888, -0.04591577500104904, 0.07998860627412796, 0.07225118577480316, -0.2156272679567337, -0.02652299962937832, 0.1276988387107849, -0.009356200695037842, 0.03971582651138306, -0.0813535749912262, -0.03941688686609268, 0.021172143518924713, 0.009966623038053513, -0.00391641678288579, -0.0237180944532156, 0.006062181666493416, -0.029210276901721954, -0.10915707796812057, 0.006141735706478357, 0.12676866352558136, 0.00768245430663228, -0.07872988283634186, -0.10875414311885834, -0.058593858033418655, -0.03962946683168411, 0.03087957203388214, -0.024735471233725548, 0.032105784863233566, 0.0497613251209259, -0.005708195734769106, -0.05533894896507263, -0.06467992067337036, -0.039844073355197906, -0.056803010404109955, 0.09126632660627365, 0.023043472319841385, 0.026669776067137718, -0.07527591288089752, 0.09551551938056946, -0.0213076863437891, -0.11828793585300446, -0.07102097570896149, -0.06846754252910614, -0.2110780030488968, -0.01225515641272068, -0.11614567041397095, -0.1259959638118744, 0.032758407294750214, 0.17687439918518066, 0.007095970679074526, 0.06652242690324783, -0.002010607160627842, 0.028787173330783844, 0.05354767292737961, 0.13736560940742493, -0.060415077954530716, -0.002238245913758874, -0.021257642656564713, 0.05231703072786331, -0.007573797833174467, -0.037986647337675095, -0.061677105724811554, -0.04001437500119209, 0.02109338343143463, 0.04897123947739601, 0.02537170611321926, 0.014566423371434212, -0.07527897506952286, -0.05470074713230133, -0.019534463062882423, -0.13625620305538177, 0.040359847247600555, 0.010849490761756897, -0.043869972229003906, 0.13095544278621674, 0.056962668895721436, -0.011650481261312962, -0.07216348499059677, -0.008361739106476307, -0.09025966376066208, 0.028068073093891144, -0.12218091636896133, -0.13411946594715118, 0.017059730365872383, 0.0033256823662668467, -0.050030454993247986, -0.08235623687505722, -0.11587823182344437, -0.06394140422344208, 0.0480373352766037, -0.03380927816033363, 0.03946206346154213, -0.08020124584436417, -0.043792419135570526, 0.02135917916893959, 0.030846720561385155, -0.08625948429107666, 0.006595269776880741, 0.006464182864874601, -0.004632262513041496, 0.03593510761857033, -0.037830982357263565, 0.011226363480091095, -0.1057412400841713, -0.01779843121767044, -0.01367349922657013, 0.17019301652908325, -0.07706235349178314, -0.03032475896179676, -0.08232045918703079, -0.028791768476366997, 0.038733985275030136, 0.002514935564249754, 0.04955456778407097, 0.1231609359383583, -0.25279495120048523, -0.016558513045310974, 0.17411406338214874, -0.1440533995628357, -0.02526102215051651, 0.08771003037691116, -0.054725196212530136, -0.014178600162267685, 0.07762700319290161, 0.04953781142830849, 0.18506020307540894, -0.0654539242386818, -0.0630742758512497, 0.028384022414684296, -0.039159536361694336, 0.08868248760700226, 0.05894562229514122, 0.024817630648612976, -0.05215814337134361, 0.01748264580965042, -0.07780946791172028, 0.004641425330191851, -0.04488604515790939, -0.06902945041656494, -0.056538451462984085, -0.04126434028148651, 0.027539560571312904, 0.0118724899366498, 0.0017194912070408463, -0.010538462549448013, -0.11391661316156387, 0.1652051955461502, 0.10532593727111816, -0.08411567658185959, 0.017406493425369263, -0.11201956123113632, -0.056013334542512894, -0.014112266711890697, -0.003420748980715871, -0.18552462756633759, -0.12056156992912292, 0.03247095271945, -0.13150183856487274, 0.048213325440883636, 0.05443410947918892, 0.011858492158353329, 0.041721902787685394, -0.04069984331727028, -0.016422860324382782, -0.03822660818696022, 0.00253804842941463, -0.053086087107658386, -0.145975723862648, -0.0744340792298317, -0.013810910284519196, 0.11950114369392395, -0.1476273387670517, 0.03934638574719429, -0.008963657543063164, 0.10053218901157379, -0.02334304340183735, -0.0712960809469223, -0.0078027271665632725, 0.03558935597538948, -0.044864337891340256, -0.05433737486600876, 0.02244669385254383, 0.047151777893304825, -0.10827724635601044, -0.012608900666236877, -0.12684570252895355, -0.07478852570056915, 0.07438035309314728, 0.03776266798377037, -0.09108209609985352, 0.023321187123656273, -0.003115458181127906, -0.017943192273378372, -0.07709351181983948, -0.05175177380442619, 0.09172356873750687, 0.03298046439886093, 0.09771651774644852, -0.0593714602291584, -0.025719748809933662, -0.02357889898121357, -0.04015195742249489, -0.010305040515959263, 0.08047933876514435, -0.06983973830938339, -0.2463240772485733, 0.10257625579833984, -0.009722452610731125, -0.0948178693652153, 0.12154147028923035, 0.007542720530182123, -0.024805154651403427, -0.11668193340301514, -0.0011946127051487565, 0.024205509573221207, 0.03602546453475952, -0.013188350014388561, 0.07781384885311127, 0.03337604179978371, 0.021276824176311493, -0.004435725975781679, -0.10908447206020355, 0.021862894296646118, 0.033628687262535095, -0.032307546585798264, -0.008865191601216793, 0.0007872581481933594, 0.021584955975413322, 0.06619110703468323, -0.024662980809807777, 0.04763133451342583, 0.03605271130800247, -0.06721260398626328, -0.13504284620285034, 0.2060798853635788, -0.12730136513710022, -0.22790519893169403, -0.16253462433815002, 0.08094564825296402, -0.08800404518842697, -0.007628193125128746, 0.02798990160226822, -0.05362257733941078, -0.03763335198163986, -0.06899643689393997, -0.05992531031370163, -0.08606456965208054, -0.03155779838562012, -0.046935420483350754, 0.038899872452020645, 0.07190898060798645, -0.15101571381092072, 0.014210899360477924, 0.02578127197921276, -0.09533428400754929, 0.020432092249393463, 0.020389767363667488, 0.009096344001591206, 0.12877605855464935, -0.012108997441828251, 0.019529975950717926, 0.008782971650362015, 0.24189293384552002, -0.02311779372394085, 0.013831663876771927, 0.1305152177810669, -0.017054205760359764, 0.07600722461938858, 0.0714179053902626, 0.036474671214818954, -0.05457156151533127, 0.024647213518619537, 0.05605526268482208, -0.02884596399962902, -0.22931574285030365, -0.020762525498867035, -0.04652857035398483, -0.06584543734788895, 0.047503963112831116, 0.005538444500416517, 0.029352053999900818, 0.04730710759758949, -0.04970710352063179, -0.009375541470944881, 0.05643647909164429, 0.09424328804016113, 0.09038220345973969, 0.04567413777112961, 0.11692193150520325, -0.056956756860017776, 0.03082604520022869, 0.05201992392539978, -0.007143000606447458, 0.1600838154554367, 0.013410461135208607, 0.04721095785498619, 0.10276498645544052, 0.03679201751947403, 0.0310200247913599, 0.014912590384483337, 0.00809069350361824, 0.028331810608506203, 0.02830437757074833, -0.07428167015314102, -0.02321471832692623, 0.04102614149451256, -0.016134409233927727, 0.042444996535778046, -0.0498187318444252, -0.037808772176504135, 0.05511840060353279, 0.13065864145755768, 0.08739162236452103, -0.2263941913843155, -0.08992251008749008, 0.02907826192677021, -0.02196105755865574, -0.05251423269510269, -0.024110721424221992, 0.0708085373044014, -0.09102767705917358, 0.07846494764089584, -0.04659658670425415, 0.09251465648412704, -0.09061575680971146, -0.008865991607308388, 0.019277319312095642, 0.09133949875831604, 0.019473060965538025, 0.08203797042369843, -0.24349677562713623, 0.06946094334125519, 0.01937716081738472, 0.07636678218841553, -0.00872777309268713, 0.030601568520069122, 0.03617939352989197, 0.010129506699740887, 0.08710088580846786, 0.041659388691186905, -0.027197865769267082, -0.1150759756565094, -0.023102223873138428, -0.007418039720505476, 0.09859456866979599, -0.017858855426311493, 0.11254573613405228, -0.02774149365723133, 0.001497284509241581, 0.014587260782718658, 0.09557417035102844, -0.2257792353630066, -0.17292897403240204, 0.07721123844385147, 0.05216822773218155, 0.005891559179872274, -0.03831576928496361, -0.06998211145401001, 0.024822402745485306, 0.19740532338619232, -0.2103462517261505, -0.07233606278896332, -0.13952100276947021, 0.09111177921295166, 0.12416928261518478, -0.05513986572623253, 0.00009735400089994073, -0.02422177419066429, 0.14852957427501678, -0.031116874888539314, -0.14463263750076294, 0.01451297476887703, -0.06030525267124176, -0.11267613619565964, -0.05059532821178436, 0.04382174462080002, 0.13375182449817657, 0.03136204183101654, 0.028089629486203194, 0.019533617421984673, 0.009650579653680325, -0.1235002875328064, -0.03506326675415039, 0.15866219997406006, 0.02247747592628002, 0.10125023126602173, -0.048202112317085266, -0.09401606768369675, -0.04505863040685654, 0.014439413323998451, 0.08238668739795685, 0.1870661973953247, -0.07968001812696457, 0.026350706815719604, 0.06339296698570251, -0.15366311371326447, -0.13537529110908508, -0.01084535475820303, 0.10304450243711472, 0.04391127824783325, 0.02795068547129631, -0.16729827225208282, 0.04932452738285065, 0.10431630909442902, -0.005228186491876841, -0.08625398576259613, -0.331917405128479, -0.09810144454240799, 0.04879440367221832, 0.07571657001972198, 0.0806494727730751, -0.10175281763076782, -0.046115290373563766, -0.07832036912441254, -0.0973736122250557, 0.2644710838794708, -0.0183671023696661, 0.12955538928508759, -0.002068682573735714, -0.07628535479307175, 0.038372885435819626, -0.03725653514266014, 0.1319662630558014, 0.05388110503554344, 0.09742793440818787, -0.013299467042088509, 0.02761622704565525, 0.0932011678814888, -0.060825806111097336, 0.10788195580244064, 0.012473334558308125, 0.08270631730556488, -0.10545656085014343, -0.03285276144742966, -0.051310889422893524, 0.01307655405253172, -0.05470138043165207, -0.023084498941898346, -0.05630824714899063, 0.08486324548721313, 0.09255378693342209, -0.021846063435077667, 0.049153704196214676, -0.0644868016242981, 0.06753507256507874, 0.1795922964811325, 0.11086799949407578, 0.10491730272769928, -0.1357363611459732, -0.00796203501522541, -0.015356073155999184, 0.07556792348623276, -0.08878416568040848, 0.061512693762779236, 0.10372717678546906, 0.01874486543238163, 0.13325223326683044, 0.08214081823825836, -0.06845542788505554, -0.013451945036649704, 0.03447776287794113, -0.06679690629243851, -0.12127379328012466, -0.03262878581881523, 0.009781312197446823, -0.04673761501908302, -0.008575957268476486, 0.15609432756900787, -0.053851038217544556, -0.01126149669289589, 0.01770956814289093, 0.017896676436066628, -0.030258458107709885, 0.12499350309371948, 0.03066054917871952, 0.06378744542598724, -0.07536640018224716, 0.13217280805110931, 0.07460890710353851, -0.041762061417102814, 0.06199248135089874, 0.07489340752363205, -0.0834018811583519, -0.052167411893606186, -0.05468985065817833, 0.0977640375494957, -0.14572970569133759, -0.03307158499956131, -0.0768420472741127, -0.08394142985343933, 0.016596071422100067, 0.13418643176555634, 0.06883121281862259, 0.11296132951974869, -0.027256835252046585, -0.0220254547894001, -0.08510271459817886, 0.05331804230809212, 0.035991642624139786, 0.03789135441184044, -0.10309403389692307, 0.10163548588752747, 0.016543392091989517, -0.017702657729387283, -0.019248230382800102, -0.03800003230571747, -0.15475058555603027, -0.013253603130578995, -0.09909892827272415, 0.017674067988991737, -0.02689705230295658, -0.011389706283807755, 0.019581230357289314, -0.030131153762340546, 0.015352029353380203, 0.04410799220204353, -0.01754850707948208, -0.06508079916238785, -0.016204819083213806, 0.07038749009370804, -0.11688005924224854, -0.020178450271487236, 0.012178920209407806, -0.08747567236423492, 0.04597383737564087, 0.04459499195218086, -0.04297042265534401, -0.005474344361573458, -0.14277306199073792, -0.008529802784323692, -0.025870036333799362, 0.05463242530822754, 0.016924383118748665, -0.14239075779914856, 0.04644409939646721, 0.011865769512951374, 0.01711282506585121, 0.03389603644609451, 0.13610851764678955, -0.11131532490253448, -0.08498310297727585, -0.009332544170320034, 0.0073852394707500935, -0.08044186234474182, 0.027333354577422142, 0.15503734350204468, 0.0781305581331253, 0.1578414887189865, -0.05805593356490135, 0.0735042467713356, -0.1932593137025833, -0.03789258003234863, 0.04128594323992729, -0.01986934058368206, 0.026910986751317978, -0.052011556923389435, 0.05672163516283035, -0.017574578523635864, 0.0958341732621193, -0.045539963990449905, 0.07110320031642914, 0.04999064281582832, -0.01866411790251732, -0.004324854351580143, 0.03556132689118385, 0.1433001607656479, 0.03574147820472717, -0.00927751138806343, 0.05565795674920082, -0.015413232147693634, -0.011262943968176842, -0.030906792730093002, 0.08786176145076752, 0.06970210373401642, -0.04754186049103737, 0.06267664581537247, 0.04087492451071739, -0.001255349605344236, -0.041630636900663376, 0.0709548071026802, -0.0212703887373209, 0.052719008177518845, -0.04906558245420456, 0.03644232824444771, 0.12547452747821808, -0.12472984194755554, 0.12890584766864777, -0.004883239511400461, -0.10244381427764893, -0.12320643663406372, -0.1686878800392151, -0.06349381804466248, -0.05562512204051018, -0.0075285779312253, -0.1425994634628296, 0.07038996368646622, 0.07254988700151443, 0.030373357236385345, -0.015587267465889454, 0.04138612002134323, 0.004951649811118841, -0.04910730943083763, 0.03400575742125511, -0.02207789197564125, 0.01334572583436966, 0.05089576542377472, -0.0008714591385796666, 0.11947233229875565, 0.06427314877510071, 0.05739249289035797, 0.05161244049668312, 0.098675936460495, 0.010708424262702465, -0.016627291217446327, -0.09647484123706818, 0.02619153819978237, -0.0020465936977416277, 0.021808089688420296, 0.09357680380344391, 0.06404224783182144, -0.026813378557562828, -0.024197036400437355, 0.21261854469776154, -0.09218671917915344, -0.10322973132133484, -0.15621735155582428, 0.22996847331523895, 0.017834262922406197, 0.04141052067279816, 0.011291940696537495, -0.10143448412418365, -0.02846521884202957, 0.12315940111875534, 0.0824773907661438, -0.046393368393182755, -0.014142340049147606, 0.041037920862436295, -0.028326867148280144, 0.02931460738182068, 0.08926624059677124, -0.001772263552993536, 0.18433497846126556, -0.04629528895020485, 0.127704456448555, -0.032595861703157425, -0.02758432924747467, -0.03689383715391159, 0.1284128874540329, -0.039467182010412216, 0.007822422310709953, -0.04500308260321617, 0.08230483531951904, -0.03602297604084015, -0.051367033272981644, 0.10093633085489273, -0.025097988545894623, -0.11398974806070328, -0.024396510794758797, 0.06002479046583176, -0.05242688208818436, 0.05477088689804077, -0.025026924908161163, 0.0375724621117115, 0.17063480615615845, -0.03723656386137009, -0.049455832690000534, -0.12045363336801529, -0.009596114978194237, -0.06487730890512466, 0.09438472986221313, 0.024705907329916954, 0.12418009340763092, 0.1092207059264183, 0.012137599289417267, -0.10241204500198364, 0.08259478211402893, -0.01986747793853283, -0.04321520775556564, 0.044761773198843, 0.15041036903858185, -0.013709241524338722, 0.16318008303642273, 0.03208295628428459, -0.12901747226715088, 0.02778017334640026, 0.002916376106441021, -0.05903704836964607, -0.06797663867473602, 0.023460211232304573, -0.05931134149432182, 0.11819195747375488, 0.1875811517238617, -0.04592861607670784, 0.007415512576699257, -0.03134014084935188, 0.04082135483622551, 0.00736593222245574, 0.056211818009614944, -0.018419325351715088, -0.1616811901330948, 0.046633459627628326, 0.04239862784743309, 0.0557604543864727, -0.16134314239025116, -0.05228674039244652, 0.010217325761914253, -0.05349377915263176, -0.01762446202337742, 0.09871968626976013, 0.011879968456923962, 0.04144349321722984, -0.04619990661740303, -0.21636387705802917, 0.01770223118364811, 0.09774301201105118, -0.08837727457284927, -0.09243623167276382 ]
null
null
transformers
# Cross-Encoder for Natural Language Inference This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. ## Training Data The model was trained on the [SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral. ## Performance For evaluation results, see [SBERT.net - Pretrained Cross-Encoder](https://www.sbert.net/docs/pretrained_cross-encoders.html#nli). ## Usage Pre-trained models can be used like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('cross-encoder/nli-distilroberta-base') scores = model.predict([('A man is eating pizza', 'A man eats something'), ('A black race car starts up in front of a crowd of people.', 'A man is driving down a lonely road.')]) #Convert scores to labels label_mapping = ['contradiction', 'entailment', 'neutral'] labels = [label_mapping[score_max] for score_max in scores.argmax(axis=1)] ``` ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('cross-encoder/nli-distilroberta-base') tokenizer = AutoTokenizer.from_pretrained('cross-encoder/nli-distilroberta-base') features = tokenizer(['A man is eating pizza', 'A black race car starts up in front of a crowd of people.'], ['A man eats something', 'A man is driving down a lonely road.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits label_mapping = ['contradiction', 'entailment', 'neutral'] labels = [label_mapping[score_max] for score_max in scores.argmax(dim=1)] print(labels) ``` ## Zero-Shot Classification This model can also be used for zero-shot-classification: ```python from transformers import pipeline classifier = pipeline("zero-shot-classification", model='cross-encoder/nli-distilroberta-base') sent = "Apple just announced the newest iPhone X" candidate_labels = ["technology", "sports", "politics"] res = classifier(sent, candidate_labels) print(res) ```
{"language": "en", "license": "apache-2.0", "tags": ["distilroberta-base"], "datasets": ["multi_nli", "snli"], "metrics": ["accuracy"], "pipeline_tag": "zero-shot-classification"}
zero-shot-classification
cross-encoder/nli-distilroberta-base
[ "transformers", "pytorch", "jax", "roberta", "text-classification", "distilroberta-base", "zero-shot-classification", "en", "dataset:multi_nli", "dataset:snli", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #roberta #text-classification #distilroberta-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
# Cross-Encoder for Natural Language Inference This model was trained using SentenceTransformers Cross-Encoder class. ## Training Data The model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral. ## Performance For evaluation results, see URL - Pretrained Cross-Encoder. ## Usage Pre-trained models can be used like this: ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ## Zero-Shot Classification This model can also be used for zero-shot-classification:
[ "# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.", "## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):", "## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ "TAGS\n#transformers #pytorch #jax #roberta #text-classification #distilroberta-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.", "## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):", "## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ 83, 30, 49, 18, 14, 33, 21 ]
[ "passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #distilroberta-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ -0.016179563477635384, 0.17483745515346527, -0.006616861559450626, 0.09181816130876541, 0.12407670170068741, 0.014483985491096973, 0.25409168004989624, 0.08052258938550949, -0.024745415896177292, -0.026574302464723587, 0.02368866838514805, 0.21927763521671295, -0.00322065781801939, 0.06625080853700638, -0.03953150287270546, -0.23719273507595062, 0.04993871971964836, 0.010431649163365364, 0.022437317296862602, 0.0974813774228096, 0.10483920574188232, -0.04197899252176285, 0.02976999618113041, -0.015116767026484013, -0.1281358003616333, 0.049260880798101425, 0.015758223831653595, -0.04626785218715668, 0.10696873068809509, 0.043159373104572296, 0.12747281789779663, 0.07645203918218613, 0.011554154567420483, -0.1631617695093155, 0.008877493441104889, 0.04404466226696968, -0.0367804691195488, 0.04261625185608864, 0.008047709241509438, -0.11447547376155853, -0.016085002571344376, -0.016497448086738586, 0.09156777709722519, 0.054332684725522995, -0.09889556467533112, -0.12051784247159958, -0.011081195436418056, 0.00017126153397839516, 0.07707807421684265, 0.09557634592056274, -0.03899674490094185, 0.1306856870651245, -0.12225210666656494, 0.0997372418642044, 0.1056051105260849, -0.23877760767936707, -0.020944200456142426, 0.07414770126342773, 0.014419755898416042, 0.08777645230293274, -0.07266884297132492, -0.0044306120835244656, -0.01774299331009388, 0.03498782590031624, 0.08532581478357315, -0.05648383870720863, -0.053472768515348434, -0.0006906320340931416, -0.1286039501428604, 0.030365698039531708, 0.18127663433551788, -0.0454140342772007, -0.059429511427879333, -0.08828768879175186, -0.056107547134160995, -0.03005872666835785, -0.02063913084566593, 0.007782226428389549, 0.019559595733880997, 0.04391113296151161, -0.055060334503650665, -0.026878872886300087, -0.10843238979578018, -0.027044519782066345, -0.08467471599578857, 0.16180631518363953, 0.044600095599889755, 0.023999322205781937, -0.10822272300720215, 0.07826224714517593, 0.006206526421010494, -0.10930239409208298, -0.02193615585565567, -0.06736277043819427, -0.16911758482456207, 0.011127190664410591, -0.13218165934085846, -0.17250461876392365, 0.05156683921813965, 0.21048834919929504, 0.002937544835731387, 0.051607150584459305, -0.018387233838438988, 0.0541221983730793, 0.0679842084646225, 0.09305501729249954, -0.020400486886501312, 0.012726773507893085, -0.0036299999337643385, 0.0016314326785504818, 0.03340854123234749, -0.058135442435741425, -0.11504131555557251, 0.009309695102274418, -0.0017239813460037112, 0.04937438666820526, 0.00951436161994934, 0.07073435187339783, -0.0335627906024456, -0.049755703657865524, -0.0006959459860809147, -0.09152618050575256, 0.009183530695736408, 0.011584545485675335, 0.03514039143919945, 0.15902100503444672, 0.031651418656110764, -0.002812829101458192, -0.048847224563360214, -0.004587486851960421, -0.10236294567584991, 0.032688502222299576, -0.08541063964366913, -0.1171928197145462, 0.012499981559813023, 0.013864015229046345, -0.040647804737091064, -0.11918497830629349, -0.08753523230552673, -0.0008315534214489162, 0.046627987176179886, -0.03458356112241745, 0.016639618203043938, -0.08538159728050232, 0.0013399961171671748, 0.03875163197517395, -0.015709562227129936, -0.09363697469234467, 0.004699713084846735, 0.03450348228216171, 0.03244491666555405, 0.07970036566257477, -0.06699655205011368, 0.016386419534683228, -0.1646519899368286, -0.03142491728067398, 0.006780876778066158, 0.1217116266489029, -0.03425361216068268, -0.0066752079874277115, -0.06473144143819809, -0.009321007877588272, 0.04376450181007385, 0.037563640624284744, -0.01458753366023302, 0.1011132001876831, -0.22586117684841156, -0.08066738396883011, 0.20015248656272888, -0.1569576859474182, -0.05629594251513481, 0.1268291175365448, -0.036664269864559174, 0.00030231449636630714, 0.0937829241156578, 0.07872416824102402, 0.12591122090816498, -0.05994310975074768, 0.004796607885509729, 0.05270844325423241, -0.0270254984498024, 0.060454316437244415, 0.07076749950647354, 0.0695013478398323, -0.052496034651994705, 0.04716191813349724, -0.06020314618945122, 0.03215261548757553, -0.011810760013759136, -0.06451310962438583, -0.02296626754105091, -0.021850228309631348, 0.11763324588537216, 0.014268080703914165, 0.05659034103155136, -0.021617628633975983, -0.11923425644636154, 0.16562096774578094, 0.08534129709005356, -0.12463154643774033, 0.06476061046123505, -0.07605014741420746, -0.05157604068517685, -0.033458955585956573, -0.004958768375217915, -0.19925834238529205, -0.0758775919675827, 0.030906612053513527, -0.0458175390958786, 0.11143489181995392, 0.06717057526111603, 0.010290004312992096, 0.07028773427009583, -0.04784758388996124, -0.008522887714207172, 0.03808912634849548, 0.024599183350801468, -0.0386495366692543, -0.15975509583950043, -0.05049788951873779, -0.038394197821617126, 0.10305216163396835, -0.0857989490032196, 0.05634041875600815, 0.012331915087997913, 0.053104277700185776, -0.011447243392467499, -0.0051216729916632175, -0.029532384127378464, 0.06958496570587158, -0.07152968645095825, 0.009618018753826618, 0.06093680486083031, 0.03378266096115112, -0.11768080294132233, 0.06514627486467361, -0.1600673645734787, -0.08432389050722122, 0.08296851068735123, -0.10564347356557846, -0.05126577988266945, 0.001399049418978393, -0.006671771872788668, -0.010683183558285236, -0.06793859601020813, -0.034471992403268814, 0.10708066076040268, 0.049419913440942764, 0.11128227412700653, -0.0504717156291008, -0.02233082242310047, -0.03787035495042801, -0.111070416867733, 0.017316877841949463, 0.06695768237113953, -0.016547420993447304, -0.18737031519412994, 0.1343526840209961, 0.03924483433365822, -0.173817977309227, 0.12946970760822296, 0.01175601501017809, -0.007131066173315048, -0.0784919485449791, -0.04020204395055771, 0.02160249836742878, -0.008550150319933891, -0.04664032161235809, 0.023650972172617912, 0.0448274239897728, 0.0415160208940506, 0.006467767059803009, -0.09996951371431351, 0.007487115915864706, 0.013942837715148926, -0.002898495877161622, -0.004599038045853376, 0.020723408088088036, 0.004215818364173174, 0.06908963620662689, -0.02101384662091732, -0.0039139119908213615, 0.07265464216470718, -0.037437908351421356, -0.11171500384807587, 0.20069730281829834, -0.11854097247123718, -0.29734691977500916, -0.13242457807064056, -0.039392098784446716, -0.10575996339321136, 0.03792177140712738, 0.08238454908132553, -0.08204706013202667, -0.021595755591988564, -0.06242850050330162, 0.020970208570361137, -0.0703187882900238, -0.023407788947224617, -0.12867413461208344, 0.05095246434211731, -0.003342570038512349, -0.12006861716508865, 0.015183577314019203, 0.015035636723041534, -0.10446028411388397, 0.06700155884027481, -0.031813845038414, 0.0019646359141916037, 0.17564554512500763, 0.02088848501443863, 0.009827089495956898, -0.011283481493592262, 0.14933401346206665, -0.006647629197686911, 0.014341549016535282, 0.15318730473518372, -0.0034220018424093723, 0.07998388260602951, 0.10431496053934097, 0.05772854760289192, -0.025488581508398056, 0.03595352545380592, 0.03833400830626488, -0.059422567486763, -0.18204456567764282, -0.05274856463074684, -0.07324843853712082, -0.027607547119259834, 0.05821212753653526, -0.000189896221854724, 0.12514004111289978, 0.07010465860366821, -0.027163533493876457, -0.0014743686188012362, 0.06147759407758713, 0.11883123964071274, 0.16548371315002441, 0.027780737727880478, 0.1472623497247696, -0.06051899120211601, -0.07565368711948395, 0.030213316902518272, 0.05498189106583595, 0.13718605041503906, 0.027757035568356514, 0.14371085166931152, 0.125014066696167, -0.003406201722100377, 0.059519775211811066, 0.04479513689875603, 0.0038814737927168608, 0.022032257169485092, -0.01044598501175642, -0.0775943249464035, -0.0037867010105401278, 0.028578635305166245, -0.020195385441184044, -0.010159359313547611, -0.03715430945158005, -0.06333133578300476, 0.10174635052680969, 0.1005399227142334, 0.09545137733221054, -0.20800025761127472, -0.04152628779411316, 0.05912453681230545, -0.028573619201779366, -0.006625225767493248, 0.02329978719353676, 0.03102215938270092, -0.09489990025758743, 0.06610623002052307, -0.03649982810020447, 0.07221157103776932, -0.07735342532396317, 0.013820097781717777, -0.01966692879796028, 0.05486895889043808, 0.035253167152404785, 0.10927782952785492, -0.2927652895450592, 0.07265476882457733, -0.00901646539568901, 0.023563170805573463, -0.037343576550483704, 0.029314659535884857, -0.002276182873174548, 0.08374850451946259, 0.05558665096759796, -0.004939447622746229, -0.08275876939296722, -0.07078203558921814, 0.003186519956216216, -0.014674581587314606, 0.08165452629327774, -0.07017985731363297, 0.10152842849493027, -0.02961593121290207, -0.008105854503810406, 0.0011123274452984333, 0.09578943997621536, -0.12093859165906906, -0.21786056458950043, 0.0880807414650917, 0.053528252989053726, -0.011638030409812927, -0.02493259124457836, -0.07641366124153137, 0.016655921936035156, 0.14715206623077393, -0.1796661764383316, -0.09044169634580612, -0.13549792766571045, 0.0122912572696805, 0.08402104675769806, -0.0789397582411766, -0.015380834229290485, -0.027268212288618088, 0.14540453255176544, -0.05334625765681267, -0.15198127925395966, 0.039407700300216675, -0.051500555127859116, -0.09440787881612778, -0.06473853439092636, 0.07421325147151947, 0.07745064049959183, 0.046454787254333496, 0.01871923729777336, 0.03982684761285782, -0.023802436888217926, -0.1094416007399559, -0.08747740834951401, 0.027760369703173637, 0.05152600258588791, 0.05364903062582016, -0.053973857313394547, -0.07973649352788925, -0.05835186690092087, 0.022482963278889656, 0.18966326117515564, 0.1650882065296173, -0.0721210464835167, 0.09591861814260483, 0.13214534521102905, -0.1299150288105011, -0.17268408834934235, -0.11367371678352356, 0.04817552492022514, -0.011846001259982586, 0.025218447670340538, -0.16305403411388397, 0.07960549741983414, 0.07957201451063156, -0.01888785883784294, -0.13352219760417938, -0.3094058334827423, -0.09645522385835648, 0.06238507851958275, 0.08344234526157379, 0.1311528980731964, -0.14314928650856018, -0.021699998527765274, -0.07430125027894974, -0.03848343715071678, 0.3309873044490814, -0.07007944583892822, 0.12766224145889282, 0.006164290010929108, -0.008706865832209587, 0.016811218112707138, -0.02183225378394127, 0.09028959274291992, 0.05097613111138344, 0.0674244835972786, -0.0019208004232496023, -0.05256814882159233, 0.12754115462303162, -0.039938367903232574, 0.0733877494931221, -0.06719009578227997, 0.04116196557879448, -0.012987282127141953, -0.04798605293035507, -0.08007510751485825, -0.005557004362344742, -0.055010173469781876, -0.06858979910612106, -0.0363316610455513, 0.04096995294094086, 0.05727098882198334, -0.030181707814335823, 0.20965923368930817, -0.09514881670475006, 0.07091905176639557, 0.17341426014900208, 0.11355318874120712, 0.08429756760597229, -0.14683124423027039, -0.040883950889110565, -0.05153617262840271, 0.07043055444955826, -0.15260078012943268, 0.09326495230197906, 0.040754154324531555, -0.003324153833091259, 0.12823991477489471, 0.10343579947948456, -0.03638547658920288, 0.006390796974301338, 0.06517700105905533, -0.0864969864487648, -0.01623670570552349, -0.012498071417212486, 0.046885937452316284, -0.03388489410281181, 0.06833715736865997, 0.17238618433475494, -0.036461807787418365, -0.0408623032271862, 0.016877571120858192, 0.029122615233063698, -0.049503110349178314, 0.08536682277917862, 0.04029679298400879, 0.05833930894732475, -0.09374278038740158, 0.09707432985305786, 0.09584259986877441, -0.005094690248370171, 0.020100897178053856, 0.11768631637096405, -0.07548633962869644, -0.09394325315952301, -0.0530935637652874, 0.06317728757858276, -0.12944823503494263, -0.017804672941565514, -0.0976978987455368, -0.09668396413326263, -0.01939958892762661, 0.12659905850887299, 0.0766846314072609, 0.11842494457960129, -0.09932445734739304, -0.019686317071318626, -0.09794259071350098, 0.051842499524354935, -0.023298628628253937, 0.06088927015662193, -0.07922925800085068, 0.12406975775957108, 0.014904997311532497, 0.002170486841350794, -0.05888098478317261, -0.06754309684038162, -0.16016803681850433, 0.01884085312485695, -0.06090020015835762, 0.014263663440942764, -0.08026371151208878, -0.0065395524725317955, 0.020119376480579376, -0.025683296844363213, 0.007969677448272705, 0.008470220491290092, -0.030647197738289833, 0.002187131904065609, -0.01459111925214529, 0.0719708725810051, -0.09134095907211304, -0.013293493539094925, 0.018665945157408714, -0.09429667890071869, 0.06254291534423828, 0.028111321851611137, -0.06703388690948486, 0.07451833039522171, -0.1378915011882782, 0.0023480693344026804, -0.024436915293335915, 0.045437440276145935, 0.009103174321353436, -0.07701527327299118, 0.007307255640625954, 0.025044677779078484, 0.05201352760195732, 0.03025892935693264, 0.057038065046072006, -0.11172782629728317, 0.004029986448585987, 0.0020826321560889482, -0.08905141055583954, -0.07296968251466751, 0.0011386439437046647, 0.07145696133375168, 0.0589909553527832, 0.17313775420188904, -0.07683166861534119, 0.08571616560220718, -0.11118802428245544, -0.014618984423577785, 0.02810252644121647, -0.0758822113275528, -0.036692064255476, -0.10213347524404526, 0.019412590190768242, -0.017477290704846382, 0.13867177069187164, -0.03479471802711487, 0.09077233076095581, 0.024683751165866852, 0.020439982414245605, 0.09082754701375961, 0.06152648478746414, 0.1565525084733963, 0.03633545711636543, -0.0259388480335474, -0.00901868287473917, 0.0023429151624441147, -0.010662316344678402, 0.0004242085851728916, 0.04680642858147621, 0.043853629380464554, -0.020808227360248566, 0.08123642206192017, 0.03400968387722969, 0.08004342764616013, -0.07147060334682465, 0.008197580464184284, 0.014643549919128418, -0.0009162737987935543, -0.05534913018345833, 0.06244191154837608, 0.14821477234363556, -0.07661014795303345, 0.10842666774988174, 0.034275300800800323, -0.10361023247241974, -0.13479052484035492, -0.19935575127601624, -0.05684429779648781, -0.06645174324512482, -0.034187037497758865, -0.14763310551643372, 0.015209724195301533, 0.09013374894857407, 0.0318148173391819, -0.014901946298778057, 0.02488572895526886, 0.09696076065301895, -0.040128953754901886, -0.0032125350553542376, -0.04694400727748871, 0.04197174683213234, 0.07110978662967682, 0.046558964997529984, 0.03762049973011017, 0.053181588649749756, 0.04261656105518341, 0.04404380917549133, 0.0778457447886467, 0.0023964443244040012, -0.09400436282157898, -0.11946669220924377, 0.016113445162773132, 0.003664941294118762, 0.01995585858821869, 0.11948590725660324, 0.044670626521110535, -0.08036690950393677, -0.0045856027863919735, 0.2301207333803177, -0.08689501136541367, -0.1251082569360733, -0.15396207571029663, 0.2857726216316223, 0.03647658973932266, 0.06389530748128891, 0.001611840445548296, -0.07280007004737854, -0.06554117798805237, 0.1745193898677826, 0.12345030903816223, -0.0024509101640433073, 0.0023756285663694143, 0.07530868798494339, -0.012673580087721348, 0.026622604578733444, 0.08768334984779358, -0.01438157632946968, 0.26289913058280945, -0.06071364879608154, 0.1048322319984436, -0.01602211594581604, -0.021167518571019173, -0.08362694829702377, 0.0885547325015068, -0.04866098240017891, -0.007364971097558737, -0.04332225024700165, 0.07167846709489822, -0.10192621499300003, 0.03093337081372738, 0.050756193697452545, -0.019278760999441147, -0.1002311110496521, -0.015285230241715908, 0.11093827337026596, -0.050110913813114166, 0.07738878577947617, 0.005788525100797415, -0.012631089426577091, 0.11618766188621521, -0.018669895827770233, -0.03229374438524246, -0.051501404494047165, 0.041841406375169754, -0.053959742188453674, 0.055369436740875244, -0.021204764023423195, 0.13008840382099152, 0.11406723409891129, 0.019049206748604774, -0.048471953719854355, 0.1211332231760025, -0.03307358920574188, -0.05321257933974266, 0.0821930542588234, 0.11083729565143585, -0.024354765191674232, 0.14556267857551575, 0.050124071538448334, -0.16575747728347778, 0.05721302703022957, -0.016143830493092537, -0.07299494743347168, -0.0363951250910759, 0.014887064695358276, -0.05211925879120827, 0.06375987082719803, 0.1320514678955078, -0.059226058423519135, -0.025236736983060837, -0.043585821986198425, 0.01620788499712944, -0.006757156923413277, -0.0285655464977026, -0.06140536814928055, -0.15907734632492065, 0.0056280698627233505, 0.08307977020740509, 0.04432760551571846, -0.21461695432662964, -0.027090881019830704, 0.019541483372449875, -0.015807973220944405, -0.09448376297950745, 0.05615435168147087, 0.042489759624004364, 0.030411502346396446, -0.048695389181375504, -0.245537668466568, 0.007639718241989613, 0.07803911715745926, -0.13509754836559296, -0.11017920821905136 ]
null
null
transformers
# Cross-Encoder for Natural Language Inference This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. ## Training Data The model was trained on the [SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral. ## Performance For evaluation results, see [SBERT.net - Pretrained Cross-Encoder](https://www.sbert.net/docs/pretrained_cross-encoders.html#nli). ## Usage Pre-trained models can be used like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('cross-encoder/nli-roberta-base') scores = model.predict([('A man is eating pizza', 'A man eats something'), ('A black race car starts up in front of a crowd of people.', 'A man is driving down a lonely road.')]) #Convert scores to labels label_mapping = ['contradiction', 'entailment', 'neutral'] labels = [label_mapping[score_max] for score_max in scores.argmax(axis=1)] ``` ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('cross-encoder/nli-roberta-base') tokenizer = AutoTokenizer.from_pretrained('cross-encoder/nli-roberta-base') features = tokenizer(['A man is eating pizza', 'A black race car starts up in front of a crowd of people.'], ['A man eats something', 'A man is driving down a lonely road.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = model(**features).logits label_mapping = ['contradiction', 'entailment', 'neutral'] labels = [label_mapping[score_max] for score_max in scores.argmax(dim=1)] print(labels) ``` ## Zero-Shot Classification This model can also be used for zero-shot-classification: ```python from transformers import pipeline classifier = pipeline("zero-shot-classification", model='cross-encoder/nli-roberta-base') sent = "Apple just announced the newest iPhone X" candidate_labels = ["technology", "sports", "politics"] res = classifier(sent, candidate_labels) print(res) ```
{"language": "en", "license": "apache-2.0", "tags": ["roberta-base"], "datasets": ["multi_nli", "snli"], "metrics": ["accuracy"], "pipeline_tag": "zero-shot-classification"}
zero-shot-classification
cross-encoder/nli-roberta-base
[ "transformers", "pytorch", "jax", "roberta", "text-classification", "roberta-base", "zero-shot-classification", "en", "dataset:multi_nli", "dataset:snli", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #roberta #text-classification #roberta-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
# Cross-Encoder for Natural Language Inference This model was trained using SentenceTransformers Cross-Encoder class. ## Training Data The model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral. ## Performance For evaluation results, see URL - Pretrained Cross-Encoder. ## Usage Pre-trained models can be used like this: ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ## Zero-Shot Classification This model can also be used for zero-shot-classification:
[ "# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.", "## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):", "## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ "TAGS\n#transformers #pytorch #jax #roberta #text-classification #roberta-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.", "## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):", "## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ 81, 30, 49, 18, 14, 33, 21 ]
[ "passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #roberta-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:" ]
[ -0.012450331822037697, 0.18565769493579865, -0.006619107909500599, 0.08923697471618652, 0.12173096835613251, 0.007555744145065546, 0.23860366642475128, 0.0946941077709198, -0.03857089951634407, -0.020934268832206726, 0.027760099619627, 0.2214883416891098, -0.0005910868640057743, 0.05267477408051491, -0.04104902222752571, -0.2249462902545929, 0.0505533404648304, 0.010125595144927502, 0.03220982849597931, 0.1024373322725296, 0.10679016262292862, -0.03439415246248245, 0.031699955463409424, -0.019713222980499268, -0.12865905463695526, 0.04705250263214111, 0.01927412860095501, -0.052999526262283325, 0.09766140580177307, 0.040177322924137115, 0.11399038881063461, 0.07511471956968307, -0.0033349243458360434, -0.16176249086856842, 0.010577261447906494, 0.044718094170093536, -0.0414261594414711, 0.03746254742145538, 0.0037557773757725954, -0.12255161255598068, 0.006816508714109659, -0.006379482802003622, 0.08544141054153442, 0.05841675028204918, -0.100577712059021, -0.12125720828771591, -0.012571580708026886, 0.0007140889065340161, 0.06652866303920746, 0.09600599110126495, -0.04178692400455475, 0.1290944516658783, -0.13174787163734436, 0.1046469658613205, 0.10754507780075073, -0.24696797132492065, -0.02518063224852085, 0.09047453850507736, 0.018305305391550064, 0.09359800070524216, -0.07329710572957993, 0.0006153449648991227, -0.025271685793995857, 0.03697926178574562, 0.08716819435358047, -0.06203833222389221, -0.07390651851892471, -0.006042410619556904, -0.12024009972810745, 0.024691374972462654, 0.1711689978837967, -0.05308464169502258, -0.052414704114198685, -0.08973576128482819, -0.044284265488386154, -0.03676522150635719, -0.024898335337638855, 0.01421323698014021, 0.017205171287059784, 0.046667974442243576, -0.05754328519105911, -0.028465790674090385, -0.09630981087684631, -0.023670313879847527, -0.07510897517204285, 0.14011380076408386, 0.040735889226198196, 0.022895149886608124, -0.10597661882638931, 0.06976048648357391, -0.02050093561410904, -0.10978848487138748, -0.026790235191583633, -0.05872846022248268, -0.16605323553085327, 0.010712525807321072, -0.13065214455127716, -0.17338524758815765, 0.06294272840023041, 0.21598485112190247, -0.004371569491922855, 0.05843447893857956, -0.03290529176592827, 0.06020529940724373, 0.05529670789837837, 0.0960676372051239, -0.02164079248905182, -0.005296574905514717, 0.004179676529020071, 0.003346445271745324, 0.032136645168066025, -0.05887006223201752, -0.11428090184926987, 0.007645449135452509, 0.01293103862553835, 0.05218927934765816, 0.01327122189104557, 0.06948358565568924, -0.029193222522735596, -0.044936228543519974, -0.009873587638139725, -0.0875915139913559, 0.007795531302690506, 0.014495502226054668, 0.031724900007247925, 0.1569458693265915, 0.031711380928754807, -0.009118339978158474, -0.05874954164028168, -0.011500927619636059, -0.10075683891773224, 0.035394199192523956, -0.07876450568437576, -0.11124494671821594, 0.01730329543352127, -0.0074829114601016045, -0.04367668181657791, -0.1113424226641655, -0.08893245458602905, -0.009352643974125385, 0.04121452569961548, -0.042052045464515686, 0.010894590057432652, -0.09175154566764832, -0.00673984969034791, 0.0372416190803051, -0.015526141040027142, -0.08178561180830002, 0.0020215539261698723, 0.03847956284880638, 0.023425810039043427, 0.07876661419868469, -0.06705465167760849, 0.020072977989912033, -0.17164160311222076, -0.03310557082295418, -0.0010835685534402728, 0.13588696718215942, -0.03681298717856407, -0.0009214031742885709, -0.07224419713020325, -0.009640789590775967, 0.043723903596401215, 0.03753501549363136, -0.019805893301963806, 0.10028775781393051, -0.2174084186553955, -0.07815679162740707, 0.2078258991241455, -0.15413279831409454, -0.0536031499505043, 0.12620580196380615, -0.03625909239053726, -0.00011715140863088891, 0.1027451753616333, 0.09397595375776291, 0.1261587291955948, -0.039543334394693375, 0.00725755887106061, 0.056654706597328186, -0.022766239941120148, 0.0521070621907711, 0.07045575231313705, 0.06330736726522446, -0.05274278670549393, 0.05019250884652138, -0.06546486914157867, 0.02483208291232586, -0.005907400045543909, -0.06826876848936081, -0.024227667599916458, -0.024844413623213768, 0.11584315448999405, 0.021924156695604324, 0.061033040285110474, -0.02400381863117218, -0.12372925877571106, 0.14643016457557678, 0.07588348537683487, -0.1250528246164322, 0.06585909426212311, -0.07270309329032898, -0.024557748809456825, -0.025992048904299736, 0.003319599200040102, -0.2118838131427765, -0.06400881707668304, 0.025655049830675125, -0.022052984684705734, 0.12084250152111053, 0.051451925188302994, 0.011525008827447891, 0.058986298739910126, -0.05141885206103325, 0.0012017883127555251, 0.043840549886226654, 0.021123921498656273, -0.035444580018520355, -0.15632383525371552, -0.04418105632066727, -0.038777485489845276, 0.10818906873464584, -0.09663290530443192, 0.054174698889255524, 0.021981481462717056, 0.050402410328388214, -0.006965021137148142, -0.010302817448973656, -0.02123831957578659, 0.07732733339071274, -0.06794185191392899, 0.009771956130862236, 0.06980151683092117, 0.0350063294172287, -0.12334111332893372, 0.06959979236125946, -0.1572062373161316, -0.10307637602090836, 0.07777225971221924, -0.10368896275758743, -0.0595441572368145, -0.004109621047973633, -0.0010901902569457889, -0.0052638971246778965, -0.06212881952524185, -0.028930626809597015, 0.11709123849868774, 0.044428709894418716, 0.11677329242229462, -0.058479633182287216, -0.02022840641438961, -0.03941849246621132, -0.1090300977230072, 0.026229269802570343, 0.0846758484840393, -0.03289739042520523, -0.18519560992717743, 0.14462050795555115, 0.03408856317400932, -0.17737914621829987, 0.1293485313653946, 0.005879031028598547, -0.0031871211249381304, -0.06671623885631561, -0.026162147521972656, 0.02710885740816593, -0.02462318167090416, -0.05220670625567436, 0.022138455882668495, 0.04719655588269234, 0.039878956973552704, 0.004247005563229322, -0.0967043787240982, 0.009573903866112232, 0.009243948385119438, -0.0066325729712843895, 0.010803228244185448, 0.025998078286647797, 0.008302241563796997, 0.07369326055049896, -0.022184044122695923, -0.00197369116358459, 0.07530448585748672, -0.036185361444950104, -0.10855858027935028, 0.2090742588043213, -0.12123976647853851, -0.29193076491355896, -0.13374346494674683, -0.046876322478055954, -0.10550013184547424, 0.03204580768942833, 0.08204121142625809, -0.07689680904150009, -0.01193046011030674, -0.06913482397794724, 0.004393591079860926, -0.06569338589906693, -0.03360237553715706, -0.14162077009677887, 0.05315481498837471, 0.002271217992529273, -0.11402598768472672, 0.012300659902393818, 0.017375318333506584, -0.11220746487379074, 0.06886997073888779, -0.03351445868611336, -0.004102560691535473, 0.1780775487422943, 0.01737040840089321, 0.008323193527758121, -0.01722915843129158, 0.14847148954868317, -0.011598735116422176, 0.0070030828937888145, 0.1525987982749939, -0.005505564622581005, 0.07806174457073212, 0.10200884938240051, 0.05584898591041565, -0.025205958634614944, 0.022778287529945374, 0.038980718702077866, -0.07139378041028976, -0.18761035799980164, -0.05153791978955269, -0.07656782120466232, -0.023912549018859863, 0.058784499764442444, 0.0059951250441372395, 0.12992975115776062, 0.0690356120467186, -0.021815136075019836, 0.0007329290965572, 0.06570333242416382, 0.11630966514348984, 0.16486039757728577, 0.03420660272240639, 0.15011534094810486, -0.05943819135427475, -0.06557590514421463, 0.027526261284947395, 0.0575735904276371, 0.13567616045475006, 0.01709792949259281, 0.14106470346450806, 0.12881769239902496, -0.008604228496551514, 0.0622728168964386, 0.026579810306429863, -0.007510789670050144, 0.023269684985280037, -0.011807480826973915, -0.08216328918933868, -0.015526183880865574, 0.025179650634527206, -0.020070306956768036, -0.014619782567024231, -0.029411379247903824, -0.07155580818653107, 0.10113569349050522, 0.10026796162128448, 0.09475132077932358, -0.20711542665958405, -0.039506491273641586, 0.0557674840092659, -0.031290240585803986, -0.005813129711896181, 0.029256368055939674, 0.03115158900618553, -0.09995120763778687, 0.07046851515769958, -0.03600506857037544, 0.07880963385105133, -0.07761731743812561, 0.012502390891313553, -0.03771689906716347, 0.05046102777123451, 0.034374356269836426, 0.10782507061958313, -0.2936159074306488, 0.06921129673719406, -0.00794208887964487, 0.02203403227031231, -0.03863126039505005, 0.02527143433690071, 0.006815529894083738, 0.0898805633187294, 0.07078266888856888, -0.0014410163275897503, -0.09332704544067383, -0.05730877444148064, -0.0034667770378291607, -0.018515843898057938, 0.06967348605394363, -0.06149952858686447, 0.09431260824203491, -0.025228282436728477, -0.006380427163094282, -0.014646177180111408, 0.08775563538074493, -0.10262458771467209, -0.21019777655601501, 0.08087897300720215, 0.059506915509700775, -0.013380584307014942, -0.018973233178257942, -0.06711675971746445, 0.015813827514648438, 0.1510452777147293, -0.17426945269107819, -0.09025180339813232, -0.13491782546043396, 0.017078332602977753, 0.07379380613565445, -0.07790034264326096, -0.020112087950110435, -0.03486156091094017, 0.13486869633197784, -0.050726428627967834, -0.152598574757576, 0.0359918475151062, -0.04495254531502724, -0.09107216447591782, -0.05870461463928223, 0.0776471197605133, 0.07092888653278351, 0.05132799595594406, 0.021823175251483917, 0.04548642784357071, -0.0171645637601614, -0.10621954500675201, -0.0885140523314476, 0.022299306467175484, 0.05113700032234192, 0.06444106251001358, -0.057388000190258026, -0.09148228168487549, -0.060147810727357864, 0.025848114863038063, 0.20653387904167175, 0.15422925353050232, -0.07067760825157166, 0.10784995555877686, 0.13928671181201935, -0.13079266250133514, -0.1883605420589447, -0.11151424050331116, 0.04821036756038666, -0.010691069066524506, 0.03179192543029785, -0.1638483852148056, 0.07557666301727295, 0.08533897250890732, -0.017466308549046516, -0.1276526302099228, -0.3185724914073944, -0.09157223999500275, 0.05842962861061096, 0.08790219575166702, 0.1381182074546814, -0.1497616022825241, -0.03312225267291069, -0.06271783262491226, -0.02897387556731701, 0.3255256712436676, -0.08629051595926285, 0.13215389847755432, 0.0007825204520486295, -0.012758033350110054, 0.01575522869825363, -0.0195584949105978, 0.08022560179233551, 0.049682386219501495, 0.06443493813276291, -0.0022013415582478046, -0.045003172010183334, 0.13290157914161682, -0.04270981252193451, 0.06555693596601486, -0.08682785928249359, 0.034320149570703506, -0.027255650609731674, -0.03942519798874855, -0.08806643635034561, -0.0024810596369206905, -0.054281752556562424, -0.05716060474514961, -0.04055728018283844, 0.048742007464170456, 0.05758865550160408, -0.025887634605169296, 0.2156621664762497, -0.11639522761106491, 0.09311468154191971, 0.1888318657875061, 0.1134427934885025, 0.08856187760829926, -0.15146325528621674, -0.04203873500227928, -0.04933140426874161, 0.07479318976402283, -0.16137710213661194, 0.08695916086435318, 0.0392037108540535, -0.0013083107769489288, 0.13433799147605896, 0.09533225744962692, -0.03444722667336464, 0.009931126609444618, 0.06690029799938202, -0.0792231410741806, -0.010253879241645336, -0.010765978135168552, 0.04333420470356941, -0.029997024685144424, 0.05569550022482872, 0.1724521964788437, -0.03874122351408005, -0.039347365498542786, 0.016536908224225044, 0.026186419650912285, -0.048673149198293686, 0.08235491812229156, 0.03646804764866829, 0.059354428201913834, -0.10045748203992844, 0.09188644587993622, 0.1002320870757103, 0.0039356364868581295, 0.029999110847711563, 0.1323169320821762, -0.0779738798737526, -0.0970979705452919, -0.05326148867607117, 0.06478368490934372, -0.11777961254119873, -0.01663161814212799, -0.09164499491453171, -0.0945427194237709, -0.012141851708292961, 0.12607525289058685, 0.07872182875871658, 0.11170949786901474, -0.09641902148723602, -0.03670111671090126, -0.09147769957780838, 0.05137421190738678, -0.034467797726392746, 0.05260249599814415, -0.061976488679647446, 0.11316274106502533, 0.01803116500377655, 0.0004827326338272542, -0.05757958069443703, -0.07269540429115295, -0.16554270684719086, 0.02317528799176216, -0.05414706841111183, 0.022160746157169342, -0.08197957277297974, -0.0036054474767297506, 0.015162445604801178, -0.014068846590816975, 0.005122476723045111, 0.0060508958995342255, -0.03442186862230301, -0.00146681338082999, -0.019080210477113724, 0.0775470957159996, -0.0888189747929573, -0.015499530360102654, 0.014998024329543114, -0.08853261917829514, 0.06202443316578865, 0.02140459604561329, -0.07560200989246368, 0.07974190264940262, -0.14136558771133423, -0.004071236588060856, -0.021847251802682877, 0.04319392517209053, 0.00724010169506073, -0.06812774389982224, 0.015139240771532059, 0.023973088711500168, 0.04937394708395004, 0.03863552585244179, 0.05129775404930115, -0.11559862643480301, -0.001832915237173438, -0.0041874004527926445, -0.08880443871021271, -0.06782576441764832, 0.015198949724435806, 0.07358159124851227, 0.05603847652673721, 0.17545641958713531, -0.08211732655763626, 0.08861978352069855, -0.10476278513669968, -0.018140899017453194, 0.030521614477038383, -0.08419627696275711, -0.05313688516616821, -0.10013669729232788, 0.01900581270456314, -0.017195412889122963, 0.13121715188026428, -0.02993791364133358, 0.08263762295246124, 0.025417737662792206, 0.024023108184337616, 0.09215352684259415, 0.05519764497876167, 0.1491793543100357, 0.04666512459516525, -0.03225264325737953, -0.0018034680979326367, 0.0037583094090223312, -0.012539030984044075, 0.004546519368886948, 0.0545489527285099, 0.04932969808578491, -0.01606171578168869, 0.07904919236898422, 0.029038414359092712, 0.09077579528093338, -0.06462176144123077, 0.028090033680200577, 0.017449606209993362, 0.008699202910065651, -0.05558589845895767, 0.06645425409078598, 0.15993380546569824, -0.07975988835096359, 0.11584963649511337, 0.037023209035396576, -0.10352065414190292, -0.13718436658382416, -0.21220919489860535, -0.0602843277156353, -0.05901140347123146, -0.03146836906671524, -0.14729511737823486, 0.01310656312853098, 0.1004943698644638, 0.032974135130643845, -0.015864668413996696, 0.022646350786089897, 0.08327309042215347, -0.04495495557785034, 0.005938476882874966, -0.04723614454269409, 0.050899773836135864, 0.06836778670549393, 0.051162924617528915, 0.04944939911365509, 0.04774516075849533, 0.05007452890276909, 0.049260515719652176, 0.08545883744955063, 0.001274654408916831, -0.10227260738611221, -0.11640217900276184, 0.020349618047475815, 0.008513372391462326, 0.02770891971886158, 0.12480239570140839, 0.04735568165779114, -0.08447862416505814, -0.002136517083272338, 0.22758834064006805, -0.0805298313498497, -0.11833160370588303, -0.156891867518425, 0.29159459471702576, 0.03818436712026596, 0.05357228219509125, -0.0035360364709049463, -0.06761692464351654, -0.05248383432626724, 0.1833062618970871, 0.12296327203512192, -0.004452371504157782, 0.0039054020307958126, 0.07213283330202103, -0.011482469737529755, 0.03249801695346832, 0.09326603263616562, -0.02214781939983368, 0.2506294846534729, -0.05969396233558655, 0.10301662981510162, -0.020685408264398575, -0.03065490908920765, -0.09430119395256042, 0.08438673615455627, -0.042060308158397675, -0.015744652599096298, -0.04170462116599083, 0.07591334730386734, -0.11736249923706055, 0.03543869033455849, 0.06528057903051376, -0.02677137404680252, -0.10847145318984985, -0.024805670604109764, 0.09372296184301376, -0.0398162417113781, 0.06957661360502243, 0.01040432695299387, -0.011880357749760151, 0.09260989725589752, -0.02058948017656803, -0.02488364651799202, -0.05536671355366707, 0.03933794051408768, -0.03530619665980339, 0.05548027530312538, -0.02345281094312668, 0.11604415625333786, 0.11743488162755966, 0.016561206430196762, -0.044704265892505646, 0.12358115613460541, -0.03522425517439842, -0.047284085303545, 0.08167476952075958, 0.10866093635559082, -0.023611299693584442, 0.1418391466140747, 0.05594569817185402, -0.14776618778705597, 0.05313095450401306, -0.021721797063946724, -0.07074429094791412, -0.032888688147068024, 0.011423781514167786, -0.04935833811759949, 0.06298163533210754, 0.14100544154644012, -0.054499056190252304, -0.028164897114038467, -0.04636334255337715, 0.012559326365590096, -0.010281362570822239, -0.04133826121687889, -0.067349374294281, -0.15483511984348297, 0.0026553256902843714, 0.09059789031744003, 0.04062337055802345, -0.23069506883621216, -0.02104339376091957, 0.01721298135817051, -0.009184980764985085, -0.09905694425106049, 0.060342442244291306, 0.04011520370841026, 0.023741453886032104, -0.048332829028367996, -0.2520994544029236, 0.011945041827857494, 0.0776926577091217, -0.1302192211151123, -0.10313306003808975 ]
null
null
transformers
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. ## Training Data Given a question and paragraph, can the question be answered by the paragraph? The models have been trained on the [GLUE QNLI](https://arxiv.org/abs/1804.07461) dataset, which transformed the [SQuAD dataset](https://rajpurkar.github.io/SQuAD-explorer/) into an NLI task. ## Performance For performance results of this model, see [SBERT.net Pre-trained Cross-Encoder][https://www.sbert.net/docs/pretrained_cross-encoders.html]. ## Usage Pre-trained models can be used like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('model_name') scores = model.predict([('Query1', 'Paragraph1'), ('Query2', 'Paragraph2')]) #e.g. scores = model.predict([('How many people live in Berlin?', 'Berlin had a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.'), ('What is the size of New York?', 'New York City is famous for the Metropolitan Museum of Art.')]) ``` ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('model_name') tokenizer = AutoTokenizer.from_pretrained('model_name') features = tokenizer(['How many people live in Berlin?', 'What is the size of New York?'], ['Berlin had a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = torch.nn.functional.sigmoid(model(**features).logits) print(scores) ```
{"license": "apache-2.0"}
text-classification
cross-encoder/qnli-distilroberta-base
[ "transformers", "pytorch", "jax", "roberta", "text-classification", "arxiv:1804.07461", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[ "1804.07461" ]
[]
TAGS #transformers #pytorch #jax #roberta #text-classification #arxiv-1804.07461 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using SentenceTransformers Cross-Encoder class. ## Training Data Given a question and paragraph, can the question be answered by the paragraph? The models have been trained on the GLUE QNLI dataset, which transformed the SQuAD dataset into an NLI task. ## Performance For performance results of this model, see [URL Pre-trained Cross-Encoder][URL ## Usage Pre-trained models can be used like this: ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library):
[ "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nGiven a question and paragraph, can the question be answered by the paragraph? The models have been trained on the GLUE QNLI dataset, which transformed the SQuAD dataset into an NLI task.", "## Performance\nFor performance results of this model, see [URL Pre-trained Cross-Encoder][URL", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):" ]
[ "TAGS\n#transformers #pytorch #jax #roberta #text-classification #arxiv-1804.07461 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nGiven a question and paragraph, can the question be answered by the paragraph? The models have been trained on the GLUE QNLI dataset, which transformed the SQuAD dataset into an NLI task.", "## Performance\nFor performance results of this model, see [URL Pre-trained Cross-Encoder][URL", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):" ]
[ 60, 34, 50, 23, 14, 33 ]
[ "passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #arxiv-1804.07461 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nGiven a question and paragraph, can the question be answered by the paragraph? The models have been trained on the GLUE QNLI dataset, which transformed the SQuAD dataset into an NLI task.## Performance\nFor performance results of this model, see [URL Pre-trained Cross-Encoder][URL## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):" ]
[ -0.06683853268623352, 0.16760697960853577, -0.0020268643274903297, 0.10583819448947906, 0.1362825334072113, -0.009169505909085274, 0.14919030666351318, 0.10903897881507874, -0.04821023717522621, -0.0023559401743113995, 0.08864589780569077, 0.13430660963058472, -0.0054634143598377705, 0.034476324915885925, 0.02518564648926258, -0.15910016000270844, 0.019955988973379135, -0.006365683861076832, -0.08433472365140915, 0.13067057728767395, 0.0988072082400322, -0.0931401327252388, 0.061112914234399796, 0.013440467417240143, -0.10522468388080597, 0.039205994457006454, -0.023723365738987923, -0.0640815943479538, 0.10041163116693497, 0.0811297595500946, 0.10844796150922775, 0.05113250017166138, 0.06541751325130463, -0.13173463940620422, 0.013813051395118237, 0.07765104621648788, -0.026058517396450043, 0.06262119859457016, 0.01992044597864151, -0.010152021422982216, -0.0005655037239193916, 0.019790975376963615, 0.037233348935842514, 0.026268666610121727, -0.06624831259250641, -0.12578657269477844, -0.0661386027932167, 0.03116391785442829, 0.11600460857152939, 0.14586001634597778, -0.03655621036887169, 0.14105455577373505, -0.11890484392642975, 0.12211491912603378, 0.06447269767522812, -0.2988509237766266, -0.039884645491838455, 0.025126058608293533, 0.08559348434209824, 0.07851284742355347, -0.05456576868891716, 0.01201643142849207, 0.04576081037521362, 0.04329817742109299, 0.014182036742568016, -0.05634068697690964, -0.12855534255504608, 0.08308294415473938, -0.16502605378627777, -0.05542537197470665, 0.1729709357023239, -0.046488940715789795, -0.06271598488092422, -0.08167596906423569, -0.06165126711130142, -0.024913186207413673, 0.0036036984529346228, -0.044971540570259094, 0.021064020693302155, 0.015857728198170662, -0.11956077069044113, -0.033619631081819534, -0.08863332867622375, -0.11785409599542618, -0.057395122945308685, 0.08514110743999481, 0.04625026509165764, 0.042303647845983505, -0.1135537326335907, 0.12018485367298126, 0.08160609751939774, -0.10745636373758316, -0.05947639420628548, -0.02424311637878418, -0.08919071406126022, 0.00890941172838211, -0.10131756961345673, -0.11463272571563721, 0.05269194021821022, 0.1965171843767166, 0.051153093576431274, 0.01824604906141758, -0.011637482792139053, 0.04784025624394417, 0.034574560821056366, 0.21728280186653137, -0.14064449071884155, -0.008191355504095554, 0.035948220640420914, 0.0076000625267624855, -0.012466652318835258, -0.03425968810915947, -0.13695009052753448, -0.0024551786482334137, 0.06762012839317322, 0.09113042056560516, 0.03413432464003563, 0.07016292959451675, 0.010430342517793179, -0.024981321766972542, 0.032159142196178436, -0.08805420994758606, 0.0058694747276604176, -0.015758614987134933, -0.08880799263715744, -0.0153798284009099, 0.09653092920780182, 0.01814018003642559, -0.07191339880228043, -0.05967153236269951, -0.09024703502655029, -0.010409367270767689, -0.07021861523389816, -0.11322527378797531, 0.009601966477930546, 0.0034246263094246387, -0.018639525398612022, -0.1537676304578781, -0.1704186499118805, -0.05199958756566048, 0.03720729053020477, -0.012784007005393505, -0.016444895416498184, -0.0595741905272007, -0.0375097431242466, 0.021855896338820457, -0.009328732267022133, 0.05659256502985954, -0.01174902357161045, 0.0787515714764595, 0.0033956000115722418, 0.07190149277448654, 0.00853913463652134, 0.050659723579883575, -0.1500507891178131, -0.0033963483292609453, -0.020674046128988266, 0.12588021159172058, -0.046421781182289124, 0.03524399548768997, -0.11872247606515884, -0.07679155468940735, 0.047609224915504456, 0.010270624421536922, 0.05977817624807358, 0.12338727712631226, -0.1544087529182434, -0.01592903584241867, 0.1283807009458542, -0.06736008822917938, -0.11897040903568268, 0.09780126065015793, -0.04853251576423645, 0.06412322074174881, 0.04699956998229027, 0.10580247640609741, 0.1538478434085846, -0.07015594840049744, -0.031635187566280365, 0.05567150563001633, -0.0678265392780304, 0.006915123667567968, 0.06452841311693192, 0.043177440762519836, -0.03597727790474892, 0.018048442900180817, -0.09559491276741028, 0.06975944340229034, -0.040747616440057755, -0.09144311398267746, -0.04836862534284592, -0.09867190569639206, 0.05501418933272362, 0.03788171708583832, 0.08055569231510162, 0.03680695220828056, -0.11269854754209518, 0.08248145133256912, 0.08990166336297989, -0.07771731168031693, 0.030897170305252075, -0.1289188414812088, 0.03713952749967575, -0.09294251352548599, 0.025099698454141617, -0.21614325046539307, -0.026741234585642815, 0.00985931046307087, 0.00942805502563715, 0.06489617377519608, 0.01667157933115959, 0.0288349911570549, 0.038976266980171204, 0.016633568331599236, -0.041141342371702194, -0.019100312143564224, -0.007467746268957853, -0.09864334017038345, -0.11688781529664993, -0.039594508707523346, -0.03511656075716019, 0.12005649507045746, -0.09536267817020416, 0.02839803695678711, -0.03473607823252678, 0.057819388806819916, -0.023539695888757706, 0.02971918135881424, 0.0070484415628015995, 0.0396498441696167, -0.06503662467002869, -0.030652714893221855, 0.03127564862370491, 0.014797977171838284, -0.11238161474466324, 0.08407531678676605, -0.07162512093782425, -0.040841154754161835, 0.07799077033996582, -0.0519942082464695, -0.06232871487736702, 0.07276387512683868, -0.03809581324458122, -0.019486282020807266, -0.052465710788965225, -0.024708691984415054, 0.24112772941589355, 0.022290779277682304, 0.1262621432542801, -0.08571378141641617, -0.04610699042677879, -0.024333016946911812, -0.059227995574474335, 0.03396028280258179, 0.091547392308712, 0.022995205596089363, -0.12426958233118057, 0.08966395258903503, 0.09900747239589691, -0.11247168481349945, 0.09898383170366287, -0.024950487539172173, -0.041310057044029236, -0.04107385873794556, 0.00023245986085385084, 0.013472733087837696, 0.02641969732940197, -0.06136687472462654, -0.0008174444083124399, 0.03396273031830788, -0.0007232127827592194, 0.03863777220249176, -0.13767066597938538, 0.013969301246106625, 0.018980583176016808, -0.030771510675549507, -0.035520996898412704, 0.05423792451620102, -0.013325664214789867, 0.0696108415722847, 0.005106380209326744, -0.02996974252164364, 0.05415669083595276, -0.028570914641022682, -0.12829607725143433, 0.22488225996494293, -0.09603536128997803, -0.22774852812290192, -0.13665509223937988, 0.024677246809005737, -0.02399257756769657, -0.00945186335593462, 0.07520220428705215, -0.10349368304014206, -0.02139909379184246, -0.0711468756198883, -0.04644950106739998, -0.016865303739905357, -0.02589220553636551, -0.07672198116779327, 0.004928263369947672, 0.027016986161470413, -0.15316550433635712, 0.042099665850400925, 0.0055327038280665874, -0.08691573143005371, 0.03167159855365753, -0.0018628028919920325, 0.11148902028799057, 0.1663927584886551, -0.019008580595254898, 0.0245186910033226, -0.015546257607638836, 0.2309579700231552, -0.08556707948446274, -0.016719231382012367, 0.16801564395427704, 0.017369689419865608, 0.04717768356204033, 0.09610187262296677, 0.009077681228518486, -0.054042570292949677, 0.04009239003062248, 0.00898192822933197, -0.06670462340116501, -0.1967061311006546, -0.08733046054840088, -0.05435071885585785, -0.0019963670056313276, 0.058085277676582336, 0.04592158645391464, 0.044658634811639786, 0.03303653001785278, 0.016391027718782425, 0.06455487012863159, -0.042085062712430954, 0.07370206713676453, 0.0832008495926857, 0.00898972898721695, 0.1295560598373413, -0.059773411601781845, -0.08394615352153778, 0.03776200860738754, 0.0699373334646225, 0.20093883574008942, -0.03501866012811661, 0.07046154886484146, 0.09120561927556992, 0.10920114815235138, 0.0713454857468605, 0.0436578243970871, -0.02143164537847042, 0.014635814353823662, 0.0036580278538167477, -0.059523992240428925, -0.012287215329706669, 0.06339888274669647, -0.016284968703985214, -0.019208626821637154, -0.05971582978963852, 0.0047233146615326405, 0.055831752717494965, 0.17679768800735474, 0.02101622335612774, -0.23761917650699615, -0.0456165075302124, 0.0186733640730381, -0.04656938090920448, -0.013333170674741268, 0.04500093683600426, 0.05143100395798683, -0.14000338315963745, -0.06039545685052872, -0.05044795200228691, 0.125654399394989, -0.026891158893704414, -0.005294478498399258, 0.021189045161008835, -0.001703764428384602, 0.02745729498565197, 0.13703376054763794, -0.2667916417121887, 0.11674847453832626, 0.009572298265993595, 0.05008187144994736, -0.053387969732284546, 0.0274791456758976, 0.02256331779062748, 0.10295333713293076, 0.12725406885147095, 0.0030123982578516006, 0.010877483524382114, -0.09620069712400436, -0.04922156780958176, 0.037676043808460236, 0.05891343578696251, 0.0051406691782176495, 0.08297355473041534, -0.020884307101368904, 0.007325014099478722, -0.005753623321652412, 0.05122080445289612, -0.08993537724018097, -0.1444888710975647, 0.031164053827524185, 0.02360234037041664, 0.07756015658378601, -0.049443237483501434, -0.05392201244831085, -0.02922237105667591, 0.1137341856956482, -0.13104788959026337, -0.08347545564174652, -0.10555845499038696, 0.042328525334596634, 0.08906083554029465, -0.08547109365463257, 0.003047098172828555, -0.012066217139363289, 0.05198335275053978, 0.0011178147979080677, -0.10362903773784637, 0.01988779567182064, -0.06434579938650131, -0.09776798635721207, -0.053710877895355225, 0.005933969747275114, 0.027272263541817665, 0.0423470102250576, 0.02857867442071438, 0.033049196004867554, -0.11053433269262314, -0.12496538460254669, -0.06572171300649643, -0.014415023848414421, -0.000938532582949847, 0.048962172120809555, -0.04123404994606972, 0.02246498130261898, -0.05426478013396263, 0.019125107675790787, 0.18611574172973633, 0.027716651558876038, -0.12143387645483017, 0.07646087557077408, 0.1571614146232605, -0.07852497696876526, -0.2750796377658844, -0.08864019811153412, 0.03706253319978714, 0.032724834978580475, -0.07439777255058289, -0.14852474629878998, 0.10069870948791504, 0.026417113840579987, -0.028291059657931328, -0.03525562956929207, -0.3468773663043976, -0.10672072321176529, 0.11028289794921875, 0.06790085136890411, 0.017017818987369537, -0.14152270555496216, -0.026377499103546143, -0.02542470395565033, -0.053526412695646286, 0.11842072755098343, -0.15185602009296417, 0.10908640921115875, -0.0077660842798650265, 0.053971562534570694, 0.0011061321711167693, -0.05203355476260185, 0.08123216032981873, 0.023336492478847504, 0.051261600106954575, -0.028821000829339027, -0.023190610110759735, 0.08404416590929031, -0.034598756581544876, 0.11941490322351456, -0.05809007212519646, 0.12276197224855423, -0.12252192944288254, 0.00017233568360097706, -0.05280499905347824, 0.009558195248246193, -0.016701793298125267, -0.07055538147687912, -0.0733523741364479, 0.04851476475596428, 0.051897041499614716, -0.0176771879196167, 0.0731092169880867, -0.036747559905052185, 0.033179160207509995, 0.16419610381126404, 0.12259941548109055, 0.013128727674484253, -0.12156126648187637, -0.013373380526900291, -0.031431205570697784, 0.14170706272125244, -0.1881362348794937, 0.08219043910503387, 0.10742058604955673, 0.0394715778529644, 0.1096542626619339, 0.07886325567960739, -0.03203693777322769, 0.02590557187795639, 0.0464790053665638, -0.12411253899335861, -0.10097771883010864, -0.02735777758061886, 0.01979765295982361, -0.0677456334233284, 0.04362274333834648, 0.13313435018062592, -0.08146915584802628, -0.030428923666477203, 0.030553290620446205, 0.0063253468833863735, -0.05826461687684059, 0.15519841015338898, 0.1079331636428833, 0.05097516253590584, -0.0825340524315834, 0.08648589998483658, 0.0829111784696579, -0.014632360078394413, 0.060698501765728, 0.06954222172498703, -0.110160693526268, -0.10051184147596359, -0.07892830669879913, 0.1455468237400055, -0.13397765159606934, -0.0733928233385086, -0.1096741259098053, -0.03487095609307289, 0.03385604918003082, 0.0701722726225853, 0.050564516335725784, 0.06091950461268425, -0.08757323026657104, -0.043117884546518326, -0.10297101736068726, 0.10449744015932083, 0.015996556729078293, 0.014125491492450237, -0.0862927958369255, 0.09464609622955322, 0.025278368964791298, 0.12302318960428238, -0.061422429978847504, -0.053639598190784454, -0.1109188124537468, 0.051465753465890884, -0.1108650267124176, 0.005700241308659315, -0.00804917886853218, -0.00954259093850851, -0.011127916164696217, -0.01819438487291336, -0.03459872305393219, 0.05927010253071785, -0.04982389509677887, -0.027976613491773605, -0.017083333805203438, 0.05314554274082184, -0.14227302372455597, -0.03162020817399025, 0.012343761511147022, -0.030025340616703033, 0.06660739332437515, 0.03452412784099579, -0.05025063455104828, 0.023977013304829597, -0.0691557377576828, -0.029691459611058235, -0.039703283458948135, 0.041693612933158875, 0.04182566702365875, -0.07525695115327835, 0.03390820696949959, 0.0129158990457654, 0.008893415331840515, 0.01731681637465954, 0.04977913200855255, -0.09505146741867065, 0.03201073408126831, -0.019522910937666893, 0.033274196088314056, -0.10332247614860535, 0.06459403783082962, 0.012841559015214443, 0.11758334934711456, 0.16421903669834137, -0.0752158910036087, 0.09582366794347763, -0.12311281263828278, -0.022933954373002052, 0.0294441320002079, -0.02162560448050499, -0.09729430824518204, -0.09793496131896973, 0.014731362462043762, -0.06328652054071426, 0.15613850951194763, 0.048196278512477875, 0.0923430398106575, 0.006857549771666527, 0.09729024022817612, 0.06867080181837082, 0.008483966812491417, 0.1500026285648346, 0.028926605358719826, 0.01021699421107769, 0.00305820326320827, 0.046558089554309845, -0.011440226808190346, 0.06908919662237167, 0.014021030627191067, 0.10226263105869293, 0.0831887423992157, 0.10516121983528137, 0.021369928494095802, -0.012843471020460129, -0.08477047830820084, -0.05878866836428642, -0.02031048759818077, 0.06092534959316254, -0.06337404251098633, 0.1027132049202919, 0.13144825398921967, -0.08360836654901505, 0.058809250593185425, 0.024886129423975945, -0.08495844900608063, -0.09705867618322372, -0.051588866859674454, -0.05258803442120552, -0.1261771321296692, 0.011857743375003338, -0.12238503992557526, 0.012698633596301079, -0.002562653971835971, 0.014114049263298512, -0.021884020417928696, 0.08085934072732925, 0.014131204225122929, -0.04646634683012962, 0.010727196000516415, -0.040407732129096985, 0.08823232352733612, 0.022766225039958954, 0.061805061995983124, 0.03990910202264786, -0.013083919882774353, 0.054742276668548584, 0.04084904491901398, 0.02544950507581234, 0.05082608386874199, -0.07746461033821106, -0.04864777997136116, -0.02018146961927414, 0.048653069883584976, 0.013535905629396439, 0.13341858983039856, 0.05444905906915665, -0.022763529792428017, -0.01780889183282852, 0.23645195364952087, -0.10533805936574936, -0.12192191928625107, -0.1653653234243393, 0.26937630772590637, 0.023285314440727234, 0.028923669829964638, -0.017611967399716377, -0.0957862064242363, -0.04759334772825241, 0.28645163774490356, 0.14203692972660065, -0.08690868318080902, -0.006080891005694866, 0.03610460087656975, -0.009376497939229012, 0.013545478694140911, 0.10087932646274567, 0.07736768573522568, 0.2724824845790863, -0.0824165940284729, 0.01594119518995285, -0.07852651923894882, -0.04358866810798645, -0.04872411862015724, 0.11386265605688095, -0.014412621967494488, -0.04961789771914482, -0.051188498735427856, 0.0750098004937172, -0.13491606712341309, -0.018695546314120293, 0.0046454379335045815, -0.05705263838171959, -0.09829151630401611, -0.01922590471804142, 0.08067891746759415, -0.015223979949951172, 0.047055572271347046, -0.01515207439661026, 0.03697894513607025, 0.13114656507968903, -0.012056717649102211, -0.06255386769771576, -0.0382801815867424, 0.05431593582034111, -0.07418250292539597, 0.13452599942684174, -0.02140597626566887, 0.11451972275972366, 0.12813307344913483, 0.039515167474746704, -0.10494865477085114, 0.08975453674793243, 0.01995856687426567, -0.10715493559837341, 0.044772952795028687, 0.04817403852939606, 0.019841190427541733, 0.058998994529247284, 0.01528418343514204, -0.07097680121660233, 0.060279808938503265, -0.06948189437389374, -0.0052827694453299046, -0.15087740123271942, 0.010146057233214378, -0.03500804305076599, 0.12241935729980469, 0.14494851231575012, -0.06073346361517906, -0.0038054739125072956, -0.046093132346868515, 0.023543426766991615, -0.008556652814149857, -0.028448930010199547, -0.019037554040551186, -0.12259189784526825, 0.005629850085824728, 0.010194981470704079, -0.00324035226367414, -0.21996444463729858, -0.0404871441423893, -0.00033505918690934777, -0.05897333845496178, 0.024518771097064018, 0.11740656942129135, 0.12440529465675354, 0.039086584001779556, -0.020194856449961662, -0.13804538547992706, -0.004218795336782932, 0.08220471441745758, -0.12070762366056442, -0.10985631495714188 ]
null
null
transformers
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. ## Training Data Given a question and paragraph, can the question be answered by the paragraph? The models have been trained on the [GLUE QNLI](https://arxiv.org/abs/1804.07461) dataset, which transformed the [SQuAD dataset](https://rajpurkar.github.io/SQuAD-explorer/) into an NLI task. ## Performance For performance results of this model, see [SBERT.net Pre-trained Cross-Encoder][https://www.sbert.net/docs/pretrained_cross-encoders.html]. ## Usage Pre-trained models can be used like this: ```python from sentence_transformers import CrossEncoder model = CrossEncoder('model_name') scores = model.predict([('Query1', 'Paragraph1'), ('Query2', 'Paragraph2')]) #e.g. scores = model.predict([('How many people live in Berlin?', 'Berlin had a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.'), ('What is the size of New York?', 'New York City is famous for the Metropolitan Museum of Art.')]) ``` ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library): ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model = AutoModelForSequenceClassification.from_pretrained('model_name') tokenizer = AutoTokenizer.from_pretrained('model_name') features = tokenizer(['How many people live in Berlin?', 'What is the size of New York?'], ['Berlin had a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt") model.eval() with torch.no_grad(): scores = torch.nn.functional.sigmoid(model(**features).logits) print(scores) ```
{"license": "apache-2.0"}
text-classification
cross-encoder/qnli-electra-base
[ "transformers", "pytorch", "electra", "text-classification", "arxiv:1804.07461", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[ "1804.07461" ]
[]
TAGS #transformers #pytorch #electra #text-classification #arxiv-1804.07461 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using SentenceTransformers Cross-Encoder class. ## Training Data Given a question and paragraph, can the question be answered by the paragraph? The models have been trained on the GLUE QNLI dataset, which transformed the SQuAD dataset into an NLI task. ## Performance For performance results of this model, see [URL Pre-trained Cross-Encoder][URL ## Usage Pre-trained models can be used like this: ## Usage with Transformers AutoModel You can use the model also directly with Transformers library (without SentenceTransformers library):
[ "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nGiven a question and paragraph, can the question be answered by the paragraph? The models have been trained on the GLUE QNLI dataset, which transformed the SQuAD dataset into an NLI task.", "## Performance\nFor performance results of this model, see [URL Pre-trained Cross-Encoder][URL", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):" ]
[ "TAGS\n#transformers #pytorch #electra #text-classification #arxiv-1804.07461 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nGiven a question and paragraph, can the question be answered by the paragraph? The models have been trained on the GLUE QNLI dataset, which transformed the SQuAD dataset into an NLI task.", "## Performance\nFor performance results of this model, see [URL Pre-trained Cross-Encoder][URL", "## Usage\n\nPre-trained models can be used like this:", "## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):" ]
[ 57, 34, 50, 23, 14, 33 ]
[ "passage: TAGS\n#transformers #pytorch #electra #text-classification #arxiv-1804.07461 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nGiven a question and paragraph, can the question be answered by the paragraph? The models have been trained on the GLUE QNLI dataset, which transformed the SQuAD dataset into an NLI task.## Performance\nFor performance results of this model, see [URL Pre-trained Cross-Encoder][URL## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):" ]
[ -0.08139800280332565, 0.16041189432144165, -0.0008130564237944782, 0.10608351975679398, 0.15591081976890564, -0.011612229980528355, 0.12859240174293518, 0.09285445511341095, -0.023193156346678734, -0.002844154601916671, 0.09732834994792938, 0.15097904205322266, -0.00794843677431345, 0.047209952026605606, 0.0248067919164896, -0.16479501128196716, 0.03332924470305443, 0.0017449820879846811, -0.09433584660291672, 0.12972529232501984, 0.10567570477724075, -0.10027066618204117, 0.05378144234418869, 0.03262406960129738, -0.10094267129898071, 0.034859634935855865, -0.032263610512018204, -0.08513280004262924, 0.1250745803117752, 0.08892641216516495, 0.10992085933685303, 0.05206308886408806, 0.08884402364492416, -0.12824943661689758, 0.013646877370774746, 0.06721924990415573, -0.035412028431892395, 0.06213344633579254, 0.007376992143690586, 0.01839068904519081, -0.0016585626872256398, 0.05138326808810234, 0.04991662874817848, 0.019196275621652603, -0.06860174238681793, -0.11976032704114914, -0.05249357968568802, 0.028090858832001686, 0.12185889482498169, 0.13583965599536896, -0.03791302070021629, 0.1663469523191452, -0.141435444355011, 0.10164438188076019, 0.040629565715789795, -0.28557613492012024, -0.027670511975884438, 0.013138815760612488, 0.10049527138471603, 0.05972932651638985, -0.03860277310013771, 0.0063815913163125515, 0.05883136764168739, 0.041739050298929214, 0.02147262915968895, -0.03436869755387306, -0.10826406627893448, 0.09091263264417648, -0.17723731696605682, -0.06637513637542725, 0.17933069169521332, -0.027695955708622932, -0.06187807396054268, -0.08125351369380951, -0.06469547003507614, -0.037126533687114716, 0.022455966100096703, -0.06079183518886566, 0.010594941675662994, 0.008495462127029896, -0.10168193280696869, -0.019162356853485107, -0.10004673898220062, -0.11831222474575043, -0.06680125743150711, 0.1075991690158844, 0.056893013417720795, 0.033461134880781174, -0.08893516659736633, 0.12554068863391876, 0.07643282413482666, -0.09495841711759567, -0.06644738465547562, -0.021701984107494354, -0.09872200340032578, -0.0042851888574659824, -0.11015177518129349, -0.09532145410776138, 0.05226999148726463, 0.21868574619293213, 0.07125167548656464, 0.008366395719349384, 0.026703793555498123, 0.046773988753557205, 0.0516340397298336, 0.2127561718225479, -0.1531628668308258, 0.02549305371940136, 0.03257477283477783, 0.009071044623851776, -0.012134011834859848, -0.04220334067940712, -0.156703919172287, -0.010980642400681973, 0.07549683004617691, 0.08833886682987213, 0.012023506686091423, 0.06494312733411789, 0.00677122687920928, -0.03589637205004692, 0.05761205032467842, -0.08330338448286057, -0.005388041026890278, -0.02135605923831463, -0.10772255808115005, -0.017310362309217453, 0.09681075811386108, 0.0329165942966938, -0.06894959509372711, -0.060483358800411224, -0.09771394729614258, -0.019486168399453163, -0.06476064771413803, -0.11602319777011871, 0.005267797503620386, 0.0018403249559924006, -0.005286057014018297, -0.17412731051445007, -0.17291617393493652, -0.0404776856303215, 0.037816789001226425, -0.003412348683923483, -0.041513774544000626, -0.062021076679229736, -0.017630161717534065, 0.009625636972486973, -0.017713600769639015, 0.02181921899318695, -0.01290993019938469, 0.08329292386770248, 0.022692644968628883, 0.06923427432775497, -0.0011288134846836329, 0.05495980381965637, -0.1454368680715561, -0.011578001081943512, -0.014261730946600437, 0.11030662059783936, -0.032760899513959885, 0.020392777398228645, -0.11172357946634293, -0.09672924131155014, 0.07478238642215729, 0.02518833801150322, 0.06091412529349327, 0.12131917476654053, -0.13158130645751953, -0.023685414344072342, 0.1157577782869339, -0.08388425409793854, -0.12303067743778229, 0.1010332778096199, -0.04843950644135475, 0.05426306650042534, 0.04906899482011795, 0.07917223125696182, 0.13783039152622223, -0.07077278196811676, -0.038807645440101624, 0.024996019899845123, -0.07197368144989014, -0.00534677691757679, 0.0662134662270546, 0.049276139587163925, -0.052924495190382004, 0.015095002017915249, -0.11117643117904663, 0.08332133293151855, -0.04751208797097206, -0.09073854237794876, -0.04353620111942291, -0.10225661098957062, 0.10020554065704346, 0.03785790503025055, 0.07416708767414093, 0.04055711254477501, -0.12684866786003113, 0.10639164596796036, 0.10764288902282715, -0.08344073593616486, 0.015291821211576462, -0.15174749493598938, 0.048125259578228, -0.1080242246389389, 0.010176953859627247, -0.20498065650463104, -0.04084979370236397, 0.00762198306620121, -0.03515848144888878, 0.07704205811023712, 0.0425899364054203, 0.011598322540521622, 0.056836869567632675, 0.022345248609781265, -0.05848044529557228, -0.05022143945097923, 0.0015535458223894238, -0.08317416906356812, -0.09390532970428467, -0.05188380926847458, -0.03076985292136669, 0.12782451510429382, -0.12303172051906586, 0.030375568196177483, -0.06593065708875656, 0.04928303509950638, -0.041594602167606354, 0.041681885719299316, -0.008600437082350254, 0.04327097907662392, -0.07353381812572479, -0.0273295845836401, 0.028205396607518196, 0.0183856263756752, -0.11378183960914612, 0.047134820371866226, -0.09562186151742935, -0.034481946378946304, 0.07185937464237213, -0.06827595084905624, -0.06740511208772659, 0.06567925214767456, -0.02990131638944149, -0.019254470244050026, -0.07681123912334442, -0.03079714998602867, 0.24600520730018616, 0.0182985570281744, 0.1291915625333786, -0.08430833369493484, -0.03867870196700096, -0.023686926811933517, -0.05832396820187569, 0.03359166532754898, 0.07285743951797485, 0.09185261279344559, -0.12377593666315079, 0.08792320638895035, 0.08277741819620132, -0.12243568897247314, 0.09036372601985931, -0.0181779395788908, -0.036820460110902786, -0.040781170129776, -0.012042832560837269, 0.005123232491314411, 0.02291913516819477, -0.07384143769741058, 0.013077628798782825, 0.03172018751502037, -0.003997522406280041, 0.037272077053785324, -0.1399163156747818, 0.00778957549482584, 0.026867931708693504, -0.03185276687145233, -0.04662272334098816, 0.054373349994421005, -0.0196001548320055, 0.06130257621407509, 0.0036240865010768175, -0.054306965321302414, 0.0679377019405365, -0.016273682937026024, -0.126485675573349, 0.2272968739271164, -0.09432802349328995, -0.23742002248764038, -0.14099344611167908, 0.049827829003334045, -0.022502955049276352, 0.00009721218521008268, 0.0777333676815033, -0.10286743938922882, -0.03366447612643242, -0.06213495135307312, -0.04504316672682762, -0.0036299435887485743, -0.027430865913629532, -0.06945542991161346, -0.010447622276842594, 0.02490893192589283, -0.15674988925457, 0.04706132784485817, 0.007023976184427738, -0.08626218140125275, 0.04185148701071739, 0.0029891638550907373, 0.1067970022559166, 0.1960018426179886, -0.02413005568087101, 0.02651907689869404, -0.00726644042879343, 0.23054414987564087, -0.08604064583778381, -0.021201562136411667, 0.18044254183769226, 0.01872771978378296, 0.043656788766384125, 0.10356315970420837, 0.008468960411846638, -0.057315099984407425, 0.049287889152765274, -0.004549901932477951, -0.06690163165330887, -0.19151751697063446, -0.11045792698860168, -0.03948434069752693, -0.021297533065080643, 0.0694139152765274, 0.04318515211343765, 0.08081472665071487, 0.04348184913396835, 0.018074190244078636, 0.03763613477349281, -0.05827733129262924, 0.06848836690187454, 0.13207189738750458, 0.02116619050502777, 0.1282666027545929, -0.04784670099616051, -0.09669385850429535, 0.03337545320391655, 0.060967229306697845, 0.19964778423309326, -0.01974724791944027, 0.07059352844953537, 0.08233581483364105, 0.11265186965465546, 0.08237448334693909, 0.0635192021727562, 0.004001186229288578, 0.010697315447032452, 0.0038907343987375498, -0.03967567905783653, -0.009146814234554768, 0.06405694037675858, 0.008191139437258244, -0.02752971649169922, -0.07786457985639572, 0.002938255900517106, 0.03741166740655899, 0.16207660734653473, 0.026552166789770126, -0.2654740810394287, -0.053377170115709305, -0.0003798084217123687, -0.05867813527584076, -0.017580129206180573, 0.0393492616713047, 0.03652188926935196, -0.1541680246591568, -0.05263804644346237, -0.048492707312107086, 0.11309026926755905, -0.020219622179865837, -0.00605420209467411, 0.020747799426317215, 0.006953320931643248, 0.031044241040945053, 0.14369075000286102, -0.2502667307853699, 0.12488118559122086, -0.007276750635355711, 0.045106690376996994, -0.05646171793341637, 0.028047095984220505, 0.02348932810127735, 0.10600898414850235, 0.10934498906135559, -0.006311903707683086, -0.007974495179951191, -0.1315472275018692, -0.03186040744185448, 0.04728011414408684, 0.030823886394500732, -0.005594308953732252, 0.08862561732530594, -0.027782952412962914, 0.01050530094653368, 0.0032611035276204348, 0.053817085921764374, -0.09527373313903809, -0.14110241830348969, 0.03590252622961998, 0.02212880179286003, 0.11458145081996918, -0.046901166439056396, -0.07091386616230011, -0.04997522383928299, 0.08300955593585968, -0.1333480328321457, -0.08625634759664536, -0.09907760471105576, 0.05868001654744148, 0.09242206811904907, -0.09274252504110336, 0.04083194583654404, -0.0013828100636601448, 0.06681860238313675, 0.0022394759580492973, -0.07944504171609879, 0.027794383466243744, -0.07028324902057648, -0.10587794333696365, -0.03641572967171669, -0.016785068437457085, 0.048428975045681, 0.030425358563661575, 0.048514340072870255, 0.03301828354597092, -0.10775123536586761, -0.12688679993152618, -0.07203947752714157, -0.034519847482442856, -0.007783565204590559, 0.057889729738235474, -0.03256085515022278, 0.02026546187698841, -0.04618046432733536, -0.0024980579037219286, 0.19505828619003296, 0.039525579661130905, -0.11683136224746704, 0.062615305185318, 0.17444665729999542, -0.06722389161586761, -0.2509252727031708, -0.11400186270475388, 0.04470352455973625, 0.0421157069504261, -0.03936753794550896, -0.16495324671268463, 0.10933112353086472, 0.04266039654612541, -0.029854783788323402, -0.035246819257736206, -0.3328753709793091, -0.10231165587902069, 0.13586847484111786, 0.0615118071436882, 0.0058561060577631, -0.11881396919488907, -0.014703931286931038, -0.033282458782196045, -0.03160635009407997, 0.12155454605817795, -0.09806844592094421, 0.10469990223646164, -0.0024012746289372444, 0.06324615329504013, 0.0019245477160438895, -0.047895826399326324, 0.0962444469332695, 0.027136772871017456, 0.040699537843465805, -0.03716219589114189, -0.028053686022758484, 0.04991461709141731, -0.025721397250890732, 0.12406432628631592, -0.0566718615591526, 0.11353502422571182, -0.12097983062267303, -0.022193288430571556, -0.04509788751602173, 0.02509419061243534, -0.005740065593272448, -0.07350139319896698, -0.07065974920988083, 0.04291513189673424, 0.04880080372095108, -0.008129912428557873, 0.050113365054130554, -0.028356170281767845, 0.03973136097192764, 0.15196964144706726, 0.13187557458877563, 0.004879108630120754, -0.12759815156459808, -0.017691902816295624, -0.026403378695249557, 0.14904849231243134, -0.14711865782737732, 0.08906105905771255, 0.11451581865549088, 0.02583991177380085, 0.12089399993419647, 0.08053065091371536, -0.03350910171866417, 0.041785646229982376, 0.041522953659296036, -0.10840277373790741, -0.09087944775819778, -0.024622449651360512, 0.04816613718867302, -0.057327330112457275, 0.05016898363828659, 0.12405502051115036, -0.08611401915550232, -0.040513891726732254, 0.025530533865094185, 0.0004927467089146376, -0.04107546806335449, 0.1607077717781067, 0.11660023033618927, 0.05957872048020363, -0.07770217210054398, 0.07625927031040192, 0.09618300199508667, -0.019504429772496223, 0.06401363760232925, 0.04222135618329048, -0.10588940978050232, -0.1103757917881012, -0.07960686087608337, 0.1384643018245697, -0.19781872630119324, -0.09469494968652725, -0.11905554682016373, -0.03372616693377495, 0.03942948579788208, 0.0783766359090805, 0.060027673840522766, 0.04003910347819328, -0.0882609635591507, -0.03351844474673271, -0.10916462540626526, 0.0971139594912529, 0.03896551579236984, 0.00898434966802597, -0.10211482644081116, 0.09407038986682892, 0.03678252920508385, 0.13014233112335205, -0.06049257144331932, -0.06628801673650742, -0.0947970300912857, 0.054597947746515274, -0.11082018911838531, 0.006630706135183573, -0.009182666428387165, -0.003364246105775237, -0.0013000423787161708, -0.022072691470384598, -0.02563532255589962, 0.05659262463450432, -0.05722765251994133, -0.016132229939103127, -0.003746338654309511, 0.05047748237848282, -0.13491074740886688, -0.03542096167802811, 0.01846247911453247, -0.030056245625019073, 0.07684505730867386, 0.016060970723628998, -0.04867076128721237, 0.03700822964310646, -0.06371543556451797, -0.026278939098119736, -0.029950659722089767, 0.052899401634931564, 0.029590507969260216, -0.10082758218050003, 0.0479368157684803, 0.010732253082096577, 0.003260029247030616, 0.004641746170818806, 0.025631263852119446, -0.09639104455709457, 0.02236383967101574, -0.021108347922563553, 0.029366178438067436, -0.10731998085975647, 0.0533750057220459, -0.0046320208348333836, 0.10615136474370956, 0.1785527467727661, -0.07071241736412048, 0.08788717538118362, -0.10957670211791992, -0.015601745806634426, 0.015768053010106087, -0.006543227005749941, -0.09936949610710144, -0.09540324658155441, 0.00806848332285881, -0.056448034942150116, 0.15897256135940552, 0.029129894450306892, 0.10313519090414047, 0.004442616365849972, 0.10726684331893921, 0.06152825430035591, 0.008106076158583164, 0.17940543591976166, 0.038186486810445786, 0.02091521956026554, -0.008555014617741108, 0.04223451390862465, -0.01013815589249134, 0.03904852271080017, 0.015664000064134598, 0.07562847435474396, 0.0323067307472229, 0.09869390726089478, 0.024889444932341576, -0.026704302057623863, -0.09224668890237808, -0.07329224795103073, -0.01579851657152176, 0.05143291503190994, -0.03889162838459015, 0.1086818054318428, 0.1208462193608284, -0.07150524109601974, 0.045061156153678894, 0.029461825266480446, -0.0804935172200203, -0.10145817697048187, -0.07401122152805328, -0.05860727280378342, -0.14500443637371063, 0.014304226264357567, -0.11121705174446106, 0.00839908141642809, 0.03192901238799095, 0.012442899867892265, -0.03689589723944664, 0.07212062925100327, 0.01241342257708311, -0.04368603974580765, -0.011692216619849205, -0.051914338022470474, 0.08563107252120972, 0.019892370328307152, 0.050665464252233505, 0.026905560865998268, 0.012094538658857346, 0.052553921937942505, 0.02468538098037243, 0.0013424432836472988, 0.058701109141111374, -0.06811324506998062, -0.04319026693701744, -0.027799414470791817, 0.04126004874706268, 0.014343496412038803, 0.11326438188552856, 0.040522441267967224, -0.020508119836449623, -0.013906619511544704, 0.21895933151245117, -0.09283414483070374, -0.1323097199201584, -0.17311827838420868, 0.2407122254371643, 0.04464066028594971, 0.03343944624066353, -0.0007162546389736235, -0.0947662889957428, -0.055648159235715866, 0.25857871770858765, 0.10800667107105255, -0.0696985051035881, -0.007954142056405544, 0.021195219829678535, -0.01281147450208664, 0.013393421657383442, 0.09773919731378555, 0.07339742034673691, 0.2580370306968689, -0.09211987257003784, 0.009003457613289356, -0.0889134630560875, -0.031529951840639114, -0.04620738700032234, 0.086461141705513, -0.016097228974103928, -0.048674214631319046, -0.05010990798473358, 0.08224178850650787, -0.12877905368804932, -0.004731197375804186, -0.0053450921550393105, -0.0320814773440361, -0.10141699016094208, -0.025547554716467857, 0.07512250542640686, -0.011001619510352612, 0.04268690571188927, -0.026301449164748192, 0.022380536422133446, 0.137967050075531, -0.0008905127760954201, -0.07643347233533859, -0.03970276564359665, 0.07141333073377609, -0.03271233290433884, 0.1386837214231491, -0.023238688707351685, 0.13844414055347443, 0.12417110800743103, 0.04076496511697769, -0.08828112483024597, 0.11496991664171219, 0.027894401922822, -0.08750739693641663, 0.07264227420091629, 0.02900698035955429, 0.023894550278782845, 0.05439811572432518, 0.00248232064768672, -0.10540831834077835, 0.0592973455786705, -0.07049877196550369, -0.0013822696637362242, -0.1395513415336609, 0.030802344903349876, -0.032772623002529144, 0.12444650381803513, 0.11269434541463852, -0.06748658418655396, -0.02025596797466278, -0.052902814000844955, 0.043637823313474655, 0.0024602997582405806, -0.029597505927085876, -0.013550655916333199, -0.13677170872688293, 0.012596950866281986, -0.04356306418776512, -0.0036949694622308016, -0.18841518461704254, -0.03744067624211311, -0.010731018148362637, -0.06564662605524063, 0.02741948887705803, 0.11450342833995819, 0.13345317542552948, 0.04000673070549965, -0.029460368677973747, -0.06212691590189934, -0.032210007309913635, 0.08422917127609253, -0.13290034234523773, -0.11524412781000137 ]
null
null
transformers
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. ## Training Data This model was trained on the [Quora Duplicate Questions](https://www.quora.com/q/quoradata/First-Quora-Dataset-Release-Question-Pairs) dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates. Note: The model is not suitable to estimate the similarity of questions, e.g. the two questions "How to learn Java" and "How to learn Python" will result in a rahter low score, as these are not duplicates. ## Usage and Performance Pre-trained models can be used like this: ``` from sentence_transformers import CrossEncoder model = CrossEncoder('model_name') scores = model.predict([('Question 1', 'Question 2'), ('Question 3', 'Question 4')]) ``` You can use this model also without sentence_transformers and by just using Transformers ``AutoModel`` class
{"license": "apache-2.0"}
text-classification
cross-encoder/quora-distilroberta-base
[ "transformers", "pytorch", "jax", "roberta", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using SentenceTransformers Cross-Encoder class. ## Training Data This model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates. Note: The model is not suitable to estimate the similarity of questions, e.g. the two questions "How to learn Java" and "How to learn Python" will result in a rahter low score, as these are not duplicates. ## Usage and Performance Pre-trained models can be used like this: You can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class
[ "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.", "## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ "TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.", "## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ 48, 34, 93, 39 ]
[ "passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ -0.028236351907253265, 0.02298957295715809, -0.0028521628119051456, 0.0646425411105156, 0.11765683442354202, 0.023299384862184525, 0.16609442234039307, 0.12721982598304749, -0.02589987777173519, -0.002497537527233362, 0.06902404129505157, 0.09041360765695572, 0.03389200568199158, 0.01736462675035, -0.0756693035364151, -0.25257617235183716, 0.05000140145421028, 0.001953172730281949, 0.005523595958948135, 0.09553001075983047, 0.1380188912153244, -0.06342750787734985, 0.06385538727045059, 0.03808078169822693, -0.1387544721364975, 0.029387259855866432, 0.007435559760779142, -0.07278154790401459, 0.1284724920988083, 0.06276270002126694, 0.12536683678627014, 0.04794098809361458, 0.06633620709180832, -0.1390024721622467, 0.010272445157170296, 0.04691065847873688, 0.030382536351680756, 0.042663365602493286, 0.05961441248655319, -0.0883123055100441, -0.015024879947304726, 0.13564015924930573, 0.11921955645084381, 0.049067553132772446, -0.11800447851419449, -0.1697719544172287, -0.09260506927967072, -0.007557870354503393, 0.19177691638469696, 0.0677613914012909, -0.019196776673197746, 0.17920920252799988, -0.16551508009433746, 0.042527250945568085, 0.11477108299732208, -0.29275941848754883, -0.004422552417963743, 0.15177641808986664, 0.027819789946079254, 0.0426139235496521, -0.035739149898290634, -0.025486495345830917, 0.04099881276488304, 0.06852063536643982, -0.023120788857340813, -0.02526349201798439, -0.027673279866576195, 0.019225681200623512, -0.13329489529132843, -0.05995437130331993, 0.16933511197566986, -0.03704995661973953, -0.05897475406527519, -0.1055462658405304, -0.039407879114151, 0.11085079610347748, -0.008900068700313568, -0.11178149282932281, -0.016698947176337242, 0.05211774632334709, -0.08724465221166611, 0.030446620658040047, -0.07607933133840561, -0.10109981149435043, -0.07195515185594559, 0.023931683972477913, 0.07561454176902771, 0.05545526742935181, -0.13705310225486755, 0.11136855185031891, 0.0016995988553389907, -0.09120044857263565, -0.040101051330566406, -0.03573271259665489, -0.09327655285596848, -0.02613293193280697, -0.13934822380542755, -0.1651875376701355, 0.0202079638838768, 0.2592458128929138, 0.08788236975669861, 0.014212816022336483, -0.03806806728243828, 0.053692687302827835, 0.04510349780321121, 0.17679710686206818, -0.05372918024659157, 0.027468273416161537, -0.01085987500846386, -0.013765137642621994, -0.021306518465280533, -0.020566361024975777, -0.1032397598028183, -0.008217532187700272, 0.1410859078168869, 0.09234005957841873, 0.03564830124378204, -0.013417183421552181, -0.0345475859940052, -0.04595864564180374, -0.012750633992254734, -0.08189065009355545, 0.014671438373625278, -0.019546320661902428, -0.08976012468338013, 0.0931757315993309, 0.008910393342375755, 0.08085379004478455, -0.09844629466533661, 0.016391411423683167, -0.03052634932100773, -0.008689015172421932, -0.051119543612003326, -0.10875950753688812, 0.003899436676874757, 0.014876306988298893, -0.007105946075171232, -0.0993533730506897, -0.1569775640964508, -0.04772467911243439, 0.08535834401845932, -0.07518336921930313, -0.023914759978652, -0.057448431849479675, 0.0000682943282299675, -0.0035771564580500126, -0.005692343693226576, 0.004734512884169817, 0.01771661452949047, 0.0641326829791069, 0.0076678963378071785, 0.07278640568256378, 0.10819133371114731, 0.07143276929855347, -0.13799850642681122, 0.02069621905684471, -0.07119812816381454, 0.11011669784784317, 0.004979265388101339, -0.013975284993648529, -0.14423952996730804, -0.05839769169688225, 0.07755563408136368, 0.024960514158010483, 0.011379744857549667, 0.1003107875585556, -0.1067274734377861, -0.02742200717329979, 0.18494172394275665, -0.11589451879262924, -0.09805846959352493, 0.17407652735710144, -0.07393323630094528, 0.09218185395002365, 0.11173079162836075, 0.04231793060898781, 0.04665884003043175, -0.03837313503026962, 0.08602046966552734, 0.029852157458662987, -0.05481569096446037, 0.09542881697416306, 0.08757112920284271, -0.003998757805675268, -0.12087751924991608, 0.008103582076728344, -0.055464938282966614, 0.008033440448343754, -0.05064034461975098, -0.09141690284013748, -0.042799897491931915, -0.039185527712106705, 0.06163324788212776, 0.03553221747279167, 0.0880303904414177, -0.02030434086918831, -0.11953816562891006, 0.0649106502532959, 0.08500145375728607, -0.08635890483856201, 0.0024070849176496267, -0.10713905841112137, 0.030678845942020416, -0.10741345584392548, 0.009620863944292068, -0.15626586973667145, 0.04919950291514397, 0.027935834601521492, -0.03920014202594757, -0.007673514541238546, 0.12937958538532257, 0.008290670812129974, 0.009866510517895222, -0.00487419543787837, 0.022477887570858, 0.05470797419548035, -0.004190203733742237, -0.09365541487932205, -0.05990401655435562, -0.03416970372200012, -0.011483747512102127, 0.11308112740516663, -0.034315500408411026, -0.010413381271064281, -0.13609528541564941, 0.09713054448366165, -0.020865246653556824, 0.0609685517847538, -0.01808066852390766, 0.0062762899324297905, -0.026004375889897346, -0.03020760416984558, 0.026992889121174812, -0.0161923598498106, -0.06696552038192749, 0.0824466347694397, -0.13160282373428345, -0.05768928676843643, 0.07812556624412537, -0.058859750628471375, -0.06345855444669724, 0.10349206626415253, -0.012653677724301815, -0.010400879196822643, -0.07582449167966843, 0.0031032441183924675, 0.23192745447158813, 0.0235385000705719, 0.1441880762577057, -0.10394033044576645, -0.03910656273365021, -0.018137700855731964, -0.09496079385280609, 0.026489682495594025, 0.06276388466358185, 0.05044737830758095, -0.16057021915912628, 0.0668400451540947, 0.08755689114332199, -0.0982009693980217, -0.049597419798374176, -0.048767510801553726, -0.04040800780057907, -0.018693367019295692, 0.020596902817487717, 0.015745369717478752, -0.046255916357040405, 0.034495946019887924, 0.02687368355691433, 0.055701132863759995, 0.05161583796143532, 0.0025433097034692764, -0.14435702562332153, 0.018420690670609474, 0.002069329610094428, -0.011670002713799477, -0.007494900841265917, 0.03560759127140045, 0.02058662660419941, 0.07043254375457764, 0.04449900612235069, -0.06807035207748413, 0.05159986391663551, -0.033485181629657745, -0.13951453566551208, 0.21684391796588898, -0.055628452450037, -0.22624246776103973, -0.10680501163005829, 0.015781333670020103, 0.008800704032182693, 0.011640649288892746, 0.07580499351024628, -0.05282660946249962, -0.06239672750234604, -0.039476968348026276, 0.07172691822052002, -0.07547223567962646, -0.07435284554958344, -0.04170399531722069, 0.012975737452507019, -0.012076806277036667, -0.1478412002325058, 0.019435778260231018, -0.017133019864559174, -0.13254156708717346, 0.07526431232690811, -0.11449142545461655, -0.03394071012735367, 0.16663315892219543, -0.02926144190132618, 0.0007279141573235393, -0.03429313376545906, 0.20161108672618866, -0.03790539130568504, -0.011347824707627296, 0.20231664180755615, -0.05146447941660881, 0.062159910798072815, 0.09932482987642288, 0.011152914725244045, -0.05724319815635681, 0.05240379646420479, 0.0005114402738399804, -0.02567918971180916, -0.2063896358013153, -0.06296171993017197, -0.013676065020263195, 0.012855847366154194, 0.1107834056019783, 0.0007182872504927218, 0.09203709661960602, 0.036990851163864136, -0.08321081846952438, 0.0317850187420845, 0.051255226135253906, 0.10255010426044464, 0.14584149420261383, 0.03280441462993622, 0.1406385451555252, -0.03090754896402359, -0.12735994160175323, 0.023079652339220047, 0.07208462059497833, 0.1642562747001648, 0.05984221771359444, -0.001967767719179392, 0.06608805805444717, 0.04556852951645851, 0.06684515625238419, 0.03215453773736954, -0.05517565459012985, 0.03261186555027962, -0.06754516810178757, -0.010860487818717957, 0.01439973246306181, 0.0679946020245552, -0.015071535483002663, -0.020678622648119926, -0.09019079059362411, -0.03930370509624481, -0.012169619090855122, 0.17837707698345184, 0.13425955176353455, -0.22493694722652435, -0.06657751649618149, 0.0727817565202713, -0.11972156912088394, -0.022309724241495132, 0.06736302375793457, -0.0696304440498352, -0.21775352954864502, -0.03360950946807861, -0.04842714965343475, 0.14884597063064575, -0.01835242100059986, 0.002515877829864621, -0.11317767202854156, -0.02717144414782524, 0.035474423319101334, 0.13648757338523865, -0.3211652338504791, 0.050579994916915894, -0.011529350653290749, 0.05429583042860031, -0.09626295417547226, 0.01913764327764511, 0.026257775723934174, 0.0498470813035965, 0.06445223838090897, -0.0350845642387867, -0.037249721586704254, -0.06635495275259018, -0.014014074578881264, 0.04204874113202095, 0.01779092289507389, -0.014398159459233284, 0.06714478880167007, -0.08821114897727966, 0.02323475293815136, 0.023213429376482964, -0.008164915256202221, -0.047604676336050034, -0.09148906171321869, 0.04724789038300514, -0.09038347005844116, 0.0661279559135437, -0.024140935391187668, -0.05117615684866905, -0.028319090604782104, 0.07918522506952286, -0.1357451230287552, -0.07783983647823334, -0.08871682733297348, 0.07960803061723709, 0.11344430595636368, -0.05970853194594383, -0.002664614701643586, -0.016676858067512512, 0.10650132596492767, -0.017847442999482155, -0.1516508162021637, 0.04764583334326744, -0.08915083855390549, -0.16881871223449707, -0.041135936975479126, 0.01868947222828865, 0.030808664858341217, 0.06020243465900421, 0.08865536749362946, 0.08448202908039093, -0.12201088666915894, -0.07252929359674454, -0.04164522513747215, -0.0001835663861129433, 0.05557839572429657, 0.04091138020157814, -0.02533305063843727, 0.001882239943370223, -0.07327264547348022, 0.02687269076704979, 0.16272500157356262, 0.09315472841262817, -0.0806574672460556, 0.11199698597192764, 0.16080211102962494, -0.03340577706694603, -0.22554153203964233, -0.11893472820520401, 0.021587781608104706, 0.031612880527973175, 0.013462396338582039, -0.10714894533157349, 0.11854009330272675, 0.030479401350021362, -0.06616199761629105, -0.17188070714473724, -0.3671005964279175, -0.07443568855524063, 0.1713097244501114, 0.031541381031274796, 0.19470879435539246, -0.10560496896505356, 0.014806372113525867, -0.07986968010663986, -0.1159261167049408, 0.10894657671451569, 0.0050005074590444565, 0.08501661568880081, -0.034093260765075684, 0.05420592054724693, 0.010197254829108715, -0.025416484102606773, 0.08226087689399719, 0.0289011113345623, 0.04439155384898186, -0.033352259546518326, 0.01492379605770111, -0.019460739567875862, -0.016228822991251945, 0.1672912836074829, -0.04890361800789833, 0.12779323756694794, -0.1419142484664917, -0.06276766210794449, -0.016046419739723206, 0.024127203971147537, -0.027066685259342194, -0.05535364896059036, -0.06470940262079239, 0.018597234040498734, 0.07261050492525101, -0.00032456679036840796, 0.04953699931502342, 0.002024851506575942, 0.10348399728536606, 0.08021503686904907, 0.09067286550998688, 0.013471505604684353, -0.07162142544984818, -0.017887728288769722, -0.023417463526129723, 0.11423425376415253, -0.1455373764038086, 0.10131525248289108, 0.06285625696182251, -0.00002391975067439489, 0.14184269309043884, 0.06915212422609329, -0.060422394424676895, 0.02636248804628849, -0.012382125481963158, -0.04637305438518524, -0.13348804414272308, 0.010848729871213436, 0.10435190796852112, -0.04518750309944153, -0.0019641646649688482, 0.11644396930932999, -0.08119875192642212, -0.0706375241279602, 0.008771682158112526, 0.057382337749004364, -0.03926582634449005, 0.10512572526931763, 0.05246136337518692, 0.057231150567531586, -0.07779885828495026, 0.10765685886144638, 0.1493096947669983, -0.018206464126706123, 0.06928034871816635, 0.017544802278280258, -0.10551436990499496, -0.06108799949288368, 0.0024780798703432083, 0.19576038420200348, -0.1448131948709488, -0.06071742996573448, -0.10743575543165207, -0.08942824602127075, 0.027615470811724663, 0.031815871596336365, 0.09662008285522461, 0.050208594650030136, -0.0582425519824028, -0.06624390184879303, -0.11565201729536057, 0.07845304906368256, 0.10126746445894241, 0.01305184792727232, -0.0042015789076685905, 0.02217823825776577, 0.043310754001140594, 0.07663457095623016, -0.06105278059840202, -0.08271121978759766, -0.12199990451335907, 0.045144807547330856, -0.17047615349292755, 0.009044838137924671, -0.036601509898900986, 0.0003022434248123318, 0.045126281678676605, -0.07448507100343704, -0.014146518893539906, 0.019420986995100975, -0.0656830221414566, -0.024723190814256668, -0.04790743812918663, 0.06352365761995316, -0.12028785049915314, -0.01956821046769619, 0.09564586728811264, -0.04970070347189903, 0.051322758197784424, 0.0371239073574543, -0.03554849326610565, 0.05435511842370033, -0.06133165955543518, -0.0969357043504715, -0.06096852943301201, 0.05549168586730957, 0.03747888654470444, -0.1624082773923874, 0.030587302520871162, 0.03868907317519188, 0.039308175444602966, 0.03676174208521843, 0.07743558287620544, -0.11606483161449432, -0.016286034137010574, -0.08082305639982224, -0.08908569067716599, -0.08358822017908096, 0.01325599942356348, 0.008200830779969692, 0.08032640069723129, 0.20025207102298737, -0.05918624997138977, 0.11930246651172638, -0.13045009970664978, -0.02069983445107937, -0.03156894072890282, -0.048923369497060776, -0.10179747641086578, -0.09497866779565811, -0.0031806458719074726, -0.02177494578063488, 0.11698029935359955, 0.0460670180618763, 0.12382645159959793, 0.03951376676559448, 0.05483991280198097, 0.10191228985786438, 0.04061557725071907, 0.14771799743175507, 0.10198737680912018, 0.017651734873652458, -0.07955329865217209, 0.026712190359830856, -0.013219586573541164, -0.0833050087094307, 0.006515745539218187, 0.16688743233680725, -0.06644535064697266, 0.047414086759090424, 0.012016004882752895, -0.004906527232378721, -0.011575120501220226, -0.035505589097738266, 0.008810104802250862, 0.006349038332700729, 0.026974722743034363, 0.09378200769424438, 0.18293404579162598, -0.09469182789325714, 0.04320510849356651, -0.050525445491075516, -0.055529188364744186, -0.16978317499160767, -0.033468253910541534, -0.10318079590797424, -0.16978172957897186, 0.0065807802602648735, -0.08645619451999664, 0.005117142107337713, 0.13488566875457764, 0.041238997131586075, -0.05799603834748268, -0.025628315284848213, -0.03854294866323471, -0.051959339529275894, -0.019285572692751884, -0.05013048276305199, 0.0983944982290268, 0.09431729465723038, 0.011599581688642502, 0.05918845534324646, 0.04862096533179283, 0.09595529735088348, 0.011941192671656609, 0.011575278826057911, 0.020041348412632942, -0.10389022529125214, -0.10060551017522812, 0.029499050229787827, 0.014725682325661182, -0.011605012230575085, 0.11036495119333267, 0.0414777547121048, 0.024291569367051125, -0.00839686393737793, 0.21794293820858002, -0.06551621854305267, -0.1474355012178421, -0.20090630650520325, 0.3832744359970093, 0.060175586491823196, 0.049961160868406296, 0.041826892644166946, -0.09354113042354584, -0.08771362900733948, 0.21245501935482025, 0.11224941909313202, -0.034066617488861084, -0.04883113503456116, 0.0555717758834362, -0.01236236747354269, 0.03185252845287323, 0.10699848085641861, 0.028393233194947243, 0.2274254858493805, -0.12071922421455383, 0.1020331010222435, -0.02440096065402031, -0.08203203231096268, -0.09326554834842682, 0.07510117441415787, -0.010312275029718876, -0.06498714536428452, -0.004234660416841507, 0.14585183560848236, -0.02664933353662491, 0.021413059905171394, -0.02548215165734291, -0.051007241010665894, -0.1135861948132515, 0.0017591587966307998, 0.03279903903603554, -0.04903649911284447, 0.07186000794172287, -0.019432999193668365, 0.035942111164331436, 0.12832286953926086, 0.009199274703860283, -0.048081398010253906, -0.05993684381246567, 0.09315922111272812, -0.007312708999961615, 0.0295732282102108, 0.029766615480184555, 0.1741655021905899, 0.09672340750694275, 0.023585068061947823, -0.02040170319378376, 0.06222079321742058, 0.04055801033973694, 0.031598370522260666, 0.11429018527269363, 0.11499851942062378, 0.0022774036042392254, 0.03847866132855415, 0.08489995449781418, -0.1526382565498352, 0.03130566328763962, -0.14398004114627838, -0.021081620827317238, -0.11639705300331116, 0.05685395747423172, -0.026789942756295204, 0.10026440769433975, 0.15534526109695435, -0.06755910068750381, 0.02107861079275608, -0.022147351875901222, 0.06704210489988327, 0.017395958304405212, -0.04797833785414696, -0.023424401879310608, -0.1702391356229782, 0.01641293615102768, -0.045574892312288284, 0.009130828082561493, -0.21578000485897064, -0.008812090381979942, -0.010528406128287315, -0.0485992357134819, -0.007692213635891676, 0.10061061382293701, 0.07568053901195526, 0.0011062530102208257, -0.06103714928030968, -0.18307805061340332, -0.04356975853443146, 0.07539752870798111, -0.0829671174287796, -0.10309106856584549 ]
null
null
transformers
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. ## Training Data This model was trained on the [Quora Duplicate Questions](https://www.quora.com/q/quoradata/First-Quora-Dataset-Release-Question-Pairs) dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates. Note: The model is not suitable to estimate the similarity of questions, e.g. the two questions "How to learn Java" and "How to learn Python" will result in a rahter low score, as these are not duplicates. ## Usage and Performance Pre-trained models can be used like this: ``` from sentence_transformers import CrossEncoder model = CrossEncoder('model_name') scores = model.predict([('Question 1', 'Question 2'), ('Question 3', 'Question 4')]) ``` You can use this model also without sentence_transformers and by just using Transformers ``AutoModel`` class
{"license": "apache-2.0"}
text-classification
cross-encoder/quora-roberta-base
[ "transformers", "pytorch", "jax", "roberta", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using SentenceTransformers Cross-Encoder class. ## Training Data This model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates. Note: The model is not suitable to estimate the similarity of questions, e.g. the two questions "How to learn Java" and "How to learn Python" will result in a rahter low score, as these are not duplicates. ## Usage and Performance Pre-trained models can be used like this: You can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class
[ "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.", "## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ "TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.", "## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ 48, 34, 93, 39 ]
[ "passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ -0.028236351907253265, 0.02298957295715809, -0.0028521628119051456, 0.0646425411105156, 0.11765683442354202, 0.023299384862184525, 0.16609442234039307, 0.12721982598304749, -0.02589987777173519, -0.002497537527233362, 0.06902404129505157, 0.09041360765695572, 0.03389200568199158, 0.01736462675035, -0.0756693035364151, -0.25257617235183716, 0.05000140145421028, 0.001953172730281949, 0.005523595958948135, 0.09553001075983047, 0.1380188912153244, -0.06342750787734985, 0.06385538727045059, 0.03808078169822693, -0.1387544721364975, 0.029387259855866432, 0.007435559760779142, -0.07278154790401459, 0.1284724920988083, 0.06276270002126694, 0.12536683678627014, 0.04794098809361458, 0.06633620709180832, -0.1390024721622467, 0.010272445157170296, 0.04691065847873688, 0.030382536351680756, 0.042663365602493286, 0.05961441248655319, -0.0883123055100441, -0.015024879947304726, 0.13564015924930573, 0.11921955645084381, 0.049067553132772446, -0.11800447851419449, -0.1697719544172287, -0.09260506927967072, -0.007557870354503393, 0.19177691638469696, 0.0677613914012909, -0.019196776673197746, 0.17920920252799988, -0.16551508009433746, 0.042527250945568085, 0.11477108299732208, -0.29275941848754883, -0.004422552417963743, 0.15177641808986664, 0.027819789946079254, 0.0426139235496521, -0.035739149898290634, -0.025486495345830917, 0.04099881276488304, 0.06852063536643982, -0.023120788857340813, -0.02526349201798439, -0.027673279866576195, 0.019225681200623512, -0.13329489529132843, -0.05995437130331993, 0.16933511197566986, -0.03704995661973953, -0.05897475406527519, -0.1055462658405304, -0.039407879114151, 0.11085079610347748, -0.008900068700313568, -0.11178149282932281, -0.016698947176337242, 0.05211774632334709, -0.08724465221166611, 0.030446620658040047, -0.07607933133840561, -0.10109981149435043, -0.07195515185594559, 0.023931683972477913, 0.07561454176902771, 0.05545526742935181, -0.13705310225486755, 0.11136855185031891, 0.0016995988553389907, -0.09120044857263565, -0.040101051330566406, -0.03573271259665489, -0.09327655285596848, -0.02613293193280697, -0.13934822380542755, -0.1651875376701355, 0.0202079638838768, 0.2592458128929138, 0.08788236975669861, 0.014212816022336483, -0.03806806728243828, 0.053692687302827835, 0.04510349780321121, 0.17679710686206818, -0.05372918024659157, 0.027468273416161537, -0.01085987500846386, -0.013765137642621994, -0.021306518465280533, -0.020566361024975777, -0.1032397598028183, -0.008217532187700272, 0.1410859078168869, 0.09234005957841873, 0.03564830124378204, -0.013417183421552181, -0.0345475859940052, -0.04595864564180374, -0.012750633992254734, -0.08189065009355545, 0.014671438373625278, -0.019546320661902428, -0.08976012468338013, 0.0931757315993309, 0.008910393342375755, 0.08085379004478455, -0.09844629466533661, 0.016391411423683167, -0.03052634932100773, -0.008689015172421932, -0.051119543612003326, -0.10875950753688812, 0.003899436676874757, 0.014876306988298893, -0.007105946075171232, -0.0993533730506897, -0.1569775640964508, -0.04772467911243439, 0.08535834401845932, -0.07518336921930313, -0.023914759978652, -0.057448431849479675, 0.0000682943282299675, -0.0035771564580500126, -0.005692343693226576, 0.004734512884169817, 0.01771661452949047, 0.0641326829791069, 0.0076678963378071785, 0.07278640568256378, 0.10819133371114731, 0.07143276929855347, -0.13799850642681122, 0.02069621905684471, -0.07119812816381454, 0.11011669784784317, 0.004979265388101339, -0.013975284993648529, -0.14423952996730804, -0.05839769169688225, 0.07755563408136368, 0.024960514158010483, 0.011379744857549667, 0.1003107875585556, -0.1067274734377861, -0.02742200717329979, 0.18494172394275665, -0.11589451879262924, -0.09805846959352493, 0.17407652735710144, -0.07393323630094528, 0.09218185395002365, 0.11173079162836075, 0.04231793060898781, 0.04665884003043175, -0.03837313503026962, 0.08602046966552734, 0.029852157458662987, -0.05481569096446037, 0.09542881697416306, 0.08757112920284271, -0.003998757805675268, -0.12087751924991608, 0.008103582076728344, -0.055464938282966614, 0.008033440448343754, -0.05064034461975098, -0.09141690284013748, -0.042799897491931915, -0.039185527712106705, 0.06163324788212776, 0.03553221747279167, 0.0880303904414177, -0.02030434086918831, -0.11953816562891006, 0.0649106502532959, 0.08500145375728607, -0.08635890483856201, 0.0024070849176496267, -0.10713905841112137, 0.030678845942020416, -0.10741345584392548, 0.009620863944292068, -0.15626586973667145, 0.04919950291514397, 0.027935834601521492, -0.03920014202594757, -0.007673514541238546, 0.12937958538532257, 0.008290670812129974, 0.009866510517895222, -0.00487419543787837, 0.022477887570858, 0.05470797419548035, -0.004190203733742237, -0.09365541487932205, -0.05990401655435562, -0.03416970372200012, -0.011483747512102127, 0.11308112740516663, -0.034315500408411026, -0.010413381271064281, -0.13609528541564941, 0.09713054448366165, -0.020865246653556824, 0.0609685517847538, -0.01808066852390766, 0.0062762899324297905, -0.026004375889897346, -0.03020760416984558, 0.026992889121174812, -0.0161923598498106, -0.06696552038192749, 0.0824466347694397, -0.13160282373428345, -0.05768928676843643, 0.07812556624412537, -0.058859750628471375, -0.06345855444669724, 0.10349206626415253, -0.012653677724301815, -0.010400879196822643, -0.07582449167966843, 0.0031032441183924675, 0.23192745447158813, 0.0235385000705719, 0.1441880762577057, -0.10394033044576645, -0.03910656273365021, -0.018137700855731964, -0.09496079385280609, 0.026489682495594025, 0.06276388466358185, 0.05044737830758095, -0.16057021915912628, 0.0668400451540947, 0.08755689114332199, -0.0982009693980217, -0.049597419798374176, -0.048767510801553726, -0.04040800780057907, -0.018693367019295692, 0.020596902817487717, 0.015745369717478752, -0.046255916357040405, 0.034495946019887924, 0.02687368355691433, 0.055701132863759995, 0.05161583796143532, 0.0025433097034692764, -0.14435702562332153, 0.018420690670609474, 0.002069329610094428, -0.011670002713799477, -0.007494900841265917, 0.03560759127140045, 0.02058662660419941, 0.07043254375457764, 0.04449900612235069, -0.06807035207748413, 0.05159986391663551, -0.033485181629657745, -0.13951453566551208, 0.21684391796588898, -0.055628452450037, -0.22624246776103973, -0.10680501163005829, 0.015781333670020103, 0.008800704032182693, 0.011640649288892746, 0.07580499351024628, -0.05282660946249962, -0.06239672750234604, -0.039476968348026276, 0.07172691822052002, -0.07547223567962646, -0.07435284554958344, -0.04170399531722069, 0.012975737452507019, -0.012076806277036667, -0.1478412002325058, 0.019435778260231018, -0.017133019864559174, -0.13254156708717346, 0.07526431232690811, -0.11449142545461655, -0.03394071012735367, 0.16663315892219543, -0.02926144190132618, 0.0007279141573235393, -0.03429313376545906, 0.20161108672618866, -0.03790539130568504, -0.011347824707627296, 0.20231664180755615, -0.05146447941660881, 0.062159910798072815, 0.09932482987642288, 0.011152914725244045, -0.05724319815635681, 0.05240379646420479, 0.0005114402738399804, -0.02567918971180916, -0.2063896358013153, -0.06296171993017197, -0.013676065020263195, 0.012855847366154194, 0.1107834056019783, 0.0007182872504927218, 0.09203709661960602, 0.036990851163864136, -0.08321081846952438, 0.0317850187420845, 0.051255226135253906, 0.10255010426044464, 0.14584149420261383, 0.03280441462993622, 0.1406385451555252, -0.03090754896402359, -0.12735994160175323, 0.023079652339220047, 0.07208462059497833, 0.1642562747001648, 0.05984221771359444, -0.001967767719179392, 0.06608805805444717, 0.04556852951645851, 0.06684515625238419, 0.03215453773736954, -0.05517565459012985, 0.03261186555027962, -0.06754516810178757, -0.010860487818717957, 0.01439973246306181, 0.0679946020245552, -0.015071535483002663, -0.020678622648119926, -0.09019079059362411, -0.03930370509624481, -0.012169619090855122, 0.17837707698345184, 0.13425955176353455, -0.22493694722652435, -0.06657751649618149, 0.0727817565202713, -0.11972156912088394, -0.022309724241495132, 0.06736302375793457, -0.0696304440498352, -0.21775352954864502, -0.03360950946807861, -0.04842714965343475, 0.14884597063064575, -0.01835242100059986, 0.002515877829864621, -0.11317767202854156, -0.02717144414782524, 0.035474423319101334, 0.13648757338523865, -0.3211652338504791, 0.050579994916915894, -0.011529350653290749, 0.05429583042860031, -0.09626295417547226, 0.01913764327764511, 0.026257775723934174, 0.0498470813035965, 0.06445223838090897, -0.0350845642387867, -0.037249721586704254, -0.06635495275259018, -0.014014074578881264, 0.04204874113202095, 0.01779092289507389, -0.014398159459233284, 0.06714478880167007, -0.08821114897727966, 0.02323475293815136, 0.023213429376482964, -0.008164915256202221, -0.047604676336050034, -0.09148906171321869, 0.04724789038300514, -0.09038347005844116, 0.0661279559135437, -0.024140935391187668, -0.05117615684866905, -0.028319090604782104, 0.07918522506952286, -0.1357451230287552, -0.07783983647823334, -0.08871682733297348, 0.07960803061723709, 0.11344430595636368, -0.05970853194594383, -0.002664614701643586, -0.016676858067512512, 0.10650132596492767, -0.017847442999482155, -0.1516508162021637, 0.04764583334326744, -0.08915083855390549, -0.16881871223449707, -0.041135936975479126, 0.01868947222828865, 0.030808664858341217, 0.06020243465900421, 0.08865536749362946, 0.08448202908039093, -0.12201088666915894, -0.07252929359674454, -0.04164522513747215, -0.0001835663861129433, 0.05557839572429657, 0.04091138020157814, -0.02533305063843727, 0.001882239943370223, -0.07327264547348022, 0.02687269076704979, 0.16272500157356262, 0.09315472841262817, -0.0806574672460556, 0.11199698597192764, 0.16080211102962494, -0.03340577706694603, -0.22554153203964233, -0.11893472820520401, 0.021587781608104706, 0.031612880527973175, 0.013462396338582039, -0.10714894533157349, 0.11854009330272675, 0.030479401350021362, -0.06616199761629105, -0.17188070714473724, -0.3671005964279175, -0.07443568855524063, 0.1713097244501114, 0.031541381031274796, 0.19470879435539246, -0.10560496896505356, 0.014806372113525867, -0.07986968010663986, -0.1159261167049408, 0.10894657671451569, 0.0050005074590444565, 0.08501661568880081, -0.034093260765075684, 0.05420592054724693, 0.010197254829108715, -0.025416484102606773, 0.08226087689399719, 0.0289011113345623, 0.04439155384898186, -0.033352259546518326, 0.01492379605770111, -0.019460739567875862, -0.016228822991251945, 0.1672912836074829, -0.04890361800789833, 0.12779323756694794, -0.1419142484664917, -0.06276766210794449, -0.016046419739723206, 0.024127203971147537, -0.027066685259342194, -0.05535364896059036, -0.06470940262079239, 0.018597234040498734, 0.07261050492525101, -0.00032456679036840796, 0.04953699931502342, 0.002024851506575942, 0.10348399728536606, 0.08021503686904907, 0.09067286550998688, 0.013471505604684353, -0.07162142544984818, -0.017887728288769722, -0.023417463526129723, 0.11423425376415253, -0.1455373764038086, 0.10131525248289108, 0.06285625696182251, -0.00002391975067439489, 0.14184269309043884, 0.06915212422609329, -0.060422394424676895, 0.02636248804628849, -0.012382125481963158, -0.04637305438518524, -0.13348804414272308, 0.010848729871213436, 0.10435190796852112, -0.04518750309944153, -0.0019641646649688482, 0.11644396930932999, -0.08119875192642212, -0.0706375241279602, 0.008771682158112526, 0.057382337749004364, -0.03926582634449005, 0.10512572526931763, 0.05246136337518692, 0.057231150567531586, -0.07779885828495026, 0.10765685886144638, 0.1493096947669983, -0.018206464126706123, 0.06928034871816635, 0.017544802278280258, -0.10551436990499496, -0.06108799949288368, 0.0024780798703432083, 0.19576038420200348, -0.1448131948709488, -0.06071742996573448, -0.10743575543165207, -0.08942824602127075, 0.027615470811724663, 0.031815871596336365, 0.09662008285522461, 0.050208594650030136, -0.0582425519824028, -0.06624390184879303, -0.11565201729536057, 0.07845304906368256, 0.10126746445894241, 0.01305184792727232, -0.0042015789076685905, 0.02217823825776577, 0.043310754001140594, 0.07663457095623016, -0.06105278059840202, -0.08271121978759766, -0.12199990451335907, 0.045144807547330856, -0.17047615349292755, 0.009044838137924671, -0.036601509898900986, 0.0003022434248123318, 0.045126281678676605, -0.07448507100343704, -0.014146518893539906, 0.019420986995100975, -0.0656830221414566, -0.024723190814256668, -0.04790743812918663, 0.06352365761995316, -0.12028785049915314, -0.01956821046769619, 0.09564586728811264, -0.04970070347189903, 0.051322758197784424, 0.0371239073574543, -0.03554849326610565, 0.05435511842370033, -0.06133165955543518, -0.0969357043504715, -0.06096852943301201, 0.05549168586730957, 0.03747888654470444, -0.1624082773923874, 0.030587302520871162, 0.03868907317519188, 0.039308175444602966, 0.03676174208521843, 0.07743558287620544, -0.11606483161449432, -0.016286034137010574, -0.08082305639982224, -0.08908569067716599, -0.08358822017908096, 0.01325599942356348, 0.008200830779969692, 0.08032640069723129, 0.20025207102298737, -0.05918624997138977, 0.11930246651172638, -0.13045009970664978, -0.02069983445107937, -0.03156894072890282, -0.048923369497060776, -0.10179747641086578, -0.09497866779565811, -0.0031806458719074726, -0.02177494578063488, 0.11698029935359955, 0.0460670180618763, 0.12382645159959793, 0.03951376676559448, 0.05483991280198097, 0.10191228985786438, 0.04061557725071907, 0.14771799743175507, 0.10198737680912018, 0.017651734873652458, -0.07955329865217209, 0.026712190359830856, -0.013219586573541164, -0.0833050087094307, 0.006515745539218187, 0.16688743233680725, -0.06644535064697266, 0.047414086759090424, 0.012016004882752895, -0.004906527232378721, -0.011575120501220226, -0.035505589097738266, 0.008810104802250862, 0.006349038332700729, 0.026974722743034363, 0.09378200769424438, 0.18293404579162598, -0.09469182789325714, 0.04320510849356651, -0.050525445491075516, -0.055529188364744186, -0.16978317499160767, -0.033468253910541534, -0.10318079590797424, -0.16978172957897186, 0.0065807802602648735, -0.08645619451999664, 0.005117142107337713, 0.13488566875457764, 0.041238997131586075, -0.05799603834748268, -0.025628315284848213, -0.03854294866323471, -0.051959339529275894, -0.019285572692751884, -0.05013048276305199, 0.0983944982290268, 0.09431729465723038, 0.011599581688642502, 0.05918845534324646, 0.04862096533179283, 0.09595529735088348, 0.011941192671656609, 0.011575278826057911, 0.020041348412632942, -0.10389022529125214, -0.10060551017522812, 0.029499050229787827, 0.014725682325661182, -0.011605012230575085, 0.11036495119333267, 0.0414777547121048, 0.024291569367051125, -0.00839686393737793, 0.21794293820858002, -0.06551621854305267, -0.1474355012178421, -0.20090630650520325, 0.3832744359970093, 0.060175586491823196, 0.049961160868406296, 0.041826892644166946, -0.09354113042354584, -0.08771362900733948, 0.21245501935482025, 0.11224941909313202, -0.034066617488861084, -0.04883113503456116, 0.0555717758834362, -0.01236236747354269, 0.03185252845287323, 0.10699848085641861, 0.028393233194947243, 0.2274254858493805, -0.12071922421455383, 0.1020331010222435, -0.02440096065402031, -0.08203203231096268, -0.09326554834842682, 0.07510117441415787, -0.010312275029718876, -0.06498714536428452, -0.004234660416841507, 0.14585183560848236, -0.02664933353662491, 0.021413059905171394, -0.02548215165734291, -0.051007241010665894, -0.1135861948132515, 0.0017591587966307998, 0.03279903903603554, -0.04903649911284447, 0.07186000794172287, -0.019432999193668365, 0.035942111164331436, 0.12832286953926086, 0.009199274703860283, -0.048081398010253906, -0.05993684381246567, 0.09315922111272812, -0.007312708999961615, 0.0295732282102108, 0.029766615480184555, 0.1741655021905899, 0.09672340750694275, 0.023585068061947823, -0.02040170319378376, 0.06222079321742058, 0.04055801033973694, 0.031598370522260666, 0.11429018527269363, 0.11499851942062378, 0.0022774036042392254, 0.03847866132855415, 0.08489995449781418, -0.1526382565498352, 0.03130566328763962, -0.14398004114627838, -0.021081620827317238, -0.11639705300331116, 0.05685395747423172, -0.026789942756295204, 0.10026440769433975, 0.15534526109695435, -0.06755910068750381, 0.02107861079275608, -0.022147351875901222, 0.06704210489988327, 0.017395958304405212, -0.04797833785414696, -0.023424401879310608, -0.1702391356229782, 0.01641293615102768, -0.045574892312288284, 0.009130828082561493, -0.21578000485897064, -0.008812090381979942, -0.010528406128287315, -0.0485992357134819, -0.007692213635891676, 0.10061061382293701, 0.07568053901195526, 0.0011062530102208257, -0.06103714928030968, -0.18307805061340332, -0.04356975853443146, 0.07539752870798111, -0.0829671174287796, -0.10309106856584549 ]
null
null
transformers
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. ## Training Data This model was trained on the [Quora Duplicate Questions](https://www.quora.com/q/quoradata/First-Quora-Dataset-Release-Question-Pairs) dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates. Note: The model is not suitable to estimate the similarity of questions, e.g. the two questions "How to learn Java" and "How to learn Python" will result in a rahter low score, as these are not duplicates. ## Usage and Performance Pre-trained models can be used like this: ``` from sentence_transformers import CrossEncoder model = CrossEncoder('model_name') scores = model.predict([('Question 1', 'Question 2'), ('Question 3', 'Question 4')]) ``` You can use this model also without sentence_transformers and by just using Transformers ``AutoModel`` class
{"license": "apache-2.0"}
text-classification
cross-encoder/quora-roberta-large
[ "transformers", "pytorch", "jax", "roberta", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using SentenceTransformers Cross-Encoder class. ## Training Data This model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates. Note: The model is not suitable to estimate the similarity of questions, e.g. the two questions "How to learn Java" and "How to learn Python" will result in a rahter low score, as these are not duplicates. ## Usage and Performance Pre-trained models can be used like this: You can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class
[ "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.", "## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ "TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.", "## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ 52, 34, 93, 39 ]
[ "passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ -0.03287089988589287, 0.03220580890774727, -0.002792539307847619, 0.0679178386926651, 0.11583749204874039, 0.01822967268526554, 0.18687804043293, 0.13437679409980774, -0.03370516374707222, 0.01239317748695612, 0.06613951176404953, 0.07354146987199783, 0.02668229676783085, 0.02977769263088703, -0.06534553319215775, -0.2713264226913452, 0.04993590712547302, 0.006637313403189182, -0.016340354457497597, 0.082265205681324, 0.14126195013523102, -0.07298634946346283, 0.06217064708471298, 0.05221877619624138, -0.1548863649368286, 0.03254738822579384, 0.011856010183691978, -0.08051485568284988, 0.14062024652957916, 0.05658416822552681, 0.14231140911579132, 0.05355057120323181, 0.06941576302051544, -0.11361847072839737, 0.011748974211513996, 0.057355280965566635, 0.028690189123153687, 0.06483281403779984, 0.07381010800600052, -0.0848153606057167, -0.029585467651486397, 0.11678961664438248, 0.12436320632696152, 0.041707661002874374, -0.12498147040605545, -0.17967434227466583, -0.08553016185760498, -0.009838024154305458, 0.1716412454843521, 0.05830838158726692, -0.026011858135461807, 0.18556058406829834, -0.16587524116039276, 0.04901837557554245, 0.1095905676484108, -0.2921205163002014, -0.0037483195774257183, 0.1480834037065506, 0.040460679680109024, 0.02314128540456295, -0.03290810436010361, -0.013682027347385883, 0.038183312863111496, 0.06915600597858429, -0.008306341245770454, -0.013202316127717495, -0.038243960589170456, 0.013941487297415733, -0.1394719034433365, -0.0653618797659874, 0.17320235073566437, -0.03870536759495735, -0.0455210916697979, -0.09346166998147964, -0.06613437831401825, 0.11054395139217377, 0.005591654684394598, -0.09999769181013107, -0.019150730222463608, 0.037361808121204376, -0.09903206676244736, 0.01328632514923811, -0.09017538279294968, -0.0847623348236084, -0.0869031473994255, 0.04317522048950195, 0.07581494003534317, 0.0557035431265831, -0.12159845978021622, 0.1135704517364502, 0.010577685199677944, -0.10575850307941437, -0.016711916774511337, -0.03581656888127327, -0.11436302214860916, -0.034662242978811264, -0.1445568948984146, -0.1624370664358139, 0.009622065350413322, 0.2746860086917877, 0.09734425693750381, -0.004394448362290859, -0.03641950711607933, 0.060459788888692856, 0.052444856613874435, 0.18199503421783447, -0.07806859910488129, 0.05308889225125313, -0.01046827994287014, 0.015972938388586044, -0.01670544221997261, -0.03589500114321709, -0.11638379096984863, -0.0042594014666974545, 0.15442943572998047, 0.08017696440219879, 0.06319478154182434, -0.0013094873866066337, -0.023120921105146408, -0.03826361522078514, -0.015064915642142296, -0.08397867530584335, 0.03448371961712837, -0.03227211534976959, -0.09382329881191254, 0.0662597268819809, 0.02178693376481533, 0.08284146338701248, -0.0862513929605484, 0.011881799437105656, -0.04020761698484421, -0.00654161861166358, -0.058926839381456375, -0.13009408116340637, 0.005502915941178799, 0.02374870888888836, -0.013810358941555023, -0.10695427656173706, -0.15747015178203583, -0.042498767375946045, 0.08429151773452759, -0.07375969737768173, -0.04254097491502762, -0.04171816632151604, -0.006774070207029581, 0.0047939070500433445, -0.005850460380315781, -0.0019866847433149815, 0.01938837766647339, 0.05480531230568886, 0.007294788025319576, 0.08186715096235275, 0.09258009493350983, 0.07079054415225983, -0.14123468101024628, 0.02159246988594532, -0.07554320245981216, 0.09733321517705917, 0.0007424764335155487, -0.03692271560430527, -0.13090606033802032, -0.05251488462090492, 0.08022475242614746, 0.03617674112319946, 0.022776678204536438, 0.08334081619977951, -0.12080594152212143, -0.030719386413693428, 0.16126753389835358, -0.1115150973200798, -0.08656304329633713, 0.176836296916008, -0.07481200248003006, 0.08404294401407242, 0.08925269544124603, 0.06167721003293991, 0.01441936008632183, -0.030046794563531876, 0.06590458005666733, 0.017781605944037437, -0.04565657302737236, 0.10647022724151611, 0.0802680179476738, 0.0016104582464322448, -0.10714733600616455, 0.0021788794547319412, -0.04201694577932358, 0.0033173300325870514, -0.06894306093454361, -0.09312140196561813, -0.04831736162304878, -0.029916822910308838, 0.07975649833679199, 0.026474790647625923, 0.08946885913610458, -0.024270592257380486, -0.12763433158397675, 0.09345117956399918, 0.08358923345804214, -0.07853730767965317, 0.0015502659371122718, -0.08790002018213272, 0.05514185503125191, -0.12374123185873032, -0.0019612046889960766, -0.16382460296154022, 0.005857852753251791, 0.035480793565511703, -0.04692165181040764, 0.015140080824494362, 0.14238341152668, -0.00022190056915860623, 0.04730154946446419, 0.0001127191717387177, 0.004371917340904474, 0.04459569230675697, 0.007726526353508234, -0.09755449742078781, -0.07664702832698822, -0.04875147342681885, -0.0175436083227396, 0.1276908814907074, -0.032725993543863297, -0.00023919263912830502, -0.13663747906684875, 0.12270230799913406, -0.02795683778822422, 0.06726392358541489, -0.010458721779286861, 0.012501533143222332, -0.04388565570116043, -0.02504308894276619, 0.029530491679906845, -0.014848093502223492, -0.049829524010419846, 0.09627728164196014, -0.15836934745311737, -0.0409531369805336, 0.08836714923381805, -0.07706864178180695, -0.06798715144395828, 0.0993131622672081, -0.018489625304937363, -0.009376898407936096, -0.09485068172216415, -0.0156826451420784, 0.20844614505767822, 0.015332008711993694, 0.1537424772977829, -0.10160452872514725, -0.021846767514944077, -0.02373281493782997, -0.11354660987854004, 0.01935139298439026, 0.04426613077521324, 0.04160051792860031, -0.14409592747688293, 0.09184277057647705, 0.09137747436761856, -0.11941323429346085, -0.029019704088568687, -0.03746531158685684, -0.03288352116942406, -0.036266423761844635, -0.008530855178833008, 0.00992781762033701, -0.027668938040733337, 0.05548539385199547, 0.022097064182162285, 0.06508249789476395, 0.0435728020966053, 0.014114669524133205, -0.15526345372200012, 0.02620646171271801, 0.008247812278568745, -0.0009513023542240262, 0.00020343544019851834, 0.03980361670255661, 0.03240915760397911, 0.08016364276409149, 0.0351797454059124, -0.08251238614320755, 0.04375332593917847, -0.0245799720287323, -0.11935722082853317, 0.20220446586608887, -0.06418857723474503, -0.23171773552894592, -0.12543325126171112, 0.04153398051857948, -0.0254649817943573, 0.0284960325807333, 0.07483630627393723, -0.047969311475753784, -0.05940685793757439, -0.04744051769375801, 0.07056485861539841, -0.06553135067224503, -0.057584840804338455, -0.06564519554376602, 0.006420007906854153, -0.0010493986774235964, -0.15580697357654572, 0.011212955228984356, -0.01985001191496849, -0.13604773581027985, 0.08270909637212753, -0.07899542897939682, -0.010255329310894012, 0.16837617754936218, -0.04011749476194382, -0.0020551157649606466, -0.035932037979364395, 0.21144583821296692, -0.03162882849574089, 0.012489858083426952, 0.20609085261821747, -0.04364995285868645, 0.07578480243682861, 0.12185834348201752, 0.014143186621367931, -0.04346136003732681, 0.055697668343782425, 0.004085979890078306, -0.020722251385450363, -0.1827675998210907, -0.07326171547174454, -0.013312481343746185, 0.0015584429493173957, 0.1118616834282875, -0.002338045509532094, 0.11703769117593765, 0.03614424914121628, -0.08995484560728073, 0.03120020404458046, 0.033848587423563004, 0.10977817326784134, 0.15997794270515442, 0.03603929281234741, 0.16397495567798615, -0.01753595843911171, -0.132488414645195, 0.01706407032907009, 0.07142373919487, 0.15959052741527557, 0.04072384163737297, -0.018242504447698593, 0.07191752642393112, 0.03405921161174774, 0.07868117094039917, 0.02640828862786293, -0.03679972514510155, 0.04370824247598648, -0.06622111797332764, -0.014968479983508587, 0.026352783665060997, 0.07653508335351944, 0.005359221249818802, -0.018195437267422676, -0.09236548840999603, -0.0386410690844059, -0.004821833688765764, 0.1912795752286911, 0.14512018859386444, -0.2419639676809311, -0.061507634818553925, 0.06962853670120239, -0.11014198511838913, -0.018605468794703484, 0.04699371010065079, -0.057219963520765305, -0.20587919652462006, -0.036483559757471085, -0.044423479586839676, 0.12873513996601105, -0.02396460250020027, 0.017407871782779694, -0.1088668704032898, -0.0037940931506454945, 0.023492345586419106, 0.1306726485490799, -0.30602124333381653, 0.07735942304134369, -0.005571290384978056, 0.028852391988039017, -0.08847087621688843, 0.02636065147817135, 0.037499554455280304, 0.05410102382302284, 0.027914877980947495, -0.03412196412682533, -0.06794535368680954, -0.1019541546702385, -0.01447890605777502, 0.03329499810934067, 0.02434415929019451, -0.01893431507050991, 0.07295048981904984, -0.08791227638721466, 0.018951719626784325, 0.031694427132606506, 0.025779711082577705, -0.061407942324876785, -0.08968307077884674, 0.053266994655132294, -0.09155987203121185, 0.05666026845574379, -0.04244745522737503, -0.0647258535027504, -0.0575680211186409, 0.07236741483211517, -0.09864015132188797, -0.06886366009712219, -0.09187743812799454, 0.10358817875385284, 0.1201443299651146, -0.058892492204904556, 0.005045982543379068, -0.03299126774072647, 0.1276741623878479, -0.030271658673882484, -0.14077235758304596, 0.04787242412567139, -0.08695114403963089, -0.18228569626808167, -0.05778321251273155, 0.016378920525312424, 0.005391437094658613, 0.06597761064767838, 0.08555072546005249, 0.1060669869184494, -0.1281929761171341, -0.05918307974934578, -0.024490704759955406, 0.010349417105317116, 0.0530129075050354, 0.05980363488197327, -0.03403376042842865, -0.0053240396082401276, -0.05730501562356949, 0.03151492029428482, 0.16266945004463196, 0.13333921134471893, -0.09335033595561981, 0.09398265928030014, 0.16118700802326202, -0.035691626369953156, -0.23807929456233978, -0.10470245033502579, 0.02114677056670189, 0.04097786173224449, 0.01596149243414402, -0.10623084008693695, 0.10559507459402084, 0.026080582290887833, -0.06810668110847473, -0.19921305775642395, -0.3535882830619812, -0.08426278084516525, 0.16295690834522247, 0.009544314816594124, 0.16835252940654755, -0.09334618598222733, 0.01882992312312126, -0.07200830429792404, -0.05674244090914726, 0.11677592247724533, -0.010123214684426785, 0.09212194383144379, -0.022549862042069435, 0.045421380549669266, 0.014774376526474953, -0.025502415373921394, 0.09999138861894608, 0.0214613676071167, 0.058467257767915726, -0.03684123978018761, 0.012596921063959599, 0.002864339156076312, -0.027969928458333015, 0.169255331158638, -0.034150782972574234, 0.12560446560382843, -0.12425871938467026, -0.0488470233976841, -0.02793426811695099, 0.04721225053071976, -0.029558653011918068, -0.06160027161240578, -0.07574033737182617, 0.022044742479920387, 0.06429511308670044, -0.0031263066921383142, 0.03413941711187363, 0.007824328728020191, 0.11757781356573105, 0.11530592292547226, 0.07052170485258102, 0.002290953416377306, -0.07644207030534744, 0.005038187373429537, -0.02865632250905037, 0.11116921156644821, -0.13136543333530426, 0.09488803893327713, 0.0625145360827446, -0.00515682203695178, 0.14277641475200653, 0.07113510370254517, -0.0691724419593811, 0.025240028277039528, -0.0022865214850753546, -0.06102057173848152, -0.12716534733772278, 0.001014350214973092, 0.1120416522026062, -0.0400361493229866, -0.01710464432835579, 0.12777604162693024, -0.07729045301675797, -0.06502936780452728, 0.01161304023116827, 0.041276223957538605, -0.02173643931746483, 0.1007012128829956, 0.06882841140031815, 0.06178441271185875, -0.06406088918447495, 0.10418111085891724, 0.1749107986688614, -0.009909868240356445, 0.07344341278076172, 0.015463700518012047, -0.08871257305145264, -0.06942931562662125, -0.01776329055428505, 0.2067471742630005, -0.15562395751476288, -0.05347033962607384, -0.12376556545495987, -0.07888049632310867, 0.016117596998810768, 0.047817569226026535, 0.09953563660383224, 0.034959256649017334, -0.06096327304840088, -0.05343448743224144, -0.12025012075901031, 0.09239010512828827, 0.11192873865365982, 0.04453551769256592, -0.013860038481652737, 0.06599844992160797, 0.032309625297784805, 0.07947007566690445, -0.06409940868616104, -0.08243674784898758, -0.12620995938777924, 0.028330039232969284, -0.15713593363761902, -0.010111291892826557, -0.03577524796128273, -0.016892528161406517, 0.04599938169121742, -0.08631021529436111, -0.013367350213229656, 0.012637665495276451, -0.07286768406629562, -0.026227736845612526, -0.048548243939876556, 0.06766729056835175, -0.13450081646442413, -0.02712484821677208, 0.09844990819692612, -0.051080815494060516, 0.05167121812701225, 0.006130293942987919, -0.02681487426161766, 0.021283909678459167, -0.040365420281887054, -0.10448890179395676, -0.06326239556074142, 0.05370960384607315, 0.04407564178109169, -0.15046574175357819, 0.034633342176675797, 0.023813530802726746, 0.04863613843917847, 0.032495491206645966, 0.09285463392734528, -0.10681762546300888, -0.011162840761244297, -0.0654686763882637, -0.09288402646780014, -0.08509011566638947, 0.015860581770539284, 0.015865439549088478, 0.07959801703691483, 0.19732770323753357, -0.06674476712942123, 0.11078502982854843, -0.14607058465480804, -0.014212640933692455, -0.045948781073093414, -0.06775068491697311, -0.09055338054895401, -0.10398133844137192, 0.0018127302173525095, -0.020655889064073563, 0.1290983110666275, 0.07949729263782501, 0.1304897964000702, 0.04222028702497482, 0.07883203029632568, 0.09613931179046631, 0.047242846339941025, 0.10612136870622635, 0.09710146486759186, 0.024568289518356323, -0.07696165144443512, 0.03487970307469368, -0.005151093006134033, -0.075327567756176, -0.009145429357886314, 0.13885553181171417, -0.04313445836305618, 0.047403838485479355, 0.027900487184524536, -0.025592822581529617, -0.03286461904644966, -0.061549026519060135, 0.010117235593497753, 0.023258639499545097, 0.01573597453534603, 0.05198691040277481, 0.17294037342071533, -0.09616678953170776, 0.05143878236413002, -0.042103033512830734, -0.05420548841357231, -0.15789414942264557, -0.03333789110183716, -0.09685321897268295, -0.15734343230724335, -0.0025480675976723433, -0.0885271281003952, 0.015913022682070732, 0.1391577422618866, 0.028112351894378662, -0.05023058131337166, -0.016297034919261932, -0.042800422757864, -0.0447847880423069, -0.03685012832283974, -0.04912824183702469, 0.10199691355228424, 0.07438888400793076, -0.011180942878127098, 0.043773967772722244, 0.08056911081075668, 0.07660676538944244, 0.012079463340342045, 0.021409275010228157, 0.01567625068128109, -0.10064703226089478, -0.10546135157346725, 0.030458437278866768, -0.00639837933704257, -0.010613745078444481, 0.09491019695997238, 0.03778057545423508, 0.02429492399096489, -0.0063088382594287395, 0.22515292465686798, -0.0718308836221695, -0.15340447425842285, -0.19897988438606262, 0.3925504982471466, 0.03735268488526344, 0.04824819415807724, 0.031650569289922714, -0.10381948202848434, -0.1141526848077774, 0.1846715360879898, 0.13957232236862183, -0.03757867217063904, -0.04579593613743782, 0.05021177977323532, -0.022542808204889297, 0.01426080334931612, 0.09667779505252838, 0.024123458191752434, 0.19574320316314697, -0.12143624573945999, 0.11033306270837784, -0.023489724844694138, -0.07747054100036621, -0.09914333373308182, 0.08757908642292023, -0.0006054493715055287, -0.055196214467287064, -0.0055865035392344, 0.12537860870361328, -0.011684657074511051, 0.016954666003584862, -0.0414222776889801, -0.06764676421880722, -0.12190880626440048, 0.021816249936819077, 0.04498741030693054, -0.05891931802034378, 0.08162081986665726, -0.031079348176717758, 0.04032259061932564, 0.12032638490200043, -0.006586451083421707, -0.03429294750094414, -0.05201276019215584, 0.1021626740694046, -0.03672444075345993, 0.04110773652791977, 0.01947266049683094, 0.17124368250370026, 0.11389374732971191, 0.019412698224186897, -0.027855420485138893, 0.05599062889814377, 0.044213857501745224, 0.01930294930934906, 0.11120204627513885, 0.11952421814203262, 0.011368711479008198, 0.03814197704195976, 0.08243228495121002, -0.18457508087158203, 0.04268735274672508, -0.14795227348804474, -0.030769649893045425, -0.10342152416706085, 0.042879872024059296, -0.01862787827849388, 0.10443770885467529, 0.15344755351543427, -0.06863529980182648, 0.013661391101777554, -0.024796264246106148, 0.05178695544600487, 0.019096791744232178, -0.04969056695699692, -0.027796681970357895, -0.1725776195526123, 0.029365232214331627, -0.05421414598822594, 0.02013411559164524, -0.20382514595985413, -0.0325913205742836, 0.001093793660402298, -0.05562740936875343, -0.00424076896160841, 0.10651307553052902, 0.07583669573068619, -0.001469014910981059, -0.06099686771631241, -0.162979394197464, -0.047733474522829056, 0.06599317491054535, -0.0958397164940834, -0.09633159637451172 ]
null
null
transformers
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. ## Training Data This model was trained on the [STS benchmark dataset](http://ixa2.si.ehu.eus/stswiki/index.php/STSbenchmark). The model will predict a score between 0 and 1 how for the semantic similarity of two sentences. ## Usage and Performance Pre-trained models can be used like this: ``` from sentence_transformers import CrossEncoder model = CrossEncoder('model_name') scores = model.predict([('Sentence 1', 'Sentence 2'), ('Sentence 3', 'Sentence 4')]) ``` The model will predict scores for the pairs `('Sentence 1', 'Sentence 2')` and `('Sentence 3', 'Sentence 4')`. You can use this model also without sentence_transformers and by just using Transformers ``AutoModel`` class
{"license": "apache-2.0"}
text-classification
cross-encoder/stsb-TinyBERT-L-4
[ "transformers", "pytorch", "jax", "bert", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using SentenceTransformers Cross-Encoder class. ## Training Data This model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences. ## Usage and Performance Pre-trained models can be used like this: The model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'. You can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class
[ "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.", "## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ "TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.", "## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ 47, 34, 38, 85 ]
[ "passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ -0.1018068939447403, 0.05181725695729256, -0.0023968915920704603, 0.059157323092222214, 0.1179395318031311, -0.01356672216206789, 0.1925230473279953, 0.06266329437494278, 0.0032281309831887484, -0.024624381214380264, 0.1129741221666336, 0.10616011172533035, 0.016521207988262177, 0.10064561665058136, -0.042440690100193024, -0.2364850789308548, 0.06007177755236626, -0.0070038302801549435, -0.028153659775853157, 0.13106979429721832, 0.10826661437749863, -0.07041341066360474, 0.06480579078197479, 0.03029913827776909, -0.1302100270986557, 0.031012598425149918, 0.006991867907345295, -0.05731632187962532, 0.10829401761293411, 0.03067704662680626, 0.10081019997596741, 0.0939921960234642, 0.07745488733053207, -0.12296551465988159, 0.007382601965218782, 0.0063626691699028015, 0.017311513423919678, 0.025654563680291176, 0.05087742954492569, -0.10948707908391953, -0.10683263838291168, 0.09693643450737, 0.08571042865514755, 0.01270565576851368, -0.12037516385316849, -0.08612044155597687, -0.025205207988619804, -0.04198529198765755, 0.16773280501365662, 0.08882227540016174, -0.05072407051920891, 0.14051233232021332, -0.16546614468097687, 0.041112933307886124, 0.03895944729447365, -0.323636531829834, 0.003483750158920884, 0.08764201402664185, 0.0034883744083344936, 0.03945263475179672, -0.027819111943244934, -0.005329758860170841, 0.041045285761356354, 0.05586545914411545, 0.03135641664266586, -0.038039617240428925, 0.03505055606365204, 0.014455686323344707, -0.17962168157100677, -0.02957088127732277, 0.2132486253976822, -0.012625647708773613, -0.05105971544981003, -0.09227420389652252, -0.03575388714671135, 0.05454649776220322, -0.027917219325900078, -0.09403304010629654, -0.0002995362738147378, 0.06360745429992676, -0.09977148473262787, 0.062051158398389816, -0.11694594472646713, -0.12564818561077118, -0.12393181025981903, 0.12397140264511108, 0.058516498655080795, 0.04854318127036095, -0.10782363265752792, 0.1164080873131752, -0.03153989091515541, -0.07901397347450256, -0.02859816700220108, -0.07697615772485733, -0.12585236132144928, -0.013163987547159195, -0.19046393036842346, -0.15082679688930511, 0.018758419901132584, 0.22904662787914276, 0.060466669499874115, 0.004972934722900391, 0.033935729414224625, 0.041198570281267166, 0.03478030487895012, 0.1743755042552948, -0.09926558285951614, 0.021922988817095757, -0.048428550362586975, -0.002462996169924736, -0.013277100399136543, -0.004015984013676643, -0.1436050832271576, -0.03460471332073212, 0.08038260042667389, 0.03557009622454643, -0.04887658357620239, 0.06058119982481003, -0.038236744701862335, -0.04881091043353081, 0.05196266248822212, -0.07350710779428482, -0.014168184250593185, -0.00074140535434708, -0.08883697539567947, 0.07080439478158951, -0.01790766790509224, 0.06728566437959671, -0.0643407553434372, -0.027974655851721764, -0.04908320680260658, -0.008619080297648907, -0.05475703999400139, -0.1105925589799881, -0.02645830437541008, 0.048080120235681534, -0.0027558491565287113, -0.11305185407400131, -0.15955372154712677, -0.058441177010536194, 0.05148078501224518, -0.03036997653543949, 0.0027453897055238485, -0.12507066130638123, -0.018036365509033203, -0.005363141186535358, -0.026156703010201454, -0.017081469297409058, 0.0020510340109467506, 0.032484348863363266, 0.012856436893343925, 0.1034550592303276, 0.006090695038437843, 0.05886368453502655, -0.17560699582099915, -0.0024692874867469072, -0.07560712844133377, 0.1525801718235016, 0.018734389916062355, 0.028917210176587105, -0.08698704093694687, -0.04454522207379341, -0.021243851631879807, 0.053984384983778, 0.040969934314489365, 0.13965754210948944, -0.16391904652118683, -0.061851002275943756, 0.2513390779495239, -0.1284327507019043, -0.11598508805036545, 0.1371479332447052, -0.03622976318001747, 0.01943778805434704, 0.13422444462776184, 0.11409961432218552, 0.056645311415195465, -0.013185767456889153, 0.08900745213031769, 0.060256052762269974, -0.08443780988454819, 0.002127943327650428, 0.1069052442908287, 0.026767587289214134, -0.10531603544950485, 0.041462890803813934, -0.07566383481025696, 0.03522966429591179, -0.05421165004372597, -0.06971017271280289, -0.02012300305068493, -0.06686417013406754, 0.06848528981208801, 0.013206487521529198, 0.08170952647924423, -0.021763892844319344, -0.09114915132522583, 0.08406414836645126, 0.08524670451879501, -0.08920326083898544, 0.02498677745461464, -0.11104071140289307, 0.012608169578015804, -0.0720713883638382, -0.00818286556750536, -0.2049560546875, 0.030590128153562546, 0.0010281912982463837, 0.026924526318907738, 0.04258265718817711, 0.0865471139550209, 0.03300182521343231, 0.027052683755755424, 0.008815101347863674, -0.017389707267284393, 0.10040396451950073, -0.005394225008785725, -0.08699692040681839, -0.09920375794172287, -0.00590997189283371, -0.018526380881667137, 0.16438403725624084, -0.013949048705399036, 0.024582549929618835, -0.10159825533628464, 0.05533580854535103, -0.050071485340595245, 0.058535899966955185, -0.0007597642834298313, 0.03356035426259041, -0.02695932239294052, 0.0003881471639033407, 0.05768725648522377, 0.01757374405860901, -0.1148795336484909, 0.15499082207679749, -0.15193228423595428, -0.04112286865711212, 0.11559657007455826, -0.10884633660316467, -0.06290720403194427, 0.013559277169406414, -0.014183246530592442, 0.014825847931206226, -0.047694068402051926, 0.0007997894426807761, 0.18626055121421814, 0.022305170074105263, 0.12479870021343231, -0.08444536477327347, -0.024937346577644348, -0.01686849072575569, -0.08763117343187332, 0.02107910066843033, 0.05984365940093994, 0.034147195518016815, -0.1403716504573822, 0.08710502088069916, 0.0974455326795578, -0.08408372104167938, -0.0122535340487957, -0.018786951899528503, -0.05176491662859917, -0.0221885833889246, -0.023430481553077698, -0.03204912319779396, -0.05046255141496658, -0.033675484359264374, -0.03023233637213707, 0.034682922065258026, 0.04521404951810837, 0.0034213250037282705, -0.155630424618721, 0.0080204326659441, 0.003638196038082242, -0.0203767828643322, -0.014345481060445309, 0.03083779849112034, -0.013281100429594517, 0.07333807647228241, 0.02015344798564911, -0.14261353015899658, 0.05891278013586998, -0.03858984634280205, -0.15263532102108002, 0.22569985687732697, -0.07110138237476349, -0.2786460220813751, -0.12235891819000244, 0.028318103402853012, -0.021125204861164093, 0.06356321275234222, 0.06646061688661575, -0.05676727741956711, -0.03817734122276306, -0.07428191602230072, -0.027466019615530968, -0.03168352693319321, 0.009854322299361229, -0.07029229402542114, -0.0037583066150546074, 0.004941050428897142, -0.08933887630701065, -0.00434132432565093, -0.06283523142337799, -0.07909105718135834, 0.043802373111248016, -0.12476077675819397, -0.024092787876725197, 0.2181084007024765, 0.01601947657763958, 0.031246930360794067, -0.05494377017021179, 0.18576189875602722, -0.009880712255835533, -0.03940189629793167, 0.16586916148662567, -0.01927039958536625, 0.03855458274483681, 0.15495267510414124, 0.011838171631097794, -0.08688342571258545, 0.07673905789852142, -0.029861317947506905, -0.027374960482120514, -0.16827046871185303, -0.08797821402549744, -0.019058354198932648, 0.0029649899806827307, 0.08240267634391785, -0.01751703955233097, 0.08579690009355545, 0.0695563480257988, -0.021018480882048607, -0.0005222564213909209, 0.04430143162608147, 0.07364543527364731, 0.16303817927837372, 0.0025723474100232124, 0.16941216588020325, -0.014031201601028442, -0.12389104068279266, 0.02420184761285782, 0.018687864765524864, 0.08982378989458084, 0.0533696673810482, 0.012395033612847328, 0.08919225633144379, 0.021671243011951447, 0.07004024088382721, 0.006178485695272684, -0.0004188550228718668, 0.006742182187736034, -0.037925828248262405, -0.0244216937571764, 0.005592566914856434, 0.0764646902680397, 0.001343207317404449, -0.02597714215517044, -0.13366010785102844, -0.009858713485300541, 0.044687867164611816, 0.1371791809797287, 0.17257051169872284, -0.3082663416862488, -0.05881521478295326, 0.036572180688381195, -0.06716003268957138, 0.0197786632925272, 0.10509262979030609, -0.024768933653831482, -0.1274256408214569, 0.02793874219059944, -0.003028233302757144, 0.09369175136089325, 0.004122328013181686, 0.026334088295698166, -0.0601830892264843, 0.008279175497591496, 0.02421797625720501, 0.12616759538650513, -0.36846140027046204, 0.09926177561283112, -0.026310620829463005, 0.020479802042245865, -0.04275868460536003, -0.016107743605971336, 0.03927360102534294, 0.15804648399353027, 0.09322899580001831, -0.03330657631158829, -0.05638880655169487, -0.031805459409952164, -0.066896952688694, 0.027491994202136993, 0.03170914947986603, -0.05940031260251999, 0.0503549762070179, -0.04554425925016403, -0.007709940429776907, 0.024181507527828217, 0.026194728910923004, -0.020046330988407135, -0.06001001596450806, -0.00879651214927435, 0.007586304564028978, 0.072652667760849, -0.007978993467986584, -0.06681016832590103, -0.13694220781326294, 0.052377160638570786, -0.10020110011100769, -0.06997738033533096, -0.0902373418211937, 0.11111807078123093, 0.07158495485782623, -0.05788290500640869, -0.00672223512083292, 0.011055811308324337, 0.12987114489078522, 0.016508178785443306, -0.13435445725917816, 0.08141613006591797, -0.07861921191215515, -0.10619295388460159, -0.059240348637104034, 0.06913714110851288, 0.04407327249646187, 0.04800542816519737, 0.07065581530332565, 0.048081379383802414, -0.09482857584953308, -0.0784943625330925, -0.03114073909819126, 0.010291383601725101, 0.05988830700516701, 0.020789729431271553, -0.015861110761761665, -0.036075446754693985, -0.07026886940002441, 0.05532866343855858, 0.1666422188282013, 0.049988165497779846, -0.07837219536304474, 0.10538347065448761, 0.19119104743003845, -0.03443188592791557, -0.2710075378417969, -0.09393182396888733, 0.052779968827962875, 0.05966762825846672, -0.0036762533709406853, -0.09317929297685623, 0.12003692984580994, 0.01758931390941143, -0.03927377238869667, -0.20792117714881897, -0.30292636156082153, -0.1059204638004303, 0.19534169137477875, 0.045353010296821594, 0.19527271389961243, -0.06996535509824753, 0.0006368850008584559, -0.0701228529214859, -0.02188224345445633, 0.13054832816123962, -0.03208127245306969, 0.07262752205133438, 0.029194189235568047, 0.026614254340529442, -0.00354999047704041, -0.012914572842419147, 0.07330986857414246, 0.11169873923063278, 0.04423724114894867, -0.0230934526771307, 0.008428706787526608, 0.01551474817097187, -0.02156813070178032, 0.1388540118932724, -0.03433184698224068, 0.10813695192337036, -0.037260331213474274, -0.09222746640443802, -0.02682902291417122, -0.007109030149877071, 0.020407337695360184, -0.045286476612091064, -0.003708033822476864, 0.01889204792678356, 0.04637684300541878, -0.008154493756592274, 0.03212818130850792, -0.06161528453230858, 0.023317275568842888, 0.08996660262346268, 0.1535702794790268, -0.007848595269024372, -0.030906114727258682, -0.018208878114819527, -0.044997502118349075, 0.10206405073404312, -0.1243642196059227, 0.09742600470781326, 0.08768989145755768, -0.023522846400737762, 0.1322346031665802, 0.10066074132919312, -0.028441043570637703, 0.007724082097411156, 0.03397630527615547, -0.08349581062793732, -0.07828566431999207, -0.0177594106644392, -0.010412252508103848, -0.009639780037105083, 0.06921340525150299, 0.12561507523059845, -0.04973725974559784, -0.049510978162288666, -0.0014399120118469, 0.03787189722061157, -0.06970417499542236, 0.1286851465702057, 0.04328149929642677, 0.02771693840622902, -0.10920175164937973, 0.08079849183559418, 0.06557505577802658, -0.02886749431490898, 0.05063258111476898, 0.002709042513743043, -0.14908482134342194, -0.07961821556091309, -0.038368768990039825, 0.16730418801307678, -0.1231829896569252, -0.1041891947388649, -0.12432868778705597, -0.1195027232170105, 0.055146776139736176, 0.11373137682676315, 0.1308000683784485, 0.07971426844596863, -0.08254364877939224, -0.06629341840744019, -0.06873215734958649, 0.07043080776929855, 0.0590834766626358, 0.0014476168435066938, -0.07249955087900162, 0.1078842356801033, 0.02953767031431198, 0.06995809078216553, -0.07426763325929642, -0.10544809699058533, -0.11137775331735611, 0.06761300563812256, -0.09696733206510544, -0.0013616761425510049, -0.07044665515422821, -0.020865771919488907, 0.03833087161183357, -0.06195796653628349, -0.02672228403389454, 0.009703190997242928, -0.06699907779693604, 0.02723001316189766, -0.046083513647317886, 0.06712814420461655, -0.11217520385980606, -0.023237034678459167, 0.026498185470700264, -0.03590751066803932, 0.04223775491118431, 0.09461784362792969, -0.03911161795258522, 0.061090435832738876, -0.09148291498422623, -0.04641120508313179, -0.001979498891159892, 0.056505050510168076, 0.030138403177261353, -0.12308505177497864, 0.007712612394243479, 0.08656618744134903, 0.012758440338075161, 0.027216410264372826, 0.05956273898482323, -0.06229201331734657, 0.024123365059494972, -0.0008471736800856888, -0.04460861161351204, -0.08894117176532745, -0.018898092210292816, 0.025442427024245262, 0.1234932616353035, 0.17448166012763977, -0.06430473178625107, 0.04541806876659393, -0.12680687010288239, 0.005586212035268545, -0.026076631620526314, -0.03762659430503845, -0.11931457370519638, -0.10948051512241364, 0.021586015820503235, 0.007657730486243963, 0.18232405185699463, 0.048868268728256226, 0.09762267023324966, 0.02667461894452572, 0.07704909890890121, 0.14989763498306274, 0.0837806686758995, 0.15277248620986938, 0.08505940437316895, -0.001308452570810914, -0.06153569743037224, 0.06000564247369766, 0.04609448090195656, 0.01871381513774395, 0.01030363142490387, 0.08698059618473053, -0.031070802360773087, 0.09347410500049591, 0.04217350855469704, 0.07338280230760574, -0.029209233820438385, -0.07380496710538864, -0.06661934405565262, -0.027354905381798744, 0.01160237193107605, 0.1700151115655899, 0.1219412311911583, -0.09097740799188614, 0.03803998976945877, 0.0007353593246079981, -0.07200080156326294, -0.1551312804222107, -0.1218455508351326, -0.0787975862622261, -0.1789652556180954, -0.020515741780400276, -0.09347008168697357, -0.03238793835043907, 0.1722450852394104, 0.061212293803691864, -0.020552337169647217, 0.036514975130558014, -0.026509378105401993, -0.026977956295013428, -0.006232324056327343, -0.08245831727981567, 0.07446432113647461, 0.06176842749118805, -0.01726594753563404, 0.07555916905403137, 0.043883468955755234, 0.059370145201683044, -0.02057473734021187, 0.01889934577047825, 0.0017426462145522237, -0.08713501691818237, -0.08469713479280472, -0.014220773242413998, 0.0037208078429102898, -0.013231090269982815, 0.024186650291085243, 0.06992467492818832, -0.03062121942639351, 0.011872920207679272, 0.19517837464809418, -0.08884906023740768, -0.13992168009281158, -0.23437249660491943, 0.35079628229141235, 0.05717453733086586, 0.07900752127170563, 0.04559033736586571, -0.0946241095662117, -0.06920360773801804, 0.25321730971336365, 0.13183432817459106, -0.08776503056287766, -0.027379212900996208, 0.06537112593650818, -0.011876639910042286, 0.06274634599685669, 0.11160000413656235, 0.006347394082695246, 0.2118706852197647, -0.09449732303619385, 0.045286647975444794, -0.042395275086164474, -0.07067117840051651, 0.00304989330470562, 0.048596229404211044, -0.025210218504071236, -0.051842279732227325, -0.03354787826538086, 0.15831102430820465, -0.08447630703449249, 0.04331846535205841, -0.02344190701842308, -0.005120161455124617, -0.08395712822675705, -0.014423919841647148, 0.03069530427455902, 0.005818168167024851, 0.08256194740533829, -0.038627512753009796, 0.0474165640771389, 0.11462542414665222, -0.0019665148574858904, -0.062286101281642914, -0.04901717230677605, 0.07197322696447372, -0.0521366111934185, 0.00411691190674901, -0.005958066321909428, 0.1981751173734665, 0.0844266265630722, 0.07160070538520813, -0.01448103692382574, 0.06654112040996552, -0.011041480116546154, 0.03469720855355263, 0.12156840413808823, 0.045243531465530396, 0.0045079803094267845, 0.04379255324602127, 0.045405611395835876, -0.16603533923625946, 0.06943680346012115, -0.10327289998531342, -0.07379655539989471, -0.11981610208749771, 0.04639599844813347, -0.020899711176753044, 0.11346854269504547, 0.10301072150468826, -0.050007306039333344, 0.015899445861577988, 0.013804791495203972, 0.018075525760650635, -0.039431195706129074, -0.09157073497772217, -0.009393940679728985, -0.18247821927070618, 0.005740893539041281, -0.007659159600734711, -0.0004097088531125337, -0.2470398247241974, -0.013012047857046127, -0.011412807740271091, -0.031997185200452805, 0.00288157700560987, 0.08741876482963562, 0.06337825208902359, 0.041867293417453766, -0.07839862257242203, -0.06604918837547302, -0.00466717267408967, 0.09431437402963638, -0.11357368528842926, -0.16285744309425354 ]
null
null
transformers
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. ## Training Data This model was trained on the [STS benchmark dataset](http://ixa2.si.ehu.eus/stswiki/index.php/STSbenchmark). The model will predict a score between 0 and 1 how for the semantic similarity of two sentences. ## Usage and Performance Pre-trained models can be used like this: ``` from sentence_transformers import CrossEncoder model = CrossEncoder('model_name') scores = model.predict([('Sentence 1', 'Sentence 2'), ('Sentence 3', 'Sentence 4')]) ``` The model will predict scores for the pairs `('Sentence 1', 'Sentence 2')` and `('Sentence 3', 'Sentence 4')`. You can use this model also without sentence_transformers and by just using Transformers ``AutoModel`` class
{"license": "apache-2.0"}
text-classification
cross-encoder/stsb-distilroberta-base
[ "transformers", "pytorch", "jax", "roberta", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using SentenceTransformers Cross-Encoder class. ## Training Data This model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences. ## Usage and Performance Pre-trained models can be used like this: The model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'. You can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class
[ "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.", "## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ "TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.", "## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ 52, 34, 38, 85 ]
[ "passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ -0.08450348675251007, 0.06658493727445602, -0.0020891379099339247, 0.06090733781456947, 0.1219165176153183, -0.010885688476264477, 0.18726801872253418, 0.06976161152124405, 0.004681415390223265, -0.032464902848005295, 0.10880614817142487, 0.10593046247959137, 0.009075884707272053, 0.1311509907245636, -0.0318918414413929, -0.2397564798593521, 0.05535270273685455, -0.021333403885364532, -0.023506149649620056, 0.1251000612974167, 0.10538756102323532, -0.08173905313014984, 0.06480519473552704, 0.036619383841753006, -0.1481119692325592, 0.0361047126352787, 0.004630628507584333, -0.06398891657590866, 0.1055711954832077, 0.029057493433356285, 0.12339451909065247, 0.09259660542011261, 0.07925070077180862, -0.12724295258522034, 0.012720810249447823, 0.016518067568540573, 0.009311544708907604, 0.03629666194319725, 0.036708857864141464, -0.10701856762170792, -0.09100929647684097, 0.08599001914262772, 0.08524163067340851, 0.0009422770235687494, -0.11783528327941895, -0.1288551539182663, -0.02296801283955574, -0.0451493114233017, 0.1464175432920456, 0.0878152996301651, -0.045602668076753616, 0.1485908329486847, -0.1755804717540741, 0.04145662114024162, 0.054525889456272125, -0.3068389296531677, 0.010614708065986633, 0.0938933864235878, 0.023070018738508224, 0.04588997736573219, -0.04204950109124184, -0.006081583444029093, 0.028909359127283096, 0.04380388930439949, 0.06281910091638565, -0.04001569375395775, -0.010886033065617085, 0.037939924746751785, -0.1775822639465332, -0.03491581231355667, 0.21934248507022858, -0.03001006878912449, -0.03818020224571228, -0.10346876084804535, -0.030659932643175125, 0.05167389288544655, -0.022362971678376198, -0.08230966329574585, 0.004080281592905521, 0.05453440919518471, -0.09994271397590637, 0.043615926057100296, -0.12259428203105927, -0.1212959885597229, -0.11897437274456024, 0.1481919288635254, 0.0519876554608345, 0.05614864081144333, -0.11817987263202667, 0.11007988452911377, -0.018013888970017433, -0.08047500997781754, -0.03508051112294197, -0.07759242504835129, -0.13628479838371277, -0.017040323466062546, -0.1888146996498108, -0.1461862325668335, 0.019468307495117188, 0.23470254242420197, 0.05425926297903061, -0.003253627335652709, 0.05223825201392174, 0.052410710602998734, 0.05260057747364044, 0.16987985372543335, -0.09921816736459732, 0.019514871761202812, -0.040489714592695236, -0.006623496767133474, -0.014924317598342896, -0.020644033327698708, -0.15668311715126038, -0.02293773740530014, 0.06836400926113129, 0.03340352699160576, -0.03544469177722931, 0.060046419501304626, -0.03591883182525635, -0.04250994324684143, 0.04778998717665672, -0.07110997289419174, 0.007199421525001526, 0.001498668105341494, -0.08984625339508057, 0.04597581550478935, 0.004727287217974663, 0.05564960837364197, -0.04794371500611305, -0.03951052203774452, -0.05796322971582413, -0.0041662254370749, -0.07994621992111206, -0.13172027468681335, -0.017280934378504753, 0.0645439475774765, 0.004875025246292353, -0.12934610247612, -0.15153756737709045, -0.06250552833080292, 0.06014043092727661, -0.025576326996088028, -0.0060349274426698685, -0.10992851108312607, -0.024137191474437714, 0.014359449967741966, -0.03408343344926834, -0.012038946151733398, -0.006700094789266586, 0.027012227103114128, -0.010106818750500679, 0.10634177178144455, 0.011723324656486511, 0.056721992790699005, -0.17293746769428253, -0.014487387612462044, -0.06987504661083221, 0.13998807966709137, 0.00621027173474431, 0.036297235637903214, -0.07619620859622955, -0.03155680000782013, -0.03006790578365326, 0.06660110503435135, 0.037052832543849945, 0.1293592005968094, -0.16096548736095428, -0.06480851024389267, 0.2344948649406433, -0.12800095975399017, -0.11015582084655762, 0.12573368847370148, -0.049440640956163406, 0.03731274977326393, 0.10660473257303238, 0.12351691722869873, 0.0702773854136467, -0.034878578037023544, 0.08157497644424438, 0.06925155967473984, -0.0970594733953476, -0.01633760891854763, 0.10197000205516815, 0.03484668955206871, -0.08979984372854233, 0.031350795179605484, -0.06290501356124878, 0.05259009823203087, -0.04869356006383896, -0.06925348937511444, -0.013377794064581394, -0.054921574890613556, 0.07530007511377335, 0.020727040246129036, 0.08574032783508301, -0.02007264457643032, -0.10619869083166122, 0.11279048770666122, 0.08791369199752808, -0.0893058329820633, 0.029166527092456818, -0.10053735971450806, 0.013915755786001682, -0.07938611507415771, -0.007064095698297024, -0.20703130960464478, 0.006485046353191137, 0.0018438141560181975, 0.014914322644472122, 0.034115370362997055, 0.10435596108436584, 0.030128762125968933, 0.022408952936530113, 0.029981357976794243, -0.018031882122159004, 0.08979107439517975, 0.009407839737832546, -0.08929740637540817, -0.10686875879764557, -0.02606169506907463, -0.03660851716995239, 0.15328125655651093, -0.013654633425176144, 0.02513095922768116, -0.09524213522672653, 0.0469084158539772, -0.029759293422102928, 0.05308831110596657, -0.004855465143918991, 0.04144696891307831, -0.04124278575181961, 0.0014224101323634386, 0.04660613834857941, 0.017619729042053223, -0.09673509001731873, 0.1709372103214264, -0.18291769921779633, -0.02089911699295044, 0.12491315603256226, -0.11515885591506958, -0.06447350978851318, 0.02345173992216587, -0.014730798080563545, 0.02202417515218258, -0.05534713342785835, 0.014598351903259754, 0.16470010578632355, 0.014831818640232086, 0.12296251952648163, -0.083440862596035, -0.026016926392912865, -0.018848687410354614, -0.09721343219280243, 0.018556935712695122, 0.04981476813554764, 0.039000771939754486, -0.1162211075425148, 0.10053393989801407, 0.10247835516929626, -0.10287507623434067, 0.015802966430783272, 0.004366625100374222, -0.04628602787852287, -0.03741034120321274, -0.04338941350579262, -0.03472225368022919, -0.027288775891065598, -0.029719965532422066, -0.042659372091293335, 0.03510940074920654, 0.029411355033516884, 0.013905813917517662, -0.15196870267391205, 0.005906535778194666, 0.013755157589912415, -0.011594702489674091, -0.01026098895817995, 0.04376592859625816, -0.008289550431072712, 0.061599910259246826, 0.028126584365963936, -0.14367526769638062, 0.058363452553749084, -0.0351489894092083, -0.14141881465911865, 0.22615239024162292, -0.07566404342651367, -0.27953410148620605, -0.12290982156991959, 0.021696176379919052, 0.006087016314268112, 0.062644362449646, 0.0708455741405487, -0.045471686869859695, -0.03838431462645531, -0.0820717066526413, -0.03139931708574295, -0.024082103744149208, 0.0024445243179798126, -0.0878162756562233, 0.00009271603630622849, -0.003847202518954873, -0.09421621263027191, -0.0031321351416409016, -0.05890399217605591, -0.06755509972572327, 0.04214369133114815, -0.09165877103805542, -0.007889453321695328, 0.2206796407699585, 0.01076422818005085, 0.03367805480957031, -0.05330446735024452, 0.17513544857501984, -0.015779828652739525, -0.03460876643657684, 0.1611550748348236, -0.012880779802799225, 0.04445143789052963, 0.17388220131397247, 0.019242996349930763, -0.0683070495724678, 0.07065898925065994, -0.03215143457055092, -0.04168960079550743, -0.1452464610338211, -0.09278552234172821, -0.038062673062086105, -0.0233317743986845, 0.07220902293920517, -0.01197992917150259, 0.08731245249509811, 0.05724746733903885, -0.003093191422522068, -0.01159081980586052, 0.03088546171784401, 0.0780472680926323, 0.1442079395055771, 0.003124864771962166, 0.17355042695999146, -0.016894925385713577, -0.11764701455831528, 0.027372056618332863, 0.02583908475935459, 0.10140083730220795, 0.05113178864121437, 0.009328491054475307, 0.10008900612592697, 0.027490992099046707, 0.07921173423528671, -0.004589700140058994, 0.004766227677464485, 0.0007489954587072134, -0.03577183932065964, -0.029451249167323112, 0.016168830916285515, 0.09111353009939194, 0.01197222899645567, -0.02815566398203373, -0.12441352754831314, -0.03104766085743904, 0.04662104323506355, 0.11217736452817917, 0.16801334917545319, -0.30731672048568726, -0.05213860049843788, 0.04639245942234993, -0.06299179047346115, 0.023163730278611183, 0.09142906963825226, -0.004638497717678547, -0.13518525660037994, 0.00644878763705492, -0.004350210539996624, 0.08985988050699234, 0.0018690075958147645, 0.02098994515836239, -0.06347128748893738, 0.02823813259601593, 0.014463197439908981, 0.12384771555662155, -0.34352806210517883, 0.10411574691534042, -0.029895853251218796, 0.011633627116680145, -0.04604942351579666, -0.006390930153429508, 0.04763137176632881, 0.14670772850513458, 0.08665851503610611, -0.02811373583972454, -0.09105808287858963, -0.06061112508177757, -0.04584410414099693, 0.022072266787290573, 0.025318568572402, -0.05517273768782616, 0.05771936476230621, -0.04354587942361832, -0.005889877211302519, 0.02616376057267189, 0.05868706852197647, -0.025132019072771072, -0.07593091577291489, -0.007170701399445534, 0.013278568163514137, 0.05211564898490906, -0.015402347780764103, -0.06917586177587509, -0.13652382791042328, 0.042136695235967636, -0.07789728045463562, -0.06028033047914505, -0.103256456553936, 0.09486180543899536, 0.07285741716623306, -0.06623293459415436, 0.0039231302216649055, 0.016175266355276108, 0.133241206407547, 0.004705662373453379, -0.1416292041540146, 0.08322334289550781, -0.07802871614694595, -0.08273158967494965, -0.06269396841526031, 0.03027677722275257, 0.04064171016216278, 0.04588952288031578, 0.0636964812874794, 0.06164630129933357, -0.11224881559610367, -0.07974357903003693, -0.02778312936425209, 0.016498364508152008, 0.05221463739871979, 0.024097010493278503, -0.005767281167209148, -0.055050041526556015, -0.03973664715886116, 0.06736788898706436, 0.17233264446258545, 0.0396297462284565, -0.09726333618164062, 0.12583638727664948, 0.17738917469978333, -0.04895620420575142, -0.283265620470047, -0.09362755715847015, 0.04195278882980347, 0.07143981009721756, 0.014062013477087021, -0.09430427849292755, 0.1031544879078865, 0.012528144754469395, -0.048478029668331146, -0.2168145477771759, -0.3028346598148346, -0.10955806821584702, 0.18833613395690918, 0.03733060136437416, 0.1766844391822815, -0.07329234480857849, 0.00888361781835556, -0.06110344082117081, 0.008218829520046711, 0.149373859167099, -0.06190621480345726, 0.0771213248372078, 0.031203610822558403, 0.02545219659805298, -0.0002090671769110486, -0.01993192359805107, 0.07104034721851349, 0.08380923420190811, 0.060775671154260635, -0.02233022265136242, -0.005947250407189131, 0.030198819935321808, -0.038693733513355255, 0.13508112728595734, -0.038966890424489975, 0.09695424884557724, -0.060185179114341736, -0.08268783241510391, -0.033626738935709, 0.0095839723944664, 0.021442634984850883, -0.047682344913482666, -0.014281190931797028, 0.0293685682117939, 0.04048173502087593, -0.010960451327264309, 0.06327590346336365, -0.06340527534484863, 0.045411452651023865, 0.11049935221672058, 0.13265644013881683, -0.01526104100048542, -0.03578300401568413, 0.0051394738256931305, -0.047414444386959076, 0.11430126428604126, -0.164499893784523, 0.0911836102604866, 0.07006113976240158, -0.022437749430537224, 0.1343110352754593, 0.0996335968375206, -0.04331466183066368, 0.018353044986724854, 0.052883587777614594, -0.0847473293542862, -0.06420569121837616, -0.015110892243683338, -0.007178271654993296, -0.005094901192933321, 0.0529211089015007, 0.11877772212028503, -0.05779349431395531, -0.04995092377066612, -0.008740345947444439, 0.035886529833078384, -0.06572326272726059, 0.11153262853622437, 0.06726455688476562, 0.032411862164735794, -0.09644010663032532, 0.06804357469081879, 0.07756434381008148, -0.02221454121172428, 0.06189196556806564, 0.018371189013123512, -0.13287536799907684, -0.09099080413579941, -0.03262708708643913, 0.17415082454681396, -0.1126992255449295, -0.09605299681425095, -0.1218004897236824, -0.10991249978542328, 0.04737558588385582, 0.11134099960327148, 0.13007406890392303, 0.07032902538776398, -0.07839404791593552, -0.06061587482690811, -0.07010961323976517, 0.0824027732014656, 0.05524514988064766, -0.002843055408447981, -0.07459040731191635, 0.11347991228103638, 0.031216993927955627, 0.0675557404756546, -0.08114106953144073, -0.10859313607215881, -0.128158301115036, 0.06127750873565674, -0.06947355717420578, -0.005555060692131519, -0.07801594585180283, -0.010933276265859604, 0.03555882349610329, -0.05734606087207794, -0.0364067368209362, 0.013651168905198574, -0.08039981126785278, 0.025344787165522575, -0.037574876099824905, 0.06731393933296204, -0.10970719158649445, -0.029593832790851593, 0.023546205833554268, -0.03511421009898186, 0.04349852725863457, 0.07940305769443512, -0.02931799553334713, 0.04794476926326752, -0.0984886959195137, -0.04244762659072876, -0.0008963823784142733, 0.05560130998492241, 0.04657924920320511, -0.11562971025705338, 0.011889439076185226, 0.08023174852132797, 0.02496957965195179, 0.01808847486972809, 0.04983857646584511, -0.06669441610574722, 0.03893890976905823, -0.005539896432310343, -0.04568590968847275, -0.08155433088541031, 0.0019124216632917523, 0.029221396893262863, 0.11990660429000854, 0.17286483943462372, -0.05838547274470329, 0.0494680181145668, -0.12196874618530273, 0.002273800317198038, -0.03104289248585701, -0.03879985958337784, -0.12439896911382675, -0.10816622525453568, 0.023254498839378357, 0.005873893387615681, 0.19243818521499634, 0.08622796088457108, 0.09015563130378723, 0.02026992477476597, 0.08280457556247711, 0.15640299022197723, 0.08169982582330704, 0.13479216396808624, 0.0753636434674263, 0.003552020760253072, -0.04583922401070595, 0.06687968969345093, 0.04342857003211975, 0.0072935158386826515, 0.0038177764508873224, 0.09528568387031555, 0.00667149294167757, 0.10526926070451736, 0.042451679706573486, 0.0500829704105854, -0.032434239983558655, -0.06250517815351486, -0.0641176775097847, -0.01888967491686344, -0.011569607071578503, 0.12127377092838287, 0.14277249574661255, -0.08079589903354645, 0.04679444059729576, 0.015447227284312248, -0.06474620848894119, -0.1609332412481308, -0.1257009506225586, -0.07168443500995636, -0.18344774842262268, -0.027470940724015236, -0.09683065861463547, -0.02620367519557476, 0.16843217611312866, 0.04557392746210098, -0.009155435487627983, 0.02763553522527218, 0.0004341208259575069, -0.02509470470249653, -0.020050279796123505, -0.08659908920526505, 0.07982605695724487, 0.04354038089513779, -0.010250295512378216, 0.07036767154932022, 0.038146670907735825, 0.051697567105293274, -0.016725361347198486, 0.016275659203529358, 0.014421421103179455, -0.0931367501616478, -0.08636751025915146, -0.014829674735665321, 0.006926656700670719, -0.012881503440439701, 0.022557280957698822, 0.07137629389762878, -0.03945473954081535, 0.001135995495133102, 0.21051138639450073, -0.09168606251478195, -0.13335897028446198, -0.22504109144210815, 0.3292695879936218, 0.06128848344087601, 0.08006542921066284, 0.01782870665192604, -0.09772481769323349, -0.08380256593227386, 0.23835910856723785, 0.15601664781570435, -0.07657447457313538, -0.014831600710749626, 0.07228817790746689, -0.016094770282506943, 0.04914749786257744, 0.11798988282680511, -0.0018936273409053683, 0.22898456454277039, -0.08471135795116425, 0.04600175470113754, -0.03982498124241829, -0.06657011061906815, -0.0016606205608695745, 0.04384338855743408, -0.028883112594485283, -0.0646972507238388, -0.0373566634953022, 0.14134474098682404, -0.09122554212808609, 0.049046847969293594, -0.0002345513494219631, -0.022823620587587357, -0.07716405391693115, -0.006424167193472385, 0.06062676012516022, 0.00020406620751600713, 0.08504018932580948, -0.038146086037158966, 0.03794025629758835, 0.11565134674310684, -0.0017569564515724778, -0.06979916244745255, -0.058996688574552536, 0.06873651593923569, -0.06620217114686966, 0.02624395117163658, -0.020865874364972115, 0.19483453035354614, 0.09904284030199051, 0.06483505666255951, -0.03644286096096039, 0.0626656711101532, -0.013008199632167816, 0.011695616878569126, 0.12002301216125488, 0.03363015130162239, 0.0020754639990627766, 0.06408432871103287, 0.044256992638111115, -0.1768614798784256, 0.07748190313577652, -0.14381177723407745, -0.06964220106601715, -0.10966911166906357, 0.02179321087896824, -0.012797866016626358, 0.1110919713973999, 0.09294867515563965, -0.04715340584516525, 0.012934789061546326, 0.018140526488423347, 0.015916945412755013, -0.041642192751169205, -0.078044094145298, -0.03142290934920311, -0.16770023107528687, 0.007813815027475357, -0.03266200050711632, 0.002701889956369996, -0.22878091037273407, -0.011267242021858692, -0.008542446419596672, -0.033176716417074203, 0.008387149311602116, 0.08422781527042389, 0.08410125225782394, 0.03710673004388809, -0.07843427360057831, -0.06219158694148064, 0.0013762605376541615, 0.09299024194478989, -0.12327927350997925, -0.15334023535251617 ]
null
null
transformers
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. ## Training Data This model was trained on the [STS benchmark dataset](http://ixa2.si.ehu.eus/stswiki/index.php/STSbenchmark). The model will predict a score between 0 and 1 how for the semantic similarity of two sentences. ## Usage and Performance Pre-trained models can be used like this: ``` from sentence_transformers import CrossEncoder model = CrossEncoder('model_name') scores = model.predict([('Sentence 1', 'Sentence 2'), ('Sentence 3', 'Sentence 4')]) ``` The model will predict scores for the pairs `('Sentence 1', 'Sentence 2')` and `('Sentence 3', 'Sentence 4')`. You can use this model also without sentence_transformers and by just using Transformers ``AutoModel`` class
{"license": "apache-2.0"}
text-classification
cross-encoder/stsb-roberta-base
[ "transformers", "pytorch", "jax", "roberta", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using SentenceTransformers Cross-Encoder class. ## Training Data This model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences. ## Usage and Performance Pre-trained models can be used like this: The model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'. You can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class
[ "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.", "## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ "TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.", "## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ 52, 34, 38, 85 ]
[ "passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ -0.08450348675251007, 0.06658493727445602, -0.0020891379099339247, 0.06090733781456947, 0.1219165176153183, -0.010885688476264477, 0.18726801872253418, 0.06976161152124405, 0.004681415390223265, -0.032464902848005295, 0.10880614817142487, 0.10593046247959137, 0.009075884707272053, 0.1311509907245636, -0.0318918414413929, -0.2397564798593521, 0.05535270273685455, -0.021333403885364532, -0.023506149649620056, 0.1251000612974167, 0.10538756102323532, -0.08173905313014984, 0.06480519473552704, 0.036619383841753006, -0.1481119692325592, 0.0361047126352787, 0.004630628507584333, -0.06398891657590866, 0.1055711954832077, 0.029057493433356285, 0.12339451909065247, 0.09259660542011261, 0.07925070077180862, -0.12724295258522034, 0.012720810249447823, 0.016518067568540573, 0.009311544708907604, 0.03629666194319725, 0.036708857864141464, -0.10701856762170792, -0.09100929647684097, 0.08599001914262772, 0.08524163067340851, 0.0009422770235687494, -0.11783528327941895, -0.1288551539182663, -0.02296801283955574, -0.0451493114233017, 0.1464175432920456, 0.0878152996301651, -0.045602668076753616, 0.1485908329486847, -0.1755804717540741, 0.04145662114024162, 0.054525889456272125, -0.3068389296531677, 0.010614708065986633, 0.0938933864235878, 0.023070018738508224, 0.04588997736573219, -0.04204950109124184, -0.006081583444029093, 0.028909359127283096, 0.04380388930439949, 0.06281910091638565, -0.04001569375395775, -0.010886033065617085, 0.037939924746751785, -0.1775822639465332, -0.03491581231355667, 0.21934248507022858, -0.03001006878912449, -0.03818020224571228, -0.10346876084804535, -0.030659932643175125, 0.05167389288544655, -0.022362971678376198, -0.08230966329574585, 0.004080281592905521, 0.05453440919518471, -0.09994271397590637, 0.043615926057100296, -0.12259428203105927, -0.1212959885597229, -0.11897437274456024, 0.1481919288635254, 0.0519876554608345, 0.05614864081144333, -0.11817987263202667, 0.11007988452911377, -0.018013888970017433, -0.08047500997781754, -0.03508051112294197, -0.07759242504835129, -0.13628479838371277, -0.017040323466062546, -0.1888146996498108, -0.1461862325668335, 0.019468307495117188, 0.23470254242420197, 0.05425926297903061, -0.003253627335652709, 0.05223825201392174, 0.052410710602998734, 0.05260057747364044, 0.16987985372543335, -0.09921816736459732, 0.019514871761202812, -0.040489714592695236, -0.006623496767133474, -0.014924317598342896, -0.020644033327698708, -0.15668311715126038, -0.02293773740530014, 0.06836400926113129, 0.03340352699160576, -0.03544469177722931, 0.060046419501304626, -0.03591883182525635, -0.04250994324684143, 0.04778998717665672, -0.07110997289419174, 0.007199421525001526, 0.001498668105341494, -0.08984625339508057, 0.04597581550478935, 0.004727287217974663, 0.05564960837364197, -0.04794371500611305, -0.03951052203774452, -0.05796322971582413, -0.0041662254370749, -0.07994621992111206, -0.13172027468681335, -0.017280934378504753, 0.0645439475774765, 0.004875025246292353, -0.12934610247612, -0.15153756737709045, -0.06250552833080292, 0.06014043092727661, -0.025576326996088028, -0.0060349274426698685, -0.10992851108312607, -0.024137191474437714, 0.014359449967741966, -0.03408343344926834, -0.012038946151733398, -0.006700094789266586, 0.027012227103114128, -0.010106818750500679, 0.10634177178144455, 0.011723324656486511, 0.056721992790699005, -0.17293746769428253, -0.014487387612462044, -0.06987504661083221, 0.13998807966709137, 0.00621027173474431, 0.036297235637903214, -0.07619620859622955, -0.03155680000782013, -0.03006790578365326, 0.06660110503435135, 0.037052832543849945, 0.1293592005968094, -0.16096548736095428, -0.06480851024389267, 0.2344948649406433, -0.12800095975399017, -0.11015582084655762, 0.12573368847370148, -0.049440640956163406, 0.03731274977326393, 0.10660473257303238, 0.12351691722869873, 0.0702773854136467, -0.034878578037023544, 0.08157497644424438, 0.06925155967473984, -0.0970594733953476, -0.01633760891854763, 0.10197000205516815, 0.03484668955206871, -0.08979984372854233, 0.031350795179605484, -0.06290501356124878, 0.05259009823203087, -0.04869356006383896, -0.06925348937511444, -0.013377794064581394, -0.054921574890613556, 0.07530007511377335, 0.020727040246129036, 0.08574032783508301, -0.02007264457643032, -0.10619869083166122, 0.11279048770666122, 0.08791369199752808, -0.0893058329820633, 0.029166527092456818, -0.10053735971450806, 0.013915755786001682, -0.07938611507415771, -0.007064095698297024, -0.20703130960464478, 0.006485046353191137, 0.0018438141560181975, 0.014914322644472122, 0.034115370362997055, 0.10435596108436584, 0.030128762125968933, 0.022408952936530113, 0.029981357976794243, -0.018031882122159004, 0.08979107439517975, 0.009407839737832546, -0.08929740637540817, -0.10686875879764557, -0.02606169506907463, -0.03660851716995239, 0.15328125655651093, -0.013654633425176144, 0.02513095922768116, -0.09524213522672653, 0.0469084158539772, -0.029759293422102928, 0.05308831110596657, -0.004855465143918991, 0.04144696891307831, -0.04124278575181961, 0.0014224101323634386, 0.04660613834857941, 0.017619729042053223, -0.09673509001731873, 0.1709372103214264, -0.18291769921779633, -0.02089911699295044, 0.12491315603256226, -0.11515885591506958, -0.06447350978851318, 0.02345173992216587, -0.014730798080563545, 0.02202417515218258, -0.05534713342785835, 0.014598351903259754, 0.16470010578632355, 0.014831818640232086, 0.12296251952648163, -0.083440862596035, -0.026016926392912865, -0.018848687410354614, -0.09721343219280243, 0.018556935712695122, 0.04981476813554764, 0.039000771939754486, -0.1162211075425148, 0.10053393989801407, 0.10247835516929626, -0.10287507623434067, 0.015802966430783272, 0.004366625100374222, -0.04628602787852287, -0.03741034120321274, -0.04338941350579262, -0.03472225368022919, -0.027288775891065598, -0.029719965532422066, -0.042659372091293335, 0.03510940074920654, 0.029411355033516884, 0.013905813917517662, -0.15196870267391205, 0.005906535778194666, 0.013755157589912415, -0.011594702489674091, -0.01026098895817995, 0.04376592859625816, -0.008289550431072712, 0.061599910259246826, 0.028126584365963936, -0.14367526769638062, 0.058363452553749084, -0.0351489894092083, -0.14141881465911865, 0.22615239024162292, -0.07566404342651367, -0.27953410148620605, -0.12290982156991959, 0.021696176379919052, 0.006087016314268112, 0.062644362449646, 0.0708455741405487, -0.045471686869859695, -0.03838431462645531, -0.0820717066526413, -0.03139931708574295, -0.024082103744149208, 0.0024445243179798126, -0.0878162756562233, 0.00009271603630622849, -0.003847202518954873, -0.09421621263027191, -0.0031321351416409016, -0.05890399217605591, -0.06755509972572327, 0.04214369133114815, -0.09165877103805542, -0.007889453321695328, 0.2206796407699585, 0.01076422818005085, 0.03367805480957031, -0.05330446735024452, 0.17513544857501984, -0.015779828652739525, -0.03460876643657684, 0.1611550748348236, -0.012880779802799225, 0.04445143789052963, 0.17388220131397247, 0.019242996349930763, -0.0683070495724678, 0.07065898925065994, -0.03215143457055092, -0.04168960079550743, -0.1452464610338211, -0.09278552234172821, -0.038062673062086105, -0.0233317743986845, 0.07220902293920517, -0.01197992917150259, 0.08731245249509811, 0.05724746733903885, -0.003093191422522068, -0.01159081980586052, 0.03088546171784401, 0.0780472680926323, 0.1442079395055771, 0.003124864771962166, 0.17355042695999146, -0.016894925385713577, -0.11764701455831528, 0.027372056618332863, 0.02583908475935459, 0.10140083730220795, 0.05113178864121437, 0.009328491054475307, 0.10008900612592697, 0.027490992099046707, 0.07921173423528671, -0.004589700140058994, 0.004766227677464485, 0.0007489954587072134, -0.03577183932065964, -0.029451249167323112, 0.016168830916285515, 0.09111353009939194, 0.01197222899645567, -0.02815566398203373, -0.12441352754831314, -0.03104766085743904, 0.04662104323506355, 0.11217736452817917, 0.16801334917545319, -0.30731672048568726, -0.05213860049843788, 0.04639245942234993, -0.06299179047346115, 0.023163730278611183, 0.09142906963825226, -0.004638497717678547, -0.13518525660037994, 0.00644878763705492, -0.004350210539996624, 0.08985988050699234, 0.0018690075958147645, 0.02098994515836239, -0.06347128748893738, 0.02823813259601593, 0.014463197439908981, 0.12384771555662155, -0.34352806210517883, 0.10411574691534042, -0.029895853251218796, 0.011633627116680145, -0.04604942351579666, -0.006390930153429508, 0.04763137176632881, 0.14670772850513458, 0.08665851503610611, -0.02811373583972454, -0.09105808287858963, -0.06061112508177757, -0.04584410414099693, 0.022072266787290573, 0.025318568572402, -0.05517273768782616, 0.05771936476230621, -0.04354587942361832, -0.005889877211302519, 0.02616376057267189, 0.05868706852197647, -0.025132019072771072, -0.07593091577291489, -0.007170701399445534, 0.013278568163514137, 0.05211564898490906, -0.015402347780764103, -0.06917586177587509, -0.13652382791042328, 0.042136695235967636, -0.07789728045463562, -0.06028033047914505, -0.103256456553936, 0.09486180543899536, 0.07285741716623306, -0.06623293459415436, 0.0039231302216649055, 0.016175266355276108, 0.133241206407547, 0.004705662373453379, -0.1416292041540146, 0.08322334289550781, -0.07802871614694595, -0.08273158967494965, -0.06269396841526031, 0.03027677722275257, 0.04064171016216278, 0.04588952288031578, 0.0636964812874794, 0.06164630129933357, -0.11224881559610367, -0.07974357903003693, -0.02778312936425209, 0.016498364508152008, 0.05221463739871979, 0.024097010493278503, -0.005767281167209148, -0.055050041526556015, -0.03973664715886116, 0.06736788898706436, 0.17233264446258545, 0.0396297462284565, -0.09726333618164062, 0.12583638727664948, 0.17738917469978333, -0.04895620420575142, -0.283265620470047, -0.09362755715847015, 0.04195278882980347, 0.07143981009721756, 0.014062013477087021, -0.09430427849292755, 0.1031544879078865, 0.012528144754469395, -0.048478029668331146, -0.2168145477771759, -0.3028346598148346, -0.10955806821584702, 0.18833613395690918, 0.03733060136437416, 0.1766844391822815, -0.07329234480857849, 0.00888361781835556, -0.06110344082117081, 0.008218829520046711, 0.149373859167099, -0.06190621480345726, 0.0771213248372078, 0.031203610822558403, 0.02545219659805298, -0.0002090671769110486, -0.01993192359805107, 0.07104034721851349, 0.08380923420190811, 0.060775671154260635, -0.02233022265136242, -0.005947250407189131, 0.030198819935321808, -0.038693733513355255, 0.13508112728595734, -0.038966890424489975, 0.09695424884557724, -0.060185179114341736, -0.08268783241510391, -0.033626738935709, 0.0095839723944664, 0.021442634984850883, -0.047682344913482666, -0.014281190931797028, 0.0293685682117939, 0.04048173502087593, -0.010960451327264309, 0.06327590346336365, -0.06340527534484863, 0.045411452651023865, 0.11049935221672058, 0.13265644013881683, -0.01526104100048542, -0.03578300401568413, 0.0051394738256931305, -0.047414444386959076, 0.11430126428604126, -0.164499893784523, 0.0911836102604866, 0.07006113976240158, -0.022437749430537224, 0.1343110352754593, 0.0996335968375206, -0.04331466183066368, 0.018353044986724854, 0.052883587777614594, -0.0847473293542862, -0.06420569121837616, -0.015110892243683338, -0.007178271654993296, -0.005094901192933321, 0.0529211089015007, 0.11877772212028503, -0.05779349431395531, -0.04995092377066612, -0.008740345947444439, 0.035886529833078384, -0.06572326272726059, 0.11153262853622437, 0.06726455688476562, 0.032411862164735794, -0.09644010663032532, 0.06804357469081879, 0.07756434381008148, -0.02221454121172428, 0.06189196556806564, 0.018371189013123512, -0.13287536799907684, -0.09099080413579941, -0.03262708708643913, 0.17415082454681396, -0.1126992255449295, -0.09605299681425095, -0.1218004897236824, -0.10991249978542328, 0.04737558588385582, 0.11134099960327148, 0.13007406890392303, 0.07032902538776398, -0.07839404791593552, -0.06061587482690811, -0.07010961323976517, 0.0824027732014656, 0.05524514988064766, -0.002843055408447981, -0.07459040731191635, 0.11347991228103638, 0.031216993927955627, 0.0675557404756546, -0.08114106953144073, -0.10859313607215881, -0.128158301115036, 0.06127750873565674, -0.06947355717420578, -0.005555060692131519, -0.07801594585180283, -0.010933276265859604, 0.03555882349610329, -0.05734606087207794, -0.0364067368209362, 0.013651168905198574, -0.08039981126785278, 0.025344787165522575, -0.037574876099824905, 0.06731393933296204, -0.10970719158649445, -0.029593832790851593, 0.023546205833554268, -0.03511421009898186, 0.04349852725863457, 0.07940305769443512, -0.02931799553334713, 0.04794476926326752, -0.0984886959195137, -0.04244762659072876, -0.0008963823784142733, 0.05560130998492241, 0.04657924920320511, -0.11562971025705338, 0.011889439076185226, 0.08023174852132797, 0.02496957965195179, 0.01808847486972809, 0.04983857646584511, -0.06669441610574722, 0.03893890976905823, -0.005539896432310343, -0.04568590968847275, -0.08155433088541031, 0.0019124216632917523, 0.029221396893262863, 0.11990660429000854, 0.17286483943462372, -0.05838547274470329, 0.0494680181145668, -0.12196874618530273, 0.002273800317198038, -0.03104289248585701, -0.03879985958337784, -0.12439896911382675, -0.10816622525453568, 0.023254498839378357, 0.005873893387615681, 0.19243818521499634, 0.08622796088457108, 0.09015563130378723, 0.02026992477476597, 0.08280457556247711, 0.15640299022197723, 0.08169982582330704, 0.13479216396808624, 0.0753636434674263, 0.003552020760253072, -0.04583922401070595, 0.06687968969345093, 0.04342857003211975, 0.0072935158386826515, 0.0038177764508873224, 0.09528568387031555, 0.00667149294167757, 0.10526926070451736, 0.042451679706573486, 0.0500829704105854, -0.032434239983558655, -0.06250517815351486, -0.0641176775097847, -0.01888967491686344, -0.011569607071578503, 0.12127377092838287, 0.14277249574661255, -0.08079589903354645, 0.04679444059729576, 0.015447227284312248, -0.06474620848894119, -0.1609332412481308, -0.1257009506225586, -0.07168443500995636, -0.18344774842262268, -0.027470940724015236, -0.09683065861463547, -0.02620367519557476, 0.16843217611312866, 0.04557392746210098, -0.009155435487627983, 0.02763553522527218, 0.0004341208259575069, -0.02509470470249653, -0.020050279796123505, -0.08659908920526505, 0.07982605695724487, 0.04354038089513779, -0.010250295512378216, 0.07036767154932022, 0.038146670907735825, 0.051697567105293274, -0.016725361347198486, 0.016275659203529358, 0.014421421103179455, -0.0931367501616478, -0.08636751025915146, -0.014829674735665321, 0.006926656700670719, -0.012881503440439701, 0.022557280957698822, 0.07137629389762878, -0.03945473954081535, 0.001135995495133102, 0.21051138639450073, -0.09168606251478195, -0.13335897028446198, -0.22504109144210815, 0.3292695879936218, 0.06128848344087601, 0.08006542921066284, 0.01782870665192604, -0.09772481769323349, -0.08380256593227386, 0.23835910856723785, 0.15601664781570435, -0.07657447457313538, -0.014831600710749626, 0.07228817790746689, -0.016094770282506943, 0.04914749786257744, 0.11798988282680511, -0.0018936273409053683, 0.22898456454277039, -0.08471135795116425, 0.04600175470113754, -0.03982498124241829, -0.06657011061906815, -0.0016606205608695745, 0.04384338855743408, -0.028883112594485283, -0.0646972507238388, -0.0373566634953022, 0.14134474098682404, -0.09122554212808609, 0.049046847969293594, -0.0002345513494219631, -0.022823620587587357, -0.07716405391693115, -0.006424167193472385, 0.06062676012516022, 0.00020406620751600713, 0.08504018932580948, -0.038146086037158966, 0.03794025629758835, 0.11565134674310684, -0.0017569564515724778, -0.06979916244745255, -0.058996688574552536, 0.06873651593923569, -0.06620217114686966, 0.02624395117163658, -0.020865874364972115, 0.19483453035354614, 0.09904284030199051, 0.06483505666255951, -0.03644286096096039, 0.0626656711101532, -0.013008199632167816, 0.011695616878569126, 0.12002301216125488, 0.03363015130162239, 0.0020754639990627766, 0.06408432871103287, 0.044256992638111115, -0.1768614798784256, 0.07748190313577652, -0.14381177723407745, -0.06964220106601715, -0.10966911166906357, 0.02179321087896824, -0.012797866016626358, 0.1110919713973999, 0.09294867515563965, -0.04715340584516525, 0.012934789061546326, 0.018140526488423347, 0.015916945412755013, -0.041642192751169205, -0.078044094145298, -0.03142290934920311, -0.16770023107528687, 0.007813815027475357, -0.03266200050711632, 0.002701889956369996, -0.22878091037273407, -0.011267242021858692, -0.008542446419596672, -0.033176716417074203, 0.008387149311602116, 0.08422781527042389, 0.08410125225782394, 0.03710673004388809, -0.07843427360057831, -0.06219158694148064, 0.0013762605376541615, 0.09299024194478989, -0.12327927350997925, -0.15334023535251617 ]
null
null
transformers
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. ## Training Data This model was trained on the [STS benchmark dataset](http://ixa2.si.ehu.eus/stswiki/index.php/STSbenchmark). The model will predict a score between 0 and 1 how for the semantic similarity of two sentences. ## Usage and Performance Pre-trained models can be used like this: ``` from sentence_transformers import CrossEncoder model = CrossEncoder('model_name') scores = model.predict([('Sentence 1', 'Sentence 2'), ('Sentence 3', 'Sentence 4')]) ``` The model will predict scores for the pairs `('Sentence 1', 'Sentence 2')` and `('Sentence 3', 'Sentence 4')`. You can use this model also without sentence_transformers and by just using Transformers ``AutoModel`` class
{"license": "apache-2.0"}
text-classification
cross-encoder/stsb-roberta-large
[ "transformers", "pytorch", "jax", "roberta", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
# Cross-Encoder for Quora Duplicate Questions Detection This model was trained using SentenceTransformers Cross-Encoder class. ## Training Data This model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences. ## Usage and Performance Pre-trained models can be used like this: The model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'. You can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class
[ "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.", "## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ "TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.", "## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.", "## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ 52, 34, 38, 85 ]
[ "passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class" ]
[ -0.08450348675251007, 0.06658493727445602, -0.0020891379099339247, 0.06090733781456947, 0.1219165176153183, -0.010885688476264477, 0.18726801872253418, 0.06976161152124405, 0.004681415390223265, -0.032464902848005295, 0.10880614817142487, 0.10593046247959137, 0.009075884707272053, 0.1311509907245636, -0.0318918414413929, -0.2397564798593521, 0.05535270273685455, -0.021333403885364532, -0.023506149649620056, 0.1251000612974167, 0.10538756102323532, -0.08173905313014984, 0.06480519473552704, 0.036619383841753006, -0.1481119692325592, 0.0361047126352787, 0.004630628507584333, -0.06398891657590866, 0.1055711954832077, 0.029057493433356285, 0.12339451909065247, 0.09259660542011261, 0.07925070077180862, -0.12724295258522034, 0.012720810249447823, 0.016518067568540573, 0.009311544708907604, 0.03629666194319725, 0.036708857864141464, -0.10701856762170792, -0.09100929647684097, 0.08599001914262772, 0.08524163067340851, 0.0009422770235687494, -0.11783528327941895, -0.1288551539182663, -0.02296801283955574, -0.0451493114233017, 0.1464175432920456, 0.0878152996301651, -0.045602668076753616, 0.1485908329486847, -0.1755804717540741, 0.04145662114024162, 0.054525889456272125, -0.3068389296531677, 0.010614708065986633, 0.0938933864235878, 0.023070018738508224, 0.04588997736573219, -0.04204950109124184, -0.006081583444029093, 0.028909359127283096, 0.04380388930439949, 0.06281910091638565, -0.04001569375395775, -0.010886033065617085, 0.037939924746751785, -0.1775822639465332, -0.03491581231355667, 0.21934248507022858, -0.03001006878912449, -0.03818020224571228, -0.10346876084804535, -0.030659932643175125, 0.05167389288544655, -0.022362971678376198, -0.08230966329574585, 0.004080281592905521, 0.05453440919518471, -0.09994271397590637, 0.043615926057100296, -0.12259428203105927, -0.1212959885597229, -0.11897437274456024, 0.1481919288635254, 0.0519876554608345, 0.05614864081144333, -0.11817987263202667, 0.11007988452911377, -0.018013888970017433, -0.08047500997781754, -0.03508051112294197, -0.07759242504835129, -0.13628479838371277, -0.017040323466062546, -0.1888146996498108, -0.1461862325668335, 0.019468307495117188, 0.23470254242420197, 0.05425926297903061, -0.003253627335652709, 0.05223825201392174, 0.052410710602998734, 0.05260057747364044, 0.16987985372543335, -0.09921816736459732, 0.019514871761202812, -0.040489714592695236, -0.006623496767133474, -0.014924317598342896, -0.020644033327698708, -0.15668311715126038, -0.02293773740530014, 0.06836400926113129, 0.03340352699160576, -0.03544469177722931, 0.060046419501304626, -0.03591883182525635, -0.04250994324684143, 0.04778998717665672, -0.07110997289419174, 0.007199421525001526, 0.001498668105341494, -0.08984625339508057, 0.04597581550478935, 0.004727287217974663, 0.05564960837364197, -0.04794371500611305, -0.03951052203774452, -0.05796322971582413, -0.0041662254370749, -0.07994621992111206, -0.13172027468681335, -0.017280934378504753, 0.0645439475774765, 0.004875025246292353, -0.12934610247612, -0.15153756737709045, -0.06250552833080292, 0.06014043092727661, -0.025576326996088028, -0.0060349274426698685, -0.10992851108312607, -0.024137191474437714, 0.014359449967741966, -0.03408343344926834, -0.012038946151733398, -0.006700094789266586, 0.027012227103114128, -0.010106818750500679, 0.10634177178144455, 0.011723324656486511, 0.056721992790699005, -0.17293746769428253, -0.014487387612462044, -0.06987504661083221, 0.13998807966709137, 0.00621027173474431, 0.036297235637903214, -0.07619620859622955, -0.03155680000782013, -0.03006790578365326, 0.06660110503435135, 0.037052832543849945, 0.1293592005968094, -0.16096548736095428, -0.06480851024389267, 0.2344948649406433, -0.12800095975399017, -0.11015582084655762, 0.12573368847370148, -0.049440640956163406, 0.03731274977326393, 0.10660473257303238, 0.12351691722869873, 0.0702773854136467, -0.034878578037023544, 0.08157497644424438, 0.06925155967473984, -0.0970594733953476, -0.01633760891854763, 0.10197000205516815, 0.03484668955206871, -0.08979984372854233, 0.031350795179605484, -0.06290501356124878, 0.05259009823203087, -0.04869356006383896, -0.06925348937511444, -0.013377794064581394, -0.054921574890613556, 0.07530007511377335, 0.020727040246129036, 0.08574032783508301, -0.02007264457643032, -0.10619869083166122, 0.11279048770666122, 0.08791369199752808, -0.0893058329820633, 0.029166527092456818, -0.10053735971450806, 0.013915755786001682, -0.07938611507415771, -0.007064095698297024, -0.20703130960464478, 0.006485046353191137, 0.0018438141560181975, 0.014914322644472122, 0.034115370362997055, 0.10435596108436584, 0.030128762125968933, 0.022408952936530113, 0.029981357976794243, -0.018031882122159004, 0.08979107439517975, 0.009407839737832546, -0.08929740637540817, -0.10686875879764557, -0.02606169506907463, -0.03660851716995239, 0.15328125655651093, -0.013654633425176144, 0.02513095922768116, -0.09524213522672653, 0.0469084158539772, -0.029759293422102928, 0.05308831110596657, -0.004855465143918991, 0.04144696891307831, -0.04124278575181961, 0.0014224101323634386, 0.04660613834857941, 0.017619729042053223, -0.09673509001731873, 0.1709372103214264, -0.18291769921779633, -0.02089911699295044, 0.12491315603256226, -0.11515885591506958, -0.06447350978851318, 0.02345173992216587, -0.014730798080563545, 0.02202417515218258, -0.05534713342785835, 0.014598351903259754, 0.16470010578632355, 0.014831818640232086, 0.12296251952648163, -0.083440862596035, -0.026016926392912865, -0.018848687410354614, -0.09721343219280243, 0.018556935712695122, 0.04981476813554764, 0.039000771939754486, -0.1162211075425148, 0.10053393989801407, 0.10247835516929626, -0.10287507623434067, 0.015802966430783272, 0.004366625100374222, -0.04628602787852287, -0.03741034120321274, -0.04338941350579262, -0.03472225368022919, -0.027288775891065598, -0.029719965532422066, -0.042659372091293335, 0.03510940074920654, 0.029411355033516884, 0.013905813917517662, -0.15196870267391205, 0.005906535778194666, 0.013755157589912415, -0.011594702489674091, -0.01026098895817995, 0.04376592859625816, -0.008289550431072712, 0.061599910259246826, 0.028126584365963936, -0.14367526769638062, 0.058363452553749084, -0.0351489894092083, -0.14141881465911865, 0.22615239024162292, -0.07566404342651367, -0.27953410148620605, -0.12290982156991959, 0.021696176379919052, 0.006087016314268112, 0.062644362449646, 0.0708455741405487, -0.045471686869859695, -0.03838431462645531, -0.0820717066526413, -0.03139931708574295, -0.024082103744149208, 0.0024445243179798126, -0.0878162756562233, 0.00009271603630622849, -0.003847202518954873, -0.09421621263027191, -0.0031321351416409016, -0.05890399217605591, -0.06755509972572327, 0.04214369133114815, -0.09165877103805542, -0.007889453321695328, 0.2206796407699585, 0.01076422818005085, 0.03367805480957031, -0.05330446735024452, 0.17513544857501984, -0.015779828652739525, -0.03460876643657684, 0.1611550748348236, -0.012880779802799225, 0.04445143789052963, 0.17388220131397247, 0.019242996349930763, -0.0683070495724678, 0.07065898925065994, -0.03215143457055092, -0.04168960079550743, -0.1452464610338211, -0.09278552234172821, -0.038062673062086105, -0.0233317743986845, 0.07220902293920517, -0.01197992917150259, 0.08731245249509811, 0.05724746733903885, -0.003093191422522068, -0.01159081980586052, 0.03088546171784401, 0.0780472680926323, 0.1442079395055771, 0.003124864771962166, 0.17355042695999146, -0.016894925385713577, -0.11764701455831528, 0.027372056618332863, 0.02583908475935459, 0.10140083730220795, 0.05113178864121437, 0.009328491054475307, 0.10008900612592697, 0.027490992099046707, 0.07921173423528671, -0.004589700140058994, 0.004766227677464485, 0.0007489954587072134, -0.03577183932065964, -0.029451249167323112, 0.016168830916285515, 0.09111353009939194, 0.01197222899645567, -0.02815566398203373, -0.12441352754831314, -0.03104766085743904, 0.04662104323506355, 0.11217736452817917, 0.16801334917545319, -0.30731672048568726, -0.05213860049843788, 0.04639245942234993, -0.06299179047346115, 0.023163730278611183, 0.09142906963825226, -0.004638497717678547, -0.13518525660037994, 0.00644878763705492, -0.004350210539996624, 0.08985988050699234, 0.0018690075958147645, 0.02098994515836239, -0.06347128748893738, 0.02823813259601593, 0.014463197439908981, 0.12384771555662155, -0.34352806210517883, 0.10411574691534042, -0.029895853251218796, 0.011633627116680145, -0.04604942351579666, -0.006390930153429508, 0.04763137176632881, 0.14670772850513458, 0.08665851503610611, -0.02811373583972454, -0.09105808287858963, -0.06061112508177757, -0.04584410414099693, 0.022072266787290573, 0.025318568572402, -0.05517273768782616, 0.05771936476230621, -0.04354587942361832, -0.005889877211302519, 0.02616376057267189, 0.05868706852197647, -0.025132019072771072, -0.07593091577291489, -0.007170701399445534, 0.013278568163514137, 0.05211564898490906, -0.015402347780764103, -0.06917586177587509, -0.13652382791042328, 0.042136695235967636, -0.07789728045463562, -0.06028033047914505, -0.103256456553936, 0.09486180543899536, 0.07285741716623306, -0.06623293459415436, 0.0039231302216649055, 0.016175266355276108, 0.133241206407547, 0.004705662373453379, -0.1416292041540146, 0.08322334289550781, -0.07802871614694595, -0.08273158967494965, -0.06269396841526031, 0.03027677722275257, 0.04064171016216278, 0.04588952288031578, 0.0636964812874794, 0.06164630129933357, -0.11224881559610367, -0.07974357903003693, -0.02778312936425209, 0.016498364508152008, 0.05221463739871979, 0.024097010493278503, -0.005767281167209148, -0.055050041526556015, -0.03973664715886116, 0.06736788898706436, 0.17233264446258545, 0.0396297462284565, -0.09726333618164062, 0.12583638727664948, 0.17738917469978333, -0.04895620420575142, -0.283265620470047, -0.09362755715847015, 0.04195278882980347, 0.07143981009721756, 0.014062013477087021, -0.09430427849292755, 0.1031544879078865, 0.012528144754469395, -0.048478029668331146, -0.2168145477771759, -0.3028346598148346, -0.10955806821584702, 0.18833613395690918, 0.03733060136437416, 0.1766844391822815, -0.07329234480857849, 0.00888361781835556, -0.06110344082117081, 0.008218829520046711, 0.149373859167099, -0.06190621480345726, 0.0771213248372078, 0.031203610822558403, 0.02545219659805298, -0.0002090671769110486, -0.01993192359805107, 0.07104034721851349, 0.08380923420190811, 0.060775671154260635, -0.02233022265136242, -0.005947250407189131, 0.030198819935321808, -0.038693733513355255, 0.13508112728595734, -0.038966890424489975, 0.09695424884557724, -0.060185179114341736, -0.08268783241510391, -0.033626738935709, 0.0095839723944664, 0.021442634984850883, -0.047682344913482666, -0.014281190931797028, 0.0293685682117939, 0.04048173502087593, -0.010960451327264309, 0.06327590346336365, -0.06340527534484863, 0.045411452651023865, 0.11049935221672058, 0.13265644013881683, -0.01526104100048542, -0.03578300401568413, 0.0051394738256931305, -0.047414444386959076, 0.11430126428604126, -0.164499893784523, 0.0911836102604866, 0.07006113976240158, -0.022437749430537224, 0.1343110352754593, 0.0996335968375206, -0.04331466183066368, 0.018353044986724854, 0.052883587777614594, -0.0847473293542862, -0.06420569121837616, -0.015110892243683338, -0.007178271654993296, -0.005094901192933321, 0.0529211089015007, 0.11877772212028503, -0.05779349431395531, -0.04995092377066612, -0.008740345947444439, 0.035886529833078384, -0.06572326272726059, 0.11153262853622437, 0.06726455688476562, 0.032411862164735794, -0.09644010663032532, 0.06804357469081879, 0.07756434381008148, -0.02221454121172428, 0.06189196556806564, 0.018371189013123512, -0.13287536799907684, -0.09099080413579941, -0.03262708708643913, 0.17415082454681396, -0.1126992255449295, -0.09605299681425095, -0.1218004897236824, -0.10991249978542328, 0.04737558588385582, 0.11134099960327148, 0.13007406890392303, 0.07032902538776398, -0.07839404791593552, -0.06061587482690811, -0.07010961323976517, 0.0824027732014656, 0.05524514988064766, -0.002843055408447981, -0.07459040731191635, 0.11347991228103638, 0.031216993927955627, 0.0675557404756546, -0.08114106953144073, -0.10859313607215881, -0.128158301115036, 0.06127750873565674, -0.06947355717420578, -0.005555060692131519, -0.07801594585180283, -0.010933276265859604, 0.03555882349610329, -0.05734606087207794, -0.0364067368209362, 0.013651168905198574, -0.08039981126785278, 0.025344787165522575, -0.037574876099824905, 0.06731393933296204, -0.10970719158649445, -0.029593832790851593, 0.023546205833554268, -0.03511421009898186, 0.04349852725863457, 0.07940305769443512, -0.02931799553334713, 0.04794476926326752, -0.0984886959195137, -0.04244762659072876, -0.0008963823784142733, 0.05560130998492241, 0.04657924920320511, -0.11562971025705338, 0.011889439076185226, 0.08023174852132797, 0.02496957965195179, 0.01808847486972809, 0.04983857646584511, -0.06669441610574722, 0.03893890976905823, -0.005539896432310343, -0.04568590968847275, -0.08155433088541031, 0.0019124216632917523, 0.029221396893262863, 0.11990660429000854, 0.17286483943462372, -0.05838547274470329, 0.0494680181145668, -0.12196874618530273, 0.002273800317198038, -0.03104289248585701, -0.03879985958337784, -0.12439896911382675, -0.10816622525453568, 0.023254498839378357, 0.005873893387615681, 0.19243818521499634, 0.08622796088457108, 0.09015563130378723, 0.02026992477476597, 0.08280457556247711, 0.15640299022197723, 0.08169982582330704, 0.13479216396808624, 0.0753636434674263, 0.003552020760253072, -0.04583922401070595, 0.06687968969345093, 0.04342857003211975, 0.0072935158386826515, 0.0038177764508873224, 0.09528568387031555, 0.00667149294167757, 0.10526926070451736, 0.042451679706573486, 0.0500829704105854, -0.032434239983558655, -0.06250517815351486, -0.0641176775097847, -0.01888967491686344, -0.011569607071578503, 0.12127377092838287, 0.14277249574661255, -0.08079589903354645, 0.04679444059729576, 0.015447227284312248, -0.06474620848894119, -0.1609332412481308, -0.1257009506225586, -0.07168443500995636, -0.18344774842262268, -0.027470940724015236, -0.09683065861463547, -0.02620367519557476, 0.16843217611312866, 0.04557392746210098, -0.009155435487627983, 0.02763553522527218, 0.0004341208259575069, -0.02509470470249653, -0.020050279796123505, -0.08659908920526505, 0.07982605695724487, 0.04354038089513779, -0.010250295512378216, 0.07036767154932022, 0.038146670907735825, 0.051697567105293274, -0.016725361347198486, 0.016275659203529358, 0.014421421103179455, -0.0931367501616478, -0.08636751025915146, -0.014829674735665321, 0.006926656700670719, -0.012881503440439701, 0.022557280957698822, 0.07137629389762878, -0.03945473954081535, 0.001135995495133102, 0.21051138639450073, -0.09168606251478195, -0.13335897028446198, -0.22504109144210815, 0.3292695879936218, 0.06128848344087601, 0.08006542921066284, 0.01782870665192604, -0.09772481769323349, -0.08380256593227386, 0.23835910856723785, 0.15601664781570435, -0.07657447457313538, -0.014831600710749626, 0.07228817790746689, -0.016094770282506943, 0.04914749786257744, 0.11798988282680511, -0.0018936273409053683, 0.22898456454277039, -0.08471135795116425, 0.04600175470113754, -0.03982498124241829, -0.06657011061906815, -0.0016606205608695745, 0.04384338855743408, -0.028883112594485283, -0.0646972507238388, -0.0373566634953022, 0.14134474098682404, -0.09122554212808609, 0.049046847969293594, -0.0002345513494219631, -0.022823620587587357, -0.07716405391693115, -0.006424167193472385, 0.06062676012516022, 0.00020406620751600713, 0.08504018932580948, -0.038146086037158966, 0.03794025629758835, 0.11565134674310684, -0.0017569564515724778, -0.06979916244745255, -0.058996688574552536, 0.06873651593923569, -0.06620217114686966, 0.02624395117163658, -0.020865874364972115, 0.19483453035354614, 0.09904284030199051, 0.06483505666255951, -0.03644286096096039, 0.0626656711101532, -0.013008199632167816, 0.011695616878569126, 0.12002301216125488, 0.03363015130162239, 0.0020754639990627766, 0.06408432871103287, 0.044256992638111115, -0.1768614798784256, 0.07748190313577652, -0.14381177723407745, -0.06964220106601715, -0.10966911166906357, 0.02179321087896824, -0.012797866016626358, 0.1110919713973999, 0.09294867515563965, -0.04715340584516525, 0.012934789061546326, 0.018140526488423347, 0.015916945412755013, -0.041642192751169205, -0.078044094145298, -0.03142290934920311, -0.16770023107528687, 0.007813815027475357, -0.03266200050711632, 0.002701889956369996, -0.22878091037273407, -0.011267242021858692, -0.008542446419596672, -0.033176716417074203, 0.008387149311602116, 0.08422781527042389, 0.08410125225782394, 0.03710673004388809, -0.07843427360057831, -0.06219158694148064, 0.0013762605376541615, 0.09299024194478989, -0.12327927350997925, -0.15334023535251617 ]
null
null
transformers
### Kw2Poem
{"language": "vi", "tags": ["gpt"], "widget": [{"text": "<s> n\u00fai nh\u00e0 xe [SEP] "}]}
text-generation
crylake/kw2poem-generation
[ "transformers", "pytorch", "gpt2", "text-generation", "gpt", "vi", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "vi" ]
TAGS #transformers #pytorch #gpt2 #text-generation #gpt #vi #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
### Kw2Poem
[ "### Kw2Poem" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #gpt #vi #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Kw2Poem" ]
[ 52, 7 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #gpt #vi #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Kw2Poem" ]
[ 0.00957817304879427, 0.06301134824752808, -0.007064964156597853, 0.0036542851012200117, 0.12309610843658447, 0.031439412385225296, 0.07243029028177261, 0.1607857048511505, 0.00964074395596981, -0.01801552250981331, 0.18630695343017578, 0.17791466414928436, 0.03313320875167847, 0.0681326687335968, -0.027575921267271042, -0.31554916501045227, 0.05325678735971451, 0.0723961889743805, -0.020475471392273903, 0.12803497910499573, 0.04001649096608162, -0.06150275841355324, 0.08209304511547089, 0.003992442507296801, -0.16980701684951782, -0.008797045797109604, 0.025250831618905067, -0.11566734313964844, 0.09016840904951096, 0.05655558779835701, 0.06918779015541077, 0.04472552239894867, -0.03827078640460968, -0.1263580322265625, 0.02615487389266491, 0.006352890748530626, -0.04630299657583237, 0.043258532881736755, 0.09118277579545975, -0.06698725372552872, 0.09973156452178955, 0.022308319807052612, -0.07163960486650467, 0.02657143771648407, -0.13541530072689056, -0.071018747985363, -0.035225287079811096, 0.07012996822595596, 0.06396427005529404, 0.06126224249601364, -0.010328901931643486, 0.1294802576303482, -0.072997085750103, 0.08923482149839401, 0.18938976526260376, -0.3316987156867981, -0.0015500655863434076, 0.07680265605449677, 0.03494839370250702, -0.01640024594962597, -0.031431008130311966, 0.08453582972288132, 0.01143955159932375, -0.0011958361137658358, -0.019267888739705086, -0.09979125112295151, -0.11628061532974243, 0.042645879089832306, -0.10252959281206131, -0.055888764560222626, 0.2122832089662552, -0.05855756998062134, 0.0569487027823925, 0.02094966731965542, -0.09387509524822235, -0.014127666130661964, 0.0008141568978317082, 0.023698817938566208, -0.07356588542461395, 0.04334239289164543, -0.004188459366559982, -0.06054595112800598, -0.12845057249069214, -0.04309607669711113, -0.1624227911233902, 0.23481477797031403, 0.03394542261958122, 0.0395079143345356, -0.15551385283470154, 0.13015125691890717, -0.06904776394367218, -0.09651973098516464, -0.0034290305338799953, -0.07348773628473282, 0.09329099208116531, 0.026412852108478546, -0.018657071515917778, -0.05315980315208435, 0.093925341963768, 0.09602594375610352, 0.031985338777303696, 0.024359751492738724, -0.002628166461363435, 0.09244594722986221, 0.06021767109632492, 0.09955921769142151, -0.002758112270385027, -0.01708148792386055, 0.05911704897880554, -0.12917882204055786, -0.0078049651347100735, -0.06700736284255981, -0.14689482748508453, -0.07990694791078568, 0.057311221957206726, 0.07181021571159363, 0.044495828449726105, 0.10255920886993408, -0.03267213702201843, -0.029797419905662537, 0.0361289456486702, -0.0445629246532917, -0.008850942365825176, -0.0107669522985816, 0.03428369387984276, 0.1568555235862732, 0.00325220450758934, 0.026154393330216408, -0.12736763060092926, 0.08676701784133911, -0.07274826616048813, -0.0031747606117278337, -0.02124539203941822, -0.03635423257946968, 0.020667510107159615, -0.10058806091547012, 0.030601467937231064, -0.1443287432193756, -0.1689336895942688, 0.013190992176532745, -0.02336847223341465, -0.013140843249857426, -0.0766136571764946, -0.030397942289710045, -0.04604308679699898, 0.05602826178073883, -0.05154234915971756, 0.019728047773241997, -0.05774667114019394, 0.11128174513578415, -0.052658792585134506, 0.07410987466573715, -0.13703618943691254, 0.053574539721012115, -0.10183457285165787, 0.006429322995245457, -0.13172046840190887, 0.04558270052075386, -0.028972763568162918, 0.04250296205282211, -0.019867192953824997, -0.039371710270643234, -0.06422477960586548, 0.03419045731425285, -0.032960034906864166, 0.18798483908176422, -0.07311984151601791, -0.13018974661827087, 0.2925363779067993, -0.06353839486837387, -0.1162010058760643, 0.08763570338487625, 0.031029900535941124, 0.03783457726240158, 0.07800211757421494, 0.2249654084444046, 0.06717246025800705, -0.00396937970072031, 0.03586290031671524, 0.1031804159283638, -0.08144260197877884, -0.02689230628311634, 0.03230758756399155, -0.024571599438786507, -0.13738372921943665, 0.03794455528259277, -0.01866414211690426, 0.052874647080898285, -0.04736626520752907, -0.013525504618883133, -0.0316149927675724, 0.0037264577113091946, 0.08108885586261749, -0.03282852843403816, 0.10934285819530487, -0.04816563427448273, -0.027070369571447372, 0.0007734297541901469, 0.025037771090865135, -0.020882830023765564, 0.047117553651332855, -0.035543374717235565, 0.12087835371494293, -0.020327946171164513, 0.06259502470493317, -0.1060953140258789, -0.02786347270011902, 0.004772518761456013, 0.08213009685277939, 0.026698121801018715, 0.0887259989976883, 0.04245677590370178, -0.0168427973985672, -0.012565472163259983, -0.030297597870230675, 0.08050557971000671, -0.01142654474824667, -0.08747817575931549, -0.07889477163553238, 0.0820138081908226, -0.029122885316610336, -0.0694442018866539, -0.027604790404438972, 0.011847486719489098, 0.12812773883342743, 0.09897224605083466, -0.02164166420698166, 0.03927278891205788, -0.005955390632152557, 0.0067389062605798244, -0.06489384919404984, 0.008974163793027401, 0.08441240340471268, -0.010470270179212093, -0.0728292241692543, 0.2262585312128067, -0.13634425401687622, 0.15140721201896667, 0.19395232200622559, -0.2116062343120575, 0.03191465511918068, -0.02867828868329525, -0.024682335555553436, -0.002660174388438463, 0.020991845056414604, -0.029479147866368294, 0.1510322242975235, -0.0025584178511053324, 0.13806121051311493, -0.05763198435306549, -0.0770176500082016, 0.00790970865637064, -0.05813254415988922, 0.017832832410931587, 0.05987262353301048, 0.1267615258693695, -0.1311156302690506, 0.1655711531639099, 0.25164952874183655, 0.026769690215587616, 0.20776516199111938, 0.019149042665958405, -0.016153262928128242, 0.04002733156085014, -0.021725956350564957, -0.038566626608371735, -0.021408110857009888, -0.20528195798397064, -0.013087124563753605, 0.07204267382621765, 0.03846413642168045, 0.07889289408922195, -0.1345205157995224, -0.0437425896525383, -0.04111592099070549, -0.01434281188994646, 0.0031131766736507416, 0.09778230637311935, 0.028985416516661644, 0.11504949629306793, 0.01989189349114895, 0.05871374532580376, 0.0693468376994133, 0.029436497017741203, -0.07359284907579422, 0.2007758468389511, -0.1341300904750824, -0.326452374458313, -0.12633070349693298, -0.08797509968280792, -0.03790261223912239, 0.034858908504247665, 0.10500060766935349, -0.11347244679927826, -0.020820388570427895, 0.024372363463044167, 0.14073021709918976, -0.12031327188014984, 0.04537583515048027, -0.05039072781801224, 0.021589763462543488, -0.11016356945037842, -0.0658922791481018, -0.05893188714981079, 0.013843301683664322, -0.09759803861379623, 0.17724177241325378, -0.06949371099472046, 0.059028707444667816, 0.1800614595413208, 0.031153270974755287, 0.03705417737364769, -0.029824350029230118, 0.17289897799491882, -0.09951750189065933, 0.02500886283814907, 0.15677721798419952, -0.06731083244085312, 0.09062217175960541, 0.046989936381578445, -0.012154776602983475, -0.07390722632408142, 0.016875628381967545, -0.011683935299515724, -0.08091358840465546, -0.20660856366157532, -0.08553552627563477, -0.0886504277586937, 0.15322765707969666, 0.05269729718565941, 0.04272379353642464, 0.0866679698228836, 0.11853060126304626, -0.03408195450901985, 0.07567588984966278, -0.009533594362437725, 0.09122149646282196, 0.1803288757801056, -0.0285966657102108, 0.10404042154550552, -0.04416600614786148, -0.11093810200691223, 0.10502474009990692, 0.08413784950971603, 0.09985124319791794, 0.030600035563111305, -0.003252935828641057, 0.0469202995300293, 0.09457972645759583, 0.14892789721488953, 0.08599001169204712, 0.019359536468982697, -0.01143420860171318, -0.039638202637434006, -0.030968688428401947, -0.035622384399175644, 0.05220633000135422, 0.05053401365876198, -0.17141735553741455, -0.03554258495569229, -0.07563856244087219, 0.08663869649171829, 0.0934019461274147, 0.10105573385953903, -0.1376226544380188, -0.008649363182485104, 0.100326769053936, -0.028425011783838272, -0.133919358253479, 0.08279795199632645, 0.01092482265084982, -0.1892797350883484, 0.08940976858139038, -0.04272071644663811, 0.10588239878416061, -0.1050056740641594, 0.06997676193714142, -0.07266444712877274, -0.08001580089330673, 0.015113139525055885, 0.12805187702178955, -0.2630266845226288, 0.1980765014886856, 0.00470179645344615, -0.05748160555958748, -0.11423757672309875, -0.02530396357178688, 0.02106613852083683, 0.07345574349164963, 0.11278049647808075, 0.03998386114835739, 0.0015202027279883623, -0.07409339398145676, -0.04619338735938072, 0.05795834958553314, 0.1431449055671692, -0.1010911613702774, 0.02264595218002796, -0.01898161880671978, -0.0020718886516988277, -0.04349042475223541, -0.05938585847616196, -0.026471013203263283, -0.1162632405757904, 0.09133803844451904, 0.004934979602694511, 0.10479361563920975, -0.013762879185378551, -0.031141597777605057, -0.1040981113910675, 0.23426604270935059, -0.12257760763168335, -0.10206079483032227, -0.09111861139535904, -0.012589845806360245, 0.03955908864736557, -0.09625139832496643, 0.0342625267803669, -0.07532605528831482, 0.014382464811205864, -0.055122461169958115, -0.17879384756088257, 0.11824013292789459, -0.08392655104398727, -0.13032107055187225, -0.020929044112563133, 0.17684286832809448, -0.03928419202566147, 0.023285819217562675, 0.003155027749016881, 0.023544492200016975, -0.07451646775007248, -0.13398806750774384, 0.05609554797410965, 0.02090577594935894, 0.0189196839928627, 0.034923769533634186, -0.02513517625629902, 0.06144357845187187, -0.04819108545780182, -0.05300992727279663, 0.3036767840385437, 0.2120986431837082, -0.03889444097876549, 0.1935625970363617, 0.06893247365951538, -0.08998647332191467, -0.367197185754776, -0.10181158781051636, -0.07947426289319992, -0.0382821150124073, -0.05219694599509239, -0.209986612200737, 0.07809457182884216, 0.03275468572974205, -0.013884881511330605, 0.12455035746097565, -0.27253809571266174, -0.09908903390169144, 0.10868138074874878, 0.025725973770022392, 0.35784900188446045, -0.10534874349832535, -0.10031116753816605, -0.03133375570178032, -0.14740042388439178, 0.16990341246128082, -0.04684435948729515, 0.1021876186132431, -0.0563608855009079, 0.12966550886631012, 0.03599758818745613, -0.031546905636787415, 0.09999095648527145, 0.01583133079111576, -0.02859887108206749, -0.1090945228934288, -0.017750374972820282, 0.055275458842515945, 0.004516797140240669, -0.0057978457771241665, -0.047339312732219696, 0.03101438842713833, -0.15586735308170319, -0.02265102230012417, -0.10903351753950119, 0.08096453547477722, 0.015084374696016312, -0.0857151448726654, -0.014661803841590881, -0.06186206638813019, -0.028775228187441826, 0.005849520210176706, 0.15451213717460632, -0.004929624032229185, 0.0800783708691597, 0.04938990995287895, 0.09341436624526978, -0.05828000232577324, 0.027915148064494133, -0.07094196230173111, -0.060766562819480896, 0.07760199159383774, -0.11375044286251068, 0.0411730594933033, 0.1161332055926323, -0.010778318159282207, 0.0207026619464159, 0.10265020281076431, -0.03212396800518036, -0.00035324503551237285, 0.11143036186695099, -0.28819185495376587, -0.01519784890115261, -0.08012323826551437, 0.0003487319918349385, 0.09877925366163254, 0.06377357989549637, 0.19824156165122986, 0.00480060838162899, -0.0733577311038971, 0.00582854775711894, 0.010526592843234539, -0.043945059180259705, 0.08654814213514328, 0.02875732257962227, 0.01958903670310974, -0.12828795611858368, 0.07599804550409317, -0.02586091123521328, -0.14057785272598267, 0.03576691821217537, 0.1520865559577942, -0.13470661640167236, -0.1196122094988823, 0.02517283894121647, 0.07615108788013458, -0.15355950593948364, -0.009285707026720047, -0.01853850856423378, -0.1146107092499733, 0.08270766586065292, 0.16739656031131744, 0.06393714994192123, 0.09372229129076004, -0.016709135845303535, 0.00033211661502718925, -0.017629390582442284, 0.0060826437547802925, 0.015956709161400795, 0.038097936660051346, -0.0988348126411438, 0.06918919086456299, -0.029006076976656914, 0.18566696345806122, -0.08008981496095657, -0.05143186077475548, -0.15967725217342377, -0.001775190932676196, -0.08343244343996048, -0.05541709065437317, -0.10140618681907654, -0.06321018189191818, -0.023288492113351822, -0.034263819456100464, -0.038036808371543884, -0.041667092591524124, -0.1316547989845276, 0.01385523658245802, -0.07998481392860413, 0.024233905598521233, -0.031159712001681328, -0.0043153963051736355, 0.12203552573919296, -0.05407611280679703, 0.13436287641525269, 0.12337575852870941, -0.04349830374121666, 0.1346055269241333, -0.08672069013118744, -0.05363473296165466, 0.10277332365512848, 0.019689666107296944, 0.009308061562478542, 0.07256065309047699, 0.02191895991563797, 0.012116474099457264, 0.05955912545323372, 0.06519178301095963, 0.09907940775156021, -0.10559206455945969, 0.02268049865961075, -0.03747428208589554, -0.1432260125875473, -0.057319775223731995, -0.03375370427966118, 0.022990288212895393, 0.02676946483552456, 0.07909193634986877, -0.05677521973848343, 0.1060420423746109, -0.04767235741019249, 0.04007749632000923, 0.0009903513127937913, -0.16581644117832184, -0.037122491747140884, -0.06677180528640747, 0.030188927426934242, 0.01579337939620018, 0.2170153260231018, -0.004767094273120165, -0.031621865928173065, 0.02309555560350418, 0.04177442193031311, -0.008157406933605671, 0.008448606356978416, 0.20548230409622192, 0.10310740768909454, -0.06846706569194794, -0.14684386551380157, 0.08422907441854477, 0.0026786397211253643, -0.06211854889988899, 0.16172944009304047, -0.01823771744966507, 0.028763005509972572, 0.08257352560758591, -0.02413180097937584, 0.015303188003599644, -0.14920642971992493, -0.10885327309370041, -0.049236614257097244, 0.07946305721998215, -0.00982586108148098, 0.12070965021848679, 0.11676641553640366, -0.04379433020949364, 0.0015108027728274465, -0.012010033242404461, -0.05686352401971817, -0.17215904593467712, -0.21417103707790375, -0.08673834055662155, -0.13437552750110626, 0.0188327394425869, -0.10698898136615753, 0.05854099616408348, -0.0021438240073621273, 0.056589897722005844, -0.049702372401952744, 0.1177777424454689, 0.04731317237019539, -0.12980766594409943, 0.08827393501996994, -0.017468633130192757, 0.02003621682524681, 0.0049504381604492664, 0.004185068421065807, -0.08732683956623077, 0.025415578857064247, 0.0007953525637276471, 0.0564325787127018, -0.06978806853294373, -0.0034391884692013264, -0.12339784950017929, -0.10002655535936356, -0.05990864336490631, 0.05073666572570801, -0.02586827427148819, 0.08469163626432419, 0.01148642785847187, -0.022830944508314133, 0.04015327990055084, 0.1898038536310196, -0.06600670516490936, -0.0655871033668518, -0.021236835047602654, 0.19605761766433716, 0.01429776381701231, 0.08746463060379028, -0.011506977491080761, 0.01829349622130394, -0.04844636097550392, 0.3601461946964264, 0.2312781661748886, -0.04722781479358673, 0.03223119303584099, 0.014052394777536392, 0.03352605551481247, 0.1355055868625641, 0.11694815009832382, 0.08269927650690079, 0.23108479380607605, -0.10545637458562851, -0.08099020272493362, -0.0423525795340538, -0.01121626514941454, -0.08901078999042511, 0.06734571605920792, 0.03955891728401184, -0.08915092796087265, -0.05525268241763115, 0.08274523913860321, -0.23970729112625122, 0.11162406951189041, -0.08171556890010834, -0.162613183259964, -0.059820499271154404, 0.026959683746099472, 0.09690212458372116, 0.015219182707369328, 0.09199827164411545, -0.014955630525946617, -0.07619146257638931, 0.05853145197033882, 0.05077424272894859, -0.2382902354001999, 0.03595710173249245, 0.05467145889997482, -0.05854099243879318, 0.007449338678270578, -0.018502935767173767, 0.06915213912725449, 0.06619628518819809, 0.04410887509584427, -0.05260109901428223, 0.015458744950592518, -0.0034508355893194675, -0.04801314324140549, -0.02780556119978428, 0.10057107359170914, 0.034579407423734665, -0.1206040009856224, 0.09023909270763397, -0.10645681619644165, 0.027654634788632393, 0.06223444640636444, -0.0004995701601728797, -0.020608041435480118, 0.021370284259319305, -0.0963246300816536, 0.05579786375164986, 0.07245944440364838, -0.015751751139760017, -0.01680879481136799, -0.06924273818731308, -0.020133692771196365, -0.03015086241066456, -0.04436968266963959, -0.06642002612352371, -0.11278419196605682, -0.10139625519514084, 0.056709371507167816, 0.02185703255236149, -0.15141497552394867, 0.015565968118607998, -0.11661290377378464, 0.0685729831457138, -0.17312651872634888, 0.12360571324825287, 0.024642599746584892, -0.00669059669598937, 0.0005565152387134731, -0.09654258191585541, 0.029081624001264572, 0.0706893652677536, -0.11944233626127243, -0.09647279977798462 ]
null
null
transformers
#Rick Dialogpt model
{"tags": ["conversational"]}
text-generation
crystalgate/DialoGPT-small-rick
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
#Rick Dialogpt model
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 51 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.009697278961539268, 0.03208012506365776, -0.007204889785498381, 0.004809224978089333, 0.16726240515708923, 0.014898733235895634, 0.09765533357858658, 0.13672804832458496, -0.007841327227652073, -0.031050153076648712, 0.14490588009357452, 0.20411323010921478, -0.006439372431486845, 0.0661218985915184, -0.07572533935308456, -0.2683109939098358, 0.05759621039032936, 0.046649303287267685, 0.016515716910362244, 0.1200079694390297, 0.08573378622531891, -0.05473608896136284, 0.08714032918214798, -0.014583407901227474, -0.150366872549057, 0.017733458429574966, 0.043394338339567184, -0.12260226160287857, 0.11910516023635864, 0.05462685227394104, 0.07063519209623337, 0.014929565601050854, -0.07541623711585999, -0.1631229966878891, 0.03031250834465027, 0.01425902172923088, -0.0594632662832737, 0.04757995903491974, 0.059961482882499695, -0.10165371745824814, 0.10819483548402786, 0.09530027210712433, -0.013078106567263603, 0.06798283755779266, -0.16849711537361145, -0.020869607105851173, -0.01446688175201416, 0.009899779222905636, 0.05550243332982063, 0.09964893013238907, -0.03413357585668564, 0.10497362166643143, -0.09214533120393753, 0.11017382889986038, 0.10932035744190216, -0.32057443261146545, -0.005767723545432091, 0.09167823940515518, 0.039358653128147125, 0.07352814823389053, -0.04467793554067612, 0.06258884817361832, 0.018015462905168533, 0.017986174672842026, -0.014015024527907372, -0.07283061742782593, -0.11612214148044586, 0.04717336222529411, -0.08668071031570435, -0.059868961572647095, 0.2244078367948532, -0.05464440956711769, 0.06881742179393768, -0.05281897634267807, -0.10522868484258652, -0.04308144748210907, -0.029833965003490448, 0.00475557055324316, -0.07660607248544693, 0.08692064881324768, 0.00869679357856512, -0.09547875821590424, -0.1376667022705078, -0.02496783249080181, -0.1776352822780609, 0.16140350699424744, 0.02465328387916088, 0.05232657864689827, -0.2027255892753601, 0.09623090922832489, 0.017906051129102707, -0.08045592904090881, 0.022091427817940712, -0.10046248883008957, 0.029131146147847176, 0.013760408386588097, -0.04754498973488808, -0.061387211084365845, 0.0843690037727356, 0.11199145019054413, -0.01731434464454651, 0.025486016646027565, -0.039331406354904175, 0.08100687712430954, 0.03553595021367073, 0.09077847748994827, 0.007288969587534666, -0.028338588774204254, 0.025842782109975815, -0.13719046115875244, -0.003647835226729512, -0.07116208970546722, -0.16572439670562744, -0.021088803187012672, 0.02994808368384838, 0.08289173990488052, 0.015449047088623047, 0.11682453751564026, -0.03272046521306038, -0.025152435526251793, 0.03602350503206253, -0.047656361013650894, -0.012649794109165668, 0.016648368909955025, 0.013163427822291851, 0.12399329990148544, -0.0022096503525972366, 0.03235051408410072, -0.13653022050857544, 0.031423524022102356, -0.06793295592069626, -0.003740974934771657, -0.03486552834510803, -0.040637075901031494, 0.009043924510478973, -0.06862333416938782, 0.003486064961180091, -0.15030112862586975, -0.15063877403736115, 0.007587034720927477, -0.007836631499230862, -0.04107699543237686, -0.06370922178030014, -0.06952770054340363, -0.013550350442528725, 0.04251532256603241, -0.07093454152345657, -0.011352915316820145, -0.06403283774852753, 0.11004766076803207, -0.03197755664587021, 0.07921615242958069, -0.11953279376029968, 0.08390819281339645, -0.11260783672332764, -0.02386913076043129, -0.060801517218351364, 0.09317506104707718, -0.0006014376995153725, 0.09549830108880997, -0.006563255097717047, -0.017931854352355003, -0.07981178909540176, 0.06445012241601944, -0.042872510850429535, 0.21701598167419434, -0.0615808479487896, -0.11181682348251343, 0.28781595826148987, -0.052628401666879654, -0.1370542049407959, 0.11647392809391022, 0.008682746440172195, 0.05777018144726753, 0.10703510791063309, 0.19733482599258423, -0.015276194550096989, 0.004040541127324104, 0.09471915662288666, 0.11263324320316315, -0.11276852339506149, -0.033160366117954254, 0.013019153848290443, -0.04081077128648758, -0.10867965966463089, 0.04689536616206169, 0.09810488671064377, 0.07090286910533905, -0.04786505550146103, -0.03377414867281914, -0.01366397924721241, 0.0052589005790650845, 0.08885077387094498, -0.007157256826758385, 0.10962837189435959, -0.05819983780384064, -0.03796621412038803, -0.029282379895448685, -0.012126247398555279, -0.03951939567923546, 0.03137664496898651, -0.043376367539167404, 0.10821941494941711, -0.011204327456653118, 0.06364280730485916, -0.16185984015464783, -0.07691477984189987, -0.017002692446112633, 0.1581239402294159, 0.024538565427064896, 0.09859629720449448, 0.0552486926317215, -0.040398042649030685, -0.0012767292791977525, 0.012792680412530899, 0.15581141412258148, -0.022091681137681007, -0.065607450902462, -0.052166227251291275, 0.08642971515655518, -0.05641226842999458, 0.04504093527793884, -0.05937713757157326, 0.012367865070700645, 0.05064384639263153, 0.10342344641685486, -0.00018274025933351368, 0.03323284164071083, -0.008164864964783192, 0.002145637758076191, -0.058205123990774155, 0.007405933458358049, 0.10799351334571838, 0.00036868182360194623, -0.07365862280130386, 0.22074243426322937, -0.17796069383621216, 0.1765957772731781, 0.1893044263124466, -0.299345999956131, 0.017949223518371582, -0.10759581625461578, -0.04561871662735939, 0.014407722279429436, 0.05567655712366104, -0.0454222597181797, 0.1703362911939621, -0.009871348738670349, 0.18874616920948029, -0.04946064203977585, -0.04464937001466751, -0.0200483538210392, -0.05118836089968681, -0.0024189651012420654, 0.07781197130680084, 0.10685696452856064, -0.13992026448249817, 0.1964332014322281, 0.1621224284172058, 0.048237916082143784, 0.19945049285888672, 0.015346456319093704, -0.011589210480451584, 0.0909530371427536, 0.005220826715230942, -0.058739423751831055, -0.07409929484128952, -0.2594851851463318, -0.030033592134714127, 0.07992640137672424, 0.0422382652759552, 0.1212305948138237, -0.11349532753229141, -0.038956157863140106, -0.01763172075152397, -0.023146281018853188, 0.021672505885362625, 0.0914369598031044, 0.06075398623943329, 0.13201528787612915, -0.001710098935291171, -0.007300339173525572, 0.10524573177099228, 0.01783694699406624, -0.09354141354560852, 0.18308524787425995, -0.13652534782886505, -0.37097251415252686, -0.13911493122577667, -0.18057456612586975, -0.05449081212282181, 0.05712554603815079, 0.11679314076900482, -0.12011238187551498, -0.018752124160528183, 0.01578843593597412, 0.10931742936372757, -0.08449502289295197, 0.0021454424131661654, -0.06880278885364532, 0.0321490578353405, -0.10310184955596924, -0.09194442629814148, -0.055416494607925415, -0.031392451375722885, -0.08001253753900528, 0.1423761546611786, -0.10777941346168518, 0.04476889222860336, 0.20262959599494934, 0.04653622955083847, 0.05625178664922714, -0.044105201959609985, 0.19377262890338898, -0.11264272034168243, -0.01661740615963936, 0.19215328991413116, -0.048360925167798996, 0.07476246356964111, 0.1232115849852562, -0.006348740309476852, -0.08765771239995956, 0.03011748194694519, -0.02085109055042267, -0.07988511025905609, -0.23219464719295502, -0.13938382267951965, -0.12429051846265793, 0.09477275609970093, 0.028005298227071762, 0.056365787982940674, 0.17219258844852448, 0.06577219814062119, -0.038416244089603424, 0.006410336587578058, 0.02959546446800232, 0.08237514644861221, 0.23417828977108002, -0.06035616248846054, 0.1364797055721283, -0.03420931473374367, -0.14982740581035614, 0.08169995993375778, 0.0713929831981659, 0.10213395953178406, 0.06678459793329239, 0.0804823637008667, 0.0149586396291852, 0.06188136339187622, 0.1311223804950714, 0.08191446959972382, 0.019586285576224327, -0.02480296604335308, -0.03388110175728798, -0.025523077696561813, -0.05937909707427025, 0.040128443390131, 0.06589099019765854, -0.16763372719287872, -0.039227183908224106, -0.09338314831256866, 0.09657008945941925, 0.0873042419552803, 0.06609832495450974, -0.1842060089111328, -0.008006223477423191, 0.08488986641168594, -0.03854905813932419, -0.13727426528930664, 0.09535189718008041, 0.01523482333868742, -0.15144726634025574, 0.03139317408204079, -0.04061909019947052, 0.12188644707202911, -0.07804752141237259, 0.09809603542089462, -0.08108244836330414, -0.07448557764291763, 0.02123199962079525, 0.1261177361011505, -0.30527687072753906, 0.20240111649036407, -0.0024993624538183212, -0.06486981362104416, -0.1243603527545929, -0.0032166161108762026, 0.002410882618278265, 0.07357452809810638, 0.10519039630889893, -0.007196315098553896, 0.001897757756523788, -0.06300821900367737, -0.01829923689365387, 0.032471053302288055, 0.13080233335494995, -0.0401318334043026, -0.021158374845981598, -0.050194524228572845, -0.001653497340157628, -0.03173094615340233, -0.06934895366430283, 0.02002747356891632, -0.19509181380271912, 0.08751901984214783, 0.04166261479258537, 0.09648149460554123, 0.029994789510965347, 0.004265148192644119, -0.09651939570903778, 0.24698667228221893, -0.07148019969463348, -0.10072879493236542, -0.10919588059186935, -0.046813901513814926, 0.03569883480668068, -0.05628936365246773, 0.04309194162487984, -0.0788632407784462, 0.028997479006648064, -0.06352769583463669, -0.19235502183437347, 0.12410202622413635, -0.09027006477117538, -0.04412810131907463, -0.02371402643620968, 0.2110891044139862, -0.05598580464720726, 0.010335659608244896, 0.02930437959730625, 0.01208863127976656, -0.11645778268575668, -0.09678568691015244, 0.031018631532788277, -0.007351789623498917, 0.050603240728378296, 0.041841957718133926, -0.05915454775094986, -0.017138581722974777, -0.052199993282556534, -0.022926922887563705, 0.3496883809566498, 0.14231905341148376, -0.043836336582899094, 0.19347235560417175, 0.12347975373268127, -0.07452994585037231, -0.3159443140029907, -0.1066238060593605, -0.10937739163637161, -0.04680149629712105, -0.07012093812227249, -0.2002030611038208, 0.06474938243627548, 0.00662544509395957, -0.013415241613984108, 0.12749312818050385, -0.2561831772327423, -0.07571036368608475, 0.15906259417533875, -0.017980827018618584, 0.3745945692062378, -0.1168576180934906, -0.10926306992769241, -0.03950892388820648, -0.14175476133823395, 0.16968177258968353, -0.01989765651524067, 0.11221715062856674, -0.009765521623194218, 0.14388824999332428, 0.05548359826207161, -0.023479344323277473, 0.08544106781482697, 0.004999885335564613, -0.03290518373250961, -0.10304180532693863, -0.05676887184381485, 0.007092386484146118, 0.02477436140179634, 0.018026655539870262, -0.041834570467472076, 0.02227151393890381, -0.11731979995965958, -0.04657655209302902, -0.08982590585947037, 0.04431166127324104, 0.03899754583835602, -0.07325074821710587, -0.002380647463724017, -0.07165111601352692, -0.012272949330508709, 0.022334342822432518, 0.20356793701648712, -0.08029330521821976, 0.16448934376239777, 0.09239562600851059, 0.12419285625219345, -0.14376309514045715, -0.00019283240544609725, -0.0762530043721199, -0.05611240118741989, 0.07737895101308823, -0.09433035552501678, 0.058893077075481415, 0.10901971161365509, -0.04567738622426987, 0.08828683942556381, 0.10377411544322968, 0.008936077356338501, 0.003213887568563223, 0.10916902124881744, -0.2667325437068939, -0.0296600554138422, -0.07532413303852081, 0.000883326749317348, 0.09092561900615692, 0.08562852442264557, 0.18840822577476501, 0.025361526757478714, -0.04293036088347435, -0.002770674182102084, 0.028597986325621605, -0.039021048694849014, 0.051667019724845886, 0.001123449532315135, 0.01947369985282421, -0.1530752182006836, 0.072522833943367, 0.01490565575659275, -0.15215420722961426, 0.021316176280379295, 0.16572684049606323, -0.11656328290700912, -0.1283872276544571, -0.06520111113786697, 0.08313824236392975, -0.11755692958831787, -0.01578943058848381, -0.03279297426342964, -0.13145680725574493, 0.07992171496152878, 0.12629036605358124, 0.05557859688997269, 0.0972496047616005, -0.06061713397502899, -0.020469192415475845, -0.018721895292401314, -0.014099318534135818, -0.012384648434817791, -0.007667020428925753, -0.055978111922740936, 0.0590752474963665, -0.026677248999476433, 0.1425808072090149, -0.09221141785383224, -0.1037059873342514, -0.16142144799232483, 0.0374140702188015, -0.11013076454401016, -0.08825794607400894, -0.08821134269237518, -0.050188567489385605, 0.002360827289521694, -0.019856395199894905, -0.04037635400891304, -0.05829505994915962, -0.12300454825162888, 0.0338277705013752, -0.040771447122097015, 0.024727050215005875, -0.07512269169092178, 0.015856385231018066, 0.08507686108350754, -0.03285100311040878, 0.15655414760112762, 0.1450488418340683, -0.1006515845656395, 0.10741901397705078, -0.14806775748729706, -0.09138492494821548, 0.11116421222686768, 0.015329592861235142, 0.0449691042304039, 0.09723787009716034, 0.013362943194806576, 0.0635865181684494, 0.032776717096567154, 0.05308786407113075, 0.027619892731308937, -0.11959987878799438, 0.06483134627342224, -0.03626115620136261, -0.14700546860694885, -0.049338050186634064, -0.05282869189977646, 0.01647452637553215, 0.013054544106125832, 0.09622690081596375, -0.05301849544048309, 0.10698331147432327, -0.04055701196193695, 0.0346808135509491, 0.017554637044668198, -0.1730053424835205, -0.03816922754049301, -0.08538098633289337, 0.03681723028421402, 0.014741539023816586, 0.25266793370246887, 0.030072299763560295, 0.012416383251547813, 0.032671261578798294, 0.08285367488861084, 0.03899408504366875, 0.010228337720036507, 0.17482228577136993, 0.1162426546216011, -0.06621865928173065, -0.10445023328065872, 0.0729617029428482, 0.016332454979419708, 0.01286179106682539, 0.13617953658103943, 0.008365051820874214, 0.005795429926365614, 0.08649782836437225, -0.016865963116288185, 0.009968153201043606, -0.10052056610584259, -0.13426925241947174, -0.022176474332809448, 0.05151832848787308, -0.04655967652797699, 0.11727844923734665, 0.1406494379043579, -0.01806013658642769, 0.03222079202532768, -0.021771740168333054, -0.05699979141354561, -0.1683429479598999, -0.1429590880870819, -0.06883849948644638, -0.13416796922683716, 0.00897989235818386, -0.11180389672517776, 0.05395037308335304, 0.06001098081469536, 0.06750501692295074, -0.06899319589138031, 0.10220931470394135, 0.04626858979463577, -0.11440542340278625, 0.06264589726924896, -0.0296088308095932, 0.09430401772260666, -0.02759445086121559, -0.019505485892295837, -0.09039592742919922, 0.014574515633285046, 0.011419114656746387, 0.06245238706469536, -0.04707273095846176, 0.007463190704584122, -0.14696238934993744, -0.08972041308879852, -0.0523175448179245, 0.0718572810292244, -0.050409089773893356, 0.14282815158367157, 0.00775480642914772, -0.0170906875282526, 0.039554283022880554, 0.22787313163280487, -0.07476283609867096, -0.04778539761900902, -0.05269690603017807, 0.20717895030975342, 0.02975541539490223, 0.1171872541308403, -0.022938819602131844, -0.006106364540755749, -0.0919521227478981, 0.3764844834804535, 0.30030161142349243, -0.09031439572572708, 0.011794124729931355, 0.02137952297925949, 0.04502861574292183, 0.1316293478012085, 0.1216534823179245, 0.10318691283464432, 0.3006802201271057, -0.07452366501092911, -0.04653361067175865, -0.012629742734134197, -0.023858042433857918, -0.09059546142816544, 0.1021224707365036, 0.04839762672781944, -0.06382183730602264, -0.03313443064689636, 0.0954432487487793, -0.25862133502960205, 0.1277991235256195, -0.12311873584985733, -0.17578600347042084, -0.06654827296733856, 0.009760108776390553, 0.10465722531080246, 0.015642458572983742, 0.0946015790104866, 0.007128213066607714, -0.11252258718013763, 0.06305865943431854, 0.03397420793771744, -0.22762253880500793, 0.0006893770187161863, 0.06642123311758041, -0.07006710022687912, -0.0024247700348496437, -0.026499588042497635, 0.05657242611050606, 0.0656052976846695, 0.054629553109407425, -0.00971333310008049, 0.03816632181406021, 0.0034184439573436975, -0.0585215799510479, 0.016623929142951965, 0.05121519789099693, 0.02472509816288948, -0.09763528406620026, 0.06927435845136642, -0.1574270874261856, 0.04766253009438515, -0.0030655991286039352, -0.04124255105853081, 0.006064958870410919, 0.008823691867291927, -0.06491616368293762, 0.05165379121899605, 0.07916834205389023, -0.0016257909592241049, -0.0062433634884655476, -0.057178743183612823, -0.02632102556526661, -0.027755750343203545, -0.09291748702526093, -0.10495562851428986, -0.14682936668395996, -0.11640441417694092, 0.09368976950645447, -0.01011267676949501, -0.1848134547472, 0.022154374048113823, -0.08606051653623581, 0.08319322764873505, -0.1670055389404297, 0.08040720224380493, 0.07041648775339127, 0.013038921169936657, -0.0031511052511632442, -0.02002427540719509, 0.054132770746946335, 0.086809903383255, -0.10407156497240067, -0.07400695979595184 ]
null
null
spacy
NER Model for 'Ministerratsprotokolle' | Feature | Description | | --- | --- | | **Name** | `de_MRP_NER` | | **Version** | `0.0.0` | | **spaCy** | `>=3.1.0,<3.2.0` | | **Default Pipeline** | `tok2vec`, `ner` | | **Components** | `tok2vec`, `ner` | | **Vectors** | 0 keys, 0 unique vectors (0 dimensions) | | **Sources** | n/a | | **License** | `cc-by` | | **Author** | [Peter Andorfer]() | ### Label Scheme <details> <summary>View label scheme (4 labels for 1 components)</summary> | Component | Labels | | --- | --- | | **`ner`** | `GPE`, `LOC`, `ORG`, `PER` | </details> ### Accuracy | Type | Score | | --- | --- | | `ENTS_F` | 88.04 | | `ENTS_P` | 90.53 | | `ENTS_R` | 85.69 | | `TOK2VEC_LOSS` | 40077.56 | | `NER_LOSS` | 77727.57 |
{"language": ["de"], "license": "cc-by-4.0", "tags": ["spacy", "token-classification"]}
token-classification
csae8092/de_MRP_NER
[ "spacy", "token-classification", "de", "license:cc-by-4.0", "model-index", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "de" ]
TAGS #spacy #token-classification #de #license-cc-by-4.0 #model-index #region-us
NER Model for 'Ministerratsprotokolle' ### Label Scheme View label scheme (4 labels for 1 components) ### Accuracy
[ "### Label Scheme\n\n\n\nView label scheme (4 labels for 1 components)", "### Accuracy" ]
[ "TAGS\n#spacy #token-classification #de #license-cc-by-4.0 #model-index #region-us \n", "### Label Scheme\n\n\n\nView label scheme (4 labels for 1 components)", "### Accuracy" ]
[ 30, 16, 5 ]
[ "passage: TAGS\n#spacy #token-classification #de #license-cc-by-4.0 #model-index #region-us \n### Label Scheme\n\n\n\nView label scheme (4 labels for 1 components)### Accuracy" ]
[ -0.06385856866836548, 0.13472512364387512, -0.003094917628914118, 0.0397038497030735, 0.07022737711668015, 0.06061854213476181, 0.2767327129840851, 0.05958044156432152, 0.2079005092382431, 0.02886129729449749, 0.061364416033029556, 0.09853294491767883, 0.04757606238126755, 0.21973487734794617, -0.09127046167850494, -0.21529531478881836, 0.0958523154258728, -0.005406170152127743, 0.06667861342430115, 0.11700994521379471, 0.08378114551305771, -0.12813547253608704, 0.06597858667373657, -0.04171180725097656, -0.22465991973876953, 0.014053160324692726, 0.03370482847094536, -0.1148187518119812, 0.0720188170671463, -0.04196806624531746, 0.17148704826831818, 0.08092565089464188, 0.07878386229276657, -0.20359699428081512, -0.0030305241234600544, -0.039836589246988297, -0.09770330041646957, 0.0686204805970192, 0.05432229861617088, 0.016963504254817963, -0.03868686407804489, -0.030729399994015694, 0.05363672226667404, 0.06719394773244858, -0.12571488320827484, -0.17105752229690552, -0.06533747911453247, 0.1414593756198883, 0.10574507713317871, -0.04827113449573517, -0.0008316130843013525, 0.04904069006443024, -0.11670505255460739, 0.03754619136452675, 0.09219346940517426, -0.3793019652366638, 0.02676801197230816, 0.25279706716537476, -0.03578133136034012, 0.10048729926347733, -0.04890355467796326, 0.13123422861099243, 0.13417236506938934, -0.011546103283762932, -0.006435838062316179, -0.018642647191882133, 0.11277974396944046, 0.03329147771000862, -0.09689734876155853, -0.06436038017272949, 0.5035092234611511, 0.08564742654561996, -0.05668892711400986, -0.10495588183403015, -0.027417441830039024, -0.12387033551931381, -0.08367842435836792, -0.03049219585955143, 0.08829320967197418, 0.04531635716557503, 0.09365047514438629, 0.11988046765327454, -0.10377410054206848, -0.06064224988222122, -0.15473124384880066, 0.163921520113945, 0.004786196164786816, 0.09316974133253098, -0.13639798760414124, 0.024215595796704292, -0.10437284409999847, -0.06111231446266174, -0.00911126658320427, -0.07900924980640411, -0.045119304209947586, -0.03997853025794029, 0.028874948620796204, 0.17723388969898224, 0.08720173686742783, 0.10498438775539398, -0.030816076323390007, 0.010638272389769554, -0.031197812408208847, 0.07542527467012405, 0.10660627484321594, 0.14602580666542053, -0.01859539933502674, -0.005969756282866001, -0.018562190234661102, -0.0687531903386116, 0.049070023000240326, -0.03357533738017082, -0.17074260115623474, 0.00444003427401185, 0.055237870663404465, 0.11786045134067535, -0.08242696523666382, -0.03722166642546654, -0.12658122181892395, -0.03671668469905853, 0.15735626220703125, -0.0961843952536583, 0.021880265325307846, -0.002741612959653139, -0.007103442680090666, 0.07422768324613571, -0.12446694821119308, -0.010836279951035976, 0.06269129365682602, 0.01369499135762453, -0.11603740602731705, -0.002407974563539028, -0.013717913068830967, -0.08804995566606522, 0.045516520738601685, -0.10310220718383789, 0.03914336487650871, -0.053654979914426804, -0.1452307552099228, -0.0009558024466969073, -0.030147681012749672, -0.07168180495500565, -0.009538908489048481, -0.013986019417643547, -0.07069491595029831, -0.01845092885196209, 0.002758969785645604, -0.07804422825574875, -0.10382562130689621, 0.0034123361110687256, -0.03263712301850319, 0.06887941062450409, -0.12672613561153412, 0.02941019833087921, -0.06512124091386795, 0.05166752263903618, -0.1610855907201767, 0.02406146377325058, -0.09226509183645248, 0.06466101109981537, -0.07589632272720337, -0.09269633889198303, 0.061340827494859695, -0.011624468490481377, -0.13421355187892914, 0.14594314992427826, -0.22205208241939545, -0.0526152066886425, 0.16571009159088135, -0.19176805019378662, -0.12826240062713623, 0.03924340754747391, -0.006549354176968336, 0.05590920150279999, 0.08420353382825851, 0.12143457680940628, -0.02370697259902954, -0.1314009130001068, -0.025085048750042915, 0.11224989593029022, -0.04297644644975662, -0.0785822868347168, 0.10934149473905563, -0.023989448323845863, -0.0004887625109404325, 0.04426784813404083, -0.014842969365417957, -0.1327861100435257, -0.05040987953543663, -0.06960400938987732, -0.019735239446163177, 0.032347410917282104, 0.03897624462842941, 0.02424178458750248, 0.019212277606129646, -0.04772020876407623, 0.012773526832461357, 0.03171473369002342, 0.05064195767045021, 0.03200134262442589, -0.058171194046735764, -0.06322992593050003, 0.12199145555496216, -0.12139715999364853, -0.0672859326004982, -0.1207939088344574, -0.1298864334821701, 0.05583832412958145, 0.023324623703956604, 0.05183057114481926, 0.15762446820735931, -0.01785740628838539, 0.007647615857422352, 0.0006017946288920939, -0.00009712190512800589, -0.0336846299469471, 0.08433259278535843, -0.02969747968018055, -0.20240919291973114, -0.048181142657995224, -0.058493755757808685, 0.07329977303743362, -0.045227501541376114, 0.01508280634880066, 0.21874359250068665, 0.04697594419121742, 0.019380390644073486, 0.06796710193157196, 0.003552958369255066, 0.04556804150342941, -0.04112231731414795, -0.03648826852440834, 0.08591944724321365, -0.06811971217393875, -0.07636383175849915, -0.009264483116567135, -0.08540673553943634, 0.10322486609220505, 0.16760317981243134, -0.10279545933008194, -0.037930283695459366, -0.0717015340924263, 0.004292863886803389, 0.007437978405505419, -0.0956098735332489, -0.003497655736282468, -0.052054256200790405, -0.04657944664359093, 0.041597481817007065, -0.1100899875164032, -0.01715380698442459, 0.041845791041851044, -0.02108917385339737, -0.14812292158603668, 0.06613586097955704, 0.0584745816886425, -0.185817688703537, 0.1686919927597046, 0.30281689763069153, 0.16255336999893188, 0.056260641664266586, -0.06412261724472046, -0.0391269214451313, -0.06661314517259598, -0.009954454377293587, -0.07246392965316772, 0.21468143165111542, -0.13147565722465515, -0.036326006054878235, 0.06489772349596024, 0.045830368995666504, 0.0034093984868377447, -0.218227356672287, -0.009527834132313728, -0.03230443596839905, -0.07123538851737976, -0.14049088954925537, -0.031165169551968575, -0.0004003677167929709, 0.11683294177055359, 0.040234919637441635, -0.2255738228559494, 0.10721158981323242, -0.0462210513651371, -0.08177534490823746, 0.1540534794330597, -0.10037878900766373, -0.19356489181518555, -0.16068489849567413, -0.0701606422662735, -0.07641502469778061, 0.05050945654511452, -0.009261485189199448, -0.06828142702579498, -0.05475330352783203, -0.001720659201964736, -0.08436471223831177, -0.16805703938007355, -0.047997571527957916, -0.05655145272612572, 0.09935460984706879, -0.12468291819095612, -0.07338842749595642, -0.09220466017723083, -0.039462942630052567, 0.04420270398259163, 0.09570270031690598, -0.14521010220050812, 0.0699942484498024, 0.2794201076030731, -0.03050101548433304, 0.0736946314573288, -0.040954940021038055, 0.11923367530107498, -0.06064528971910477, 0.02051813155412674, 0.11583442986011505, 0.0643831118941307, 0.030563564971089363, 0.277820348739624, 0.10009436309337616, -0.15038803219795227, -0.0595962256193161, -0.07407922297716141, -0.09704037010669708, -0.16420117020606995, -0.10058043897151947, -0.046560510993003845, 0.004016268998384476, 0.036852769553661346, 0.044215861707925797, 0.05284356325864792, 0.04906097427010536, 0.015118678100407124, 0.0204082690179348, 0.06604676693677902, 0.06324142217636108, 0.15622709691524506, -0.04276072233915329, 0.09965383261442184, -0.06289228051900864, -0.025226188823580742, 0.0894511267542839, 0.12093429267406464, 0.1966082602739334, 0.20855025947093964, 0.09481177479028702, 0.11651195585727692, 0.04046018421649933, 0.1559753566980362, 0.04380505904555321, 0.15898334980010986, -0.013063622638583183, -0.02398163080215454, -0.07263662666082382, 0.0375189334154129, 0.07477036118507385, -0.0328284353017807, -0.0816185399889946, -0.06038087233901024, -0.10970363020896912, 0.07390500605106354, 0.017855538055300713, 0.2346818596124649, -0.2395005077123642, 0.022497406229376793, 0.10488820821046829, 0.0983944833278656, -0.054567813873291016, 0.10608436167240143, -0.02306150086224079, -0.09193559736013412, 0.06983160227537155, 0.00089328020112589, 0.10323382169008255, -0.09055528044700623, -0.008957095444202423, -0.005206885281950235, -0.06979008764028549, 0.00707263732329011, 0.07507967203855515, -0.12020961195230484, 0.29240480065345764, 0.0426936149597168, -0.028578734025359154, -0.06788825243711472, -0.008076646365225315, -0.0007517561316490173, 0.25961899757385254, 0.22250711917877197, 0.04293669015169144, -0.2164505124092102, -0.2013159841299057, -0.032605595886707306, -0.026546606793999672, 0.1415209025144577, -0.04416179656982422, 0.01667141355574131, 0.021114172413945198, 0.014406885951757431, -0.017347004264593124, 0.022742174565792084, -0.07771234214305878, -0.03618045151233673, 0.031241199001669884, 0.09718374162912369, -0.050463031977415085, -0.03520617634057999, -0.05626020208001137, -0.18548449873924255, 0.11872873455286026, -0.09660688787698746, -0.09831982105970383, -0.08909325301647186, -0.035723503679037094, 0.07262872904539108, 0.00034884762135334313, -0.003604019060730934, -0.051312677562236786, 0.11295121163129807, 0.0037790383212268353, -0.09101162105798721, 0.10814189165830612, -0.044469308108091354, -0.05510375276207924, -0.05687025561928749, 0.15071026980876923, -0.05432809889316559, -0.0062538031488657, 0.08159001916646957, 0.08229915052652359, -0.007715534884482622, -0.1229311153292656, 0.10220275074243546, -0.007388601545244455, 0.061741188168525696, 0.2304355949163437, -0.1039058044552803, -0.13919441401958466, -0.029209161177277565, 0.043926987797021866, 0.07515872269868851, 0.2951124608516693, -0.10481518507003784, 0.10352829098701477, 0.09339039772748947, -0.04291057214140892, -0.19017571210861206, -0.034780971705913544, -0.17624826729297638, -0.001757705700583756, -0.015936270356178284, -0.06501487642526627, 0.1285591572523117, 0.0648246556520462, -0.07398983091115952, 0.11633309721946716, -0.2414579540491104, -0.08161857724189758, 0.1729169636964798, 0.04181941971182823, 0.16291682422161102, -0.0755450427532196, -0.11555517464876175, -0.07985201478004456, -0.19725216925144196, 0.14310023188591003, 0.0005399496294558048, 0.08442310988903046, -0.04212398827075958, -0.005843115039169788, -0.0065605719573795795, 0.001037825015373528, 0.24139390885829926, 0.12887810170650482, 0.11884041875600815, 0.03965285047888756, -0.17397071421146393, 0.24620196223258972, 0.011072013527154922, 0.021278368309140205, 0.11777789145708084, 0.0003030449151992798, -0.08528976142406464, -0.013182105496525764, 0.01791825145483017, 0.015533043071627617, -0.08250680565834045, -0.0635489895939827, -0.09078391641378403, 0.011602002196013927, -0.08558391034603119, -0.07019177079200745, 0.2477424442768097, -0.055860526859760284, 0.05654608830809593, 0.17641717195510864, 0.0071913874708116055, -0.10680430382490158, -0.056847911328077316, -0.07235399633646011, -0.08549944311380386, 0.05462358519434929, -0.1716119945049286, 0.04486902058124542, 0.11605849117040634, 0.03764433413743973, 0.1458284556865692, 0.11573556065559387, -0.025558974593877792, -0.03858017176389694, 0.11410942673683167, -0.0871037021279335, -0.16330817341804504, 0.01956498622894287, -0.19243831932544708, 0.03317764028906822, 0.07748282700777054, 0.07294271886348724, 0.007732516620308161, -0.025997038930654526, 0.022729454562067986, 0.038057610392570496, -0.06932798027992249, 0.08340159058570862, 0.010080228559672832, 0.0530225895345211, -0.15094584226608276, 0.16154132783412933, 0.08157075196504593, 0.015968071296811104, -0.08788306266069412, -0.08414801955223083, -0.1479039192199707, -0.05553516745567322, -0.043264977633953094, 0.11175774782896042, -0.14274948835372925, -0.13104818761348724, -0.10925167798995972, -0.19520559906959534, 0.016354035586118698, 0.0734907016158104, 0.14293748140335083, 0.12293282151222229, 0.02050197124481201, -0.10516117513179779, 0.03622541204094887, 0.05650341510772705, -0.09742678701877594, 0.024806447327136993, -0.21353350579738617, 0.04280906170606613, -0.023841651156544685, 0.08768045157194138, -0.09157887101173401, -0.0540555939078331, -0.1144741028547287, 0.008720414713025093, -0.07954091578722, 0.07619575411081314, -0.05949022248387337, 0.00260352180339396, 0.01619233749806881, 0.005654802080243826, -0.050586551427841187, 0.0011580080026760697, -0.09682515263557434, 0.051886383444070816, 0.02870391309261322, 0.17768704891204834, -0.09211369603872299, -0.02339390479028225, 0.06155183166265488, -0.036514125764369965, 0.07221446186304092, 0.023940272629261017, 0.025443673133850098, 0.04981645569205284, -0.14346326887607574, 0.01962137594819069, 0.09380312263965607, 0.02325369603931904, 0.07448268681764603, -0.12157528847455978, -0.007448112592101097, 0.05225950852036476, -0.08137289434671402, 0.08750279992818832, -0.08615986257791519, -0.11477933824062347, -0.14613686501979828, -0.1110873743891716, -0.11810288578271866, -0.03426399827003479, 0.030515458434820175, 0.22328586876392365, 0.035684701055288315, 0.05442927032709122, 0.03418609872460365, -0.0015771803446114063, -0.04475044459104538, -0.019530853256583214, -0.07041628658771515, -0.08664001524448395, -0.07220181077718735, -0.01607312634587288, -0.013425714336335659, -0.031976502388715744, 0.36454659700393677, 0.05732744559645653, -0.011970520950853825, 0.05629134178161621, 0.19290320575237274, -0.017126457765698433, 0.0449921153485775, 0.2833068370819092, 0.061492159962654114, -0.031086483970284462, 0.018635911867022514, 0.037992510944604874, -0.005333461798727512, 0.015527223236858845, 0.1400713324546814, 0.10933893918991089, -0.08730760216712952, 0.05873912200331688, 0.0780143216252327, -0.03330259770154953, -0.03762426599860191, 0.10519281774759293, 0.07444004714488983, 0.042111918330192566, 0.03479110077023506, -0.07125592976808548, 0.12115585058927536, -0.1791415512561798, 0.09288692474365234, -0.006179723888635635, -0.08739826083183289, -0.17860956490039825, -0.10152742266654968, -0.08148526400327682, -0.07028895616531372, 0.02150050736963749, -0.10600727796554565, -0.03871189057826996, 0.17212988436222076, 0.020535040646791458, -0.023896463215351105, -0.023610737174749374, -0.26378002762794495, 0.006225386634469032, 0.10995002835988998, 0.013883452862501144, -0.006323289126157761, -0.07754497975111008, -0.011102365329861641, 0.013827439397573471, -0.09627052396535873, -0.05095994099974632, -0.029241446405649185, 0.06731701642274857, -0.043965455144643784, -0.15324409306049347, -0.05358757823705673, -0.0366339348256588, -0.006860018707811832, -0.013555695302784443, -0.02414824068546295, 0.02549726888537407, -0.017940975725650787, 0.04175939783453941, 0.198669895529747, -0.04961869493126869, 0.0036279840860515833, -0.0596187561750412, 0.23645314574241638, -0.04969308525323868, 0.08447844535112381, 0.071415014564991, -0.07115527242422104, 0.002547794720157981, 0.08144151419401169, 0.1945222020149231, -0.011031047441065311, -0.012736010365188122, 0.017325596883893013, -0.0033969159703701735, 0.045961786061525345, 0.01874379627406597, -0.07100801914930344, 0.12705321609973907, -0.060855668038129807, 0.07399261742830276, -0.10312953591346741, -0.00028696656227111816, -0.05199654400348663, -0.03880583867430687, 0.1205589696764946, -0.08417676389217377, -0.12917931377887726, 0.18815116584300995, -0.04628484323620796, 0.02685202844440937, 0.2880057990550995, -0.0838184803724289, -0.08413383364677429, 0.0018586480291560292, -0.0019251869525760412, -0.010824977420270443, 0.04467420279979706, -0.14644813537597656, 0.013314886018633842, 0.06302494555711746, 0.017491688951849937, -0.19339530169963837, -0.05360937491059303, 0.03869786486029625, -0.01130244042724371, 0.09033926576375961, -0.00819533970206976, 0.2109563946723938, 0.09916825592517853, -0.00021478664712049067, -0.07667802274227142, 0.0918726995587349, -0.0004323174653109163, 0.036239538341760635, -0.0003783024149015546, 0.023019962012767792, -0.009633944369852543, -0.055316261947155, 0.09364426881074905, -0.0695633813738823, 0.010846465826034546, 0.017939193174242973, -0.1260785311460495, -0.07138343900442123, 0.03084569238126278, -0.1096898540854454, 0.08493559807538986, 0.052276045083999634, -0.027786660939455032, -0.029946936294436455, 0.018320240080356598, 0.07753024995326996, 0.10255879908800125, -0.10511687397956848, 0.027729913592338562, 0.01991782896220684, 0.01782114803791046, 0.10816572606563568, -0.015320327132940292, -0.1585107147693634, -0.028531674295663834, -0.05843463912606239, 0.030427468940615654, -0.07957793027162552, 0.10330179333686829, 0.13538327813148499, 0.05102115124464035, -0.05644628405570984, -0.19549737870693207, 0.03676711767911911, 0.0685453936457634, -0.0640287771821022, -0.08285749703645706 ]
null
null
spacy
Regensburger Reichstag von 1576 | Feature | Description | | --- | --- | | **Name** | `de_RTA_NER` | | **Version** | `0.0.0` | | **spaCy** | `>=3.1.0,<3.2.0` | | **Default Pipeline** | `tok2vec`, `ner` | | **Components** | `tok2vec`, `ner` | | **Vectors** | 0 keys, 0 unique vectors (0 dimensions) | | **Sources** | n/a | | **License** | `https://creativecommons.org/licenses/by-nc/4.0/` | | **Author** | [n/a](https://reichstagsakten-1576.uni-graz.at) | ### Label Scheme <details> <summary>View label scheme (4 labels for 1 components)</summary> | Component | Labels | | --- | --- | | **`ner`** | `DATE`, `LOC`, `PER`, `TIME` | </details> ### Accuracy | Type | Score | | --- | --- | | `ENTS_F` | 86.86 | | `ENTS_P` | 86.30 | | `ENTS_R` | 87.43 | | `TOK2VEC_LOSS` | 43588.74 | | `NER_LOSS` | 95573.96 |
{"language": ["de"], "license": "cc-by-nc-4.0", "tags": ["spacy", "token-classification"]}
token-classification
csae8092/de_RTA_NER
[ "spacy", "token-classification", "de", "license:cc-by-nc-4.0", "model-index", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "de" ]
TAGS #spacy #token-classification #de #license-cc-by-nc-4.0 #model-index #region-us
Regensburger Reichstag von 1576 ### Label Scheme View label scheme (4 labels for 1 components) ### Accuracy
[ "### Label Scheme\n\n\n\nView label scheme (4 labels for 1 components)", "### Accuracy" ]
[ "TAGS\n#spacy #token-classification #de #license-cc-by-nc-4.0 #model-index #region-us \n", "### Label Scheme\n\n\n\nView label scheme (4 labels for 1 components)", "### Accuracy" ]
[ 32, 16, 5 ]
[ "passage: TAGS\n#spacy #token-classification #de #license-cc-by-nc-4.0 #model-index #region-us \n### Label Scheme\n\n\n\nView label scheme (4 labels for 1 components)### Accuracy" ]
[ -0.06110462173819542, 0.14928708970546722, -0.0028338648844510317, 0.038839101791381836, 0.06316200643777847, 0.0646965429186821, 0.2631303369998932, 0.06015896797180176, 0.2008504867553711, 0.03368014469742775, 0.06354188919067383, 0.1015159860253334, 0.05020608380436897, 0.21348270773887634, -0.09233225882053375, -0.19019414484500885, 0.09389578551054001, 0.0021890117786824703, 0.06711212545633316, 0.12087611854076385, 0.07616858929395676, -0.12225784361362457, 0.06479385495185852, -0.040295619517564774, -0.2286669909954071, 0.010682852938771248, 0.036758389323949814, -0.11454617977142334, 0.06631971895694733, -0.04869807884097099, 0.17253954708576202, 0.08619572222232819, 0.07923801243305206, -0.20459318161010742, -0.004246303346008062, -0.03376244381070137, -0.10251011699438095, 0.07040094584226608, 0.04750203713774681, 0.025502661243081093, -0.02476944588124752, -0.0324251651763916, 0.04968176409602165, 0.06787426024675369, -0.12092024832963943, -0.17420317232608795, -0.056716226041316986, 0.14966116845607758, 0.09733416885137558, -0.04583374410867691, -0.0024321943055838346, 0.05968613177537918, -0.11434418708086014, 0.03933097794651985, 0.0816856399178505, -0.36326414346694946, 0.02808782458305359, 0.2629767060279846, -0.046797014772892, 0.10794921219348907, -0.04286104440689087, 0.12650255858898163, 0.1369887739419937, -0.01352559681981802, -0.0023801145143806934, -0.01901690475642681, 0.11377502977848053, 0.03303033858537674, -0.1014999970793724, -0.0666455402970314, 0.4941031336784363, 0.09218547493219376, -0.04978824034333229, -0.09308982640504837, -0.028182676061987877, -0.1282017081975937, -0.08176214247941971, -0.03292077034711838, 0.08395244181156158, 0.041040655225515366, 0.0940835028886795, 0.12142830342054367, -0.09550467878580093, -0.06849364936351776, -0.14241454005241394, 0.15228673815727234, 0.0019091127905994654, 0.09312522411346436, -0.12582635879516602, 0.024333614856004715, -0.10191629827022552, -0.06097405403852463, -0.010760028846561909, -0.07445398718118668, -0.047203291207551956, -0.03998209163546562, 0.036510732024908066, 0.2015213966369629, 0.08162354677915573, 0.08816076815128326, -0.03832700476050377, 0.010092937387526035, -0.04223581776022911, 0.07210152596235275, 0.09304200112819672, 0.1339450627565384, -0.027612878009676933, 0.01115256268531084, -0.017841944471001625, -0.06511490046977997, 0.04480023309588432, -0.03720713406801224, -0.16685254871845245, 0.0025890308897942305, 0.05136943608522415, 0.11517107486724854, -0.08015115559101105, -0.03272959589958191, -0.1186399981379509, -0.03685823827981949, 0.1764293760061264, -0.08968512713909149, 0.019757559522986412, 0.010118530131876469, -0.0032089592423290014, 0.06719992309808731, -0.11482031643390656, -0.015261069871485233, 0.061030495911836624, 0.007496617268770933, -0.1205148696899414, -0.005136375781148672, -0.010322950780391693, -0.08767126500606537, 0.05129232257604599, -0.09417334944009781, 0.028626572340726852, -0.06181781738996506, -0.1451798677444458, 0.0002215041604358703, -0.03896224498748779, -0.06725562363862991, -0.013918896205723286, -0.016590874642133713, -0.08126276731491089, -0.013168801553547382, 0.006633809767663479, -0.08834993839263916, -0.1028381809592247, 0.001417609048075974, -0.028458813205361366, 0.07016754895448685, -0.11654146015644073, 0.026977211236953735, -0.06332800537347794, 0.05549534782767296, -0.14475572109222412, 0.020050646737217903, -0.09326419979333878, 0.0570247508585453, -0.07852895557880402, -0.10106373578310013, 0.05096474662423134, -0.02017456851899624, -0.13295289874076843, 0.14724178612232208, -0.2153913974761963, -0.05250290781259537, 0.16301921010017395, -0.18967191874980927, -0.13751102983951569, 0.0379890538752079, -0.0036378768272697926, 0.04848620668053627, 0.0791160985827446, 0.10962922126054764, -0.0026807901449501514, -0.12871235609054565, -0.03102375566959381, 0.11038882285356522, -0.041480742394924164, -0.08714690059423447, 0.10417705029249191, -0.015529789961874485, -0.021610897034406662, 0.04044807329773903, -0.018335646018385887, -0.1302129179239273, -0.05070408806204796, -0.07474495470523834, -0.025225086137652397, 0.032680902630090714, 0.04096785932779312, 0.025761742144823074, 0.007920372299849987, -0.046142857521772385, 0.008001288399100304, 0.0034461612813174725, 0.05432545393705368, 0.020896684378385544, -0.052211694419384, -0.06544879823923111, 0.11389130353927612, -0.12881024181842804, -0.06618030369281769, -0.11233045160770416, -0.13547652959823608, 0.057874150574207306, 0.01331031508743763, 0.051483966410160065, 0.1475301831960678, -0.026410100981593132, 0.008838362991809845, 0.002357106190174818, -0.00048384544788859785, -0.03306188806891441, 0.08313758671283722, -0.016073917970061302, -0.20119942724704742, -0.0413375161588192, -0.05324540287256241, 0.08237910270690918, -0.06827397644519806, 0.017301950603723526, 0.21818594634532928, 0.03783643618226051, 0.011769582517445087, 0.06608058512210846, 0.004398121498525143, 0.05121941119432449, -0.0392167754471302, -0.034536831080913544, 0.0866045132279396, -0.06819479167461395, -0.08293548971414566, -0.006530133541673422, -0.0914611741900444, 0.08176618069410324, 0.16157542169094086, -0.10185445845127106, -0.038921061903238297, -0.0576113797724247, 0.0074025653302669525, 0.0017722424818202853, -0.08556929975748062, -0.0013578023063018918, -0.05503223091363907, -0.05103278532624245, 0.039725158363580704, -0.11120767891407013, -0.014453868381679058, 0.048862479627132416, -0.01609986647963524, -0.1501091569662094, 0.06784731149673462, 0.0806848406791687, -0.18246974050998688, 0.16650362312793732, 0.3054373264312744, 0.14630679786205292, 0.050898145884275436, -0.056826647371053696, -0.03993844613432884, -0.07074804604053497, -0.010058927349746227, -0.06517490744590759, 0.2051897794008255, -0.12545111775398254, -0.025860246270895004, 0.06273581832647324, 0.046096645295619965, 0.005877739749848843, -0.21709392964839935, -0.011881912127137184, -0.032071635127067566, -0.07879411429166794, -0.15016750991344452, -0.03440835326910019, 0.003392346203327179, 0.11854957789182663, 0.034024305641651154, -0.22931793332099915, 0.10609198361635208, -0.04737434163689613, -0.09509562700986862, 0.16003082692623138, -0.10406653583049774, -0.20107436180114746, -0.16413523256778717, -0.06666447967290878, -0.0779612585902214, 0.04569277912378311, -0.01195988804101944, -0.0711042508482933, -0.054942209273576736, -0.00915894191712141, -0.09555239230394363, -0.16586478054523468, -0.04950440302491188, -0.04515838623046875, 0.09681696444749832, -0.11735314875841141, -0.07337035983800888, -0.08632862567901611, -0.03518729656934738, 0.034721218049526215, 0.09551652520895004, -0.14903251826763153, 0.08019978553056717, 0.29893213510513306, -0.029134975746273994, 0.07034244388341904, -0.03767318278551102, 0.11524485051631927, -0.06462199985980988, 0.015605909749865532, 0.10782861709594727, 0.06749647855758667, 0.030609358102083206, 0.2699412405490875, 0.10237937420606613, -0.14466895163059235, -0.053586822003126144, -0.07584010064601898, -0.10207401216030121, -0.15878231823444366, -0.10914254933595657, -0.047386735677719116, 0.004906166810542345, 0.03935186192393303, 0.04487299174070358, 0.06150816008448601, 0.04912425950169563, 0.01224262360483408, 0.008874048478901386, 0.0661998763680458, 0.06143568083643913, 0.15972647070884705, -0.033401161432266235, 0.09321679174900055, -0.0627286359667778, -0.01842958852648735, 0.08260802924633026, 0.12441718578338623, 0.18458519876003265, 0.20124678313732147, 0.11431397497653961, 0.11447865515947342, 0.050407566130161285, 0.16001783311367035, 0.040465738624334335, 0.1573100984096527, -0.012489864602684975, -0.013086059130728245, -0.07840822637081146, 0.042192794382572174, 0.07250076532363892, -0.036377135664224625, -0.07762982696294785, -0.05500749126076698, -0.11353377997875214, 0.07224692404270172, -0.0010787948267534375, 0.23510460555553436, -0.25581225752830505, 0.022347936406731606, 0.0959784984588623, 0.11728867143392563, -0.05414513126015663, 0.1044125109910965, -0.01692565716803074, -0.08377499878406525, 0.07198834419250488, 0.00019440639880485833, 0.10272329300642014, -0.0847085490822792, -0.01359903160482645, -0.0002808326098602265, -0.06344937533140182, 0.004871460143476725, 0.07742992788553238, -0.1053648516535759, 0.28623726963996887, 0.04499635845422745, -0.033125706017017365, -0.06514368206262589, -0.015923339873552322, -0.005727894604206085, 0.2691815495491028, 0.22284995019435883, 0.04262508824467659, -0.21850423514842987, -0.21181504428386688, -0.03627162054181099, -0.022259915247559547, 0.143771693110466, -0.03929362818598747, 0.020101236179471016, 0.031348682940006256, 0.012494059279561043, -0.006826250813901424, 0.02048342488706112, -0.08275233954191208, -0.03255539387464523, 0.033041760325431824, 0.1218532994389534, -0.05907479301095009, -0.03297272324562073, -0.05490074306726456, -0.18863923847675323, 0.12267446517944336, -0.11230623722076416, -0.08918671309947968, -0.08554235845804214, -0.038757938891649246, 0.07661259919404984, -0.0028275330550968647, -0.00882504042237997, -0.05145658925175667, 0.11679480969905853, 0.005665498785674572, -0.09478648751974106, 0.1060485690832138, -0.04567953944206238, -0.05723625421524048, -0.05143343284726143, 0.14832210540771484, -0.04571009427309036, -0.004954999778419733, 0.08285273611545563, 0.07894118875265121, -0.0044600991532206535, -0.12646590173244476, 0.09886225312948227, -0.0099251763895154, 0.061704132705926895, 0.23459993302822113, -0.10100562125444412, -0.13306733965873718, -0.035452552139759064, 0.049234360456466675, 0.07172832638025284, 0.2987561523914337, -0.11066237837076187, 0.09777367115020752, 0.09469505399465561, -0.04181015491485596, -0.18932203948497772, -0.02026222087442875, -0.17316946387290955, -0.00574108911678195, -0.020051151514053345, -0.06167563423514366, 0.12931960821151733, 0.06526342779397964, -0.07319558411836624, 0.12036222219467163, -0.23816736042499542, -0.08278535306453705, 0.16413938999176025, 0.04290027171373367, 0.14626993238925934, -0.06810710579156876, -0.10985154658555984, -0.07606552541255951, -0.2023620456457138, 0.15823830664157867, 0.008313131518661976, 0.08137176930904388, -0.04466485604643822, -0.01031293161213398, -0.005764672998338938, 0.0016942763468250632, 0.24252279102802277, 0.13000307977199554, 0.11575871706008911, 0.0410989448428154, -0.1560717523097992, 0.2534436583518982, 0.011263403110206127, 0.020132796838879585, 0.11813158541917801, 0.00565307354554534, -0.08344988524913788, -0.014656865037977695, 0.009873991832137108, 0.016077779233455658, -0.08585617691278458, -0.0618777871131897, -0.09066876024007797, 0.011224602349102497, -0.07304469496011734, -0.07567139714956284, 0.23793651163578033, -0.057095106691122055, 0.06352704018354416, 0.1880899965763092, 0.005279981531202793, -0.10675178468227386, -0.049573712050914764, -0.060629893094301224, -0.08667755872011185, 0.05587523430585861, -0.17109231650829315, 0.05056922882795334, 0.11474195867776871, 0.03820998594164848, 0.14317600429058075, 0.11862756311893463, -0.02345028892159462, -0.04831225052475929, 0.11662789434194565, -0.08561655133962631, -0.14909638464450836, 0.014233003370463848, -0.2027336061000824, 0.027250483632087708, 0.07886254042387009, 0.07179734855890274, -0.0018246157560497522, -0.016222795471549034, 0.024982329457998276, 0.035005856305360794, -0.06894157826900482, 0.07725698500871658, 0.024850904941558838, 0.05899198725819588, -0.14741376042366028, 0.15153467655181885, 0.08140043914318085, 0.01461124885827303, -0.09131469577550888, -0.0851060003042221, -0.15023864805698395, -0.05527853965759277, -0.03901746869087219, 0.10751599073410034, -0.1442699134349823, -0.13946859538555145, -0.10302101820707321, -0.18507981300354004, 0.019559036940336227, 0.0825081393122673, 0.14364397525787354, 0.11854439228773117, 0.022650552913546562, -0.10586965084075928, 0.020916512235999107, 0.05989677459001541, -0.11545216292142868, 0.025937216356396675, -0.21120066940784454, 0.05260515213012695, -0.021542245522141457, 0.09287124127149582, -0.0909704640507698, -0.050593018531799316, -0.10576430708169937, 0.01215311698615551, -0.06243650242686272, 0.0779610201716423, -0.06788360327482224, 0.005036062560975552, 0.018205055966973305, 0.015125119127333164, -0.05468454957008362, 0.002838734071701765, -0.0881517305970192, 0.05076843872666359, 0.02580079436302185, 0.17738373577594757, -0.09735121577978134, -0.020686699077486992, 0.05904453992843628, -0.035131461918354034, 0.06808776408433914, 0.0172868724912405, 0.0209550429135561, 0.058924052864313126, -0.15162110328674316, 0.031100193038582802, 0.09912572801113129, 0.03231531009078026, 0.0649973601102829, -0.12496054172515869, -0.0009505844791419804, 0.05332842841744423, -0.09204840660095215, 0.08818327635526657, -0.08908951282501221, -0.10885732620954514, -0.148936465382576, -0.11772937327623367, -0.11350806802511215, -0.031287409365177155, 0.03786865621805191, 0.21015797555446625, 0.03687160462141037, 0.0513434074819088, 0.03282148018479347, -0.004413660150021315, -0.04382277652621269, -0.029208706691861153, -0.06632871180772781, -0.0825536772608757, -0.06193843483924866, -0.01185196079313755, -0.008238728158175945, -0.028696352615952492, 0.35530662536621094, 0.04186559468507767, -0.019017742946743965, 0.05373983830213547, 0.17581380903720856, -0.01903216727077961, 0.03825109824538231, 0.2920352816581726, 0.06690358370542526, -0.034791313111782074, 0.0385039821267128, 0.03505776450037956, -0.005083543714135885, 0.026611438021063805, 0.13967959582805634, 0.09902851283550262, -0.08417165279388428, 0.05233171209692955, 0.08394066244363785, -0.03554443269968033, -0.024233177304267883, 0.11782915145158768, 0.05672597512602806, 0.04202960804104805, 0.03877565264701843, -0.07952756434679031, 0.1245117262005806, -0.17514373362064362, 0.08291848003864288, -0.001869208994321525, -0.09116273373365402, -0.18642351031303406, -0.11220096051692963, -0.08741098642349243, -0.07877695560455322, 0.025062382221221924, -0.10904214531183243, -0.04557626694440842, 0.17665907740592957, 0.019587984308600426, -0.0223850104957819, -0.018440401181578636, -0.24619726836681366, 0.011170857585966587, 0.10871850699186325, 0.015622579492628574, -0.009099767543375492, -0.07968004047870636, -0.005052314605563879, 0.024557968601584435, -0.10349047183990479, -0.0413210503757, -0.02892024628818035, 0.06758610904216766, -0.03424226865172386, -0.15372827649116516, -0.057961177080869675, -0.038767993450164795, -0.004906147718429565, -0.01300619263201952, -0.023529192432761192, 0.025360742583870888, -0.02226540446281433, 0.04073664918541908, 0.18833975493907928, -0.048294443637132645, 0.008473499678075314, -0.06769437342882156, 0.23921804130077362, -0.055905360728502274, 0.07767180353403091, 0.07663410156965256, -0.07191310822963715, 0.0003459442814346403, 0.0792202353477478, 0.18814904987812042, -0.012818766757845879, -0.008260911330580711, 0.015916144475340843, -0.003277712734416127, 0.04123755916953087, 0.024818619713187218, -0.06559041887521744, 0.12435705959796906, -0.05413991957902908, 0.05450914055109024, -0.10693985968828201, 0.006451353430747986, -0.044598665088415146, -0.04943462461233139, 0.11579426378011703, -0.08825626224279404, -0.13360080122947693, 0.18862922489643097, -0.04589084908366203, 0.02840382792055607, 0.2960452437400818, -0.07968685775995255, -0.07816416025161743, -0.0003751165058929473, -0.01642579771578312, -0.023035269230604172, 0.034523699432611465, -0.15269382297992706, 0.0059292674995958805, 0.05508046597242355, 0.01407061330974102, -0.18646740913391113, -0.05368676781654358, 0.02799980342388153, 0.001904219388961792, 0.09223173558712006, -0.0017765635857358575, 0.21953818202018738, 0.09995222836732864, -0.0035242121666669846, -0.08077842742204666, 0.10203532874584198, 0.0035727459471672773, 0.0376279316842556, -0.001758163794875145, 0.02231457643210888, -0.016706448048353195, -0.04367652162909508, 0.09478902816772461, -0.07720743119716644, 0.013284144923090935, 0.016123885288834572, -0.1176999881863594, -0.07376014441251755, 0.01919964887201786, -0.10536607354879379, 0.08765166252851486, 0.05095658078789711, -0.02849440462887287, -0.03391394391655922, 0.02172129787504673, 0.07538878917694092, 0.09047824889421463, -0.09260491281747818, 0.036674220114946365, 0.014517667703330517, 0.0216269101947546, 0.10169968754053116, -0.01017606258392334, -0.16540388762950897, -0.02997240051627159, -0.051250360906124115, 0.02315935492515564, -0.07713884115219116, 0.10741882771253586, 0.12350884824991226, 0.04750902205705643, -0.05969291180372238, -0.19981643557548523, 0.03945329040288925, 0.0740835890173912, -0.0726441964507103, -0.08894511312246323 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-base-bne-finetuned-amazon_reviews_multi This model is a fine-tuned version of [BSC-TeMU/roberta-base-bne](https://huggingface.co/BSC-TeMU/roberta-base-bne) on the amazon_reviews_multi dataset. It achieves the following results on the evaluation set: - Loss: 0.2303 - Accuracy: 0.9325 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.1942 | 1.0 | 1250 | 0.1751 | 0.932 | | 0.0935 | 2.0 | 2500 | 0.2303 | 0.9325 | ### Framework versions - Transformers 4.10.2 - Pytorch 1.9.0+cu102 - Datasets 1.12.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["amazon_reviews_multi"], "metrics": ["accuracy"], "model-index": [{"name": "roberta-base-bne-finetuned-amazon_reviews_multi", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "amazon_reviews_multi", "type": "amazon_reviews_multi", "args": "es"}, "metrics": [{"type": "accuracy", "value": 0.9325, "name": "Accuracy"}]}]}]}
text-classification
csalamea/roberta-base-bne-finetuned-amazon_reviews_multi
[ "transformers", "pytorch", "tensorboard", "roberta", "text-classification", "generated_from_trainer", "dataset:amazon_reviews_multi", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #roberta #text-classification #generated_from_trainer #dataset-amazon_reviews_multi #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
roberta-base-bne-finetuned-amazon\_reviews\_multi ================================================= This model is a fine-tuned version of BSC-TeMU/roberta-base-bne on the amazon\_reviews\_multi dataset. It achieves the following results on the evaluation set: * Loss: 0.2303 * Accuracy: 0.9325 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 2 ### Training results ### Framework versions * Transformers 4.10.2 * Pytorch 1.9.0+cu102 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2", "### Training results", "### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #roberta #text-classification #generated_from_trainer #dataset-amazon_reviews_multi #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2", "### Training results", "### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 71, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #roberta #text-classification #generated_from_trainer #dataset-amazon_reviews_multi #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.08748505264520645, 0.10135051608085632, -0.002402752870693803, 0.12601561844348907, 0.1706485152244568, 0.03687635064125061, 0.1622617095708847, 0.11569225788116455, -0.08599758893251419, -0.005865089595317841, 0.11542874574661255, 0.1548892855644226, 0.02126530557870865, 0.12217552959918976, -0.05617023631930351, -0.24667197465896606, -0.0027746891137212515, 0.031354084610939026, -0.04765567183494568, 0.14976318180561066, 0.10777784883975983, -0.12571990489959717, 0.10260860621929169, 0.006307870149612427, -0.16979162395000458, -0.014389348216354847, 0.024533210322260857, -0.0696016252040863, 0.13328847289085388, 0.02715180069208145, 0.12315018475055695, -0.001376794883981347, 0.07599072903394699, -0.19718699157238007, 0.018568968400359154, 0.04323352873325348, 0.007994817569851875, 0.09346025437116623, 0.02965611405670643, -0.017832037061452866, 0.11385732144117355, -0.05450711026787758, 0.07709876447916031, 0.017052244395017624, -0.12038940191268921, -0.23559437692165375, -0.09075293689966202, 0.04976227506995201, 0.05859314277768135, 0.09819751977920532, -0.005986036732792854, 0.15179461240768433, -0.08127760142087936, 0.08983530849218369, 0.23404285311698914, -0.28485214710235596, -0.0688856989145279, 0.021996930241584778, 0.030939679592847824, 0.08619936555624008, -0.09751075506210327, -0.033980995416641235, 0.04781979322433472, 0.05363927781581879, 0.12734320759773254, -0.0339665561914444, -0.09865715354681015, 0.014300784096121788, -0.14419035613536835, -0.03703422099351883, 0.2110384702682495, 0.05224103480577469, -0.04936090111732483, -0.033593520522117615, -0.033585041761398315, -0.13613587617874146, -0.043771665543317795, 0.00799341220408678, 0.06177709624171257, -0.047818057239055634, -0.08494832366704941, -0.01645578444004059, -0.1082940623164177, -0.04342750832438469, -0.06920549273490906, 0.12287094444036484, 0.039220601320266724, 0.02422584593296051, -0.04213975742459297, 0.0993783175945282, 0.016896851360797882, -0.11286630481481552, 0.016236115247011185, 0.007824072614312172, -0.005469383206218481, -0.04410327225923538, -0.055700261145830154, -0.07245267927646637, 0.01179128885269165, 0.1487048864364624, -0.03140841796994209, 0.029435962438583374, 0.03786361217498779, 0.058199744671583176, -0.08600983023643494, 0.2011789083480835, -0.03808872774243355, -0.019710825756192207, 0.0025984770618379116, 0.06378598511219025, 0.0366097018122673, -0.01270147506147623, -0.1338830590248108, 0.021989855915308, 0.08388182520866394, 0.012604576535522938, -0.05425940081477165, 0.05256333574652672, -0.07947677373886108, -0.051174573600292206, -0.00332809635438025, -0.07323624938726425, 0.01931113190948963, -0.012507437728345394, -0.06862229853868484, -0.023805459961295128, 0.019980812445282936, 0.038462378084659576, -0.004955208860337734, 0.12952286005020142, -0.08990225940942764, 0.027578171342611313, -0.08297199010848999, -0.09065180271863937, 0.030825134366750717, -0.08147244155406952, 0.04217829927802086, -0.10889053344726562, -0.18562471866607666, -0.022130921483039856, 0.06331305205821991, -0.01936176046729088, -0.08627636730670929, -0.031135743483901024, -0.06069953739643097, 0.007166009396314621, -0.012592478655278683, 0.14891761541366577, -0.07661006599664688, 0.10016986727714539, 0.018237467855215073, 0.05291328206658363, -0.04553815349936485, 0.04563181847333908, -0.10529284179210663, -0.0009554275893606246, -0.14402805268764496, 0.03092854842543602, -0.038881681859493256, 0.07289519160985947, -0.08744402229785919, -0.10114515572786331, 0.01812109909951687, 0.0007858672179281712, 0.043990358710289, 0.08458218723535538, -0.16604909300804138, -0.0736570879817009, 0.1392926573753357, -0.05869176238775253, -0.13591888546943665, 0.12868312001228333, -0.07782956212759018, 0.054878707975149155, 0.0784146636724472, 0.16067740321159363, 0.06189722940325737, -0.06682360172271729, 0.03472161293029785, -0.00491720624268055, 0.042379315942525864, -0.06637439131736755, 0.09732348471879959, 0.008913968689739704, -0.017364542931318283, 0.027644045650959015, -0.05057794228196144, 0.04496327042579651, -0.08496037125587463, -0.09797971695661545, -0.03988826647400856, -0.10726158320903778, 0.06420964747667313, 0.06216416135430336, 0.06328140944242477, -0.10556590557098389, -0.07531622052192688, 0.05133743956685066, 0.09141959249973297, -0.03835410252213478, 0.019969282671809196, -0.054281849414110184, 0.07662713527679443, -0.04022379219532013, -0.0186300165951252, -0.17294596135616302, -0.020581988617777824, 0.014201127924025059, -0.021982483565807343, 0.04035462439060211, 0.01553034782409668, 0.052458200603723526, 0.034605711698532104, -0.06715890020132065, -0.0013398109003901482, -0.05249806120991707, -0.004969918634742498, -0.11456017941236496, -0.21327568590641022, -0.022570785135030746, -0.015485681593418121, 0.15145157277584076, -0.20356561243534088, 0.03070143051445484, -0.05047479271888733, 0.07912503182888031, 0.038957469165325165, -0.008191968314349651, -0.023654483258724213, 0.06938077509403229, -0.029641080647706985, -0.041676975786685944, 0.07319847494363785, 0.015923120081424713, -0.11104981601238251, -0.004017536528408527, -0.07411255687475204, 0.18232916295528412, 0.12914419174194336, -0.09189614653587341, -0.07028195261955261, 0.02698177844285965, -0.04606953263282776, -0.033419687300920486, -0.07677464187145233, 0.04015912488102913, 0.1786910742521286, 0.0123989786952734, 0.1393493115901947, -0.09585434198379517, -0.046626247465610504, 0.026534689590334892, -0.04766058921813965, 0.029939081519842148, 0.1388590931892395, 0.11947977542877197, -0.08978938311338425, 0.1408824771642685, 0.1801275759935379, -0.08513115346431732, 0.13522493839263916, -0.04518233612179756, -0.057722315192222595, -0.027640100568532944, -0.046595100313425064, -0.014140930958092213, 0.10802013427019119, -0.12211117148399353, 0.008783064782619476, 0.038905952125787735, 0.01179005578160286, 0.007891636341810226, -0.220276340842247, -0.04456881061196327, 0.03768514469265938, -0.03834724798798561, -0.016397174447774887, -0.0017061536200344563, 0.01423916406929493, 0.1071462482213974, 0.013389743864536285, -0.07625076174736023, 0.04752439633011818, 0.007178463041782379, -0.08612769842147827, 0.21150490641593933, -0.07518668472766876, -0.1776520013809204, -0.13706344366073608, -0.0545949712395668, -0.05295511707663536, -0.003149912226945162, 0.0614364892244339, -0.07413402199745178, -0.03180314600467682, -0.07279524952173233, -0.0024537749122828245, -0.0038665372412651777, 0.008661623112857342, -0.011396964080631733, 0.014962404035031796, 0.05599657818675041, -0.09647130221128464, -0.012222180142998695, -0.04970064014196396, -0.02174299955368042, 0.039092689752578735, 0.040774259716272354, 0.10311493277549744, 0.15333402156829834, -0.014502475969493389, -0.008167930878698826, -0.02158825471997261, 0.23083464801311493, -0.08114238828420639, -0.043072085827589035, 0.14223170280456543, -0.01715271733701229, 0.0431506372988224, 0.13593509793281555, 0.07226701825857162, -0.08730296790599823, 0.015815844759345055, 0.02846592850983143, -0.03927796334028244, -0.268171101808548, -0.033646851778030396, -0.0507233627140522, -0.007956546731293201, 0.0783345103263855, 0.019632715731859207, 0.003466376569122076, 0.06746618449687958, 0.036469560116529465, 0.07697193324565887, -0.024515310302376747, 0.0725059062242508, 0.12220457196235657, 0.047124069184064865, 0.13204236328601837, -0.051729895174503326, -0.05593999847769737, 0.06652961671352386, -0.005725420080125332, 0.2295539677143097, 0.018175311386585236, 0.14277383685112, 0.07377737015485764, 0.1386025846004486, 0.008008481003344059, 0.04504062235355377, 0.017918335273861885, -0.019810272380709648, -0.03349725902080536, -0.024137189611792564, -0.03777434304356575, 0.026318218559026718, -0.04787589609622955, 0.05510532483458519, -0.1216469258069992, -0.011227552779018879, 0.05953656882047653, 0.26264089345932007, 0.02338678576052189, -0.31383731961250305, -0.10186455398797989, 0.016122763976454735, -0.05420726165175438, -0.0018863070290535688, 0.028928078711032867, 0.06222778186202049, -0.13198141753673553, 0.041389141231775284, -0.07491769641637802, 0.10316328704357147, -0.08767863363027573, 0.04334351792931557, 0.0641695037484169, 0.07428254187107086, 0.0013433737913146615, 0.08148038387298584, -0.29981619119644165, 0.27460289001464844, -0.008198552764952183, 0.048941873013973236, -0.06314488500356674, -0.027548687532544136, 0.035845763981342316, 0.054587073624134064, 0.05386405438184738, 0.0038666280452162027, -0.0410676933825016, -0.15771129727363586, -0.040473781526088715, 0.02706184983253479, 0.06241348013281822, -0.02723764255642891, 0.08145689219236374, -0.03775765746831894, 0.004442242439836264, 0.05066261440515518, -0.000723798293620348, -0.05325597897171974, -0.09176760166883469, -0.0010839176829904318, 0.02408144436776638, -0.048524193465709686, -0.06867502629756927, -0.1267763376235962, -0.07732057571411133, 0.11763031780719757, -0.018511952832341194, -0.04903564229607582, -0.09150055795907974, 0.06172432377934456, 0.08089186996221542, -0.08031509816646576, 0.03645249828696251, -0.002867340575903654, 0.09249304234981537, 0.028154658153653145, -0.04747135564684868, 0.09116265922784805, -0.0564410425722599, -0.19131174683570862, -0.06531649827957153, 0.1091921254992485, 0.02420763298869133, 0.06942029297351837, -0.02535104751586914, 0.010509409941732883, -0.06197625771164894, -0.08357580006122589, 0.01257217675447464, 0.00821640994399786, 0.07106112688779831, 0.04573948308825493, -0.036858171224594116, 0.012068760581314564, -0.08304876834154129, -0.06253399699926376, 0.18849623203277588, 0.2214498221874237, -0.08738681674003601, 0.04031470790505409, 0.036602821201086044, -0.07414983212947845, -0.15455880761146545, 0.011297655291855335, 0.05963510647416115, 0.0006671959417872131, 0.06797772645950317, -0.14696437120437622, 0.11902517825365067, 0.09333305060863495, -0.02068314328789711, 0.11209964007139206, -0.3217496871948242, -0.13395251333713531, 0.10134933888912201, 0.1372978538274765, 0.13894343376159668, -0.13172198832035065, -0.02082870528101921, -0.04572247341275215, -0.146324023604393, 0.1378229707479477, -0.10358641296625137, 0.12806963920593262, -0.03092869371175766, 0.1106255054473877, 0.0037294209469109774, -0.047299932688474655, 0.12950772047042847, 0.019552117213606834, 0.09933143854141235, -0.057583749294281006, -0.028160307556390762, 0.03231542930006981, -0.04104400798678398, 0.016753770411014557, -0.09342203289270401, 0.0307865459471941, -0.09329255670309067, -0.030915401875972748, -0.07106336951255798, 0.023760778829455376, -0.040977418422698975, -0.048031218349933624, -0.04165596887469292, 0.03756199777126312, 0.034384556114673615, -0.012077606283128262, 0.1650484949350357, 0.024534137919545174, 0.13603870570659637, 0.07593537122011185, 0.09001367539167404, -0.062306176871061325, -0.09832610934972763, -0.04641509801149368, -0.030273811891674995, 0.048144929111003876, -0.17123320698738098, 0.02887239120900631, 0.13160167634487152, 0.011140162125229836, 0.15611597895622253, 0.06286188215017319, -0.03973785787820816, 0.007328100968152285, 0.060210712254047394, -0.1473669558763504, -0.09190437942743301, -0.007199518382549286, -0.055258866399526596, -0.13529729843139648, 0.03206604719161987, 0.12203475832939148, -0.06728495657444, -0.03177840635180473, -0.006757003255188465, 0.017898136749863625, -0.05075324699282646, 0.17833282053470612, 0.07896671444177628, 0.047494325786828995, -0.10676582902669907, 0.1078370213508606, 0.07093336433172226, -0.0688825398683548, 0.004336613230407238, 0.05840630084276199, -0.09213374555110931, -0.056954748928546906, 0.0604483038187027, 0.1810775101184845, -0.08093660324811935, -0.05632931366562843, -0.14352649450302124, -0.12352213263511658, 0.08281289786100388, 0.1308985948562622, 0.11646679043769836, 0.011451567523181438, -0.03895536810159683, -0.02378356270492077, -0.08104820549488068, 0.10755844414234161, 0.07420925796031952, 0.06614638864994049, -0.1541292667388916, 0.07461540400981903, 0.02957092598080635, 0.05183961242437363, -0.01501353271305561, 0.037376768887043, -0.10405530780553818, 0.014517159201204777, -0.14131805300712585, 0.0029443365056067705, -0.01665409840643406, 0.023352298885583878, -0.0036167106591165066, -0.06881476938724518, -0.07280328869819641, 0.013769056648015976, -0.12243233621120453, -0.023349132388830185, 0.037886541336774826, 0.07895298302173615, -0.09635411202907562, -0.03891822323203087, 0.04241982474923134, -0.04902099817991257, 0.07320401817560196, 0.04164178669452667, 0.014727015048265457, 0.06638604402542114, -0.12203842401504517, 0.031979624181985855, 0.03355097770690918, 0.020873717963695526, 0.047238755971193314, -0.131159245967865, -0.0029366749804466963, 0.0008752434514462948, 0.06524120271205902, 0.023719217628240585, 0.07833414524793625, -0.15566109120845795, -0.011630808934569359, 0.008436531759798527, -0.08374901115894318, -0.04931286349892616, 0.013993371278047562, 0.0644172802567482, 0.03146012872457504, 0.2301100641489029, -0.06705774366855621, 0.03818197548389435, -0.19638380408287048, 0.011850639246404171, -0.02603587694466114, -0.12201353162527084, -0.14749230444431305, -0.07459408044815063, 0.04430370405316353, -0.06755319982767105, 0.16799087822437286, 0.03966196998953819, 0.07443663477897644, 0.029320290312170982, 0.019833926111459732, -0.0012724585831165314, 0.01750226877629757, 0.15605953335762024, 0.017472129315137863, -0.0385797917842865, 0.0553094781935215, 0.029253395274281502, 0.10055885463953018, 0.09190146625041962, 0.1844479888677597, 0.17639435827732086, 0.029132237657904625, 0.08518511056900024, 0.03540175408124924, -0.019055966287851334, -0.1243986189365387, 0.04176418483257294, -0.01574995554983616, 0.11440835893154144, -0.013763918541371822, 0.21950466930866241, 0.07746719568967819, -0.16513484716415405, 0.046616800129413605, -0.06316568702459335, -0.0753798633813858, -0.1084527000784874, -0.0666743814945221, -0.09666218608617783, -0.150067538022995, 0.0005522034480236471, -0.11988969147205353, 0.0009473717072978616, 0.11587756127119064, 0.0018456995021551847, -0.043480828404426575, 0.09077279269695282, 0.009991518221795559, 0.0034860013984143734, 0.08070904016494751, 0.009907604195177555, -0.04668661952018738, -0.08920586854219437, -0.06421113014221191, -0.021299738436937332, -0.012637236155569553, 0.021962594240903854, -0.05690223351120949, -0.05890936031937599, 0.0226567555218935, -0.028140811249613762, -0.10631848871707916, 0.021962102502584457, 0.009738780558109283, 0.07327093183994293, 0.032759103924036026, 0.01137351430952549, 0.026509786024689674, 0.007413798477500677, 0.26563242077827454, -0.06026383861899376, -0.07014074176549911, -0.12706468999385834, 0.22582225501537323, 0.02476755902171135, -0.04121285676956177, 0.03648244962096214, -0.06837393343448639, 0.0009176467428915203, 0.24895620346069336, 0.2247450053691864, -0.09457942843437195, -0.01954338513314724, 0.011463322676718235, -0.007764757145196199, -0.013498657383024693, 0.10854029655456543, 0.10903426259756088, 0.006476501002907753, -0.08121499419212341, -0.014552807435393333, -0.05886773392558098, -0.0011637939605861902, -0.016025392338633537, 0.06969863176345825, 0.03950197994709015, -0.0072053768672049046, -0.04185018688440323, 0.07962869107723236, -0.0905042514204979, -0.11997628957033157, 0.032596368342638016, -0.20977969467639923, -0.18324455618858337, -0.02352200448513031, 0.09226783365011215, 0.017664752900600433, 0.058241747319698334, -0.01702686958014965, -0.012324018403887749, 0.07180946320295334, -0.016307998448610306, -0.10270724445581436, -0.10217396169900894, 0.1081206202507019, -0.09479250013828278, 0.20147524774074554, -0.04637618735432625, 0.05569562315940857, 0.12050214409828186, 0.06346482783555984, -0.07389548420906067, 0.05974552035331726, 0.044635169208049774, -0.024630477651953697, 0.04403224587440491, 0.0851675420999527, -0.025177715346217155, 0.06843596696853638, 0.06200204789638519, -0.10778521746397018, 0.0037345902528613806, -0.06963280588388443, -0.04284678027033806, -0.05494605004787445, -0.01213314663618803, -0.07727371901273727, 0.1269443780183792, 0.22313793003559113, -0.05046144500374794, -0.014237524941563606, -0.06106938049197197, 0.02333267591893673, 0.06654650717973709, 0.01524131279438734, -0.05319321155548096, -0.2105739861726761, 0.00934837106615305, 0.06471182405948639, -0.008181453682482243, -0.26207032799720764, -0.07192134112119675, -0.0029018190689384937, -0.06448803097009659, -0.06932234764099121, 0.08557683229446411, 0.07956942170858383, 0.03840881958603859, -0.06028938665986061, -0.04749587923288345, -0.07239669561386108, 0.15115003287792206, -0.15023668110370636, -0.09064730256795883 ]
null
null
transformers
## BERT-base uncased model fine-tuned on SQuAD v1 This model was fine-tuned from the HuggingFace [BERT](https://www.aclweb.org/anthology/N19-1423/) base uncased checkpoint on [SQuAD1.1](https://rajpurkar.github.io/SQuAD-explorer). This model is case-insensitive: it does not make a difference between english and English. ## Details | Dataset | Split | # samples | | -------- | ----- | --------- | | SQuAD1.1 | train | 90.6K | | SQuAD1.1 | eval | 11.1k | ### Fine-tuning - Python: `3.7.5` - Machine specs: `CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz` `Memory: 32 GiB` `GPUs: 2 GeForce GTX 1070, each with 8GiB memory` `GPU driver: 418.87.01, CUDA: 10.1` - script: ```shell # after install https://github.com/huggingface/transformers cd examples/question-answering mkdir -p data wget -O data/train-v1.1.json https://rajpurkar.github.io/SQuAD-explorer/dataset/train-v1.1.json wget -O data/dev-v1.1.json https://rajpurkar.github.io/SQuAD-explorer/dataset/dev-v1.1.json python run_squad.py \ --model_type bert \ --model_name_or_path bert-base-uncased \ --do_train \ --do_eval \ --do_lower_case \ --train_file train-v1.1.json \ --predict_file dev-v1.1.json \ --per_gpu_train_batch_size 12 \ --per_gpu_eval_batch_size=16 \ --learning_rate 3e-5 \ --num_train_epochs 2.0 \ --max_seq_length 320 \ --doc_stride 128 \ --data_dir data \ --output_dir data/bert-base-uncased-squad-v1 2>&1 | tee train-energy-bert-base-squad-v1.log ``` It took about 2 hours to finish. ### Results **Model size**: `418M` | Metric | # Value | # Original ([Table 2](https://www.aclweb.org/anthology/N19-1423.pdf))| | ------ | --------- | --------- | | **EM** | **80.9** | **80.8** | | **F1** | **88.2** | **88.5** | Note that the above results didn't involve any hyperparameter search. ## Example Usage ```python from transformers import pipeline qa_pipeline = pipeline( "question-answering", model="csarron/bert-base-uncased-squad-v1", tokenizer="csarron/bert-base-uncased-squad-v1" ) predictions = qa_pipeline({ 'context': "The game was played on February 7, 2016 at Levi's Stadium in the San Francisco Bay Area at Santa Clara, California.", 'question': "What day was the game played on?" }) print(predictions) # output: # {'score': 0.8730505704879761, 'start': 23, 'end': 39, 'answer': 'February 7, 2016'} ``` > Created by [Qingqing Cao](https://awk.ai/) | [GitHub](https://github.com/csarron) | [Twitter](https://twitter.com/sysnlp) > Made with ❤️ in New York.
{"language": "en", "license": "mit", "tags": ["question-answering", "bert", "bert-base"], "datasets": ["squad"], "metrics": ["squad"], "widget": [{"text": "Which name is also used to describe the Amazon rainforest in English?", "context": "The Amazon rainforest (Portuguese: Floresta Amaz\u00f4nica or Amaz\u00f4nia; Spanish: Selva Amaz\u00f3nica, Amazon\u00eda or usually Amazonia; French: For\u00eat amazonienne; Dutch: Amazoneregenwoud), also known in English as Amazonia or the Amazon Jungle, is a moist broadleaf forest that covers most of the Amazon basin of South America. This basin encompasses 7,000,000 square kilometres (2,700,000 sq mi), of which 5,500,000 square kilometres (2,100,000 sq mi) are covered by the rainforest. This region includes territory belonging to nine nations. The majority of the forest is contained within Brazil, with 60% of the rainforest, followed by Peru with 13%, Colombia with 10%, and with minor amounts in Venezuela, Ecuador, Bolivia, Guyana, Suriname and French Guiana. States or departments in four nations contain \"Amazonas\" in their names. The Amazon represents over half of the planet's remaining rainforests, and comprises the largest and most biodiverse tract of tropical rainforest in the world, with an estimated 390 billion individual trees divided into 16,000 species."}, {"text": "How many square kilometers of rainforest is covered in the basin?", "context": "The Amazon rainforest (Portuguese: Floresta Amaz\u00f4nica or Amaz\u00f4nia; Spanish: Selva Amaz\u00f3nica, Amazon\u00eda or usually Amazonia; French: For\u00eat amazonienne; Dutch: Amazoneregenwoud), also known in English as Amazonia or the Amazon Jungle, is a moist broadleaf forest that covers most of the Amazon basin of South America. This basin encompasses 7,000,000 square kilometres (2,700,000 sq mi), of which 5,500,000 square kilometres (2,100,000 sq mi) are covered by the rainforest. This region includes territory belonging to nine nations. The majority of the forest is contained within Brazil, with 60% of the rainforest, followed by Peru with 13%, Colombia with 10%, and with minor amounts in Venezuela, Ecuador, Bolivia, Guyana, Suriname and French Guiana. States or departments in four nations contain \"Amazonas\" in their names. The Amazon represents over half of the planet's remaining rainforests, and comprises the largest and most biodiverse tract of tropical rainforest in the world, with an estimated 390 billion individual trees divided into 16,000 species."}], "model-index": [{"name": "csarron/bert-base-uncased-squad-v1", "results": [{"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad", "type": "squad", "config": "plain_text", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 80.9104, "name": "Exact Match", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiNDJlOWQ0OTE0ZjRhMTQwNDY5MjVhYmZiN2RmYzY0OWJiOWUyNjcyMWU5N2I3YmU0OThjZTVjNTc2MjM2Yzg5NiIsInZlcnNpb24iOjF9.cuJ34B-ngUur5wKGhfhVP8FM6NX4IFrIJEdXypbLQJw1i8M5Bb2EeIs-0M5n35YIx2PfqSQcnVj_jP8vLUk4Dg"}, {"type": "f1", "value": 88.2302, "name": "F1", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiYmE4NzFmNDA3MDRiODk3ZDg5NWYyNjczOGE5YjdkZWQ0ZmEzNWU5YjFjMzc1ODA2OGRjYzU0Y2M5MmU0NGNhYSIsInZlcnNpb24iOjF9.phmkVWF3I-rl2xrHW0EW9OQqzfuefoqNjWplOpFdzJuW8d2C4sJ8snW0Ikw9kQqZaBCdwdkmsf5VTgOupHb8Dw"}]}]}]}
question-answering
csarron/bert-base-uncased-squad-v1
[ "transformers", "pytorch", "jax", "safetensors", "bert", "question-answering", "bert-base", "en", "dataset:squad", "license:mit", "model-index", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #safetensors #bert #question-answering #bert-base #en #dataset-squad #license-mit #model-index #endpoints_compatible #has_space #region-us
BERT-base uncased model fine-tuned on SQuAD v1 ---------------------------------------------- This model was fine-tuned from the HuggingFace BERT base uncased checkpoint on SQuAD1.1. This model is case-insensitive: it does not make a difference between english and English. Details ------- Dataset: SQuAD1.1, Split: train, # samples: 90.6K Dataset: SQuAD1.1, Split: eval, # samples: 11.1k ### Fine-tuning * Python: '3.7.5' * Machine specs: 'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz' 'Memory: 32 GiB' 'GPUs: 2 GeForce GTX 1070, each with 8GiB memory' 'GPU driver: 418.87.01, CUDA: 10.1' * script: It took about 2 hours to finish. ### Results Model size: '418M' Metric: EM, # Value: 80.9, # Original (Table 2): 80.8 Metric: F1, # Value: 88.2, # Original (Table 2): 88.5 Note that the above results didn't involve any hyperparameter search. Example Usage ------------- > > Created by Qingqing Cao | GitHub | Twitter > > > > > Made with ️ in New York. > > >
[ "# samples: 90.6K\nDataset: SQuAD1.1, Split: eval, # samples: 11.1k", "### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 2 hours to finish.", "### Results\n\n\nModel size: '418M'\n\n\nMetric: EM, # Value: 80.9, # Original (Table 2): 80.8\nMetric: F1, # Value: 88.2, # Original (Table 2): 88.5\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>" ]
[ "TAGS\n#transformers #pytorch #jax #safetensors #bert #question-answering #bert-base #en #dataset-squad #license-mit #model-index #endpoints_compatible #has_space #region-us \n", "# samples: 90.6K\nDataset: SQuAD1.1, Split: eval, # samples: 11.1k", "### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 2 hours to finish.", "### Results\n\n\nModel size: '418M'\n\n\nMetric: EM, # Value: 80.9, # Original (Table 2): 80.8\nMetric: F1, # Value: 88.2, # Original (Table 2): 88.5\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>" ]
[ 62, 27, 95, 107 ]
[ "passage: TAGS\n#transformers #pytorch #jax #safetensors #bert #question-answering #bert-base #en #dataset-squad #license-mit #model-index #endpoints_compatible #has_space #region-us \n# samples: 90.6K\nDataset: SQuAD1.1, Split: eval, # samples: 11.1k### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 2 hours to finish.### Results\n\n\nModel size: '418M'\n\n\nMetric: EM, # Value: 80.9, # Original (Table 2): 80.8\nMetric: F1, # Value: 88.2, # Original (Table 2): 88.5\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>" ]
[ -0.12932617962360382, 0.09527844190597534, -0.0013612484326586127, 0.06507304310798645, 0.10248512774705887, 0.04070766642689705, 0.06183772161602974, 0.12735673785209656, 0.09208894520998001, 0.16582350432872772, 0.10345950722694397, 0.029340939596295357, 0.08973577618598938, 0.1347506195306778, -0.02124198153614998, -0.08617104589939117, 0.04407426714897156, -0.002348623936995864, -0.010493583045899868, 0.1197686716914177, 0.07891575247049332, -0.0649980753660202, 0.11728579550981522, -0.008787203580141068, -0.0997900515794754, -0.021268481388688087, 0.029858607798814774, -0.033813219517469406, 0.09710841625928879, 0.03504723682999611, 0.02357751876115799, 0.059879980981349945, 0.03427023068070412, -0.10387212038040161, 0.0220886692404747, 0.12788905203342438, -0.04108845070004463, 0.04301576688885689, 0.03201032429933548, 0.022163953632116318, 0.06649830937385559, -0.01784614659845829, 0.01861419901251793, 0.05478544533252716, -0.13121047616004944, -0.17107513546943665, -0.16496899724006653, 0.03450269252061844, 0.09538573771715164, 0.07215755432844162, -0.019059259444475174, 0.1349763572216034, -0.03178787603974342, 0.06628777831792831, 0.2513565421104431, -0.25441160798072815, -0.04934193566441536, 0.06875239312648773, 0.05772488936781883, 0.07069052010774612, -0.008942680433392525, 0.0030745866242796183, 0.037165652960538864, 0.04026574268937111, 0.018368229269981384, -0.011301624588668346, 0.037891414016485214, 0.013231619261205196, -0.1438325196504593, -0.055574920028448105, 0.09523468464612961, 0.044664040207862854, -0.05262686312198639, -0.03419601172208786, -0.1115698590874672, -0.2100444734096527, -0.03550101816654205, 0.015954680740833282, -0.02222253754734993, 0.016859620809555054, -0.03507425636053085, -0.0001171863914350979, -0.06001238152384758, -0.09210889041423798, -0.04184725135564804, 0.0303509384393692, 0.051970552653074265, 0.08323269337415695, 0.009682801552116871, 0.06922407448291779, -0.17697985470294952, -0.09222684800624847, -0.04422200098633766, 0.002766762161627412, -0.07759823650121689, 0.021901553496718407, 0.018398014828562737, 0.03636583313345909, 0.04918463155627251, 0.1884605437517166, -0.07648123055696487, 0.03903239220380783, 0.08343237638473511, -0.026919430121779442, -0.03417009860277176, 0.13793939352035522, -0.14397084712982178, -0.08262433856725693, 0.03556523472070694, 0.05724022909998894, -0.019722728058695793, -0.04963428154587746, -0.011491496115922928, 0.012771762907505035, 0.09927424043416977, 0.012383983470499516, 0.08180970698595047, 0.056715212762355804, -0.030198555439710617, -0.010293140076100826, 0.208576500415802, -0.066241055727005, 0.01800600253045559, 0.05172057822346687, -0.07481939345598221, -0.03429774194955826, 0.013649354688823223, -0.02496403641998768, -0.0826924666762352, -0.015320496633648872, -0.10423612594604492, -0.049669377505779266, -0.04095999151468277, -0.09072230756282806, 0.04646654427051544, -0.09507085382938385, -0.0520758181810379, -0.12032681703567505, -0.15330259501934052, -0.046840932220220566, 0.045259781181812286, -0.07434764504432678, 0.01927323266863823, 0.05593917518854141, -0.05355292558670044, 0.06662744283676147, -0.022739820182323456, 0.02561947889626026, -0.07218604534864426, 0.09486978501081467, 0.014563975855708122, 0.012159562669694424, -0.03955208882689476, 0.03204038366675377, -0.10713828355073929, 0.07743111997842789, -0.17043885588645935, 0.017071088775992393, -0.08562745153903961, 0.05940651521086693, -0.1303657740354538, -0.0423809252679348, 0.008841387927532196, -0.02623651549220085, 0.10586827248334885, 0.10929633677005768, -0.11141173541545868, -0.0180997122079134, 0.08645106852054596, -0.043725159019231796, -0.09467173367738724, 0.17957501113414764, 0.03210274130105972, -0.07073090970516205, 0.08281669020652771, 0.12327904999256134, 0.04736149311065674, -0.17229290306568146, -0.07397118955850601, -0.03778606280684471, 0.030503613874316216, -0.0380096398293972, 0.08641120791435242, -0.03312172740697861, 0.045339446514844894, 0.022093959152698517, -0.018005773425102234, -0.024212105199694633, -0.087443508207798, -0.03774942085146904, -0.06327178329229355, -0.07364396750926971, -0.0759449377655983, 0.006780123338103294, -0.005224363878369331, -0.0646403580904007, -0.14615841209888458, -0.13747233152389526, 0.08642414212226868, -0.0337863489985466, -0.015177717432379723, -0.09014178812503815, 0.17376376688480377, 0.0041328370571136475, 0.015759138390421867, -0.0973181426525116, -0.090190090239048, 0.08734408766031265, -0.08121494948863983, -0.02467106282711029, -0.048713475465774536, 0.016061602160334587, 0.028067341074347496, 0.01211546454578638, 0.0012994634453207254, 0.004790047183632851, -0.039438702166080475, -0.08580918610095978, -0.08651985973119736, -0.02045409195125103, -0.015500565059483051, 0.06863606721162796, -0.07683993875980377, 0.02783958986401558, 0.07245545834302902, 0.09130562841892242, -0.04443483427166939, -0.08367706090211868, -0.06577539443969727, -0.02606881782412529, -0.029845556244254112, -0.0854300707578659, 0.010224880650639534, 0.011755453422665596, 0.001415895065292716, 0.08473777025938034, -0.18132515251636505, 0.030505485832691193, 0.09651084989309311, 0.10809248685836792, -0.056114379316568375, 0.020022228360176086, -0.024860672652721405, -0.03609909862279892, -0.02343451790511608, -0.030398186296224594, 0.06134729087352753, 0.024503111839294434, 0.07866086810827255, -0.09249512106180191, -0.014115246944129467, 0.06663544476032257, 0.021277667954564095, -0.036545757204294205, 0.1199679970741272, 0.124179907143116, -0.17760860919952393, 0.10457179695367813, 0.007183239329606295, 0.02590322680771351, 0.1096678227186203, 0.0007817743462510407, -0.05227208137512207, -0.037927597761154175, 0.029676642268896103, -0.00868771132081747, 0.17837980389595032, -0.05755097046494484, 0.012424414046108723, 0.027332276105880737, 0.010513942688703537, 0.036651451140642166, -0.13346660137176514, 0.006395295727998018, -0.03637048974633217, -0.08158035576343536, -0.05590479075908661, 0.016145136207342148, -0.003693752223625779, 0.07553227245807648, 0.03879564628005028, -0.028251830488443375, -0.0032108912710100412, 0.011166098527610302, -0.07539361715316772, 0.21376265585422516, -0.09589456021785736, -0.17223761975765228, -0.07606903463602066, -0.01859196461737156, -0.009305106475949287, -0.018327180296182632, 0.038644663989543915, -0.09579513221979141, -0.056125544011592865, -0.06396704167127609, 0.08249752223491669, 0.03345334157347679, 0.02200450748205185, 0.0077803898602724075, -0.03610680624842644, 0.03229643777012825, -0.118328757584095, -0.010133707895874977, -0.017586831003427505, -0.10040958970785141, 0.04639869183301926, -0.026295650750398636, 0.04173017665743828, 0.14622443914413452, 0.0452050119638443, 0.0015140373725444078, 0.027688605710864067, 0.2774498760700226, -0.06704521179199219, 0.07738552987575531, 0.16200071573257446, 0.12889213860034943, 0.07562616467475891, 0.15063956379890442, -0.002533361781388521, -0.09068287163972855, 0.00244868453592062, 0.06264893710613251, -0.07677565515041351, -0.21540533006191254, -0.04452720656991005, -0.019984522834420204, 0.09027890861034393, 0.07620132714509964, 0.06188764423131943, -0.09501797705888748, 0.03200150281190872, -0.08787587285041809, 0.07669968158006668, 0.005339606199413538, 0.02209189534187317, -0.010765799321234226, 0.06733442097902298, 0.07942456007003784, 0.007119955029338598, 0.06614141166210175, 0.1135859563946724, 0.023987429216504097, 0.1170167624950409, -0.07698395103216171, 0.19237540662288666, 0.006574857980012894, 0.18717753887176514, 0.025094442069530487, 0.00577850267291069, -0.021351328119635582, 0.020077720284461975, 0.010450529865920544, -0.03229645639657974, -0.05047442764043808, 0.007041223347187042, 0.02773929014801979, -0.07420109212398529, -0.04454195499420166, 0.09565652906894684, 0.04478701949119568, 0.3267490267753601, 0.10555431246757507, -0.2863011956214905, -0.1254839301109314, -0.02379368618130684, -0.03326467424631119, -0.12474375218153, -0.0035115585196763277, 0.06531558185815811, -0.07023275643587112, -0.0002598424907773733, -0.04390999302268028, 0.08272963762283325, 0.015896636992692947, 0.0015774033963680267, 0.14153090119361877, 0.08048204332590103, -0.023524004966020584, 0.0024177886079996824, -0.19370588660240173, 0.16873973608016968, 0.04334590956568718, 0.05687454715371132, -0.04651058837771416, 0.00499389972537756, -0.029942290857434273, 0.013174455612897873, 0.060067590326070786, 0.030757341533899307, -0.018587857484817505, -0.1730903834104538, -0.2010306864976883, 0.01880086399614811, 0.0698627382516861, -0.030739570036530495, 0.07550223916769028, -0.012653414160013199, -0.002842299174517393, -0.02818796969950199, -0.01779618114233017, -0.06289578974246979, -0.07896735519170761, 0.09204927086830139, -0.0646408274769783, -0.06364371627569199, -0.09798726439476013, 0.001398950582370162, -0.06287092715501785, 0.11378558725118637, -0.0728965476155281, -0.05949185788631439, -0.124782994389534, 0.043033916503190994, 0.15005597472190857, -0.09903857111930847, 0.05495788902044296, -0.06900211423635483, 0.0258889552205801, -0.01744743436574936, -0.10786356776952744, 0.11907324939966202, -0.09382147341966629, -0.20237857103347778, -0.06414800137281418, 0.11716429889202118, 0.022805094718933105, 0.019682085141539574, 0.005316729657351971, 0.023569585755467415, -0.06737852841615677, -0.08101378381252289, 0.024897318333387375, -0.05043058097362518, 0.04812312126159668, 0.06835107505321503, 0.05658997967839241, -0.010469536297023296, -0.035923928022384644, -0.03820358216762543, -0.05829101428389549, 0.27774739265441895, -0.062495507299900055, 0.013879083096981049, 0.11518797278404236, 0.0075990138575434685, -0.23922105133533478, -0.0016476581804454327, 0.01800476387143135, -0.00012216651521157473, -0.08546695113182068, -0.11977250128984451, 0.03824513778090477, 0.12445946037769318, -0.08318526297807693, 0.17846983671188354, -0.20790702104568481, -0.11226828396320343, 0.04439050331711769, -0.0008177252602763474, 0.10079948604106903, -0.20941157639026642, -0.05367864668369293, 0.01868799887597561, -0.19879794120788574, 0.1039763018488884, -0.07879771292209625, 0.08610179275274277, -0.015094593167304993, 0.05863004922866821, 0.002865842543542385, -0.09638675302267075, 0.10485311597585678, -0.0163038931787014, 0.03271135315299034, -0.032476458698511124, 0.026555394753813744, 0.026328830048441887, -0.08129409700632095, 0.08705757558345795, -0.035646941512823105, 0.08425428718328476, -0.07042881846427917, -0.007827936671674252, -0.09780098497867584, 0.03795607015490532, -0.044521354138851166, -0.01841883920133114, -0.11000855267047882, 0.05181330814957619, -0.012078003026545048, -0.03644152358174324, 0.07850531488656998, 0.04626961052417755, 0.07334157824516296, 0.14747683703899384, -0.05192997679114342, -0.09210719913244247, -0.06993280351161957, -0.012951300479471684, -0.010630782693624496, 0.05954656004905701, -0.15827077627182007, 0.043712686747312546, 0.07327432930469513, 0.017811408266425133, 0.07086441665887833, -0.00670740008354187, -0.12166193127632141, 0.0001820592297008261, 0.0484573133289814, -0.16202224791049957, -0.11536379903554916, 0.01999085023999214, -0.04161614552140236, -0.04615718126296997, 0.026548651978373528, 0.12576673924922943, -0.007774354424327612, -0.021777305752038956, -0.013110833242535591, 0.09747102111577988, 0.013260024599730968, 0.21484580636024475, 0.050696227699518204, 0.07128537446260452, -0.13175718486309052, 0.08107136189937592, 0.0773545578122139, -0.06263001263141632, 0.030007489025592804, 0.11144791543483734, -0.13466878235340118, -0.06305393576622009, 0.01989286206662655, -0.031073221936821938, -0.07784814387559891, 0.012524311430752277, -0.11539239436388016, -0.03730837255716324, 0.06213248148560524, 0.08425556868314743, 0.017399102449417114, 0.052630722522735596, 0.09923559427261353, 0.0013254339573904872, -0.05080118775367737, 0.049693763256073, 0.046267468482255936, 0.04266437515616417, -0.09634417295455933, 0.09012728184461594, -0.004267234355211258, 0.08602356165647507, -0.008259821683168411, 0.06455008685588837, -0.1345210075378418, -0.015798933804035187, -0.13249854743480682, 0.0060732197016477585, -0.03859938681125641, -0.017760921269655228, -0.023370856419205666, -0.04220357537269592, -0.07364487648010254, 0.033379871398210526, -0.06896509975194931, -0.06256455183029175, -0.047020796686410904, -0.007466315291821957, -0.12909282743930817, -0.005067525897175074, 0.07694672048091888, -0.06463082879781723, 0.11302319169044495, 0.07883470505475998, 0.030646968632936478, 0.030827077105641365, -0.07003810256719589, 0.0005729474942199886, -0.03760702535510063, 0.0477917306125164, 0.04510211944580078, -0.014299989677965641, 0.02804255299270153, -0.006830951198935509, 0.03241370990872383, 0.0063338205218315125, 0.049419380724430084, -0.11992432922124863, -0.052759770303964615, -0.05979601666331291, 0.0033262036740779877, -0.03818093612790108, 0.012415079399943352, 0.07944846153259277, 0.05271385237574577, 0.09385941922664642, -0.053359854966402054, 0.010163498111069202, -0.21470162272453308, -0.015684720128774643, 0.009073643945157528, -0.048723284155130386, -0.017766118049621582, 0.049573808908462524, 0.11256293952465057, -0.0294329896569252, 0.1386604756116867, 0.029574772343039513, -0.023953964933753014, 0.03662450984120369, 0.012155307456851006, 0.03533734381198883, 0.028888585045933723, 0.059271231293678284, 0.024946298450231552, 0.019903771579265594, 0.023763328790664673, 0.012679914012551308, 0.04626740142703056, 0.0600520595908165, 0.10694878548383713, 0.11977499723434448, 0.15467426180839539, 0.08992307633161545, 0.05354926735162735, -0.1643214225769043, -0.07271148264408112, 0.023060232400894165, -0.18472512066364288, 0.10327903181314468, -0.03754295036196709, 0.07302463054656982, 0.11120422184467316, -0.1365637332201004, 0.02942361868917942, -0.060693975538015366, -0.08106401562690735, -0.11766768246889114, -0.07531900703907013, -0.07759025692939758, -0.13369013369083405, 0.017119938507676125, -0.04789062589406967, 0.05972621217370033, 0.09250277280807495, 0.05700019747018814, -0.003426644951105118, 0.09932531416416168, 0.07109908759593964, -0.03786993399262428, 0.05934887006878853, 0.0565064400434494, -0.03790393844246864, 0.11956938356161118, -0.07211889326572418, 0.06485005468130112, -0.0018039783462882042, 0.10646461695432663, 0.016450515016913414, -0.04287562891840935, 0.118175208568573, 0.005369361490011215, -0.09227381646633148, -0.03193233534693718, -0.0031115449965000153, -0.020456036552786827, 0.08403573930263519, 0.03995326906442642, 0.06060263141989708, 0.014632970094680786, 0.18239018321037292, -0.02117457240819931, 0.024229977279901505, -0.14063924551010132, 0.09944065660238266, -0.053309399634599686, 0.018946899101138115, 0.017884422093629837, -0.10395421087741852, -0.020231151953339577, 0.17358942329883575, 0.14569272100925446, -0.06285636872053146, -0.040539514273405075, 0.0033664791844785213, -0.01698381081223488, -0.029102787375450134, 0.09110759943723679, 0.07068585604429245, 0.16642719507217407, -0.026012800633907318, -0.03267159312963486, -0.015093696303665638, -0.06282123178243637, -0.09358847886323929, 0.07317916303873062, -0.005398460663855076, 0.04963167384266853, -0.04629317671060562, 0.03400987759232521, -0.005078617483377457, -0.17134995758533478, 0.0022273699287325144, -0.10637053847312927, -0.11648929119110107, 0.010992459021508694, -0.004520716145634651, -0.007831824012100697, 0.07725109905004501, -0.01251812744885683, 0.033222075551748276, 0.02468113601207733, -0.03642919287085533, -0.14312398433685303, -0.13234375417232513, 0.09422384947538376, 0.06893356144428253, 0.25350576639175415, -0.020365510135889053, 0.04684649780392647, 0.1550900936126709, -0.008533555082976818, -0.10625056177377701, 0.06462197750806808, 0.05231251195073128, -0.14851787686347961, -0.010933483019471169, 0.06213158741593361, -0.007628923282027245, 0.05762145668268204, 0.04692940413951874, -0.03998645395040512, -0.017307454720139503, 0.02902400679886341, 0.015097782015800476, -0.12131704390048981, -0.01181893702596426, -0.07924219965934753, 0.14206863939762115, 0.164865180850029, -0.05862019211053848, 0.021843407303094864, -0.06718625873327255, 0.03355111554265022, -0.001337337540462613, 0.08674121648073196, 0.002132061868906021, -0.26697319746017456, 0.05938104912638664, -0.007998572662472725, 0.03831733390688896, -0.17423830926418304, -0.01004110835492611, -0.003022170625627041, -0.019566942006349564, -0.07184203714132309, 0.14447835087776184, 0.056312188506126404, 0.020911935716867447, -0.06863939017057419, -0.021335572004318237, -0.0742020234465599, 0.11285985261201859, -0.1407293826341629, -0.12538515031337738 ]
null
null
transformers
## MobileBERT fine-tuned on SQuAD v1 [MobileBERT](https://arxiv.org/abs/2004.02984) is a thin version of BERT_LARGE, while equipped with bottleneck structures and a carefully designed balance between self-attentions and feed-forward networks. This model was fine-tuned from the HuggingFace checkpoint `google/mobilebert-uncased` on [SQuAD1.1](https://rajpurkar.github.io/SQuAD-explorer). ## Details | Dataset | Split | # samples | | -------- | ----- | --------- | | SQuAD1.1 | train | 90.6K | | SQuAD1.1 | eval | 11.1k | ### Fine-tuning - Python: `3.7.5` - Machine specs: `CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz` `Memory: 32 GiB` `GPUs: 2 GeForce GTX 1070, each with 8GiB memory` `GPU driver: 418.87.01, CUDA: 10.1` - script: ```shell # after install https://github.com/huggingface/transformers cd examples/question-answering mkdir -p data wget -O data/train-v1.1.json https://rajpurkar.github.io/SQuAD-explorer/dataset/train-v1.1.json wget -O data/dev-v1.1.json https://rajpurkar.github.io/SQuAD-explorer/dataset/dev-v1.1.json export SQUAD_DIR=`pwd`/data python run_squad.py \ --model_type mobilebert \ --model_name_or_path google/mobilebert-uncased \ --do_train \ --do_eval \ --do_lower_case \ --train_file $SQUAD_DIR/train-v1.1.json \ --predict_file $SQUAD_DIR/dev-v1.1.json \ --per_gpu_train_batch_size 16 \ --per_gpu_eval_batch_size 16 \ --learning_rate 4e-5 \ --num_train_epochs 5.0 \ --max_seq_length 320 \ --doc_stride 128 \ --warmup_steps 1400 \ --output_dir $SQUAD_DIR/mobilebert-uncased-warmup-squad_v1 2>&1 | tee train-mobilebert-warmup-squad_v1.log ``` It took about 3 hours to finish. ### Results **Model size**: `95M` | Metric | # Value | # Original ([Table 5](https://arxiv.org/pdf/2004.02984.pdf))| | ------ | --------- | --------- | | **EM** | **82.6** | **82.9** | | **F1** | **90.0** | **90.0** | Note that the above results didn't involve any hyperparameter search. ## Example Usage ```python from transformers import pipeline qa_pipeline = pipeline( "question-answering", model="csarron/mobilebert-uncased-squad-v1", tokenizer="csarron/mobilebert-uncased-squad-v1" ) predictions = qa_pipeline({ 'context': "The game was played on February 7, 2016 at Levi's Stadium in the San Francisco Bay Area at Santa Clara, California.", 'question': "What day was the game played on?" }) print(predictions) # output: # {'score': 0.7754058241844177, 'start': 23, 'end': 39, 'answer': 'February 7, 2016'} ``` > Created by [Qingqing Cao](https://awk.ai/) | [GitHub](https://github.com/csarron) | [Twitter](https://twitter.com/sysnlp) > Made with ❤️ in New York.
{"language": "en", "license": "mit", "tags": ["question-answering", "mobilebert"], "datasets": ["squad"], "metrics": ["squad"], "widget": [{"text": "Which name is also used to describe the Amazon rainforest in English?", "context": "The Amazon rainforest (Portuguese: Floresta Amaz\u00f4nica or Amaz\u00f4nia; Spanish: Selva Amaz\u00f3nica, Amazon\u00eda or usually Amazonia; French: For\u00eat amazonienne; Dutch: Amazoneregenwoud), also known in English as Amazonia or the Amazon Jungle, is a moist broadleaf forest that covers most of the Amazon basin of South America. This basin encompasses 7,000,000 square kilometres (2,700,000 sq mi), of which 5,500,000 square kilometres (2,100,000 sq mi) are covered by the rainforest. This region includes territory belonging to nine nations. The majority of the forest is contained within Brazil, with 60% of the rainforest, followed by Peru with 13%, Colombia with 10%, and with minor amounts in Venezuela, Ecuador, Bolivia, Guyana, Suriname and French Guiana. States or departments in four nations contain \"Amazonas\" in their names. The Amazon represents over half of the planet's remaining rainforests, and comprises the largest and most biodiverse tract of tropical rainforest in the world, with an estimated 390 billion individual trees divided into 16,000 species."}, {"text": "How many square kilometers of rainforest is covered in the basin?", "context": "The Amazon rainforest (Portuguese: Floresta Amaz\u00f4nica or Amaz\u00f4nia; Spanish: Selva Amaz\u00f3nica, Amazon\u00eda or usually Amazonia; French: For\u00eat amazonienne; Dutch: Amazoneregenwoud), also known in English as Amazonia or the Amazon Jungle, is a moist broadleaf forest that covers most of the Amazon basin of South America. This basin encompasses 7,000,000 square kilometres (2,700,000 sq mi), of which 5,500,000 square kilometres (2,100,000 sq mi) are covered by the rainforest. This region includes territory belonging to nine nations. The majority of the forest is contained within Brazil, with 60% of the rainforest, followed by Peru with 13%, Colombia with 10%, and with minor amounts in Venezuela, Ecuador, Bolivia, Guyana, Suriname and French Guiana. States or departments in four nations contain \"Amazonas\" in their names. The Amazon represents over half of the planet's remaining rainforests, and comprises the largest and most biodiverse tract of tropical rainforest in the world, with an estimated 390 billion individual trees divided into 16,000 species."}]}
question-answering
csarron/mobilebert-uncased-squad-v1
[ "transformers", "pytorch", "safetensors", "mobilebert", "question-answering", "en", "dataset:squad", "arxiv:2004.02984", "license:mit", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2004.02984" ]
[ "en" ]
TAGS #transformers #pytorch #safetensors #mobilebert #question-answering #en #dataset-squad #arxiv-2004.02984 #license-mit #endpoints_compatible #region-us
MobileBERT fine-tuned on SQuAD v1 --------------------------------- MobileBERT is a thin version of BERT\_LARGE, while equipped with bottleneck structures and a carefully designed balance between self-attentions and feed-forward networks. This model was fine-tuned from the HuggingFace checkpoint 'google/mobilebert-uncased' on SQuAD1.1. Details ------- Dataset: SQuAD1.1, Split: train, # samples: 90.6K Dataset: SQuAD1.1, Split: eval, # samples: 11.1k ### Fine-tuning * Python: '3.7.5' * Machine specs: 'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz' 'Memory: 32 GiB' 'GPUs: 2 GeForce GTX 1070, each with 8GiB memory' 'GPU driver: 418.87.01, CUDA: 10.1' * script: It took about 3 hours to finish. ### Results Model size: '95M' Metric: EM, # Value: 82.6, # Original (Table 5): 82.9 Metric: F1, # Value: 90.0, # Original (Table 5): 90.0 Note that the above results didn't involve any hyperparameter search. Example Usage ------------- > > Created by Qingqing Cao | GitHub | Twitter > > > > > Made with ️ in New York. > > >
[ "# samples: 90.6K\nDataset: SQuAD1.1, Split: eval, # samples: 11.1k", "### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 3 hours to finish.", "### Results\n\n\nModel size: '95M'\n\n\nMetric: EM, # Value: 82.6, # Original (Table 5): 82.9\nMetric: F1, # Value: 90.0, # Original (Table 5): 90.0\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>" ]
[ "TAGS\n#transformers #pytorch #safetensors #mobilebert #question-answering #en #dataset-squad #arxiv-2004.02984 #license-mit #endpoints_compatible #region-us \n", "# samples: 90.6K\nDataset: SQuAD1.1, Split: eval, # samples: 11.1k", "### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 3 hours to finish.", "### Results\n\n\nModel size: '95M'\n\n\nMetric: EM, # Value: 82.6, # Original (Table 5): 82.9\nMetric: F1, # Value: 90.0, # Original (Table 5): 90.0\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>" ]
[ 56, 27, 95, 107 ]
[ "passage: TAGS\n#transformers #pytorch #safetensors #mobilebert #question-answering #en #dataset-squad #arxiv-2004.02984 #license-mit #endpoints_compatible #region-us \n# samples: 90.6K\nDataset: SQuAD1.1, Split: eval, # samples: 11.1k### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 3 hours to finish.### Results\n\n\nModel size: '95M'\n\n\nMetric: EM, # Value: 82.6, # Original (Table 5): 82.9\nMetric: F1, # Value: 90.0, # Original (Table 5): 90.0\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>" ]
[ -0.1281723529100418, 0.08570630848407745, -0.0007291804067790508, 0.058129169046878815, 0.13153044879436493, 0.02889830246567726, 0.055642496794462204, 0.1125917062163353, 0.028485996648669243, 0.15931548178195953, 0.09990927577018738, 0.07868606597185135, 0.08968527615070343, 0.13260599970817566, -0.023596402257680893, -0.07598183304071426, 0.03766999766230583, 0.00540190702304244, 0.02326218970119953, 0.13577726483345032, 0.06668034940958023, -0.08778902143239975, 0.12380661070346832, 0.006110794376581907, -0.10720532387495041, -0.01668737269937992, 0.022113298997282982, -0.03610038012266159, 0.10813000798225403, 0.0002679723547771573, 0.035003796219825745, 0.053422000259160995, 0.03375904634594917, -0.10011233389377594, 0.02191036008298397, 0.1293012499809265, -0.023997837677598, 0.03726348653435707, 0.0698721781373024, 0.002741941250860691, 0.0740412250161171, -0.022071989253163338, 0.016623208299279213, 0.0493050143122673, -0.11501078307628632, -0.21026688814163208, -0.1677536815404892, 0.03447209671139717, 0.09948986023664474, 0.07222483307123184, -0.013021879829466343, 0.16908496618270874, -0.02829725854098797, 0.07727405428886414, 0.2995966076850891, -0.25542306900024414, -0.05047981068491936, 0.07663514465093613, 0.0649174377322197, 0.0335833840072155, -0.008733865804970264, -0.00558442110195756, 0.022315863519906998, 0.026988821104168892, 0.013991720043122768, -0.0003156281018164009, -0.028195297345519066, 0.0020319544710218906, -0.1470910906791687, -0.03017105907201767, 0.07265578955411911, 0.029706785455346107, -0.03695593774318695, -0.05276186391711235, -0.11573825776576996, -0.17500729858875275, -0.0404442623257637, 0.012291195802390575, -0.015016064047813416, -0.005138043779879808, -0.02790842577815056, 0.013943352736532688, -0.0468619242310524, -0.0932098925113678, -0.05141409859061241, 0.03996976464986801, 0.06724973767995834, 0.06818445026874542, 0.01520901545882225, 0.07512777298688889, -0.1344144344329834, -0.0852445662021637, -0.03591204062104225, -0.004272595047950745, -0.054465122520923615, 0.023538779467344284, 0.031145447865128517, 0.023193122819066048, 0.04457341879606247, 0.1584451049566269, -0.047634486109018326, 0.04557894542813301, 0.0634578987956047, -0.03687847778201103, -0.07176114618778229, 0.13696007430553436, -0.13393165171146393, -0.10100404918193817, 0.039891745895147324, 0.07395326346158981, 0.021401159465312958, -0.024462493136525154, -0.009353662841022015, 0.022815696895122528, 0.09428108483552933, 0.03622260317206383, 0.09403444081544876, 0.06257787346839905, -0.036091167479753494, -0.0052993036806583405, 0.2115626484155655, -0.061565618962049484, 0.005831829737871885, 0.047650039196014404, -0.06899625062942505, -0.0157284215092659, 0.010910937562584877, -0.034126799553632736, -0.09318246692419052, -0.017250049859285355, -0.10285595059394836, -0.04104690998792648, -0.04929700866341591, -0.08563736826181412, 0.03660774230957031, -0.09507399052381516, -0.06429053097963333, -0.11651943624019623, -0.1718679964542389, -0.0368206724524498, 0.040904451161623, -0.06794393062591553, 0.036658186465501785, 0.04344703257083893, -0.06590315699577332, 0.04145476222038269, -0.02024911344051361, 0.0347946360707283, -0.07209000736474991, 0.10292915999889374, 0.03053976595401764, 0.016056643798947334, -0.004664186388254166, 0.03721780329942703, -0.11739791929721832, 0.0700545459985733, -0.18013177812099457, 0.008250212296843529, -0.07943408936262131, 0.05128416046500206, -0.12314870208501816, -0.07076875120401382, -0.001000970951281488, -0.03582946956157684, 0.09961288422346115, 0.12101729959249496, -0.08869898319244385, -0.03366914764046669, 0.10395041108131409, -0.05375559255480766, -0.10695614665746689, 0.16707858443260193, 0.03487902879714966, -0.06567992269992828, 0.09279084950685501, 0.11923854798078537, 0.01427658274769783, -0.17230567336082458, -0.09733448177576065, -0.049640778452157974, 0.005916693713515997, -0.05898142606019974, 0.07115781307220459, -0.045078303664922714, 0.045956891030073166, 0.028009792789816856, -0.021049970760941505, -0.016271909698843956, -0.08914881944656372, -0.03156176581978798, -0.06911934167146683, -0.07593926042318344, -0.05653279647231102, 0.028567517176270485, -0.021402940154075623, -0.07829848676919937, -0.1466924101114273, -0.13491366803646088, 0.09685006737709045, -0.04715485870838165, -0.016939274966716766, -0.10246147960424423, 0.19299627840518951, -0.00618429621681571, 0.006170296110212803, -0.10496208816766739, -0.11436451971530914, 0.08220595866441727, -0.12115510553121567, -0.04876110702753067, -0.06052029877901077, 0.01727294735610485, 0.057163652032613754, -0.004294142127037048, -0.007466957438737154, 0.007648894097656012, -0.0301689263433218, -0.06709866225719452, -0.09125685691833496, -0.007296395022422075, -0.0030296319164335728, 0.08431290090084076, -0.0856182873249054, 0.029514366760849953, 0.050681665539741516, 0.10191141068935394, -0.04173027351498604, -0.08968625962734222, -0.05828815698623657, -0.01856853999197483, -0.03836235776543617, -0.09503632038831711, 0.03256338834762573, 0.007338738068938255, 0.03438485413789749, 0.0745897889137268, -0.2134910672903061, 0.02690223604440689, 0.11342931538820267, 0.06320236623287201, -0.04176324978470802, 0.03831552341580391, -0.022973692044615746, -0.050698172301054, -0.030209219083189964, -0.0013064505765214562, 0.02755667082965374, 0.035415809601545334, 0.09186898916959763, -0.10780512541532516, -0.017602333799004555, 0.05697731673717499, 0.027748124673962593, -0.022255197167396545, 0.12205840647220612, 0.15334376692771912, -0.18240001797676086, 0.09751949459314346, 0.013116379268467426, 0.001293693669140339, 0.09467928111553192, 0.013034231029450893, -0.03866375610232353, -0.02766496129333973, 0.03197164833545685, -0.0030305134132504463, 0.16470380127429962, -0.044636406004428864, 0.027084486559033394, 0.030735762789845467, -0.01670413464307785, 0.047287192195653915, -0.14480258524417877, 0.014055841602385044, -0.050755392760038376, -0.07440098375082016, -0.053076814860105515, 0.03994777053594589, 0.010386054404079914, 0.07374117523431778, 0.01863345317542553, -0.03932831063866615, 0.0015412588836625218, 0.017800409346818924, -0.06252558529376984, 0.21975776553153992, -0.09289627522230148, -0.20707975327968597, -0.06715793162584305, -0.05012746527791023, -0.0439496673643589, -0.024305136874318123, 0.03428393229842186, -0.10480619221925735, -0.060947809368371964, -0.06252685189247131, 0.07711748033761978, 0.05261329561471939, 0.005283801816403866, 0.00417301757261157, -0.026435611769557, 0.03452940285205841, -0.12087853252887726, 0.0064681158401072025, -0.032104481011629105, -0.11673321574926376, 0.03435057774186134, -0.02594999223947525, 0.04384561628103256, 0.15907885134220123, 0.040934041142463684, -0.0058520641177892685, 0.017080971971154213, 0.2309456467628479, -0.05171231925487518, 0.0636562928557396, 0.19525891542434692, 0.1372450739145279, 0.08158185333013535, 0.13251665234565735, -0.00923473946750164, -0.08812344074249268, -0.009057295508682728, 0.0698154866695404, -0.0677889958024025, -0.22707077860832214, -0.056107353419065475, -0.018411843106150627, 0.09428385645151138, 0.058355145156383514, 0.05123889818787575, -0.08617939800024033, 0.03039383888244629, -0.0904838889837265, 0.030735397711396217, 0.00956042855978012, 0.035126883536577225, 0.030641967430710793, 0.06873127818107605, 0.09426084905862808, -0.01068099308758974, 0.04511015862226486, 0.0947004035115242, 0.010015550069510937, 0.14250726997852325, -0.0810835137963295, 0.16522741317749023, 0.008300893008708954, 0.1753038465976715, 0.013530018739402294, 0.021893661469221115, -0.01621248945593834, 0.015208140946924686, 0.00942090805619955, -0.038081757724285126, -0.050812017172575, -0.011833248659968376, 0.037510547786951065, -0.06472085416316986, -0.018620213493704796, 0.11899451166391373, 0.05507862940430641, 0.27987438440322876, 0.09625507891178131, -0.29798170924186707, -0.09379517287015915, -0.037213586270809174, -0.030533239245414734, -0.12500028312206268, 0.0019429032690823078, 0.02439350076019764, -0.0769483670592308, 0.03657359257340431, -0.06980489194393158, 0.07449367642402649, 0.006943781394511461, -0.006429658737033606, 0.12749989330768585, 0.10870843380689621, -0.006578861735761166, 0.020624175667762756, -0.1773792803287506, 0.17109785974025726, 0.03335776925086975, 0.06511598825454712, -0.04550167918205261, -0.0033497512340545654, -0.025622954592108727, 0.00887361727654934, 0.06360466033220291, 0.012896018102765083, 0.0033115334808826447, -0.140901580452919, -0.1992248296737671, 0.005682358518242836, 0.08233020454645157, 0.009668082930147648, 0.0808180421590805, -0.028004685416817665, -0.01823950931429863, -0.02792060375213623, -0.06394444406032562, -0.07124938815832138, -0.07706892490386963, 0.08425886183977127, -0.07859651744365692, -0.056871041655540466, -0.10622875392436981, -0.020317038521170616, -0.08686867356300354, 0.12209145724773407, -0.08225228637456894, -0.07095641642808914, -0.1013856828212738, 0.0200240109115839, 0.15064887702465057, -0.09309768676757812, 0.07915309816598892, -0.081821970641613, 0.029298007488250732, -0.00005200473970035091, -0.1097412034869194, 0.11141075193881989, -0.11384148895740509, -0.21799024939537048, -0.054595697671175, 0.12195868790149689, 0.003964398987591267, 0.007741871755570173, -0.00441429577767849, 0.02720395289361477, -0.07571899890899658, -0.07730959355831146, 0.016544535756111145, -0.0325239934027195, 0.08348660171031952, 0.10715679824352264, 0.041884277015924454, -0.05047295615077019, -0.042925167828798294, -0.043793682008981705, -0.04999789968132973, 0.28265380859375, -0.0481351763010025, 0.00926921796053648, 0.10793404281139374, 0.003547027939930558, -0.25197210907936096, 0.0009037457639351487, 0.014119905419647694, -0.022735752165317535, -0.06953179836273193, -0.11491650342941284, 0.052741799503564835, 0.11640908569097519, -0.08863838016986847, 0.16266939043998718, -0.22235481441020966, -0.12469402700662613, 0.05224720761179924, 0.01629801094532013, 0.12703457474708557, -0.19542984664440155, -0.043366797268390656, 0.004029137082397938, -0.16769176721572876, 0.10334132611751556, -0.06091926619410515, 0.09582805633544922, -0.03310491517186165, 0.01410752348601818, -0.0023097428493201733, -0.1168244257569313, 0.08990463614463806, -0.02854992263019085, 0.04087362438440323, -0.027594804763793945, 0.03812519088387489, 0.056707713752985, -0.06615842878818512, 0.08156998455524445, -0.013484043069183826, 0.09649258106946945, -0.037955593317747116, -0.010178606025874615, -0.10893367230892181, 0.050075385719537735, -0.03372151032090187, -0.007498697843402624, -0.11171559244394302, 0.04828931763768196, -0.010799835436046124, -0.03058622032403946, 0.07617335021495819, 0.05343976989388466, 0.0794675424695015, 0.11737023293972015, -0.027566751465201378, -0.06896141171455383, -0.0881236270070076, -0.03030143305659294, -0.007797908969223499, 0.03603983297944069, -0.09966365247964859, 0.033539071679115295, 0.0722324550151825, 0.016332218423485756, 0.04284796491265297, 0.0024117755237966776, -0.10813689976930618, -0.0038608673494309187, 0.060640327632427216, -0.1737251728773117, -0.12167509645223618, 0.020879363641142845, -0.023205138742923737, -0.04747404158115387, 0.050773587077856064, 0.10199900716543198, 0.00639977864921093, -0.028274646028876305, -0.004613880068063736, 0.10485051572322845, 0.0384281724691391, 0.21457092463970184, 0.061469461768865585, 0.06322800368070602, -0.141827791929245, 0.0894114151597023, 0.06735550612211227, -0.05811115726828575, 0.004670657217502594, 0.09585600346326828, -0.1471938043832779, -0.04481268674135208, 0.05564812943339348, -0.005988295190036297, -0.06881692260503769, -0.01187820266932249, -0.13564494252204895, -0.03324778005480766, 0.061516884714365005, 0.08317626267671585, 0.030020352452993393, 0.052598338574171066, 0.09150780737400055, -0.008144129998981953, -0.07207709550857544, 0.07137487083673477, 0.0580412894487381, 0.055953413248062134, -0.1128140389919281, 0.09627275913953781, 0.0236677099019289, 0.07107191532850266, -0.010865801945328712, 0.07522078603506088, -0.10224664956331253, -0.01184140332043171, -0.1198311373591423, 0.0034043455962091684, -0.042027588933706284, -0.010304766707122326, -0.03033776581287384, -0.025352830067276955, -0.07378429919481277, 0.033859193325042725, -0.07028207182884216, -0.05243883281946182, -0.03154654800891876, -0.016927173361182213, -0.1425936371088028, -0.00009854932432062924, 0.08165675401687622, -0.07249492406845093, 0.10469141602516174, 0.0783856138586998, 0.03225061297416687, 0.04668701812624931, -0.050853896886110306, -0.03174346685409546, -0.030845602974295616, 0.06681918352842331, 0.015550102107226849, -0.04061933234333992, 0.033559445291757584, 0.019686130806803703, 0.03179789334535599, 0.004230717197060585, 0.04821373149752617, -0.12702491879463196, -0.04484424740076065, -0.07396277785301208, -0.0032591246999800205, -0.01785336807370186, 0.013548623770475388, 0.07078985869884491, 0.03773192688822746, 0.1296970695257187, -0.056063808500766754, -0.003697392763569951, -0.2067391723394394, -0.012151308357715607, -0.007883132435381413, -0.08597986400127411, -0.04033525660634041, 0.05562194436788559, 0.1184370219707489, -0.0066884104162454605, 0.1383175402879715, 0.026844263076782227, -0.04270349442958832, 0.02684815227985382, 0.04740207642316818, 0.030365651473402977, 0.023847846314311028, 0.06245436146855354, 0.02349221706390381, 0.0297593604773283, 0.05056219920516014, 0.007757210172712803, 0.0591050460934639, 0.04905487224459648, 0.15195126831531525, 0.13920824229717255, 0.13268937170505524, 0.08602630347013474, 0.0540211945772171, -0.1908472180366516, -0.09493941068649292, 0.038596488535404205, -0.18510907888412476, 0.11085380613803864, -0.030573217198252678, 0.079927459359169, 0.10857335478067398, -0.11250822246074677, 0.01929665543138981, -0.04850829020142555, -0.07133303582668304, -0.1307152807712555, -0.025466151535511017, -0.08851084858179092, -0.13747437298297882, 0.02306281588971615, -0.0349297933280468, 0.052755262702703476, 0.11643386632204056, 0.049224864691495895, -0.009821922518312931, 0.1349688023328781, 0.0654420256614685, -0.045275963842868805, 0.03503873571753502, 0.04409567639231682, -0.04400062561035156, 0.1207452043890953, -0.07290308177471161, 0.048154424875974655, -0.004736609756946564, 0.12080845981836319, 0.022443482652306557, -0.07319579273462296, 0.1231035515666008, -0.0019160052761435509, -0.08430390805006027, -0.03154836967587471, 0.015265172347426414, -0.010493306443095207, 0.10382641851902008, 0.025247404351830482, 0.06459442526102066, 0.012375871650874615, 0.16448810696601868, -0.013946505263447762, -0.006885244511067867, -0.09986419230699539, 0.1234905868768692, -0.022958559915423393, 0.01940230093896389, 0.026480359956622124, -0.08674603700637817, -0.012049615383148193, 0.18189582228660583, 0.1332205981016159, -0.040074028074741364, -0.050708696246147156, 0.014535804279148579, -0.016546206548810005, -0.04235183075070381, 0.07896623015403748, 0.08299049735069275, 0.17539039254188538, -0.0300071332603693, -0.0511050783097744, -0.010319283232092857, -0.048886075615882874, -0.09443975239992142, 0.061108872294425964, -0.002485140459612012, 0.028373125940561295, -0.038449086248874664, 0.032238930463790894, -0.000900645274668932, -0.144352525472641, -0.011073320172727108, -0.10436812788248062, -0.1364927589893341, 0.007674311753362417, 0.04124966636300087, -0.016291802749037743, 0.06499557942152023, -0.009521308355033398, 0.018702669069170952, 0.023437708616256714, -0.032494254410266876, -0.15516406297683716, -0.1355285346508026, 0.08219451457262039, 0.049585483968257904, 0.25397977232933044, -0.024880874902009964, 0.05850524082779884, 0.16060583293437958, -0.03746908903121948, -0.10895667970180511, 0.07481741160154343, 0.03347165510058403, -0.14713945984840393, 0.023371858522295952, 0.05064348131418228, -0.006386316381394863, 0.04198093339800835, 0.02564511075615883, -0.011862942948937416, -0.007177518215030432, 0.047512806951999664, -0.008516724221408367, -0.10942783206701279, 0.004614336881786585, -0.08490052819252014, 0.13803359866142273, 0.1545625925064087, -0.04673907905817032, 0.03756178542971611, -0.06723552197217941, 0.0291450172662735, -0.001575352973304689, 0.10009229928255081, 0.02550787851214409, -0.29713162779808044, 0.05295230448246002, -0.04489642754197121, 0.042872972786426544, -0.15767212212085724, -0.0098408292979002, 0.0008992935181595385, -0.03541203588247299, -0.11183057725429535, 0.1330677717924118, 0.04871894046664238, 0.018283400684595108, -0.06487303972244263, 0.014872150495648384, -0.08986767381429672, 0.10973016172647476, -0.15408168733119965, -0.12841732800006866 ]
null
null
transformers
## MobileBERT fine-tuned on SQuAD v2 [MobileBERT](https://arxiv.org/abs/2004.02984) is a thin version of BERT_LARGE, while equipped with bottleneck structures and a carefully designed balance between self-attentions and feed-forward networks. This model was fine-tuned from the HuggingFace checkpoint `google/mobilebert-uncased` on [SQuAD2.0](https://rajpurkar.github.io/SQuAD-explorer). ## Details | Dataset | Split | # samples | | -------- | ----- | --------- | | SQuAD2.0 | train | 130k | | SQuAD2.0 | eval | 12.3k | ### Fine-tuning - Python: `3.7.5` - Machine specs: `CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz` `Memory: 32 GiB` `GPUs: 2 GeForce GTX 1070, each with 8GiB memory` `GPU driver: 418.87.01, CUDA: 10.1` - script: ```shell # after install https://github.com/huggingface/transformers cd examples/question-answering mkdir -p data wget -O data/train-v2.0.json https://rajpurkar.github.io/SQuAD-explorer/dataset/train-v2.0.json wget -O data/dev-v2.0.json https://rajpurkar.github.io/SQuAD-explorer/dataset/dev-v2.0.json export SQUAD_DIR=`pwd`/data python run_squad.py \ --model_type mobilebert \ --model_name_or_path google/mobilebert-uncased \ --do_train \ --do_eval \ --do_lower_case \ --version_2_with_negative \ --train_file $SQUAD_DIR/train-v2.0.json \ --predict_file $SQUAD_DIR/dev-v2.0.json \ --per_gpu_train_batch_size 16 \ --per_gpu_eval_batch_size 16 \ --learning_rate 4e-5 \ --num_train_epochs 5.0 \ --max_seq_length 320 \ --doc_stride 128 \ --warmup_steps 1400 \ --save_steps 2000 \ --output_dir $SQUAD_DIR/mobilebert-uncased-warmup-squad_v2 2>&1 | tee train-mobilebert-warmup-squad_v2.log ``` It took about 3.5 hours to finish. ### Results **Model size**: `95M` | Metric | # Value | # Original ([Table 5](https://arxiv.org/pdf/2004.02984.pdf))| | ------ | --------- | --------- | | **EM** | **75.2** | **76.2** | | **F1** | **78.8** | **79.2** | Note that the above results didn't involve any hyperparameter search. ## Example Usage ```python from transformers import pipeline qa_pipeline = pipeline( "question-answering", model="csarron/mobilebert-uncased-squad-v2", tokenizer="csarron/mobilebert-uncased-squad-v2" ) predictions = qa_pipeline({ 'context': "The game was played on February 7, 2016 at Levi's Stadium in the San Francisco Bay Area at Santa Clara, California.", 'question': "What day was the game played on?" }) print(predictions) # output: # {'score': 0.71434086561203, 'start': 23, 'end': 39, 'answer': 'February 7, 2016'} ``` > Created by [Qingqing Cao](https://awk.ai/) | [GitHub](https://github.com/csarron) | [Twitter](https://twitter.com/sysnlp) > Made with ❤️ in New York.
{"language": "en", "license": "mit", "tags": ["question-answering", "mobilebert"], "datasets": ["squad_v2"], "metrics": ["squad_v2"], "widget": [{"text": "Which name is also used to describe the Amazon rainforest in English?", "context": "The Amazon rainforest (Portuguese: Floresta Amaz\u00f4nica or Amaz\u00f4nia; Spanish: Selva Amaz\u00f3nica, Amazon\u00eda or usually Amazonia; French: For\u00eat amazonienne; Dutch: Amazoneregenwoud), also known in English as Amazonia or the Amazon Jungle, is a moist broadleaf forest that covers most of the Amazon basin of South America. This basin encompasses 7,000,000 square kilometres (2,700,000 sq mi), of which 5,500,000 square kilometres (2,100,000 sq mi) are covered by the rainforest. This region includes territory belonging to nine nations. The majority of the forest is contained within Brazil, with 60% of the rainforest, followed by Peru with 13%, Colombia with 10%, and with minor amounts in Venezuela, Ecuador, Bolivia, Guyana, Suriname and French Guiana. States or departments in four nations contain \"Amazonas\" in their names. The Amazon represents over half of the planet's remaining rainforests, and comprises the largest and most biodiverse tract of tropical rainforest in the world, with an estimated 390 billion individual trees divided into 16,000 species."}, {"text": "How many square kilometers of rainforest is covered in the basin?", "context": "The Amazon rainforest (Portuguese: Floresta Amaz\u00f4nica or Amaz\u00f4nia; Spanish: Selva Amaz\u00f3nica, Amazon\u00eda or usually Amazonia; French: For\u00eat amazonienne; Dutch: Amazoneregenwoud), also known in English as Amazonia or the Amazon Jungle, is a moist broadleaf forest that covers most of the Amazon basin of South America. This basin encompasses 7,000,000 square kilometres (2,700,000 sq mi), of which 5,500,000 square kilometres (2,100,000 sq mi) are covered by the rainforest. This region includes territory belonging to nine nations. The majority of the forest is contained within Brazil, with 60% of the rainforest, followed by Peru with 13%, Colombia with 10%, and with minor amounts in Venezuela, Ecuador, Bolivia, Guyana, Suriname and French Guiana. States or departments in four nations contain \"Amazonas\" in their names. The Amazon represents over half of the planet's remaining rainforests, and comprises the largest and most biodiverse tract of tropical rainforest in the world, with an estimated 390 billion individual trees divided into 16,000 species."}]}
question-answering
csarron/mobilebert-uncased-squad-v2
[ "transformers", "pytorch", "onnx", "safetensors", "mobilebert", "question-answering", "en", "dataset:squad_v2", "arxiv:2004.02984", "license:mit", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2004.02984" ]
[ "en" ]
TAGS #transformers #pytorch #onnx #safetensors #mobilebert #question-answering #en #dataset-squad_v2 #arxiv-2004.02984 #license-mit #endpoints_compatible #region-us
MobileBERT fine-tuned on SQuAD v2 --------------------------------- MobileBERT is a thin version of BERT\_LARGE, while equipped with bottleneck structures and a carefully designed balance between self-attentions and feed-forward networks. This model was fine-tuned from the HuggingFace checkpoint 'google/mobilebert-uncased' on SQuAD2.0. Details ------- Dataset: SQuAD2.0, Split: train, # samples: 130k Dataset: SQuAD2.0, Split: eval, # samples: 12.3k ### Fine-tuning * Python: '3.7.5' * Machine specs: 'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz' 'Memory: 32 GiB' 'GPUs: 2 GeForce GTX 1070, each with 8GiB memory' 'GPU driver: 418.87.01, CUDA: 10.1' * script: It took about 3.5 hours to finish. ### Results Model size: '95M' Metric: EM, # Value: 75.2, # Original (Table 5): 76.2 Metric: F1, # Value: 78.8, # Original (Table 5): 79.2 Note that the above results didn't involve any hyperparameter search. Example Usage ------------- > > Created by Qingqing Cao | GitHub | Twitter > > > > > Made with ️ in New York. > > >
[ "# samples: 130k\nDataset: SQuAD2.0, Split: eval, # samples: 12.3k", "### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 3.5 hours to finish.", "### Results\n\n\nModel size: '95M'\n\n\nMetric: EM, # Value: 75.2, # Original (Table 5): 76.2\nMetric: F1, # Value: 78.8, # Original (Table 5): 79.2\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>" ]
[ "TAGS\n#transformers #pytorch #onnx #safetensors #mobilebert #question-answering #en #dataset-squad_v2 #arxiv-2004.02984 #license-mit #endpoints_compatible #region-us \n", "# samples: 130k\nDataset: SQuAD2.0, Split: eval, # samples: 12.3k", "### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 3.5 hours to finish.", "### Results\n\n\nModel size: '95M'\n\n\nMetric: EM, # Value: 75.2, # Original (Table 5): 76.2\nMetric: F1, # Value: 78.8, # Original (Table 5): 79.2\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>" ]
[ 63, 26, 95, 106 ]
[ "passage: TAGS\n#transformers #pytorch #onnx #safetensors #mobilebert #question-answering #en #dataset-squad_v2 #arxiv-2004.02984 #license-mit #endpoints_compatible #region-us \n# samples: 130k\nDataset: SQuAD2.0, Split: eval, # samples: 12.3k### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 3.5 hours to finish.### Results\n\n\nModel size: '95M'\n\n\nMetric: EM, # Value: 75.2, # Original (Table 5): 76.2\nMetric: F1, # Value: 78.8, # Original (Table 5): 79.2\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>" ]
[ -0.1348331719636917, 0.06482125073671341, -0.0014147701440379024, 0.06846626847982407, 0.12749825417995453, 0.030476873740553856, 0.07556270807981491, 0.12080681324005127, 0.036692678928375244, 0.15933430194854736, 0.11493188887834549, 0.05672044679522514, 0.10260391235351562, 0.12507805228233337, -0.029870033264160156, -0.06451567262411118, 0.021088002249598503, -0.01894855685532093, 0.009883414953947067, 0.12182489037513733, 0.07178754359483719, -0.05509971082210541, 0.12925876677036285, -0.008578754030168056, -0.11878717690706253, -0.02193247154355049, 0.039938703179359436, -0.037402622401714325, 0.09524746984243393, 0.024694446474313736, 0.01329679973423481, 0.05307583138346672, 0.031040508300065994, -0.10887689888477325, 0.025260739028453827, 0.13255377113819122, -0.04389951750636101, 0.02870817296206951, 0.05680600181221962, 0.01113729178905487, 0.07766890525817871, -0.01704060472548008, 0.01566094346344471, 0.06325140595436096, -0.11228108406066895, -0.1691054403781891, -0.1785532832145691, 0.0019119702046737075, 0.07045222818851471, 0.07205919176340103, -0.00417752331122756, 0.159361332654953, -0.012000841088593006, 0.09209232777357101, 0.2546135485172272, -0.2456066757440567, -0.05730119347572327, 0.07755997776985168, 0.06694239377975464, 0.055682044476270676, -0.0016493768198415637, -0.003205580171197653, 0.015256049111485481, 0.045784372836351395, 0.013953780755400658, -0.015520048327744007, 0.014174536801874638, -0.0012585826916620135, -0.13256551325321198, -0.04595695808529854, 0.07159682363271713, 0.0425277017056942, -0.040034741163253784, -0.05335162580013275, -0.12929567694664001, -0.19027842581272125, -0.02746642753481865, 0.010770496912300587, -0.03878052905201912, -0.0017314390279352665, -0.045077428221702576, -0.0067637222819030285, -0.045885514467954636, -0.09043674916028976, -0.036997199058532715, 0.02699427865445614, 0.05244852229952812, 0.08181840926408768, -0.007198757492005825, 0.07145199924707413, -0.14432387053966522, -0.09506437182426453, -0.025633305311203003, -0.009152333252131939, -0.08169374614953995, 0.02605416625738144, 0.04428299888968468, -0.005838743411004543, 0.04107629507780075, 0.15827420353889465, -0.0756428986787796, 0.0501379631459713, 0.061086505651474, -0.04298532009124756, -0.050424348562955856, 0.1408945620059967, -0.1301269382238388, -0.11197088658809662, 0.040574487298727036, 0.05276830121874809, -0.022335905581712723, -0.04079902917146683, -0.017672916874289513, 0.02624223381280899, 0.0794583410024643, 0.01769668608903885, 0.07901889085769653, 0.04687368869781494, -0.033657144755125046, -0.016714414581656456, 0.20920021831989288, -0.057883478701114655, 0.014124120585620403, 0.05103038251399994, -0.07178646326065063, -0.01674412377178669, 0.020453309640288353, -0.041247446089982986, -0.08401820063591003, -0.03955862671136856, -0.09400021284818649, -0.0515160895884037, -0.040960583835840225, -0.07933327555656433, 0.04499272629618645, -0.07496897876262665, -0.05790998414158821, -0.13436418771743774, -0.16269788146018982, -0.04720320925116539, 0.041711095720529556, -0.07363785058259964, 0.04282372444868088, 0.050076693296432495, -0.06806588917970657, 0.051332298666238785, -0.009975780732929707, 0.02994430810213089, -0.06768705695867538, 0.11284131556749344, 0.04012124985456467, 0.00847654789686203, -0.00893340166658163, 0.03750036656856537, -0.09667719155550003, 0.06026172637939453, -0.1816948801279068, 0.01857839524745941, -0.08067227900028229, 0.06629940122365952, -0.11262200027704239, -0.053286630660295486, -0.007275879383087158, -0.04713151976466179, 0.10793141275644302, 0.11355511099100113, -0.10911983996629715, -0.030858850106596947, 0.09205437451601028, -0.038860175758600235, -0.10065856575965881, 0.1669652909040451, 0.03425586223602295, -0.048420388251543045, 0.08131548762321472, 0.11273399740457535, 0.060543328523635864, -0.18551842868328094, -0.08912960439920425, -0.04549495130777359, -0.01306143682450056, -0.07019457221031189, 0.07148536294698715, -0.019573615863919258, 0.03382408991456032, 0.026583347469568253, 0.015403376892209053, -0.00665037939324975, -0.08712419867515564, -0.02743547223508358, -0.07597742229700089, -0.06695222854614258, -0.09308359771966934, 0.025030553340911865, -0.025997286662459373, -0.06496132910251617, -0.13205105066299438, -0.1531618982553482, 0.0884876623749733, -0.03446047753095627, -0.024294409900903702, -0.09503377974033356, 0.17841362953186035, 0.007844019681215286, 0.02558610588312149, -0.09224912524223328, -0.11786813288927078, 0.08336535096168518, -0.11645854264497757, -0.03619533032178879, -0.0810890793800354, 0.013332301750779152, 0.05141584575176239, 0.0060492572374641895, 0.010488479398190975, -0.00045097761903889477, -0.038483861833810806, -0.07371237874031067, -0.08706994354724884, -0.0025193863548338413, -0.014986516907811165, 0.05740660801529884, -0.07987397164106369, 0.03559320420026779, 0.08772893249988556, 0.10393025726079941, -0.033877357840538025, -0.09221091121435165, -0.06881581246852875, -0.034234486520290375, -0.031748272478580475, -0.09717584401369095, 0.008375917561352253, 0.01791330799460411, 0.03939003124833107, 0.09181573987007141, -0.20963048934936523, 0.03842736780643463, 0.09461328387260437, 0.09233810752630234, -0.04869237169623375, 0.024879785254597664, -0.021061638370156288, -0.04723455384373665, -0.04114806279540062, -0.010948087088763714, 0.03749584034085274, 0.03328258916735649, 0.09700644761323929, -0.10765132308006287, -0.015010151080787182, 0.08089877665042877, 0.028543680906295776, -0.03131740540266037, 0.1362035572528839, 0.13764113187789917, -0.17782799899578094, 0.09005396068096161, 0.01817108504474163, 0.039103999733924866, 0.11150537431240082, 0.016077082604169846, -0.03561709076166153, -0.05347372964024544, 0.03240615874528885, -0.0004040376516059041, 0.17523877322673798, -0.03929611295461655, 0.03522917628288269, 0.02535278908908367, -0.0007227034657262266, 0.05142279341816902, -0.12031596899032593, 0.009594643488526344, -0.04006841033697128, -0.07719156891107559, -0.036157846450805664, 0.027170758694410324, 0.014480270445346832, 0.06661684066057205, 0.014588924124836922, -0.053454943001270294, 0.012162400409579277, 0.02064933069050312, -0.06787529587745667, 0.21314795315265656, -0.11286848038434982, -0.1915399432182312, -0.07094243168830872, -0.05911662429571152, -0.041433777660131454, -0.029216717928647995, 0.032878417521715164, -0.10690892487764359, -0.04690791666507721, -0.05461728200316429, 0.08041893690824509, 0.04633680731058121, 0.02125677280128002, 0.002876556944102049, -0.013798462226986885, 0.02894367277622223, -0.11821004748344421, 0.000486676930449903, -0.011212819255888462, -0.11579986661672592, 0.04483799263834953, -0.019701629877090454, 0.02930363640189171, 0.16491924226284027, 0.03916148096323013, 0.0036076658871024847, 0.02743173949420452, 0.239219531416893, -0.06583308428525925, 0.08032689988613129, 0.16792848706245422, 0.16190990805625916, 0.07864197343587875, 0.132242813706398, -0.001750884810462594, -0.10026516765356064, 0.011645302176475525, 0.06375906616449356, -0.08152353018522263, -0.23826782405376434, -0.05428393930196762, -0.006162494886666536, 0.1147189736366272, 0.04669208452105522, 0.06412835419178009, -0.06057362258434296, 0.0204917062073946, -0.06766138970851898, 0.05642900615930557, 0.030629243701696396, 0.024660352617502213, 0.007690927013754845, 0.06506771594285965, 0.07981478422880173, 0.0036779153160750866, 0.06977567076683044, 0.12005521357059479, 0.027814548462629318, 0.12022709846496582, -0.07321034371852875, 0.15584062039852142, 0.00012641778448596597, 0.16988520324230194, 0.008701481856405735, -0.0030554889235645533, -0.0017406785627827048, 0.018089009448885918, 0.017327893525362015, -0.03799505531787872, -0.03523990139365196, -0.005917562637478113, 0.04574687406420708, -0.08235104382038116, -0.043926119804382324, 0.12684576213359833, 0.057203829288482666, 0.29495662450790405, 0.08559902757406235, -0.27912768721580505, -0.10359197109937668, -0.039050642400979996, -0.04867742583155632, -0.13135594129562378, 0.010564575903117657, 0.07138548791408539, -0.08911094814538956, 0.00949674379080534, -0.06849472969770432, 0.08223137259483337, 0.007352953776717186, 0.002886299742385745, 0.1708400547504425, 0.0947020873427391, -0.014461755752563477, -0.00701071647927165, -0.1915665864944458, 0.16142480075359344, 0.04890453815460205, 0.06213477626442909, -0.042807117104530334, 0.005097328685224056, -0.028616761788725853, 0.01615128107368946, 0.05213978886604309, 0.017008261755108833, -0.015444912016391754, -0.1449717879295349, -0.19390732049942017, 0.002401783363893628, 0.08573709428310394, 0.006111904978752136, 0.0556677021086216, -0.010607185773551464, -0.016208332031965256, -0.03444278612732887, -0.07309979945421219, -0.06998365372419357, -0.06117863208055496, 0.08949273824691772, -0.06029640510678291, -0.08626436442136765, -0.09719620645046234, 0.005781920161098242, -0.07849756628274918, 0.12409244477748871, -0.09111499786376953, -0.05115395039319992, -0.1178598552942276, 0.021216977387666702, 0.13457083702087402, -0.07800544798374176, 0.05396252125501633, -0.08349763602018356, 0.008259491994976997, -0.019733255729079247, -0.10515950620174408, 0.10955928266048431, -0.10117296874523163, -0.20965059101581573, -0.06121477857232094, 0.11762060225009918, 0.02683568000793457, 0.0033774988260120153, -0.009682273492217064, 0.009889221750199795, -0.07606826722621918, -0.08948666602373123, 0.016378726810216904, -0.03482753410935402, 0.06635941565036774, 0.08401720225811005, 0.06251971423625946, -0.03782954439520836, -0.03013736568391323, -0.020831899717450142, -0.08156714588403702, 0.2861442565917969, -0.045257698744535446, 0.004515094216912985, 0.11972567439079285, 0.009480423294007778, -0.23643815517425537, 0.014848031103610992, 0.013037081807851791, -0.010838235728442669, -0.05731187388300896, -0.11363107711076736, 0.04471229016780853, 0.12186188995838165, -0.08061345666646957, 0.177566796541214, -0.24130786955356598, -0.11332201212644577, 0.025751888751983643, -0.009498639963567257, 0.11684077978134155, -0.18273937702178955, -0.05599500238895416, 0.010425654239952564, -0.17193861305713654, 0.08758438378572464, -0.10144858062267303, 0.09202231466770172, -0.011393621563911438, 0.04051385447382927, -0.0034467962104827166, -0.10318537056446075, 0.08977499604225159, -0.011953071691095829, 0.04142184555530548, -0.026326585561037064, 0.047118179500103, 0.049301162362098694, -0.08018404990434647, 0.0762656107544899, -0.008490418083965778, 0.1032635048031807, -0.0807107537984848, -0.0045647406950592995, -0.09575337916612625, 0.06319194287061691, -0.034005049616098404, 0.0036333089228719473, -0.11007029563188553, 0.05066094174981117, -0.017046483233571053, -0.0361638106405735, 0.06617727875709534, 0.04584668576717377, 0.04376015067100525, 0.16472257673740387, -0.047864221036434174, -0.08122091740369797, -0.0646846741437912, -0.025044599547982216, -0.002922005718573928, 0.043508049100637436, -0.11412105709314346, 0.03890938311815262, 0.06889235228300095, 0.0025144501123577356, 0.033639635890722275, -0.005509801208972931, -0.11023804545402527, -0.006482742261141539, 0.0479883998632431, -0.17342379689216614, -0.09715636819601059, 0.009011887945234776, -0.025444040074944496, -0.028121475130319595, 0.03979288786649704, 0.1278742104768753, 0.011957037262618542, -0.011896476149559021, -0.012359658256173134, 0.10895814746618271, 0.015564044937491417, 0.21250635385513306, 0.05453051999211311, 0.058698419481515884, -0.13887713849544525, 0.09257350116968155, 0.07872849702835083, -0.06607192009687424, 0.015396300703287125, 0.12810003757476807, -0.1401439756155014, -0.05803163722157478, 0.03277469426393509, -0.027342375367879868, -0.06558695435523987, -0.00017469850718043745, -0.11368627101182938, -0.01413644663989544, 0.06158934533596039, 0.12229891866445541, 0.012485681101679802, 0.055413179099559784, 0.0889621376991272, -0.00279720826074481, -0.07900621742010117, 0.06112065538764, 0.042381804436445236, 0.036552801728248596, -0.10627780109643936, 0.07612224668264389, 0.011221976019442081, 0.10241568833589554, -0.013680223375558853, 0.07153818011283875, -0.11272856593132019, -0.01952824927866459, -0.1011662483215332, -0.013392903842031956, -0.03870587423443794, -0.012457243166863918, -0.02432967536151409, -0.030372900888323784, -0.07373017817735672, 0.034489743411540985, -0.05381782352924347, -0.05298604816198349, -0.030663490295410156, -0.010751071386039257, -0.1287216991186142, -0.010115094482898712, 0.07274964451789856, -0.071330726146698, 0.11637204140424728, 0.08879709243774414, 0.025012675672769547, 0.020578008145093918, -0.044556792825460434, -0.009849697351455688, -0.03921699523925781, 0.05847616121172905, 0.034982483834028244, -0.029393797740340233, 0.04927792772650719, 0.007790009491145611, 0.04141681268811226, 0.014301362447440624, 0.05990172177553177, -0.1288014054298401, -0.02681107632815838, -0.07664389163255692, -0.0006017145933583379, -0.035163599997758865, 0.008696618489921093, 0.0648256316781044, 0.04504961520433426, 0.10366618633270264, -0.05499552935361862, 0.01007228996604681, -0.20486584305763245, -0.015757201239466667, 0.007604541257023811, -0.07557160407304764, -0.01936223916709423, 0.06769633293151855, 0.11954747885465622, -0.01569502428174019, 0.14435000717639923, 0.03363832086324692, -0.03140869736671448, 0.038460295647382736, 0.018929725512862206, 0.012130686081945896, 0.022173769772052765, 0.016558963805437088, 0.024685552343726158, 0.037998802959918976, 0.04189207777380943, 0.016971372067928314, 0.05063599720597267, 0.031988419592380524, 0.13192886114120483, 0.12721692025661469, 0.13379476964473724, 0.09398150444030762, 0.05600174143910408, -0.17644278705120087, -0.08785908669233322, 0.03692740947008133, -0.21091385185718536, 0.10223986208438873, -0.05952521413564682, 0.04713758826255798, 0.10146670788526535, -0.10837873071432114, 0.03011152893304825, -0.033397335559129715, -0.08545713871717453, -0.12700052559375763, -0.0363888218998909, -0.07634913921356201, -0.14513979852199554, 0.017944911494851112, -0.029961377382278442, 0.06056534871459007, 0.0862906277179718, 0.05556677281856537, -0.010644372552633286, 0.12834478914737701, 0.06456902623176575, -0.025848396122455597, 0.05730942636728287, 0.05232679843902588, -0.028290536254644394, 0.10821722447872162, -0.07394231855869293, 0.04306493327021599, 0.00023573647195007652, 0.11424470692873001, 0.029095470905303955, -0.050496943295001984, 0.10935445129871368, -0.004281180910766125, -0.08561962842941284, -0.04095667600631714, 0.002849854528903961, -0.040309786796569824, 0.10641604661941528, 0.042498864233493805, 0.04758431017398834, 0.01770045794546604, 0.17011375725269318, -0.02489110454916954, 0.020001256838440895, -0.13337738811969757, 0.10221792012453079, -0.028469432145357132, 0.029426055029034615, 0.00783777516335249, -0.09318839013576508, -0.035255637019872665, 0.1573617160320282, 0.12456817924976349, -0.05204317718744278, -0.04881466552615166, 0.024862607941031456, -0.017035435885190964, -0.04018436744809151, 0.08199281245470047, 0.06230364739894867, 0.18624477088451385, -0.010487762279808521, -0.0632333904504776, -0.005669459700584412, -0.06465958058834076, -0.08288388699293137, 0.04772495850920677, -0.013835538178682327, 0.04348573833703995, -0.022935904562473297, 0.04936932027339935, 0.003533120732754469, -0.14932803809642792, 0.015919514000415802, -0.08627418428659439, -0.1107056513428688, 0.024906247854232788, 0.0002245309588033706, -0.03202281519770622, 0.06339973956346512, -0.007371486164629459, 0.026014108210802078, 0.03481810539960861, -0.04214536398649216, -0.14475028216838837, -0.13627216219902039, 0.0790550708770752, 0.021123366430401802, 0.2686983346939087, -0.03309524059295654, 0.062195658683776855, 0.163852721452713, -0.026210501790046692, -0.10252328217029572, 0.07071689516305923, 0.037577830255031586, -0.15975546836853027, -0.015147584490478039, 0.06202227622270584, -0.00675982004031539, 0.06121225282549858, 0.028044072911143303, -0.016561077907681465, -0.016836773604154587, 0.024781635031104088, 0.027185121551156044, -0.11046166718006134, -0.000894260301720351, -0.06741277873516083, 0.13456477224826813, 0.14954042434692383, -0.045499350875616074, 0.04774820804595947, -0.07158298790454865, 0.03987516835331917, -0.01164600346237421, 0.11438732594251633, 0.016446711495518684, -0.3007405698299408, 0.05721544101834297, -0.03360287472605705, 0.05472428724169731, -0.15194830298423767, -0.02172429859638214, 0.014130598865449429, -0.030473848804831505, -0.09200019389390945, 0.14946943521499634, 0.04485694319009781, 0.017781635746359825, -0.07117582112550735, 0.013169560581445694, -0.09068875759840012, 0.1158563569188118, -0.16052468121051788, -0.12758013606071472 ]
null
null
transformers
## RoBERTa-base fine-tuned on SQuAD v1 This model was fine-tuned from the HuggingFace [RoBERTa](https://arxiv.org/abs/1907.11692) base checkpoint on [SQuAD1.1](https://rajpurkar.github.io/SQuAD-explorer). This model is case-sensitive: it makes a difference between english and English. ## Details | Dataset | Split | # samples | | -------- | ----- | --------- | | SQuAD1.1 | train | 96.8K | | SQuAD1.1 | eval | 11.8k | ### Fine-tuning - Python: `3.7.5` - Machine specs: `CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz` `Memory: 32 GiB` `GPUs: 2 GeForce GTX 1070, each with 8GiB memory` `GPU driver: 418.87.01, CUDA: 10.1` - script: ```shell # after install https://github.com/huggingface/transformers cd examples/question-answering mkdir -p data wget -O data/train-v1.1.json https://rajpurkar.github.io/SQuAD-explorer/dataset/train-v1.1.json wget -O data/dev-v1.1.json https://rajpurkar.github.io/SQuAD-explorer/dataset/dev-v1.1.json python run_energy_squad.py \ --model_type roberta \ --model_name_or_path roberta-base \ --do_train \ --do_eval \ --train_file train-v1.1.json \ --predict_file dev-v1.1.json \ --per_gpu_train_batch_size 12 \ --per_gpu_eval_batch_size 16 \ --learning_rate 3e-5 \ --num_train_epochs 2.0 \ --max_seq_length 320 \ --doc_stride 128 \ --data_dir data \ --output_dir data/roberta-base-squad-v1 2>&1 | tee train-roberta-base-squad-v1.log ``` It took about 2 hours to finish. ### Results **Model size**: `477M` | Metric | # Value | | ------ | --------- | | **EM** | **83.0** | | **F1** | **90.4** | Note that the above results didn't involve any hyperparameter search. ## Example Usage ```python from transformers import pipeline qa_pipeline = pipeline( "question-answering", model="csarron/roberta-base-squad-v1", tokenizer="csarron/roberta-base-squad-v1" ) predictions = qa_pipeline({ 'context': "The game was played on February 7, 2016 at Levi's Stadium in the San Francisco Bay Area at Santa Clara, California.", 'question': "What day was the game played on?" }) print(predictions) # output: # {'score': 0.8625259399414062, 'start': 23, 'end': 39, 'answer': 'February 7, 2016'} ``` > Created by [Qingqing Cao](https://awk.ai/) | [GitHub](https://github.com/csarron) | [Twitter](https://twitter.com/sysnlp) > Made with ❤️ in New York.
{"language": "en", "license": "mit", "tags": ["question-answering", "roberta", "roberta-base"], "datasets": ["squad"], "metrics": ["squad"], "widget": [{"text": "Which name is also used to describe the Amazon rainforest in English?", "context": "The Amazon rainforest (Portuguese: Floresta Amaz\u00f4nica or Amaz\u00f4nia; Spanish: Selva Amaz\u00f3nica, Amazon\u00eda or usually Amazonia; French: For\u00eat amazonienne; Dutch: Amazoneregenwoud), also known in English as Amazonia or the Amazon Jungle, is a moist broadleaf forest that covers most of the Amazon basin of South America. This basin encompasses 7,000,000 square kilometres (2,700,000 sq mi), of which 5,500,000 square kilometres (2,100,000 sq mi) are covered by the rainforest. This region includes territory belonging to nine nations. The majority of the forest is contained within Brazil, with 60% of the rainforest, followed by Peru with 13%, Colombia with 10%, and with minor amounts in Venezuela, Ecuador, Bolivia, Guyana, Suriname and French Guiana. States or departments in four nations contain \"Amazonas\" in their names. The Amazon represents over half of the planet's remaining rainforests, and comprises the largest and most biodiverse tract of tropical rainforest in the world, with an estimated 390 billion individual trees divided into 16,000 species."}, {"text": "How many square kilometers of rainforest is covered in the basin?", "context": "The Amazon rainforest (Portuguese: Floresta Amaz\u00f4nica or Amaz\u00f4nia; Spanish: Selva Amaz\u00f3nica, Amazon\u00eda or usually Amazonia; French: For\u00eat amazonienne; Dutch: Amazoneregenwoud), also known in English as Amazonia or the Amazon Jungle, is a moist broadleaf forest that covers most of the Amazon basin of South America. This basin encompasses 7,000,000 square kilometres (2,700,000 sq mi), of which 5,500,000 square kilometres (2,100,000 sq mi) are covered by the rainforest. This region includes territory belonging to nine nations. The majority of the forest is contained within Brazil, with 60% of the rainforest, followed by Peru with 13%, Colombia with 10%, and with minor amounts in Venezuela, Ecuador, Bolivia, Guyana, Suriname and French Guiana. States or departments in four nations contain \"Amazonas\" in their names. The Amazon represents over half of the planet's remaining rainforests, and comprises the largest and most biodiverse tract of tropical rainforest in the world, with an estimated 390 billion individual trees divided into 16,000 species."}]}
question-answering
csarron/roberta-base-squad-v1
[ "transformers", "pytorch", "jax", "safetensors", "roberta", "question-answering", "roberta-base", "en", "dataset:squad", "arxiv:1907.11692", "license:mit", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "1907.11692" ]
[ "en" ]
TAGS #transformers #pytorch #jax #safetensors #roberta #question-answering #roberta-base #en #dataset-squad #arxiv-1907.11692 #license-mit #endpoints_compatible #region-us
RoBERTa-base fine-tuned on SQuAD v1 ----------------------------------- This model was fine-tuned from the HuggingFace RoBERTa base checkpoint on SQuAD1.1. This model is case-sensitive: it makes a difference between english and English. Details ------- Dataset: SQuAD1.1, Split: train, # samples: 96.8K Dataset: SQuAD1.1, Split: eval, # samples: 11.8k ### Fine-tuning * Python: '3.7.5' * Machine specs: 'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz' 'Memory: 32 GiB' 'GPUs: 2 GeForce GTX 1070, each with 8GiB memory' 'GPU driver: 418.87.01, CUDA: 10.1' * script: It took about 2 hours to finish. ### Results Model size: '477M' Note that the above results didn't involve any hyperparameter search. Example Usage ------------- > > Created by Qingqing Cao | GitHub | Twitter > > > > > Made with ️ in New York. > > >
[ "# samples: 96.8K\nDataset: SQuAD1.1, Split: eval, # samples: 11.8k", "### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 2 hours to finish.", "### Results\n\n\nModel size: '477M'\n\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>" ]
[ "TAGS\n#transformers #pytorch #jax #safetensors #roberta #question-answering #roberta-base #en #dataset-squad #arxiv-1907.11692 #license-mit #endpoints_compatible #region-us \n", "# samples: 96.8K\nDataset: SQuAD1.1, Split: eval, # samples: 11.8k", "### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 2 hours to finish.", "### Results\n\n\nModel size: '477M'\n\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>" ]
[ 64, 27, 95, 65 ]
[ "passage: TAGS\n#transformers #pytorch #jax #safetensors #roberta #question-answering #roberta-base #en #dataset-squad #arxiv-1907.11692 #license-mit #endpoints_compatible #region-us \n# samples: 96.8K\nDataset: SQuAD1.1, Split: eval, # samples: 11.8k### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 2 hours to finish.### Results\n\n\nModel size: '477M'\n\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>" ]
[ -0.12511669099330902, 0.08033295720815659, 0.0019898004829883575, 0.09229458123445511, 0.1241801381111145, 0.0372810885310173, 0.06103558838367462, 0.11210767179727554, 0.08790256828069687, 0.15083837509155273, 0.12080906331539154, 0.005513670854270458, 0.11062385141849518, 0.07094967365264893, -0.020809201523661613, -0.055015310645103455, 0.07802557945251465, 0.0058367252349853516, -0.03937360271811485, 0.11245864629745483, 0.05890525504946709, -0.07193668186664581, 0.14353559911251068, 0.01708204858005047, -0.14267095923423767, -0.03910968452692032, 0.0027513992972671986, -0.02401791885495186, 0.10150570422410965, 0.06479541212320328, 0.03067050687968731, 0.037357281893491745, 0.058527640998363495, -0.07764778286218643, 0.02674601413309574, 0.08750122785568237, -0.019844358786940575, 0.04380087926983833, 0.016718478873372078, 0.11056305468082428, 0.07935472577810287, 0.02526266686618328, -0.033494096249341965, 0.05109177902340889, -0.07219788432121277, -0.07746578007936478, -0.13002565503120422, 0.0421563982963562, 0.09030601382255554, 0.06259012967348099, 0.0025327084586024284, 0.15781839191913605, -0.046115074306726456, 0.026457538828253746, 0.20671707391738892, -0.23454728722572327, -0.07349208742380142, 0.07540369778871536, 0.102747343480587, 0.031172743067145348, -0.0004600492538884282, -0.029989758506417274, 0.04059353843331337, 0.009943129494786263, -0.08794315159320831, -0.005039576906710863, -0.06306574493646622, 0.0196515042334795, -0.07946407794952393, -0.03540051355957985, 0.09555551409721375, 0.04006609320640564, -0.014046581462025642, -0.015940312296152115, -0.11723894625902176, -0.1892227679491043, 0.008109575137495995, 0.04782066121697426, -0.05414870008826256, 0.01991982012987137, 0.004263496957719326, 0.02871181257069111, -0.03715760260820389, -0.15082497894763947, -0.027636803686618805, 0.006126039661467075, 0.09185240417718887, 0.08717487007379532, 0.058489035815000534, 0.07142117619514465, -0.11532342433929443, -0.05882543697953224, -0.05904599279165268, -0.021439990028738976, -0.09162328392267227, 0.060554083436727524, 0.018693430349230766, 0.10993645340204239, 0.07199402898550034, 0.17090792953968048, -0.04285818710923195, 0.032326605170965195, 0.11752071231603622, -0.03693123534321785, -0.09242262691259384, 0.07903185486793518, -0.18633343279361725, -0.0550951287150383, 0.04635453596711159, 0.046942565590143204, -0.015975231304764748, -0.01013822853565216, -0.05387534573674202, -0.032813090831041336, 0.04820476099848747, -0.0071702697314321995, 0.04553172364830971, 0.03875013813376427, -0.045253705233335495, 0.0033331005834043026, 0.24754223227500916, -0.00763206509873271, -0.010957153514027596, 0.053292497992515564, -0.08705583214759827, -0.06002170965075493, -0.008156083524227142, -0.01931033656001091, -0.060257527977228165, -0.07482248544692993, -0.10467620939016342, -0.05596429482102394, -0.07085558772087097, -0.0681738331913948, 0.035996414721012115, -0.10007650405168533, -0.014087649993598461, -0.11745559424161911, -0.19888950884342194, -0.010803153738379478, 0.06995881348848343, -0.11074025183916092, 0.012133174575865269, 0.06954643875360489, -0.03539124131202698, 0.048242200165987015, -0.016835786402225494, 0.09176085889339447, -0.06114637851715088, 0.09504581242799759, 0.060890913009643555, 0.0359007902443409, -0.06572005152702332, 0.02077685482800007, -0.10984192788600922, 0.08759152889251709, -0.09512759745121002, -0.013207699172198772, -0.061888258904218674, 0.06689146161079407, -0.10570560395717621, -0.029922939836978912, -0.00878959521651268, -0.02714766189455986, 0.1265120953321457, 0.10358907282352448, -0.07845049351453781, 0.026599399745464325, 0.04275236278772354, -0.019014161080121994, -0.09372232109308243, 0.2047242522239685, 0.03612635284662247, -0.05605677142739296, 0.026818227022886276, 0.07597742974758148, -0.00478793028742075, -0.14531858265399933, -0.07063446938991547, -0.015051867812871933, 0.0007983158575370908, -0.11781530827283859, 0.09988489001989365, 0.014427870512008667, 0.030393710359930992, 0.0360155925154686, -0.020405836403369904, 0.02032954804599285, -0.08551383763551712, -0.04193012788891792, -0.09364071488380432, -0.11558616906404495, -0.11212100833654404, 0.018420027568936348, 0.01317423116415739, -0.031230665743350983, -0.13267022371292114, -0.19822393357753754, 0.108881376683712, -0.04496333748102188, -0.019832564517855644, -0.11805010586977005, 0.18342095613479614, -0.007990180514752865, 0.013897573575377464, -0.05748526751995087, -0.04004311189055443, 0.08797436207532883, -0.029886268079280853, -0.0474105142056942, -0.04384811222553253, -0.0005638161092065275, 0.039642587304115295, 0.008677268400788307, 0.025555090978741646, 0.0335683599114418, -0.04327473044395447, -0.08873657137155533, -0.05981065332889557, -0.003827781416475773, 0.006992635317146778, 0.07947976142168045, -0.09589002281427383, 0.027093978598713875, 0.043423283845186234, 0.07513155788183212, -0.06706930696964264, -0.05478530004620552, -0.053690288215875626, -0.018272899091243744, -0.08302430808544159, -0.08566063642501831, 0.028892314061522484, 0.00534641882404685, 0.02174641564488411, 0.08916537463665009, -0.11509377509355545, -0.018907733261585236, 0.10824289172887802, 0.14130403101444244, -0.05108734220266342, -0.0017057252116501331, -0.006014936603605747, -0.10383453220129013, 0.019973112270236015, -0.022486049681901932, 0.05275988206267357, 0.005689411889761686, 0.08618691563606262, -0.0910477414727211, 0.043556202203035355, 0.054551612585783005, 0.04429495334625244, -0.02007310464978218, 0.12531161308288574, 0.19901205599308014, -0.12131654471158981, 0.11471742391586304, -0.035755306482315063, -0.018425974994897842, 0.07511133700609207, 0.03355373442173004, -0.05960497260093689, -0.003382425056770444, 0.06652391701936722, -0.004454551264643669, 0.2018878012895584, -0.08638107031583786, 0.03406859189271927, 0.027198564261198044, -0.0364646278321743, 0.04740689694881439, -0.15912029147148132, 0.0060183703899383545, -0.0766647532582283, -0.06641428172588348, -0.02427137829363346, -0.008818353526294231, -0.04270834103226662, 0.054251834750175476, 0.049719348549842834, -0.08830530941486359, 0.03144653141498566, 0.02722499892115593, -0.032417796552181244, 0.19478629529476166, -0.06447888910770416, -0.17093664407730103, -0.09211451560258865, -0.04901131987571716, 0.0033918346744030714, 0.003964693751186132, -0.004340191371738911, -0.10574639588594437, -0.08041159808635712, -0.03640498220920563, 0.05086902901530266, 0.07833154499530792, 0.011693866923451424, 0.05860462412238121, 0.019833015277981758, 0.010167497210204601, -0.06187431886792183, 0.05483057349920273, -0.021275650709867477, -0.0767352506518364, 0.053011804819107056, -0.06795322149991989, 0.06328283995389938, 0.10066501051187515, 0.06260813772678375, -0.045625340193510056, 0.06242063269019127, 0.19418513774871826, -0.04049207270145416, 0.07347402721643448, 0.17389319837093353, 0.09094569087028503, 0.025887642055749893, 0.14873231947422028, -0.024499299004673958, -0.08302769064903259, 0.0008359103812836111, 0.01466398872435093, -0.09926797449588776, -0.1955271065235138, -0.031239181756973267, -0.05715440958738327, 0.08077002316713333, 0.05198691785335541, 0.058936748653650284, -0.14919129014015198, 0.07356786727905273, -0.06528079509735107, 0.10455609112977982, -0.03696320578455925, 0.019746161997318268, -0.027643907815217972, 0.06367212533950806, 0.047492511570453644, -0.006610110402107239, 0.06932542473077774, 0.08948501944541931, 0.12045704573392868, 0.14302316308021545, -0.07024543732404709, 0.24662064015865326, -0.043091677129268646, 0.259014755487442, 0.05838152766227722, 0.020290188491344452, -0.03471509367227554, 0.0006411668146029115, -0.004206801764667034, 0.010167845524847507, -0.08642220497131348, 0.05286344140768051, -0.0038494945038110018, -0.07576483488082886, 0.03027421422302723, 0.17603209614753723, 0.028710227459669113, 0.3026382327079773, 0.11073609441518784, -0.27055028080940247, -0.13840560615062714, -0.023689229041337967, -0.03465264290571213, -0.16146674752235413, 0.03150816261768341, 0.10258857905864716, -0.03568821772933006, -0.08290619403123856, -0.06370226293802261, 0.0812639445066452, 0.024654017761349678, -0.017327532172203064, 0.13018231093883514, 0.10912411659955978, -0.0009388708858750761, 0.024454697966575623, -0.16978655755519867, 0.1553334891796112, 0.04372698441147804, 0.04458607733249664, -0.08061715960502625, 0.014444938860833645, -0.026292147114872932, -0.0176131222397089, 0.05844777449965477, 0.0002727328974287957, -0.04820246249437332, -0.13236001133918762, -0.24677640199661255, 0.02975097857415676, 0.024892905727028847, 0.04185381904244423, 0.05057726055383682, -0.02294851839542389, 0.010208538733422756, -0.030934102833271027, -0.023952048271894455, -0.04057185351848602, -0.09576497226953506, 0.07212257385253906, -0.03861511871218681, -0.03230797126889229, -0.09196159243583679, 0.02196114882826805, -0.07865983992815018, 0.17134541273117065, -0.1518375724554062, -0.05053498223423958, -0.06666160374879837, 0.05308578908443451, 0.17036136984825134, -0.0744728147983551, 0.049510031938552856, -0.08705455809831619, -0.02973516471683979, 0.00902554951608181, -0.06262864917516708, 0.12830311059951782, -0.1057542935013771, -0.20205965638160706, -0.08982089906930923, 0.05073537304997444, -0.007667015306651592, 0.03443003445863724, -0.009770984761416912, 0.003097095061093569, -0.1204366609454155, -0.049665067344903946, -0.028278980404138565, -0.08083072304725647, 0.06325042992830276, 0.08110218495130539, 0.01416532602161169, -0.01849510334432125, -0.013932867906987667, -0.11190661042928696, -0.05248258262872696, 0.20970696210861206, -0.027542611584067345, 0.0011010878952220082, 0.13777825236320496, 0.04194314405322075, -0.23418192565441132, 0.0065557630732655525, -0.002301809610798955, 0.034975651651620865, -0.15125438570976257, -0.13100066781044006, 0.027029965072870255, 0.150539368391037, -0.08454665541648865, 0.21434111893177032, -0.1551101803779602, -0.11628198623657227, 0.03139668330550194, -0.014003731310367584, 0.10807280242443085, -0.1819053441286087, -0.02874867245554924, 0.014214541763067245, -0.20172472298145294, 0.08854002505540848, -0.00259042507968843, 0.09018535166978836, -0.053911324590444565, 0.034253861755132675, -0.016581688076257706, -0.1306401640176773, 0.12408355623483658, -0.03623967245221138, 0.030617985874414444, -0.04651077836751938, 0.026393504813313484, 0.05095502734184265, -0.05539441108703613, 0.07239196449518204, -0.0570080392062664, 0.08435958623886108, -0.04568242281675339, -0.027311792597174644, -0.06917482614517212, -0.01485166884958744, -0.00266720331273973, -0.011463218368589878, -0.11992849409580231, 0.0794827789068222, -0.09165371209383011, -0.018978308886289597, 0.09243986010551453, 0.06480126082897186, -0.02737390249967575, 0.1479605734348297, -0.05159126594662666, -0.07133834064006805, 0.046789031475782394, -0.00289475847966969, 0.014812774024903774, 0.08224822580814362, -0.12524281442165375, 0.014033900573849678, 0.08335968852043152, -0.03036402352154255, 0.019021442160010338, -0.007445489056408405, -0.07257555425167084, 0.04784915968775749, 0.04531199485063553, -0.15284332633018494, -0.08199446648359299, 0.024116355925798416, -0.03212409093976021, -0.04062014818191528, -0.024483639746904373, 0.05469662696123123, -0.020036380738019943, -0.02064058743417263, -0.009202237240970135, 0.10542626678943634, 0.01933528482913971, 0.20677368342876434, 0.03864389657974243, 0.05326426774263382, -0.1357565075159073, 0.04147309437394142, 0.06236249953508377, -0.04085729643702507, 0.015650272369384766, 0.10731633007526398, -0.151676207780838, -0.0619879812002182, -0.026816826313734055, -0.06956535577774048, -0.10000719875097275, -0.027059605345129967, -0.128329798579216, -0.02235296741127968, 0.03737792372703552, -0.015455869026482105, 0.03303945064544678, 0.03658042848110199, 0.11329052597284317, -0.023293767124414444, -0.09780408442020416, 0.040934883058071136, 0.00021820818074047565, 0.0589040108025074, -0.09443497657775879, 0.02195035107433796, 0.010961705818772316, 0.07474999129772186, -0.0006352117634378374, 0.09432482719421387, -0.1353064328432083, 0.018234169110655785, -0.19460001587867737, 0.019177168607711792, -0.0722452700138092, -0.023983508348464966, -0.009800449945032597, -0.032140132039785385, -0.08423460274934769, 0.03833796828985214, -0.04931497201323509, -0.022340063005685806, -0.05051322281360626, -0.021595532074570656, -0.06236673891544342, -0.0062759872525930405, 0.09809009730815887, -0.05501163750886917, 0.09967754036188126, 0.08458871394395828, 0.03015897609293461, -0.006599153857678175, -0.0967446118593216, -0.002036202931776643, -0.042262688279151917, 0.08061495423316956, 0.017470696941018105, -0.001912513398565352, 0.048478417098522186, 0.026227768510580063, 0.05060058459639549, -0.025658614933490753, 0.1094214916229248, -0.09858562797307968, -0.08258114010095596, -0.10390038788318634, -0.009037399664521217, -0.019261043518781662, -0.003923955839127302, 0.09572352468967438, 0.053531575947999954, 0.09135178476572037, -0.011255043558776379, -0.0074316831305623055, -0.12908023595809937, 0.003583968849852681, 0.008582305163145065, -0.04731256887316704, -0.0483899861574173, 0.01617048680782318, 0.10602815449237823, 0.003197113983333111, 0.1649995744228363, -0.010968852788209915, -0.06925515085458755, 0.01575944945216179, 0.027072738856077194, 0.05333217605948448, 0.020296916365623474, 0.0925002470612526, 0.0488501638174057, 0.0602470263838768, 0.013809454627335072, 0.03261339291930199, 0.06401683390140533, 0.08154385536909103, 0.0799485519528389, 0.13307936489582062, 0.14818289875984192, 0.1316453516483307, 0.038804661482572556, -0.19494791328907013, -0.02414006181061268, 0.07603884488344193, -0.257575660943985, 0.05336753651499748, -0.03908374533057213, 0.05981828272342682, 0.10352659225463867, -0.1122567430138588, -0.043688077479600906, -0.09422563761472702, -0.0944831594824791, -0.0839943066239357, -0.02210305444896221, -0.09246519953012466, -0.12575215101242065, 0.044706761837005615, 0.006127920933067799, -0.009363592602312565, 0.131329745054245, 0.04864474758505821, -0.019439680501818657, 0.11931996047496796, 0.13070684671401978, -0.01802360638976097, -0.019541461020708084, 0.06929042935371399, -0.0387980081140995, 0.10683926939964294, -0.13577552139759064, 0.04999830201268196, -0.050973113626241684, 0.10991673916578293, 0.018798300996422768, -0.028212809935212135, 0.11582256108522415, 0.029997780919075012, -0.05081027373671532, -0.05795849859714508, 0.011383144184947014, -0.010886295698583126, 0.12010978907346725, 0.01860242336988449, 0.0991823598742485, 0.018175166100263596, 0.09196852892637253, -0.002361285500228405, 0.03892827406525612, -0.14392554759979248, 0.044723086059093475, -0.09725549072027206, -0.004538796842098236, 0.02490386553108692, -0.06712735444307327, -0.019337791949510574, 0.2266186773777008, 0.12285266071557999, -0.12832112610340118, -0.08526937663555145, 0.007128546014428139, -0.02024785988032818, -0.09512249380350113, 0.10485120117664337, 0.12377376109361649, 0.1605490893125534, -0.033363793045282364, -0.1296795904636383, -0.0054885284043848515, -0.07179886847734451, -0.1096532791852951, 0.002744883531704545, -0.014071911573410034, 0.036331307142972946, -0.05776062607765198, -0.01828037016093731, -0.02075415477156639, -0.10121777653694153, 0.06690993160009384, -0.0630887821316719, -0.11967804282903671, 0.028874170035123825, 0.004759683273732662, -0.019093530252575874, 0.03328077495098114, -0.061331916600465775, 0.07537327706813812, -0.02189522050321102, -0.034466084092855453, -0.1554774045944214, -0.09291493147611618, 0.10461296886205673, 0.12922148406505585, 0.22783328592777252, -0.03200850635766983, 0.09445347636938095, 0.1615506112575531, -0.011533587239682674, -0.16101686656475067, 0.06943288445472717, 0.029604170471429825, -0.17241813242435455, 0.006542276591062546, 0.01661098562180996, -0.011806783266365528, 0.06984207779169083, 0.04075060412287712, 0.032504938542842865, -0.01129087433218956, 0.033277466893196106, 0.06496579200029373, -0.13294020295143127, 0.016417672857642174, -0.09708263725042343, 0.14337758719921112, 0.14947152137756348, -0.06567158550024033, -0.0007600170793011785, -0.0518001914024353, 0.05922393873333931, -0.03952344134449959, 0.08298414945602417, -0.008808977901935577, -0.2958131730556488, 0.047758832573890686, -0.007834800519049168, 0.021065223962068558, -0.15410815179347992, -0.011737802997231483, -0.05935872718691826, -0.05107859894633293, -0.094682477414608, 0.12004440277814865, 0.07197826355695724, 0.03033703751862049, -0.06493119895458221, -0.008059477433562279, -0.09701120853424072, 0.06263981014490128, -0.12052959948778152, -0.17217421531677246 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-emotion This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset. It achieves the following results on the evaluation set: - Loss: 0.2175 - Accuracy: 0.923 - F1: 0.9233 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.8352 | 1.0 | 250 | 0.3079 | 0.91 | 0.9086 | | 0.247 | 2.0 | 500 | 0.2175 | 0.923 | 0.9233 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["emotion"], "metrics": ["accuracy", "f1"], "model-index": [{"name": "distilbert-base-uncased-finetuned-emotion", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "emotion", "type": "emotion", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.923, "name": "Accuracy"}, {"type": "f1", "value": 0.9232542847906783, "name": "F1"}]}]}]}
text-classification
cscottp27/distilbert-base-uncased-finetuned-emotion
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "dataset:emotion", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-emotion ========================================= This model is a fine-tuned version of distilbert-base-uncased on the emotion dataset. It achieves the following results on the evaluation set: * Loss: 0.2175 * Accuracy: 0.923 * F1: 0.9233 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 64 * eval\_batch\_size: 64 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 2 ### Training results ### Framework versions * Transformers 4.11.3 * Pytorch 1.10.0+cu111 * Datasets 1.16.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ 67, 98, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ -0.10365526378154755, 0.11108539253473282, -0.0026109113823622465, 0.1317654550075531, 0.16546793282032013, 0.045472968369722366, 0.1148209348320961, 0.12493137270212173, -0.08185860514640808, 0.032128069549798965, 0.10837704688310623, 0.1617085337638855, 0.02285127155482769, 0.09674810618162155, -0.05789001286029816, -0.2760312557220459, -0.013312868773937225, 0.05033736675977707, -0.015638014301657677, 0.13228055834770203, 0.09348677843809128, -0.12362140417098999, 0.09650623053312302, 0.003427819348871708, -0.17494015395641327, 0.003667527576908469, 0.0020847665145993233, -0.04351482167840004, 0.1478479653596878, 0.0203389972448349, 0.10677221417427063, 0.008356831967830658, 0.08522238582372665, -0.2250649482011795, 0.018061498180031776, 0.03874968737363815, 0.0002561478759162128, 0.08761061728000641, 0.03720410540699959, -0.015301057137548923, 0.15303263068199158, -0.06405466049909592, 0.05445051193237305, 0.021798426285386086, -0.11284452676773071, -0.2199874073266983, -0.08077996224164963, 0.041833750903606415, 0.06329819560050964, 0.11930888146162033, -0.019842874258756638, 0.12837383151054382, -0.09601787477731705, 0.09610112756490707, 0.23596636950969696, -0.2448965162038803, -0.06858085840940475, 0.020751558244228363, 0.014530576765537262, 0.043537914752960205, -0.11989867687225342, -0.039740413427352905, 0.050211407244205475, 0.05117671564221382, 0.11863920837640762, -0.03312069922685623, -0.0999940037727356, 0.00910378061234951, -0.12911130487918854, -0.04672069475054741, 0.16751150786876678, 0.0594206377863884, -0.02597641386091709, -0.05425291508436203, -0.056162502616643906, -0.1675000786781311, -0.031035728752613068, -0.016450364142656326, 0.054442595690488815, -0.01592334546148777, -0.06372737884521484, 0.010441784746944904, -0.12011896073818207, -0.04583032429218292, -0.06421378999948502, 0.10641534626483917, 0.021976949647068977, 0.006818883586674929, -0.017207900062203407, 0.10352788120508194, 0.0009800537955015898, -0.12149646133184433, 0.0210917666554451, 0.02099326066672802, 0.026333261281251907, -0.03047778643667698, -0.06905151158571243, -0.055389244109392166, -0.004919437691569328, 0.10190358012914658, -0.06666407734155655, 0.045935261994600296, 0.04524286091327667, 0.037766341120004654, -0.06852026283740997, 0.19527418911457062, -0.03294919803738594, -0.03212658688426018, -0.012501158751547337, 0.06016761064529419, 0.020768651738762856, -0.006436762399971485, -0.12171690911054611, 0.02350432053208351, 0.08950570225715637, -0.00007678163092350587, -0.09379716217517853, 0.08152900636196136, -0.07605183124542236, -0.01949211396276951, -0.021253660321235657, -0.0765935555100441, 0.028752928599715233, 0.02075762115418911, -0.07241667807102203, 0.0030450790654867887, 0.030792895704507828, 0.008529079146683216, -0.015946075320243835, 0.0917603150010109, -0.0787631943821907, 0.025490665808320045, -0.0946962907910347, -0.1045740470290184, 0.02897842787206173, -0.09482478350400925, 0.033921513706445694, -0.09247273951768875, -0.19394199550151825, -0.024185366928577423, 0.0686044692993164, -0.021704984828829765, -0.0471469946205616, -0.07331191003322601, -0.06313978135585785, 0.01850917749106884, -0.002557029016315937, 0.09711819887161255, -0.06576484441757202, 0.09101833403110504, 0.027181919664144516, 0.08311154693365097, -0.031909260898828506, 0.056395988911390305, -0.11463139951229095, 0.004114741925150156, -0.13587668538093567, 0.049997471272945404, -0.047578345984220505, 0.07202952355146408, -0.06358253955841064, -0.11173929274082184, 0.015572934411466122, -0.006378253921866417, 0.06443363428115845, 0.10948903113603592, -0.19403833150863647, -0.09464851766824722, 0.16610102355480194, -0.07067948579788208, -0.10680301487445831, 0.12891265749931335, -0.06742480397224426, 0.06976816058158875, 0.06998448073863983, 0.17836694419384003, 0.05879383906722069, -0.07357979565858841, -0.016340071335434914, 0.011599463410675526, 0.050163306295871735, -0.031123390421271324, 0.052572306245565414, 0.02624650113284588, 0.03453371301293373, 0.03880535066127777, -0.01196232344955206, 0.07060033828020096, -0.09326792508363724, -0.10062386095523834, -0.03427024930715561, -0.09134482592344284, 0.04848787933588028, 0.09255674481391907, 0.06272212415933609, -0.10717325657606125, -0.073724165558815, 0.028808284550905228, 0.09422781318426132, -0.06488901376724243, 0.028829196467995644, -0.0596344880759716, 0.0622246228158474, 0.0027249802369624376, -0.015573904849588871, -0.17582817375659943, 0.013070456683635712, 0.006194745190441608, 0.027181124314665794, 0.006623989902436733, 0.03696899861097336, 0.06610392779111862, 0.04349454119801521, -0.05583404377102852, -0.024697057902812958, -0.04591572657227516, -0.002911053132265806, -0.11217767000198364, -0.222952201962471, -0.01753520965576172, -0.024884168058633804, 0.17654204368591309, -0.21061822772026062, 0.04510215297341347, -0.006178705487400293, 0.055832020938396454, 0.014611254446208477, -0.01919422671198845, -0.033304933458566666, 0.06556599587202072, -0.054479777812957764, -0.042237140238285065, 0.07778333127498627, 0.01036946102976799, -0.08646371960639954, -0.037160664796829224, -0.10661646723747253, 0.14299476146697998, 0.13010098040103912, -0.11318360269069672, -0.07107231020927429, -0.016780277714133263, -0.06628571450710297, -0.01899905502796173, -0.03863293677568436, 0.03865957632660866, 0.19597330689430237, -0.007075474597513676, 0.13832490146160126, -0.06229453533887863, -0.02452687919139862, 0.02397817187011242, -0.04459702596068382, 0.005227315239608288, 0.13486889004707336, 0.12143418937921524, -0.06046079099178314, 0.15041641891002655, 0.13522854447364807, -0.08931641280651093, 0.16353555023670197, -0.03590400516986847, -0.05880224332213402, -0.025102846324443817, -0.04856446385383606, -0.018963899463415146, 0.10565771162509918, -0.18460632860660553, -0.011765911243855953, 0.02307305857539177, 0.0011658030562102795, 0.006093000527471304, -0.2260960191488266, -0.05179370567202568, 0.04856545850634575, -0.04331237077713013, -0.006302523892372847, -0.010050495155155659, 0.00567513657733798, 0.1050773411989212, -0.0037724007852375507, -0.08540020883083344, 0.030277138575911522, -0.001458055805414915, -0.08583482354879379, 0.20442630350589752, -0.0918363556265831, -0.17275619506835938, -0.1108991801738739, -0.07250522077083588, -0.047569986432790756, 0.00643935427069664, 0.07166393846273422, -0.11759302020072937, -0.019171783700585365, -0.07828955352306366, 0.0264158733189106, 0.011945844627916813, 0.020029593259096146, 0.028738269582390785, -0.0024876517709344625, 0.047073788940906525, -0.10855920612812042, -0.019495608285069466, -0.06457715481519699, -0.04853179305791855, 0.054971616715192795, 0.019171450287103653, 0.11925762891769409, 0.16973093152046204, -0.005927021149545908, 0.011735835112631321, -0.03866672143340111, 0.22641333937644958, -0.072452612221241, -0.019722480326890945, 0.13654112815856934, -0.012618577107787132, 0.05264660716056824, 0.11521118134260178, 0.06805716454982758, -0.09154286980628967, 0.014333990402519703, 0.04578052833676338, -0.037165895104408264, -0.22064699232578278, -0.04134365916252136, -0.04843695089221001, 0.025485754013061523, 0.06965936720371246, 0.021163061261177063, 0.0463443286716938, 0.07631208002567291, 0.04127596318721771, 0.04955020546913147, -0.04838881269097328, 0.05194804072380066, 0.1304904669523239, 0.018573161214590073, 0.10156048089265823, -0.03687027096748352, -0.05285428836941719, 0.05776938423514366, -0.019366687163710594, 0.21297587454319, 0.001861072494648397, 0.14540186524391174, 0.05809827521443367, 0.16863800585269928, -0.03068048506975174, 0.07258651405572891, -0.014496046118438244, -0.04065464809536934, -0.03182343393564224, -0.028970252722501755, -0.06382753700017929, 0.03464755788445473, -0.05723574012517929, 0.08380259573459625, -0.13994887471199036, 0.01241863239556551, 0.06259757280349731, 0.2786177396774292, 0.027485299855470657, -0.31776177883148193, -0.11263156682252884, 0.005084906704723835, -0.03786880522966385, -0.005435932893306017, 0.02235650084912777, 0.09276892989873886, -0.09591041505336761, 0.0349600613117218, -0.06057644262909889, 0.08381333947181702, -0.07126171886920929, 0.06418641656637192, 0.046352777630090714, 0.0721132755279541, 0.010117967613041401, 0.0875658169388771, -0.2863155007362366, 0.26768237352371216, -0.010248835198581219, 0.05953718349337578, -0.08695551007986069, -0.0006002221489325166, 0.06266553699970245, 0.06651800870895386, 0.06717213243246078, -0.007731396239250898, 0.0020543483551591635, -0.182090163230896, -0.03915363922715187, 0.029793573543429375, 0.06323503702878952, -0.03606581315398216, 0.08719656616449356, -0.02526232600212097, 0.008744700811803341, 0.07756873965263367, 0.03433313965797424, -0.04849676042795181, -0.10168127715587616, -0.010243501514196396, 0.033136118203401566, -0.053475815802812576, -0.05309470370411873, -0.128390833735466, -0.10907188057899475, 0.14396999776363373, -0.003665628144517541, -0.023246966302394867, -0.10303903371095657, 0.0824100524187088, 0.04088747873902321, -0.088455930352211, 0.027437539771199226, 0.009255582466721535, 0.08022436499595642, 0.020462913438677788, -0.06966151297092438, 0.106208935379982, -0.07627078890800476, -0.1729302853345871, -0.06904780864715576, 0.09420742839574814, 0.05349254980683327, 0.07726606726646423, -0.0060472674667835236, -0.010492062196135521, -0.04896660894155502, -0.08469203859567642, 0.03883951157331467, 0.030602402985095978, 0.060137249529361725, 0.015432341024279594, -0.05169805884361267, 0.006301424000412226, -0.07002141326665878, -0.037205908447504044, 0.19957755506038666, 0.23401984572410583, -0.088113933801651, 0.030145883560180664, 0.032210662961006165, -0.0741310715675354, -0.1937704235315323, 0.04973391443490982, 0.059615232050418854, 0.009831813164055347, 0.041455693542957306, -0.19466006755828857, 0.12152144312858582, 0.08430802822113037, -0.01133981067687273, 0.09791947156190872, -0.30057990550994873, -0.11297563463449478, 0.13992652297019958, 0.1440286636352539, 0.12114910036325455, -0.14069239795207977, -0.002414488699287176, -0.030839765444397926, -0.12448009848594666, 0.11540801078081131, -0.08373696357011795, 0.1239585131406784, -0.024087201803922653, 0.11903462558984756, 0.008925092406570911, -0.046338435262441635, 0.11534544825553894, 0.021625714376568794, 0.09919611364603043, -0.07103241235017776, -0.0303041934967041, 0.022858066484332085, -0.040134724229574203, 0.02944108098745346, -0.09900438040494919, 0.017833665013313293, -0.11792459338903427, -0.03237378969788551, -0.08853314816951752, 0.03513661399483681, -0.040174245834350586, -0.07397525757551193, -0.050041165202856064, 0.02816113457083702, 0.07622185349464417, -0.004667900502681732, 0.08376561850309372, 0.01950220949947834, 0.11442000418901443, 0.09925757348537445, 0.09666679054498672, -0.05519923195242882, -0.07141568511724472, -0.022381464019417763, -0.009976433590054512, 0.048478029668331146, -0.14822426438331604, 0.01608099974691868, 0.13894839584827423, 0.019480889663100243, 0.1673259288072586, 0.08592981100082397, -0.0385739766061306, 0.017456399276852608, 0.06034190207719803, -0.1507411003112793, -0.08740292489528656, -0.02028888277709484, -0.07070736587047577, -0.12298166751861572, 0.031328827142715454, 0.08302098512649536, -0.07266725599765778, -0.00016929450794123113, -0.015392606146633625, 0.01725717820227146, -0.04057107865810394, 0.16434220969676971, 0.048006441444158554, 0.029739059507846832, -0.10324610769748688, 0.07677920162677765, 0.02069034054875374, -0.10905935615301132, 0.029330000281333923, 0.07366786152124405, -0.07665736973285675, -0.05674770846962929, 0.06659182161092758, 0.21424585580825806, -0.060438938438892365, -0.049563173204660416, -0.1491101235151291, -0.12783226370811462, 0.08513522893190384, 0.1479852944612503, 0.1134083941578865, 0.008781511336565018, -0.08621153980493546, 0.024828435853123665, -0.11682089418172836, 0.08967143297195435, 0.05856127291917801, 0.04151647537946701, -0.13298381865024567, 0.12173733115196228, 0.00987596157938242, 0.04105108231306076, -0.020841378718614578, 0.010980355553328991, -0.09201997518539429, 0.008116669952869415, -0.11870959401130676, -0.027281617745757103, -0.04016470909118652, 0.011824891902506351, 0.0021165378857403994, -0.04331228882074356, -0.0449797622859478, 0.003677732776850462, -0.11604902893304825, -0.015674734488129616, 0.03586804121732712, 0.07664954662322998, -0.113495834171772, -0.03808976709842682, 0.028106754645705223, -0.06515103578567505, 0.09211961925029755, 0.06284593045711517, 0.013682783581316471, 0.05738198012113571, -0.16451308131217957, 0.02591823786497116, 0.09183235466480255, 0.015367215499281883, 0.05428994446992874, -0.08173952251672745, -0.01165669783949852, -0.010551336221396923, 0.039947234094142914, 0.016846122220158577, 0.08105680346488953, -0.12655404210090637, 0.018837958574295044, 0.005064732860773802, -0.08740627020597458, -0.06847129762172699, 0.03200390934944153, 0.08091627061367035, 0.009806549176573753, 0.19735103845596313, -0.07809913903474808, 0.046744346618652344, -0.21774250268936157, 0.007651552092283964, 0.00039223834755830467, -0.10069233924150467, -0.12886367738246918, -0.07508815824985504, 0.05666074901819229, -0.05609523504972458, 0.1320609599351883, 0.04614754393696785, 0.009946395643055439, 0.010845249518752098, -0.009018459357321262, 0.023153048008680344, 0.003420208115130663, 0.18353857100009918, 0.035507675260305405, -0.05026087164878845, 0.06071584299206734, 0.05424007400870323, 0.11871401220560074, 0.12723422050476074, 0.19787128269672394, 0.1401710957288742, 0.025031769648194313, 0.10930082201957703, 0.03284158185124397, -0.03587397560477257, -0.14979085326194763, 0.030222611501812935, -0.0520317442715168, 0.11461912095546722, -0.017922502011060715, 0.24583737552165985, 0.06318626552820206, -0.15751786530017853, 0.0627346932888031, -0.0622653029859066, -0.080485038459301, -0.10323148965835571, -0.062278345227241516, -0.07988351583480835, -0.14291183650493622, 0.0025185495615005493, -0.1344880908727646, 0.005178164690732956, 0.0950811505317688, 0.010578589513897896, -0.041072338819503784, 0.13897496461868286, 0.01453063078224659, 0.020731834694743156, 0.08987818658351898, 0.008688630536198616, -0.06495719403028488, -0.13327042758464813, -0.0563986711204052, -0.012872141785919666, -0.01658868044614792, 0.040306270122528076, -0.050967987626791, -0.06254465132951736, 0.0255191158503294, -0.017901567742228508, -0.1019640564918518, 0.008450163528323174, 0.00674017146229744, 0.06198660656809807, 0.045322615653276443, 0.0007392247207462788, 0.02236340567469597, 0.0022965685930103064, 0.19070175290107727, -0.07466679811477661, -0.02867077849805355, -0.10520850121974945, 0.22406406700611115, 0.02177613228559494, -0.014602554962038994, 0.03250817582011223, -0.0718652755022049, -0.005044568330049515, 0.24970677495002747, 0.20883415639400482, -0.08701977133750916, -0.005430365912616253, 0.0030840749386698008, 0.002857531188055873, -0.046956535428762436, 0.0954923927783966, 0.15175102651119232, 0.02246778830885887, -0.09816353768110275, -0.02384945936501026, -0.058102697134017944, -0.023486129939556122, -0.016953278332948685, 0.05730053782463074, 0.062264759093523026, 0.012224104255437851, -0.04443337395787239, 0.050532374531030655, -0.08826623111963272, -0.10048242658376694, 0.07598186284303665, -0.21897344291210175, -0.15388427674770355, -0.017186596989631653, 0.09836289286613464, 0.028858967125415802, 0.07270942628383636, -0.017048347741365433, -0.0037730636540800333, 0.1151106208562851, -0.02014276571571827, -0.11828415095806122, -0.07104312628507614, 0.09810949116945267, -0.13000449538230896, 0.20294146239757538, -0.06483820080757141, 0.0402655303478241, 0.12451666593551636, 0.07169801741838455, -0.05337971821427345, 0.07340917736291885, 0.048565737903118134, -0.055177778005599976, 0.006028510630130768, 0.10180963575839996, -0.0313725620508194, 0.07693233340978622, 0.04867885634303093, -0.15307050943374634, 0.025778576731681824, -0.0403456836938858, -0.06710763275623322, -0.04520198702812195, -0.00809670053422451, -0.06513381749391556, 0.12103945016860962, 0.22164419293403625, -0.02433830127120018, -0.002707727486267686, -0.07023309171199799, 0.006103829480707645, 0.04813467711210251, 0.00954868271946907, -0.05603432282805443, -0.20643669366836548, 0.012327268719673157, 0.06683766096830368, -0.013608798384666443, -0.2591656446456909, -0.10369356721639633, 0.0030699747148901224, -0.06895381212234497, -0.09046507626771927, 0.061003975570201874, 0.06755086034536362, 0.059811607003211975, -0.04745471104979515, -0.05766627937555313, -0.06149698793888092, 0.1691356599330902, -0.13927510380744934, -0.08477252721786499 ]
null
null
transformers
# BanglaBERT This repository contains the pretrained discriminator checkpoint of the model **BanglaBERT**. This is an [ELECTRA](https://openreview.net/pdf?id=r1xMH1BtvB) discriminator model pretrained with the Replaced Token Detection (RTD) objective. Finetuned models using this checkpoint achieve state-of-the-art results on many of the NLP tasks in bengali. For finetuning on different downstream tasks such as `Sentiment classification`, `Named Entity Recognition`, `Natural Language Inference` etc., refer to the scripts in the official GitHub [repository](https://github.com/csebuetnlp/banglabert). **Note**: This model was pretrained using a specific normalization pipeline available [here](https://github.com/csebuetnlp/normalizer). All finetuning scripts in the official GitHub repository uses this normalization by default. If you need to adapt the pretrained model for a different task make sure the text units are normalized using this pipeline before tokenizing to get best results. A basic example is given below: ## Using this model as a discriminator in `transformers` (tested on 4.11.0.dev0) ```python from transformers import AutoModelForPreTraining, AutoTokenizer from normalizer import normalize # pip install git+https://github.com/csebuetnlp/normalizer import torch model = AutoModelForPreTraining.from_pretrained("csebuetnlp/banglabert") tokenizer = AutoTokenizer.from_pretrained("csebuetnlp/banglabert") original_sentence = "আমি কৃতজ্ঞ কারণ আপনি আমার জন্য অনেক কিছু করেছেন।" fake_sentence = "আমি হতাশ কারণ আপনি আমার জন্য অনেক কিছু করেছেন।" fake_sentence = normalize(fake_sentence) # this normalization step is required before tokenizing the text fake_tokens = tokenizer.tokenize(fake_sentence) fake_inputs = tokenizer.encode(fake_sentence, return_tensors="pt") discriminator_outputs = model(fake_inputs).logits predictions = torch.round((torch.sign(discriminator_outputs) + 1) / 2) [print("%7s" % token, end="") for token in fake_tokens] print("\n" + "-" * 50) [print("%7s" % int(prediction), end="") for prediction in predictions.squeeze().tolist()[1:-1]] print("\n" + "-" * 50) ``` ## Benchmarks * Zero-shot cross-lingual transfer-learning | Model | Params | SC (macro-F1) | NLI (accuracy) | NER (micro-F1) | QA (EM/F1) | BangLUE score | |----------------|-----------|-----------|-----------|-----------|-----------|-----------| |[mBERT](https://huggingface.co/bert-base-multilingual-cased) | 180M | 27.05 | 62.22 | 39.27 | 59.01/64.18 | 50.35 | |[XLM-R (base)](https://huggingface.co/xlm-roberta-base) | 270M | 42.03 | 72.18 | 45.37 | 55.03/61.83 | 55.29 | |[XLM-R (large)](https://huggingface.co/xlm-roberta-large) | 550M | 49.49 | 78.13 | 56.48 | 71.13/77.70 | 66.59 | |[BanglishBERT](https://huggingface.co/csebuetnlp/banglishbert) | 110M | 48.39 | 75.26 | 55.56 | 72.87/78.63 | 66.14 | * Supervised fine-tuning | Model | Params | SC (macro-F1) | NLI (accuracy) | NER (micro-F1) | QA (EM/F1) | BangLUE score | |----------------|-----------|-----------|-----------|-----------|-----------|-----------| |[mBERT](https://huggingface.co/bert-base-multilingual-cased) | 180M | 67.59 | 75.13 | 68.97 | 67.12/72.64 | 70.29 | |[XLM-R (base)](https://huggingface.co/xlm-roberta-base) | 270M | 69.54 | 78.46 | 73.32 | 68.09/74.27 | 72.82 | |[XLM-R (large)](https://huggingface.co/xlm-roberta-large) | 550M | 70.97 | 82.40 | 78.39 | 73.15/79.06 | 76.79 | |[sahajBERT](https://huggingface.co/neuropark/sahajBERT) | 18M | 71.12 | 76.92 | 70.94 | 65.48/70.69 | 71.03 | |[BanglishBERT](https://huggingface.co/csebuetnlp/banglishbert) | 110M | 70.61 | 80.95 | 76.28 | 72.43/78.40 | 75.73 | |[BanglaBERT](https://huggingface.co/csebuetnlp/banglabert) | 110M | 72.89 | 82.80 | 77.78 | 72.63/79.34 | **77.09** | The benchmarking datasets are as follows: * **SC:** **[Sentiment Classification](https://aclanthology.org/2021.findings-emnlp.278)** * **NER:** **[Named Entity Recognition](https://multiconer.github.io/competition)** * **NLI:** **[Natural Language Inference](https://github.com/csebuetnlp/banglabert/#datasets)** * **QA:** **[Question Answering](https://github.com/csebuetnlp/banglabert/#datasets)** ## Citation If you use this model, please cite the following paper: ``` @inproceedings{bhattacharjee-etal-2022-banglabert, title = "{B}angla{BERT}: Language Model Pretraining and Benchmarks for Low-Resource Language Understanding Evaluation in {B}angla", author = "Bhattacharjee, Abhik and Hasan, Tahmid and Ahmad, Wasi and Mubasshir, Kazi Samin and Islam, Md Saiful and Iqbal, Anindya and Rahman, M. Sohel and Shahriyar, Rifat", booktitle = "Findings of the Association for Computational Linguistics: NAACL 2022", month = jul, year = "2022", address = "Seattle, United States", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2022.findings-naacl.98", pages = "1318--1327", abstract = "In this work, we introduce BanglaBERT, a BERT-based Natural Language Understanding (NLU) model pretrained in Bangla, a widely spoken yet low-resource language in the NLP literature. To pretrain BanglaBERT, we collect 27.5 GB of Bangla pretraining data (dubbed {`}Bangla2B+{'}) by crawling 110 popular Bangla sites. We introduce two downstream task datasets on natural language inference and question answering and benchmark on four diverse NLU tasks covering text classification, sequence labeling, and span prediction. In the process, we bring them under the first-ever Bangla Language Understanding Benchmark (BLUB). BanglaBERT achieves state-of-the-art results outperforming multilingual and monolingual models. We are making the models, datasets, and a leaderboard publicly available at \url{https://github.com/csebuetnlp/banglabert} to advance Bangla NLP.", } ``` If you use the normalization module, please cite the following paper: ``` @inproceedings{hasan-etal-2020-low, title = "Not Low-Resource Anymore: Aligner Ensembling, Batch Filtering, and New Datasets for {B}engali-{E}nglish Machine Translation", author = "Hasan, Tahmid and Bhattacharjee, Abhik and Samin, Kazi and Hasan, Masum and Basak, Madhusudan and Rahman, M. Sohel and Shahriyar, Rifat", booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)", month = nov, year = "2020", address = "Online", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/2020.emnlp-main.207", doi = "10.18653/v1/2020.emnlp-main.207", pages = "2612--2623", abstract = "Despite being the seventh most widely spoken language in the world, Bengali has received much less attention in machine translation literature due to being low in resources. Most publicly available parallel corpora for Bengali are not large enough; and have rather poor quality, mostly because of incorrect sentence alignments resulting from erroneous sentence segmentation, and also because of a high volume of noise present in them. In this work, we build a customized sentence segmenter for Bengali and propose two novel methods for parallel corpus creation on low-resource setups: aligner ensembling and batch filtering. With the segmenter and the two methods combined, we compile a high-quality Bengali-English parallel corpus comprising of 2.75 million sentence pairs, more than 2 million of which were not available before. Training on neural models, we achieve an improvement of more than 9 BLEU score over previous approaches to Bengali-English machine translation. We also evaluate on a new test set of 1000 pairs made with extensive quality control. We release the segmenter, parallel corpus, and the evaluation set, thus elevating Bengali from its low-resource status. To the best of our knowledge, this is the first ever large scale study on Bengali-English machine translation. We believe our study will pave the way for future research on Bengali-English machine translation as well as other low-resource languages. Our data and code are available at https://github.com/csebuetnlp/banglanmt.", } ```
{"language": ["bn"], "licenses": ["cc-by-nc-sa-4.0"]}
null
csebuetnlp/banglabert
[ "transformers", "pytorch", "electra", "pretraining", "bn", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "bn" ]
TAGS #transformers #pytorch #electra #pretraining #bn #endpoints_compatible #has_space #region-us
BanglaBERT ========== This repository contains the pretrained discriminator checkpoint of the model BanglaBERT. This is an ELECTRA discriminator model pretrained with the Replaced Token Detection (RTD) objective. Finetuned models using this checkpoint achieve state-of-the-art results on many of the NLP tasks in bengali. For finetuning on different downstream tasks such as 'Sentiment classification', 'Named Entity Recognition', 'Natural Language Inference' etc., refer to the scripts in the official GitHub repository. Note: This model was pretrained using a specific normalization pipeline available here. All finetuning scripts in the official GitHub repository uses this normalization by default. If you need to adapt the pretrained model for a different task make sure the text units are normalized using this pipeline before tokenizing to get best results. A basic example is given below: Using this model as a discriminator in 'transformers' (tested on 4.11.0.dev0) ----------------------------------------------------------------------------- Benchmarks ---------- * Zero-shot cross-lingual transfer-learning * Supervised fine-tuning The benchmarking datasets are as follows: * SC: Sentiment Classification * NER: Named Entity Recognition * NLI: Natural Language Inference * QA: Question Answering If you use this model, please cite the following paper: If you use the normalization module, please cite the following paper:
[]
[ "TAGS\n#transformers #pytorch #electra #pretraining #bn #endpoints_compatible #has_space #region-us \n" ]
[ 33 ]
[ "passage: TAGS\n#transformers #pytorch #electra #pretraining #bn #endpoints_compatible #has_space #region-us \n" ]
[ -0.0004112394235562533, 0.041354142129421234, -0.008528155274689198, -0.03102412074804306, 0.07681901752948761, 0.02007954567670822, 0.012598119676113129, 0.0646592229604721, 0.04147854819893837, 0.043214455246925354, 0.18332809209823608, 0.09383107721805573, -0.06503050029277802, 0.05114077031612396, -0.01736832968890667, -0.2509968876838684, 0.08372611552476883, 0.06038285046815872, -0.08393190056085587, 0.10378177464008331, 0.047619886696338654, -0.14532780647277832, 0.03572488948702812, -0.017941409721970558, -0.09317171573638916, 0.04069788381457329, -0.04251347482204437, -0.0805816501379013, 0.1405310183763504, 0.018291758373379707, 0.19978167116641998, 0.035918258130550385, -0.08932095021009445, -0.15195386111736298, 0.04345838353037834, 0.005574648734182119, -0.08992258459329605, 0.0498429574072361, -0.03951637074351311, -0.06969453394412994, 0.11775260418653488, 0.0358562096953392, 0.003754012519493699, 0.007837334647774696, -0.20779648423194885, -0.14565764367580414, -0.03541754558682442, 0.028747469186782837, -0.006246470846235752, 0.05760728940367699, -0.020497648045420647, 0.16387692093849182, -0.1811470091342926, 0.033113595098257065, 0.16855975985527039, -0.3143838346004486, -0.008360587060451508, 0.10539181530475616, 0.14951273798942566, 0.08691398799419403, -0.017013538628816605, 0.07504642754793167, 0.0594821572303772, 0.011466129682958126, 0.020147845149040222, -0.05267513915896416, -0.014599094167351723, 0.10996408760547638, -0.14509563148021698, -0.11043968796730042, 0.254686176776886, -0.029262498021125793, 0.07830287516117096, 0.013772690668702126, -0.07884658128023148, -0.14590929448604584, 0.04520614072680473, 0.0026330274995416403, -0.0049156006425619125, 0.038008399307727814, 0.047033220529556274, -0.019544513896107674, -0.15124498307704926, 0.06107987463474274, -0.18533262610435486, 0.2643808126449585, 0.0003308326122350991, 0.0615878626704216, -0.1735832691192627, 0.07805110514163971, 0.0036724291276186705, -0.0839289128780365, 0.03248788043856621, -0.10621437430381775, 0.02779177948832512, 0.02349206991493702, -0.1280091106891632, 0.07184570282697678, 0.05408177524805069, 0.14981554448604584, 0.015863966196775436, 0.009631853550672531, 0.07638216763734818, 0.15075933933258057, 0.07397035509347916, 0.05969224125146866, -0.05560861900448799, 0.00012562645133584738, -0.03007574938237667, -0.060922712087631226, 0.028188431635499, -0.03684408962726593, -0.11114306002855301, -0.08288317918777466, 0.005051752086728811, 0.05996648594737053, 0.03790896385908127, 0.01578747294843197, -0.08183763921260834, 0.05077175796031952, -0.0016833044355735183, -0.01786896027624607, -0.026684697717428207, -0.020236963406205177, 0.005471234209835529, 0.06323514878749847, 0.020098969340324402, 0.0027798283845186234, 0.036347974091768265, 0.049077894538640976, -0.1082748994231224, -0.014954476617276669, -0.0567358173429966, -0.06285220384597778, 0.05307726189494133, -0.14453305304050446, 0.09401827305555344, -0.16278260946273804, 0.006573964841663837, 0.009190497919917107, 0.07161904871463776, -0.0009785270085558295, 0.002985185943543911, 0.04520949721336365, -0.05650826916098595, 0.006915666162967682, -0.07078620791435242, -0.02853657677769661, -0.07965071499347687, 0.08913742750883102, -0.0038500959053635597, 0.10972831398248672, -0.11476588994264603, 0.05240604653954506, -0.0772668719291687, 0.04463693127036095, -0.10606101900339127, -0.017576292157173157, -0.0167388953268528, 0.11604676395654678, 0.03217073529958725, -0.059753261506557465, -0.15440422296524048, 0.07252475619316101, -0.010512527078390121, 0.13824570178985596, -0.07674701511859894, -0.06294940412044525, 0.15393252670764923, -0.07869051396846771, -0.13528120517730713, 0.034883566200733185, -0.005420091561973095, 0.008535950444638729, -0.012329958379268646, 0.18301606178283691, -0.037178490310907364, -0.07195069640874863, -0.024386776611208916, 0.06593277305364609, -0.10258170962333679, -0.07009898126125336, 0.05975327640771866, 0.04804649204015732, -0.012106447480618954, -0.02111349068582058, 0.05216153711080551, 0.08940333127975464, -0.08064303547143936, -0.03747595474123955, -0.016401177272200584, -0.03268832713365555, 0.10689026117324829, 0.06618049740791321, 0.053169745951890945, -0.09390737861394882, -0.0634048730134964, 0.06981851905584335, 0.029447780922055244, 0.06398533284664154, 0.04899004474282265, -0.049374427646398544, 0.09855900704860687, -0.12223415076732635, -0.04745597392320633, -0.19901502132415771, -0.09100515395402908, -0.08051510900259018, 0.07951728999614716, 0.023484904319047928, 0.31558695435523987, 0.06812369078397751, -0.10139766335487366, -0.00983357708901167, -0.04641420766711235, 0.039105188101530075, 0.0354425348341465, -0.027270440012216568, -0.06384865194559097, -0.01288951002061367, -0.09591913223266602, -0.03103673830628395, -0.08334320038557053, 0.04777867719531059, 0.08772335946559906, 0.1395852416753769, -0.005023861303925514, 0.049047261476516724, 0.02013910561800003, 0.048047978430986404, -0.05950259044766426, 0.01673496514558792, 0.058578941971063614, -0.015111393295228481, -0.08375831693410873, 0.08678558468818665, -0.13623882830142975, 0.33996742963790894, 0.205155611038208, -0.2672557234764099, 0.002807808807119727, 0.053086791187524796, -0.052053097635507584, 0.03480076789855957, 0.08317737281322479, -0.012364029884338379, 0.026694465428590775, -0.013916105963289738, 0.08156011253595352, -0.024807492271065712, -0.054136257618665695, -0.013335944153368473, -0.05617549642920494, -0.10000460594892502, 0.09495817869901657, 0.05787186324596405, -0.10804971307516098, 0.18190988898277283, 0.3140229880809784, -0.02174604870378971, 0.16751913726329803, 0.013977342285215855, -0.012260744348168373, 0.0006993251736275852, -0.03958621248602867, -0.08525049686431885, 0.10000360757112503, -0.26294824481010437, -0.0648106038570404, 0.06170731037855148, -0.012238415889441967, 0.0876331701874733, -0.14857515692710876, -0.0970139354467392, 0.033501897007226944, 0.06423191726207733, -0.05850604921579361, 0.11584218591451645, 0.005187953822314739, 0.08774769306182861, 0.020191524177789688, -0.06272605806589127, 0.05924614146351814, 0.0031453159172087908, -0.015201346017420292, 0.12217328697443008, -0.10419262945652008, -0.20651744306087494, -0.04602757841348648, -0.023447461426258087, 0.054314497858285904, 0.011447029188275337, 0.07058338075876236, -0.06666956841945648, -0.0005737446481361985, 0.04733463004231453, 0.03820301592350006, -0.17506633698940277, 0.04615093395113945, -0.025863971561193466, 0.040316272526979446, -0.09765812009572983, -0.09337668120861053, -0.0578727200627327, -0.06581471115350723, -0.024746540933847427, 0.11771070957183838, -0.02752642333507538, 0.07919833809137344, 0.1304364949464798, 0.010744374245405197, 0.034686166793107986, -0.0037712978664785624, 0.22229883074760437, -0.11996106803417206, -0.03056221641600132, 0.14757944643497467, -0.00862937979400158, 0.048175305128097534, 0.15896180272102356, 0.040623728185892105, -0.05521335452795029, -0.006256227381527424, -0.06589237600564957, -0.10465794056653976, -0.1532374918460846, -0.11455032229423523, -0.12561888992786407, -0.025711558759212494, -0.025202125310897827, 0.05679804086685181, 0.055055927485227585, 0.07254693657159805, 0.0174062829464674, -0.12727132439613342, -0.08102244138717651, 0.03782955929636955, 0.18572333455085754, -0.06862401217222214, 0.09615868330001831, -0.06657487899065018, -0.09396278858184814, 0.047296226024627686, 0.06638147681951523, 0.1123565211892128, 0.11809718608856201, -0.04697784036397934, 0.08952423185110092, 0.13554565608501434, 0.14420604705810547, 0.08188251405954361, 0.0027293458115309477, -0.057943664491176605, -0.026515068486332893, 0.0139937624335289, -0.05774539336562157, 0.08537470549345016, 0.1401548534631729, -0.10509054362773895, -0.03348548337817192, -0.25013354420661926, 0.06178712844848633, 0.005267721600830555, 0.09275928139686584, -0.16262611746788025, -0.017163557931780815, 0.10569140315055847, -0.023304540663957596, -0.06608039885759354, 0.06885390728712082, 0.08767291903495789, -0.07681910693645477, 0.03659747913479805, 0.030028099194169044, 0.07109272480010986, 0.04431986063718796, 0.09253036230802536, -0.10267242044210434, -0.17551718652248383, 0.020433368161320686, 0.04239816218614578, -0.2604841887950897, 0.26600635051727295, -0.023440929129719734, -0.11570149660110474, -0.025261640548706055, -0.05132715031504631, -0.011573508381843567, 0.1544002890586853, 0.09223347902297974, 0.021953308954834938, -0.12646235525608063, -0.07862329483032227, 0.10081269592046738, 0.016526537016034126, 0.09337138384580612, -0.05808178335428238, 0.0037325306329876184, 0.02074800617992878, 0.03850021958351135, -0.0031783725135028362, 0.12884917855262756, 0.0013324243482202291, -0.1212742030620575, 0.050439853221178055, 0.05078117176890373, 0.04463677108287811, -0.006078772246837616, -0.03601044416427612, -0.09443075209856033, 0.06372044235467911, -0.11305639892816544, -0.04032048583030701, -0.058239858597517014, -0.10084503889083862, 0.1518031656742096, -0.04940143600106239, 0.08152024447917938, -0.03635422885417938, -0.03719133138656616, -0.10007737576961517, -0.09648727625608444, 0.14003127813339233, -0.10627901554107666, -0.003970146644860506, -0.04681471735239029, 0.12069917470216751, -0.016997195780277252, 0.06503621488809586, -0.03367353603243828, 0.08610087633132935, -0.17201074957847595, -0.061618171632289886, 0.04618167504668236, -0.07060904055833817, 0.10353291034698486, 0.04016603156924248, 0.004439698066562414, 0.05417109653353691, 0.08670372515916824, -0.007406284566968679, 0.19429141283035278, 0.2469898909330368, -0.07810001820325851, 0.08597534149885178, 0.11952795833349228, 0.005204856861382723, -0.2476789653301239, -0.04672439396381378, -0.16207727789878845, 0.002510861726477742, 0.020795919001102448, -0.1389521062374115, 0.0340413935482502, 0.05091669037938118, -0.06674852967262268, 0.10389768332242966, -0.2622458338737488, -0.04165757820010185, 0.14714911580085754, -0.06934729218482971, 0.45425960421562195, -0.13867256045341492, -0.003335164161399007, 0.0226359311491251, -0.15998339653015137, 0.11179657280445099, -0.028145764023065567, 0.07109943777322769, -0.021546967327594757, 0.04505598172545433, 0.02337975800037384, -0.06924790143966675, 0.17911212146282196, -0.01044986117631197, 0.027927592396736145, -0.1006297692656517, -0.15987257659435272, 0.08414573967456818, -0.035950928926467896, -0.04594564810395241, 0.014850537292659283, -0.02157275192439556, -0.2374611347913742, 0.015326544642448425, -0.14154188334941864, 0.0829271599650383, 0.017750253900885582, -0.05502743273973465, -0.06905731558799744, 0.010341349989175797, -0.016178030520677567, 0.035247739404439926, 0.29359692335128784, -0.06546547263860703, 0.19652362167835236, 0.0784636065363884, 0.10050506889820099, -0.15145273506641388, -0.0012021649163216352, 0.015278860926628113, -0.02782091312110424, 0.10278802365064621, -0.12286192923784256, -0.0009501700988039374, 0.12574073672294617, -0.030363017693161964, 0.011328809894621372, 0.1143098920583725, -0.011135652661323547, 0.03182440251111984, 0.1584305316209793, -0.20599408447742462, -0.09467098116874695, -0.021394137293100357, 0.022021712735295296, 0.100815050303936, 0.07741749286651611, 0.08930154889822006, 0.006564377807080746, -0.019130578264594078, -0.0024126730859279633, -0.05760333314538002, -0.10254362225532532, 0.005163176916539669, 0.10288561135530472, 0.06162402406334877, -0.07596814632415771, 0.029851263388991356, 0.029713930562138557, -0.22709839046001434, -0.01061203796416521, 0.09122344851493835, -0.060750652104616165, -0.15511764585971832, -0.1457691341638565, 0.0071221888065338135, -0.1545080691576004, -0.021792348474264145, 0.005229126662015915, -0.09271065145730972, 0.06839679926633835, 0.2846852242946625, 0.10493139922618866, 0.053527623414993286, 0.01305050402879715, 0.019522447139024734, 0.05965500324964523, -0.05649113655090332, -0.020886793732643127, 0.014745530672371387, -0.09378314018249512, 0.04321848601102829, -0.019735414534807205, 0.14989104866981506, -0.09426362812519073, -0.06038996949791908, -0.14122411608695984, 0.06486719101667404, -0.08878141641616821, -0.13833411037921906, -0.09733043611049652, -0.08664953708648682, 0.007047866005450487, -0.12046467512845993, -0.06667592376470566, -0.05373881012201309, -0.13978376984596252, 0.07399661093950272, 0.027466297149658203, 0.03951488062739372, -0.054180826991796494, -0.017092091962695122, 0.1238366961479187, -0.035174816846847534, 0.10916559398174286, 0.10287265479564667, -0.04405292123556137, 0.06832891702651978, 0.0008871106547303498, -0.10612627863883972, 0.09389644116163254, -0.011457288637757301, 0.04233083128929138, 0.06217039003968239, -0.02167568914592266, -0.005965042859315872, 0.030982255935668945, 0.04551161825656891, -0.04783423990011215, -0.0701371431350708, 0.039051853120326996, 0.04166662320494652, -0.11476851999759674, -0.018522121012210846, -0.08909609913825989, 0.1428602784872055, 0.008310544304549694, 0.0758599042892456, 0.047473419457674026, 0.02336987294256687, -0.06503959000110626, 0.04340346157550812, -0.021439045667648315, -0.16291072964668274, 0.053254906088113785, -0.04498935118317604, 0.010616892017424107, -0.023154599592089653, 0.2873225510120392, -0.02796005830168724, -0.06096808984875679, 0.04890015348792076, 0.07178494334220886, -0.05299663171172142, 0.041642915457487106, 0.1934858113527298, 0.11185499280691147, -0.05809216946363449, -0.10931473970413208, 0.0647563636302948, 0.012289133854210377, -0.06262043863534927, 0.10341612994670868, 0.1374136507511139, 0.09263501316308975, 0.10938360542058945, -0.037829913198947906, 0.013174560852348804, -0.07008972018957138, -0.12207391113042831, 0.03411508724093437, 0.011765895411372185, -0.04041648283600807, 0.08072693645954132, 0.21589288115501404, -0.02846800908446312, 0.03526124730706215, -0.042620472609996796, 0.04816947504878044, -0.11881574243307114, -0.09438898414373398, -0.009078496135771275, -0.08662469685077667, 0.0022848232183605433, -0.07049301266670227, 0.016432665288448334, 0.24730068445205688, 0.03331257775425911, -0.012623870745301247, 0.1030958890914917, 0.10206084698438644, -0.04290137439966202, 0.003899678122252226, 0.014852939173579216, 0.06601212173700333, -0.05442531779408455, -0.010322327725589275, -0.09317927807569504, -0.06140200421214104, -0.06263742595911026, 0.02630556933581829, -0.13982219994068146, -0.043467819690704346, -0.1103200912475586, -0.09290221333503723, -0.0724717527627945, 0.07025430351495743, 0.0006160566699691117, 0.08965422958135605, -0.03241516649723053, 0.010654966346919537, -0.0060256412252783775, 0.2343865931034088, -0.08350787311792374, -0.009714961051940918, 0.0026803791988641024, 0.07741864025592804, 0.02488582581281662, 0.12121956795454025, -0.045554663985967636, 0.00017290985852014273, -0.11001236736774445, 0.20291759073734283, 0.2568155527114868, -0.08930061012506485, 0.08132679760456085, 0.06461839377880096, 0.03932706266641617, 0.06092759221792221, 0.01150329690426588, 0.11975438892841339, 0.20511725544929504, -0.15357105433940887, -0.034683819860219955, -0.06375471502542496, 0.026345783844590187, -0.051612306386232376, 0.030848387628793716, 0.03294576704502106, -0.07466058433055878, -0.06982763111591339, 0.044428497552871704, -0.1759749948978424, 0.05744984373450279, 0.043016642332077026, -0.25491923093795776, -0.0645657330751419, -0.010884488932788372, 0.1871432065963745, -0.0246000736951828, 0.11718714237213135, -0.05348975583910942, -0.16744530200958252, -0.00975964404642582, 0.02840207703411579, -0.19237692654132843, -0.07878107577562332, 0.14237932860851288, 0.07446548342704773, 0.012301168404519558, -0.06267943978309631, -0.02141178771853447, 0.09839288890361786, 0.04352439567446709, -0.037776343524456024, 0.021074019372463226, 0.07845572382211685, -0.03732490539550781, -0.13300707936286926, 0.017899388447403908, 0.0317610539495945, -0.040884777903556824, 0.09369798749685287, -0.2153923064470291, 0.05646141618490219, -0.014853804372251034, -0.014459695667028427, 0.009441089816391468, 0.03283824771642685, -0.013573242351412773, 0.07440705597400665, 0.021787064149975777, -0.019397106021642685, -0.04749598354101181, -0.02434224635362625, -0.040392909198999405, 0.06951574236154556, 0.04201528802514076, -0.14545252919197083, -0.013198891654610634, -0.057603951543569565, 0.0432538278400898, -0.013354936614632607, -0.07987949997186661, -0.022308066487312317, -0.023925410583615303, 0.06286203861236572, -0.05050002411007881, 0.026798952370882034, 0.06403107941150665, 0.027687007561326027, -0.019890379160642624, -0.01794293522834778, 0.04868467524647713, 0.05412912368774414, -0.14384233951568604, -0.055877987295389175 ]
null
null
transformers
# mT5-m2o-english-CrossSum This repository contains the many-to-one (m2o) mT5 checkpoint finetuned on all cross-lingual pairs of the [CrossSum](https://huggingface.co/datasets/csebuetnlp/CrossSum) dataset, where the target summary was in **english**, i.e. this model tries to **summarize text written in any language in English.** For finetuning details and scripts, see the [paper](https://arxiv.org/abs/2112.08804) and the [official repository](https://github.com/csebuetnlp/CrossSum). ## Using this model in `transformers` (tested on 4.11.0.dev0) ```python import re from transformers import AutoTokenizer, AutoModelForSeq2SeqLM WHITESPACE_HANDLER = lambda k: re.sub('\s+', ' ', re.sub('\n+', ' ', k.strip())) article_text = """Videos that say approved vaccines are dangerous and cause autism, cancer or infertility are among those that will be taken down, the company said. The policy includes the termination of accounts of anti-vaccine influencers. Tech giants have been criticised for not doing more to counter false health information on their sites. In July, US President Joe Biden said social media platforms were largely responsible for people's scepticism in getting vaccinated by spreading misinformation, and appealed for them to address the issue. YouTube, which is owned by Google, said 130,000 videos were removed from its platform since last year, when it implemented a ban on content spreading misinformation about Covid vaccines. In a blog post, the company said it had seen false claims about Covid jabs "spill over into misinformation about vaccines in general". The new policy covers long-approved vaccines, such as those against measles or hepatitis B. "We're expanding our medical misinformation policies on YouTube with new guidelines on currently administered vaccines that are approved and confirmed to be safe and effective by local health authorities and the WHO," the post said, referring to the World Health Organization.""" model_name = "csebuetnlp/mT5_m2o_english_crossSum" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForSeq2SeqLM.from_pretrained(model_name) input_ids = tokenizer( [WHITESPACE_HANDLER(article_text)], return_tensors="pt", padding="max_length", truncation=True, max_length=512 )["input_ids"] output_ids = model.generate( input_ids=input_ids, max_length=84, no_repeat_ngram_size=2, num_beams=4 )[0] summary = tokenizer.decode( output_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False ) print(summary) ``` ## Citation If you use this model, please cite the following paper: ``` @article{hasan2021crosssum, author = {Tahmid Hasan and Abhik Bhattacharjee and Wasi Uddin Ahmad and Yuan-Fang Li and Yong-bin Kang and Rifat Shahriyar}, title = {CrossSum: Beyond English-Centric Cross-Lingual Abstractive Text Summarization for 1500+ Language Pairs}, journal = {CoRR}, volume = {abs/2112.08804}, year = {2021}, url = {https://arxiv.org/abs/2112.08804}, eprinttype = {arXiv}, eprint = {2112.08804} } ```
{"language": ["am", "ar", "az", "bn", "my", "zh", "en", "fr", "gu", "ha", "hi", "ig", "id", "ja", "rn", "ko", "ky", "mr", "ne", "om", "ps", "fa", "pcm", "pt", "pa", "ru", "gd", "sr", "si", "so", "es", "sw", "ta", "te", "th", "ti", "tr", "uk", "ur", "uz", "vi", "cy", "yo"], "tags": ["summarization", "mT5"], "licenses": ["cc-by-nc-sa-4.0"], "widget": [{"text": "Videos that say approved vaccines are dangerous and cause autism, cancer or infertility are among those that will be taken down, the company said. The policy includes the termination of accounts of anti-vaccine influencers. Tech giants have been criticised for not doing more to counter false health information on their sites. In July, US President Joe Biden said social media platforms were largely responsible for people's scepticism in getting vaccinated by spreading misinformation, and appealed for them to address the issue. YouTube, which is owned by Google, said 130,000 videos were removed from its platform since last year, when it implemented a ban on content spreading misinformation about Covid vaccines. In a blog post, the company said it had seen false claims about Covid jabs \"spill over into misinformation about vaccines in general\". The new policy covers long-approved vaccines, such as those against measles or hepatitis B. \"We're expanding our medical misinformation policies on YouTube with new guidelines on currently administered vaccines that are approved and confirmed to be safe and effective by local health authorities and the WHO,\" the post said, referring to the World Health Organization."}]}
summarization
csebuetnlp/mT5_m2o_english_crossSum
[ "transformers", "pytorch", "mt5", "text2text-generation", "summarization", "mT5", "am", "ar", "az", "bn", "my", "zh", "en", "fr", "gu", "ha", "hi", "ig", "id", "ja", "rn", "ko", "ky", "mr", "ne", "om", "ps", "fa", "pcm", "pt", "pa", "ru", "gd", "sr", "si", "so", "es", "sw", "ta", "te", "th", "ti", "tr", "uk", "ur", "uz", "vi", "cy", "yo", "arxiv:2112.08804", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2112.08804" ]
[ "am", "ar", "az", "bn", "my", "zh", "en", "fr", "gu", "ha", "hi", "ig", "id", "ja", "rn", "ko", "ky", "mr", "ne", "om", "ps", "fa", "pcm", "pt", "pa", "ru", "gd", "sr", "si", "so", "es", "sw", "ta", "te", "th", "ti", "tr", "uk", "ur", "uz", "vi", "cy", "yo" ]
TAGS #transformers #pytorch #mt5 #text2text-generation #summarization #mT5 #am #ar #az #bn #my #zh #en #fr #gu #ha #hi #ig #id #ja #rn #ko #ky #mr #ne #om #ps #fa #pcm #pt #pa #ru #gd #sr #si #so #es #sw #ta #te #th #ti #tr #uk #ur #uz #vi #cy #yo #arxiv-2112.08804 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# mT5-m2o-english-CrossSum This repository contains the many-to-one (m2o) mT5 checkpoint finetuned on all cross-lingual pairs of the CrossSum dataset, where the target summary was in english, i.e. this model tries to summarize text written in any language in English. For finetuning details and scripts, see the paper and the official repository. ## Using this model in 'transformers' (tested on 4.11.0.dev0) If you use this model, please cite the following paper:
[ "# mT5-m2o-english-CrossSum\n\nThis repository contains the many-to-one (m2o) mT5 checkpoint finetuned on all cross-lingual pairs of the CrossSum dataset, where the target summary was in english, i.e. this model tries to summarize text written in any language in English. For finetuning details and scripts, see the paper and the official repository.", "## Using this model in 'transformers' (tested on 4.11.0.dev0)\n\n\n\n\n\n\nIf you use this model, please cite the following paper:" ]
[ "TAGS\n#transformers #pytorch #mt5 #text2text-generation #summarization #mT5 #am #ar #az #bn #my #zh #en #fr #gu #ha #hi #ig #id #ja #rn #ko #ky #mr #ne #om #ps #fa #pcm #pt #pa #ru #gd #sr #si #so #es #sw #ta #te #th #ti #tr #uk #ur #uz #vi #cy #yo #arxiv-2112.08804 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# mT5-m2o-english-CrossSum\n\nThis repository contains the many-to-one (m2o) mT5 checkpoint finetuned on all cross-lingual pairs of the CrossSum dataset, where the target summary was in english, i.e. this model tries to summarize text written in any language in English. For finetuning details and scripts, see the paper and the official repository.", "## Using this model in 'transformers' (tested on 4.11.0.dev0)\n\n\n\n\n\n\nIf you use this model, please cite the following paper:" ]
[ 157, 104, 31 ]
[ "passage: TAGS\n#transformers #pytorch #mt5 #text2text-generation #summarization #mT5 #am #ar #az #bn #my #zh #en #fr #gu #ha #hi #ig #id #ja #rn #ko #ky #mr #ne #om #ps #fa #pcm #pt #pa #ru #gd #sr #si #so #es #sw #ta #te #th #ti #tr #uk #ur #uz #vi #cy #yo #arxiv-2112.08804 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# mT5-m2o-english-CrossSum\n\nThis repository contains the many-to-one (m2o) mT5 checkpoint finetuned on all cross-lingual pairs of the CrossSum dataset, where the target summary was in english, i.e. this model tries to summarize text written in any language in English. For finetuning details and scripts, see the paper and the official repository.## Using this model in 'transformers' (tested on 4.11.0.dev0)\n\n\n\n\n\n\nIf you use this model, please cite the following paper:" ]
[ -0.09608187526464462, -0.03249843046069145, -0.0047204685397446156, -0.01687534712255001, 0.06920012831687927, 0.006863716058433056, 0.10838908702135086, 0.10223676264286041, 0.020791489630937576, 0.050613172352313995, 0.06287726014852524, 0.11210094392299652, 0.06030525639653206, 0.05553257837891579, -0.00011314707808196545, -0.24768857657909393, 0.025020962581038475, -0.0036083338782191277, -0.05488017573952675, 0.14682047069072723, 0.1400417536497116, -0.07035364210605621, 0.09357466548681259, -0.0022418475709855556, -0.027669604867696762, 0.07798011600971222, -0.006785335950553417, -0.047171901911497116, 0.13866941630840302, 0.11156458407640457, 0.0483054481446743, 0.06813783198595047, 0.055489782243967056, -0.22453677654266357, 0.03028535656630993, 0.019703790545463562, -0.028052760288119316, 0.035020142793655396, 0.07017799466848373, -0.03019106201827526, 0.17540346086025238, -0.08139634132385254, -0.05520252883434296, 0.08105015009641647, -0.09385492652654648, -0.13911405205726624, -0.05919887498021126, 0.0802159532904625, 0.0880703553557396, 0.10516814142465591, -0.07043712586164474, 0.11682868748903275, -0.032962020486593246, 0.07597774267196655, 0.2404051572084427, -0.27106037735939026, -0.03808446228504181, 0.14333556592464447, 0.08687426894903183, 0.07947811484336853, -0.037566523998975754, 0.05621614679694176, 0.04284936562180519, 0.012089194729924202, -0.010167432017624378, -0.07211910933256149, 0.09086395800113678, 0.048144228756427765, -0.16866950690746307, -0.04668673127889633, 0.23690523207187653, 0.04394909366965294, -0.040766216814517975, -0.017200199887156487, -0.077814020216465, 0.03711554780602455, 0.0010816180147230625, -0.0724213719367981, 0.030734390020370483, -0.028897827491164207, 0.04810824990272522, -0.04536290466785431, -0.1073082685470581, -0.05375918745994568, -0.08834025263786316, 0.03744835779070854, -0.0024967321660369635, 0.022974226623773575, -0.044697415083646774, 0.07605062425136566, -0.2249334305524826, -0.08256937563419342, -0.07205382734537125, -0.05913691967725754, -0.07020602375268936, -0.051276691257953644, -0.014143708162009716, -0.15310733020305634, 0.052139706909656525, 0.16254988312721252, 0.041910022497177124, 0.019965626299381256, 0.030309252440929413, 0.03223820775747299, 0.10016662627458572, 0.10463674366474152, -0.0996498391032219, -0.08614253997802734, 0.00412668427452445, 0.005526178982108831, 0.003925191704183817, 0.017036376520991325, -0.04538211598992348, -0.030004402622580528, 0.04249507188796997, 0.06136051565408707, 0.01573854871094227, 0.04617364704608917, -0.02365618757903576, -0.02772068977355957, 0.07869653403759003, -0.14465086162090302, -0.03225550428032875, 0.009502166882157326, -0.030133754014968872, 0.05407314747571945, -0.022606993094086647, 0.025267889723181725, -0.09484191238880157, 0.05275505781173706, -0.09309876710176468, 0.015341418795287609, -0.03994688764214516, -0.13512912392616272, 0.039457790553569794, -0.0036687457468360662, -0.01406197901815176, -0.12786255776882172, -0.1884189248085022, -0.06911540031433105, -0.018809707835316658, -0.02908303402364254, 0.0009287158027291298, -0.04284804314374924, -0.03230789676308632, -0.0009852354414761066, -0.00907447375357151, 0.07453096657991409, -0.07364937663078308, 0.03672296553850174, 0.01885894313454628, 0.03538491949439049, 0.048528797924518585, 0.02300771325826645, -0.08855453878641129, 0.007859001867473125, -0.19586919248104095, 0.07917758822441101, -0.09202078729867935, -0.028630372136831284, -0.1681988388299942, -0.07167384773492813, 0.004165246617048979, 0.036687977612018585, 0.011195043101906776, 0.19063664972782135, -0.15422306954860687, -0.07232622057199478, 0.17202290892601013, -0.1366683542728424, -0.07409042119979858, 0.09872192144393921, -0.0024950657971203327, 0.05295003205537796, 0.06852338463068008, 0.09493208676576614, 0.014046325348317623, -0.10236319154500961, -0.1042974665760994, 0.038755785673856735, -0.04669509455561638, 0.03383014723658562, 0.14717692136764526, -0.04979848861694336, -0.007033463567495346, -0.03301052376627922, -0.07229963690042496, -0.061372313648462296, -0.06917070597410202, -0.014514966867864132, 0.014452483505010605, 0.024546636268496513, -0.017484430223703384, -0.011045522056519985, 0.02824225276708603, -0.02245512418448925, -0.08469744771718979, 0.10468625277280807, 0.08053454011678696, -0.013624956831336021, 0.02989398129284382, -0.06481290608644485, 0.07545405626296997, -0.14655271172523499, 0.011904168874025345, -0.17814652621746063, 0.003918274771422148, -0.035195495933294296, -0.01831330917775631, 0.08493191003799438, 0.11567158997058868, 0.03843705728650093, -0.0003502104082144797, -0.05371128395199776, -0.017226506024599075, -0.020814118906855583, 0.012539055198431015, -0.04840230941772461, -0.14780540764331818, -0.0603923462331295, -0.047830671072006226, 0.038078270852565765, -0.1947949230670929, 0.048610471189022064, 0.13488981127738953, 0.09240087866783142, -0.02021961659193039, -0.01544780284166336, 0.037636157125234604, 0.010542750358581543, -0.0391099713742733, -0.006070731673389673, 0.03888712450861931, -0.030429869890213013, -0.09433295577764511, 0.12121254205703735, -0.06951451301574707, 0.04550592228770256, 0.09678824245929718, -0.09512206166982651, -0.04221123456954956, -0.0015357627999037504, -0.04806359484791756, -0.018145881593227386, -0.06481865793466568, -0.10257669538259506, 0.08736004680395126, 0.03818660229444504, 0.08947106450796127, -0.12211883813142776, -0.05619807541370392, -0.011669391766190529, -0.04097792133688927, -0.038024209439754486, 0.14730796217918396, 0.13614684343338013, -0.22474130988121033, 0.09459280967712402, 0.09634040296077728, 0.056217752397060394, 0.11632884293794632, 0.03079274669289589, -0.09302976727485657, -0.03665049374103546, -0.037349577993154526, -0.011092516593635082, 0.022355396300554276, -0.13813695311546326, 0.018614528700709343, 0.03626350685954094, 0.06274086982011795, 0.048110686242580414, -0.08430922031402588, 0.023102769628167152, 0.034098707139492035, -0.054320819675922394, -0.06708180159330368, 0.05114087834954262, -0.01752246357500553, 0.11275601387023926, -0.013112070970237255, 0.054923027753829956, -0.02163437381386757, -0.03853237256407738, -0.1621776521205902, 0.24190585315227509, -0.07230739295482635, -0.1990939825773239, -0.14092570543289185, -0.03290741890668869, -0.004290411714464426, 0.006632637698203325, 0.04710689187049866, -0.06208227202296257, -0.057030946016311646, -0.05827367678284645, 0.15698522329330444, -0.05020911246538162, -0.025685248896479607, -0.026260916143655777, 0.013282785192131996, -0.034448087215423584, -0.11714126169681549, -0.03257855027914047, 0.000629653048235923, -0.03999554365873337, 0.05400373041629791, -0.09403865784406662, 0.08558535575866699, 0.14661449193954468, 0.0055513703264296055, 0.013002083636820316, -0.0360204242169857, 0.2374764382839203, -0.09815853834152222, 0.08083433657884598, 0.126445934176445, 0.02106372080743313, 0.04833296686410904, 0.136552095413208, 0.03837266191840172, -0.04840388149023056, 0.0078066689893603325, 0.02648116834461689, -0.028180690482258797, -0.2976570725440979, -0.03613896667957306, -0.040955014526844025, 0.006232014391571283, -0.023737939074635506, 0.03058096021413803, 0.0006151704001240432, 0.03875765949487686, 0.00014052286860533059, -0.06797593086957932, 0.07905148714780807, 0.08120037615299225, 0.17933610081672668, -0.03187602758407593, 0.1018315926194191, -0.05022238567471504, 0.014464686624705791, 0.0988161489367485, -0.0317094512283802, 0.07845721393823624, 0.03673075512051582, 0.10731338709592819, 0.0854557529091835, -0.06932787597179413, 0.060933079570531845, 0.09500563889741898, 0.03541182726621628, -0.01575794629752636, -0.05141366273164749, -0.07150494307279587, -0.013204546645283699, 0.07754538208246231, -0.0006494509871117771, 0.030169246718287468, -0.05510684847831726, 0.00025268536410294473, 0.0796356201171875, 0.12233056128025055, 0.11905698478221893, -0.18897558748722076, -0.0825224369764328, 0.03925305977463722, -0.08570335805416107, -0.028067778795957565, 0.013382963836193085, 0.050236549228429794, -0.12715069949626923, 0.08691269159317017, -0.003926199395209551, 0.0817563533782959, -0.04231928661465645, 0.013595031574368477, -0.05688280612230301, 0.04757945239543915, -0.06635307520627975, 0.10871340334415436, -0.21348366141319275, 0.24860087037086487, 0.028948677703738213, 0.003447059541940689, -0.032121725380420685, 0.002351206261664629, 0.027420859783887863, 0.12465158104896545, 0.16898320615291595, 0.038206346333026886, -0.11525604128837585, -0.06141266971826553, -0.057370707392692566, 0.019095610827207565, 0.055444035679101944, -0.055309657007455826, 0.10468124598264694, -0.01123357005417347, -0.0286799855530262, -0.05152622610330582, 0.08912047743797302, -0.08574827760457993, -0.059975869953632355, 0.055604953318834305, -0.04861712083220482, 0.0570690780878067, -0.039388760924339294, -0.07933654636144638, -0.008551347069442272, 0.20244550704956055, -0.09379859268665314, -0.08874478191137314, -0.1119011789560318, 0.10880919545888901, 0.1179039403796196, -0.10153573006391525, -0.01801026239991188, -0.00041161212720908225, 0.07553088665008545, -0.06341224163770676, -0.027070054784417152, 0.05060676857829094, -0.09900392591953278, -0.1290867030620575, -0.006477785762399435, 0.17706312239170074, 0.028443120419979095, 0.07298622280359268, -0.01846177503466606, 0.04566960036754608, 0.005181921645998955, -0.09748733788728714, 0.031605977565050125, 0.07200680673122406, 0.03631093353033066, 0.10945894569158554, -0.014866901561617851, -0.1495436280965805, -0.08084730058908463, -0.08226050436496735, 0.13337749242782593, 0.2415807992219925, -0.022731944918632507, 0.0059094661846756935, 0.14988526701927185, -0.12702897191047668, -0.17947930097579956, -0.025609849020838737, -0.017975356429815292, 0.03621434420347214, -0.03056873381137848, -0.0697714239358902, 0.09296654909849167, 0.11502140760421753, 0.01461400929838419, 0.012918460182845592, -0.2774164080619812, -0.14455026388168335, 0.05357075855135918, 0.04011279344558716, 0.06414824724197388, -0.12101832777261734, -0.08832836896181107, -0.017887132242321968, -0.15037941932678223, -0.046370457857847214, 0.024574926123023033, 0.07341130077838898, -0.0012781245168298483, 0.003961221314966679, -0.0020799252670258284, -0.062235649675130844, 0.1423482745885849, 0.07378239184617996, -0.021500710397958755, -0.049922551959753036, -0.0834885984659195, 0.046477679163217545, -0.01795426569879055, 0.0944494754076004, 0.028150035068392754, 0.030609557405114174, -0.1693129688501358, -0.06600363552570343, -0.09165352582931519, 0.04092852771282196, -0.04155939444899559, -0.05823942646384239, -0.05174296349287033, 0.05033516883850098, -0.015236235223710537, -0.02891722321510315, 0.0866067111492157, -0.18337032198905945, 0.08640597015619278, 0.12675884366035461, 0.1816687285900116, -0.051380548626184464, -0.08014682680368423, 0.0003993920690845698, -0.028780536726117134, 0.020264262333512306, -0.14086386561393738, 0.01714113913476467, 0.1891166865825653, 0.03878457844257355, 0.1366960108280182, 0.03663665056228638, -0.12230265885591507, -0.0070900991559028625, 0.05075749382376671, -0.10571947693824768, -0.13921815156936646, -0.061642084270715714, -0.06681475043296814, -0.06401066482067108, 0.011244140565395355, 0.18681524693965912, -0.05103781446814537, -0.03454102575778961, 0.013388270512223244, 0.04745827987790108, -0.034712500870227814, 0.14446942508220673, 0.043610721826553345, 0.10284263640642166, -0.04293866083025932, 0.06577582657337189, 0.04928402230143547, -0.14590518176555634, -0.0013801095774397254, 0.13444030284881592, -0.10592292994260788, -0.10968507081270218, -0.08309195190668106, 0.04468484967947006, -0.13626843690872192, -0.040553636848926544, -0.05217250436544418, -0.09318459033966064, 0.11313868314027786, 0.26966196298599243, 0.07630538195371628, 0.005832205060869455, 0.002672710223123431, -0.04942847415804863, -0.03453722968697548, 0.07400050014257431, 0.06338871270418167, 0.0057105072773993015, -0.05523018166422844, 0.04828348383307457, 0.005404562223702669, 0.06822384148836136, -0.03664863854646683, -0.009829435497522354, -0.11073896288871765, -0.00211991835385561, -0.12346631288528442, 0.02706395648419857, -0.11411421746015549, -0.0448688268661499, -0.0232267864048481, -0.0666583776473999, -0.06368216872215271, 0.001143145840615034, -0.06493087112903595, 0.00845089741051197, -0.03337518870830536, 0.07265631854534149, -0.11740630120038986, -0.01084820181131363, 0.0683927983045578, -0.04303618520498276, 0.08845889568328857, 0.05756441876292229, -0.025138691067695618, 0.04297187179327011, -0.03983324021100998, 0.0012445427710190415, 0.03553915396332741, 0.027929389849305153, 0.010498234070837498, -0.16496944427490234, 0.03163580223917961, 0.014075500890612602, 0.010047223418951035, 0.02202426828444004, -0.020190052688121796, -0.06885429471731186, 0.08063716441392899, -0.015062405727803707, -0.009842400439083576, -0.0858205035328865, 0.03400293365120888, 0.06472918391227722, 0.011953581124544144, 0.09104674309492111, -0.08441045135259628, 0.08935285359621048, -0.19628871977329254, 0.06105251982808113, -0.02586461789906025, -0.06568820029497147, 0.007893064059317112, -0.04067995399236679, 0.05260491371154785, -0.07871285825967789, 0.13803057372570038, -0.03530099615454674, 0.09094929695129395, 0.0954563170671463, -0.01816081814467907, -0.0009340798133052886, 0.01109381951391697, 0.08991771936416626, 0.035627368837594986, -0.017645999789237976, -0.006560726091265678, 0.05828696861863136, 0.01686730422079563, -0.026306962594389915, 0.18915627896785736, 0.15456297993659973, 0.06830284744501114, 0.11857038736343384, -0.020357679575681686, -0.053671762347221375, -0.06374875456094742, -0.060787394642829895, 0.006244149059057236, 0.028175652027130127, -0.011636503040790558, 0.13011476397514343, 0.24591515958309174, -0.1355893313884735, 0.055136773735284805, -0.0183431226760149, -0.058465663343667984, -0.15457986295223236, -0.15160225331783295, -0.09146776795387268, -0.052706457674503326, 0.01212689932435751, -0.14993134140968323, 0.031088154762983322, 0.0800442025065422, 0.08113407343626022, 0.03167780116200447, 0.10042259097099304, -0.04176720231771469, -0.10184504091739655, 0.054506558924913406, -0.03298872336745262, 0.01669522561132908, -0.01713918149471283, 0.05466693639755249, 0.0629722997546196, -0.031763188540935516, 0.004489050712436438, 0.044671256095170975, 0.013491661287844181, 0.0033857894595712423, -0.056724876165390015, -0.0799194872379303, -0.037689439952373505, 0.08722206205129623, 0.019582660868763924, 0.040729399770498276, 0.043964482843875885, -0.03370688483119011, 0.02483776956796646, 0.1747635006904602, -0.04714681953191757, -0.14979234337806702, -0.10011957585811615, 0.27194100618362427, 0.017094040289521217, 0.060714516788721085, 0.011510791257023811, -0.08254165202379227, 0.04262172058224678, 0.17997372150421143, 0.25481027364730835, 0.022738616913557053, 0.002361099235713482, 0.040480490773916245, 0.02064366266131401, 0.08538824319839478, -0.04714485630393028, 0.08051390945911407, 0.20523370802402496, -0.10947464406490326, 0.06159772723913193, -0.05394895002245903, 0.011531834490597248, 0.05711811035871506, 0.0914478749036789, 0.05253611505031586, -0.02448766864836216, 0.01770862191915512, 0.11310328543186188, -0.0520821250975132, -0.14464548230171204, -0.046666719019412994, -0.07635517418384552, -0.07667813450098038, -0.017557887360453606, -0.08716671168804169, 0.04671185836195946, 0.04439183697104454, -0.005941525101661682, -0.06776446849107742, 0.09240076690912247, 0.05372226610779762, -0.15006721019744873, -0.08835455030202866, 0.11770972609519958, -0.051096998155117035, 0.04880569130182266, -0.0333821177482605, 0.09975002706050873, 0.12282567471265793, -0.02882368303835392, -0.07001163065433502, 0.09133550524711609, 0.04952778294682503, -0.008223874494433403, 0.02949834056198597, 0.06134914234280586, 0.035606276243925095, 0.13139943778514862, 0.03864523023366928, -0.17284081876277924, 0.0519588403403759, 0.04460575431585312, -0.04622754454612732, -0.06262288242578506, 0.10816153138875961, -0.1159312054514885, 0.11276450753211975, 0.1413266509771347, -0.0021738146897405386, 0.002547777956351638, -0.05706258490681648, 0.05900374799966812, 0.02661932073533535, 0.08303573727607727, -0.0012838876573368907, -0.13310682773590088, 0.047784414142370224, -0.09077069908380508, 0.01665882207453251, -0.25229036808013916, -0.06172454357147217, 0.010941039770841599, 0.002646885346621275, -0.0419132262468338, 0.09119468182325363, 0.07950460910797119, 0.045082613825798035, -0.02342906780540943, -0.13387075066566467, -0.0014576348476111889, 0.1547166109085083, -0.15400967001914978, -0.05059441924095154 ]
null
null
transformers
# mT5-multilingual-XLSum This repository contains the mT5 checkpoint finetuned on the 45 languages of [XL-Sum](https://huggingface.co/datasets/csebuetnlp/xlsum) dataset. For finetuning details and scripts, see the [paper](https://aclanthology.org/2021.findings-acl.413/) and the [official repository](https://github.com/csebuetnlp/xl-sum). ## Using this model in `transformers` (tested on 4.11.0.dev0) ```python import re from transformers import AutoTokenizer, AutoModelForSeq2SeqLM WHITESPACE_HANDLER = lambda k: re.sub('\s+', ' ', re.sub('\n+', ' ', k.strip())) article_text = """Videos that say approved vaccines are dangerous and cause autism, cancer or infertility are among those that will be taken down, the company said. The policy includes the termination of accounts of anti-vaccine influencers. Tech giants have been criticised for not doing more to counter false health information on their sites. In July, US President Joe Biden said social media platforms were largely responsible for people's scepticism in getting vaccinated by spreading misinformation, and appealed for them to address the issue. YouTube, which is owned by Google, said 130,000 videos were removed from its platform since last year, when it implemented a ban on content spreading misinformation about Covid vaccines. In a blog post, the company said it had seen false claims about Covid jabs "spill over into misinformation about vaccines in general". The new policy covers long-approved vaccines, such as those against measles or hepatitis B. "We're expanding our medical misinformation policies on YouTube with new guidelines on currently administered vaccines that are approved and confirmed to be safe and effective by local health authorities and the WHO," the post said, referring to the World Health Organization.""" model_name = "csebuetnlp/mT5_multilingual_XLSum" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForSeq2SeqLM.from_pretrained(model_name) input_ids = tokenizer( [WHITESPACE_HANDLER(article_text)], return_tensors="pt", padding="max_length", truncation=True, max_length=512 )["input_ids"] output_ids = model.generate( input_ids=input_ids, max_length=84, no_repeat_ngram_size=2, num_beams=4 )[0] summary = tokenizer.decode( output_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False ) print(summary) ``` ## Benchmarks Scores on the XL-Sum test sets are as follows: Language | ROUGE-1 / ROUGE-2 / ROUGE-L ---------|---------------------------- Amharic | 20.0485 / 7.4111 / 18.0753 Arabic | 34.9107 / 14.7937 / 29.1623 Azerbaijani | 21.4227 / 9.5214 / 19.3331 Bengali | 29.5653 / 12.1095 / 25.1315 Burmese | 15.9626 / 5.1477 / 14.1819 Chinese (Simplified) | 39.4071 / 17.7913 / 33.406 Chinese (Traditional) | 37.1866 / 17.1432 / 31.6184 English | 37.601 / 15.1536 / 29.8817 French | 35.3398 / 16.1739 / 28.2041 Gujarati | 21.9619 / 7.7417 / 19.86 Hausa | 39.4375 / 17.6786 / 31.6667 Hindi | 38.5882 / 16.8802 / 32.0132 Igbo | 31.6148 / 10.1605 / 24.5309 Indonesian | 37.0049 / 17.0181 / 30.7561 Japanese | 48.1544 / 23.8482 / 37.3636 Kirundi | 31.9907 / 14.3685 / 25.8305 Korean | 23.6745 / 11.4478 / 22.3619 Kyrgyz | 18.3751 / 7.9608 / 16.5033 Marathi | 22.0141 / 9.5439 / 19.9208 Nepali | 26.6547 / 10.2479 / 24.2847 Oromo | 18.7025 / 6.1694 / 16.1862 Pashto | 38.4743 / 15.5475 / 31.9065 Persian | 36.9425 / 16.1934 / 30.0701 Pidgin | 37.9574 / 15.1234 / 29.872 Portuguese | 37.1676 / 15.9022 / 28.5586 Punjabi | 30.6973 / 12.2058 / 25.515 Russian | 32.2164 / 13.6386 / 26.1689 Scottish Gaelic | 29.0231 / 10.9893 / 22.8814 Serbian (Cyrillic) | 23.7841 / 7.9816 / 20.1379 Serbian (Latin) | 21.6443 / 6.6573 / 18.2336 Sinhala | 27.2901 / 13.3815 / 23.4699 Somali | 31.5563 / 11.5818 / 24.2232 Spanish | 31.5071 / 11.8767 / 24.0746 Swahili | 37.6673 / 17.8534 / 30.9146 Tamil | 24.3326 / 11.0553 / 22.0741 Telugu | 19.8571 / 7.0337 / 17.6101 Thai | 37.3951 / 17.275 / 28.8796 Tigrinya | 25.321 / 8.0157 / 21.1729 Turkish | 32.9304 / 15.5709 / 29.2622 Ukrainian | 23.9908 / 10.1431 / 20.9199 Urdu | 39.5579 / 18.3733 / 32.8442 Uzbek | 16.8281 / 6.3406 / 15.4055 Vietnamese | 32.8826 / 16.2247 / 26.0844 Welsh | 32.6599 / 11.596 / 26.1164 Yoruba | 31.6595 / 11.6599 / 25.0898 ## Citation If you use this model, please cite the following paper: ``` @inproceedings{hasan-etal-2021-xl, title = "{XL}-Sum: Large-Scale Multilingual Abstractive Summarization for 44 Languages", author = "Hasan, Tahmid and Bhattacharjee, Abhik and Islam, Md. Saiful and Mubasshir, Kazi and Li, Yuan-Fang and Kang, Yong-Bin and Rahman, M. Sohel and Shahriyar, Rifat", booktitle = "Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021", month = aug, year = "2021", address = "Online", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2021.findings-acl.413", pages = "4693--4703", } ```
{"language": ["am", "ar", "az", "bn", "my", "zh", "en", "fr", "gu", "ha", "hi", "ig", "id", "ja", "rn", "ko", "ky", "mr", "ne", "om", "ps", "fa", "pcm", "pt", "pa", "ru", "gd", "sr", "si", "so", "es", "sw", "ta", "te", "th", "ti", "tr", "uk", "ur", "uz", "vi", "cy", "yo"], "tags": ["summarization", "mT5"], "datasets": ["csebuetnlp/xlsum"], "licenses": ["cc-by-nc-sa-4.0"], "widget": [{"text": "Videos that say approved vaccines are dangerous and cause autism, cancer or infertility are among those that will be taken down, the company said. The policy includes the termination of accounts of anti-vaccine influencers. Tech giants have been criticised for not doing more to counter false health information on their sites. In July, US President Joe Biden said social media platforms were largely responsible for people's scepticism in getting vaccinated by spreading misinformation, and appealed for them to address the issue. YouTube, which is owned by Google, said 130,000 videos were removed from its platform since last year, when it implemented a ban on content spreading misinformation about Covid vaccines. In a blog post, the company said it had seen false claims about Covid jabs \"spill over into misinformation about vaccines in general\". The new policy covers long-approved vaccines, such as those against measles or hepatitis B. \"We're expanding our medical misinformation policies on YouTube with new guidelines on currently administered vaccines that are approved and confirmed to be safe and effective by local health authorities and the WHO,\" the post said, referring to the World Health Organization."}], "model-index": [{"name": "csebuetnlp/mT5_multilingual_XLSum", "results": [{"task": {"type": "summarization", "name": "Summarization"}, "dataset": {"name": "xsum", "type": "xsum", "config": "default", "split": "test"}, "metrics": [{"type": "rouge", "value": 36.5002, "name": "ROUGE-1", "verified": true}, {"type": "rouge", "value": 13.934, "name": "ROUGE-2", "verified": true}, {"type": "rouge", "value": 28.9876, "name": "ROUGE-L", "verified": true}, {"type": "rouge", "value": 28.9958, "name": "ROUGE-LSUM", "verified": true}, {"type": "loss", "value": 2.0674800872802734, "name": "loss", "verified": true}, {"type": "gen_len", "value": 26.9733, "name": "gen_len", "verified": true}]}]}]}
summarization
csebuetnlp/mT5_multilingual_XLSum
[ "transformers", "pytorch", "mt5", "text2text-generation", "summarization", "mT5", "am", "ar", "az", "bn", "my", "zh", "en", "fr", "gu", "ha", "hi", "ig", "id", "ja", "rn", "ko", "ky", "mr", "ne", "om", "ps", "fa", "pcm", "pt", "pa", "ru", "gd", "sr", "si", "so", "es", "sw", "ta", "te", "th", "ti", "tr", "uk", "ur", "uz", "vi", "cy", "yo", "dataset:csebuetnlp/xlsum", "model-index", "autotrain_compatible", "endpoints_compatible", "has_space", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "am", "ar", "az", "bn", "my", "zh", "en", "fr", "gu", "ha", "hi", "ig", "id", "ja", "rn", "ko", "ky", "mr", "ne", "om", "ps", "fa", "pcm", "pt", "pa", "ru", "gd", "sr", "si", "so", "es", "sw", "ta", "te", "th", "ti", "tr", "uk", "ur", "uz", "vi", "cy", "yo" ]
TAGS #transformers #pytorch #mt5 #text2text-generation #summarization #mT5 #am #ar #az #bn #my #zh #en #fr #gu #ha #hi #ig #id #ja #rn #ko #ky #mr #ne #om #ps #fa #pcm #pt #pa #ru #gd #sr #si #so #es #sw #ta #te #th #ti #tr #uk #ur #uz #vi #cy #yo #dataset-csebuetnlp/xlsum #model-index #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us
mT5-multilingual-XLSum ====================== This repository contains the mT5 checkpoint finetuned on the 45 languages of XL-Sum dataset. For finetuning details and scripts, see the paper and the official repository. Using this model in 'transformers' (tested on 4.11.0.dev0) ---------------------------------------------------------- Benchmarks ---------- Scores on the XL-Sum test sets are as follows: If you use this model, please cite the following paper:
[]
[ "TAGS\n#transformers #pytorch #mt5 #text2text-generation #summarization #mT5 #am #ar #az #bn #my #zh #en #fr #gu #ha #hi #ig #id #ja #rn #ko #ky #mr #ne #om #ps #fa #pcm #pt #pa #ru #gd #sr #si #so #es #sw #ta #te #th #ti #tr #uk #ur #uz #vi #cy #yo #dataset-csebuetnlp/xlsum #model-index #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n" ]
[ 168 ]
[ "passage: TAGS\n#transformers #pytorch #mt5 #text2text-generation #summarization #mT5 #am #ar #az #bn #my #zh #en #fr #gu #ha #hi #ig #id #ja #rn #ko #ky #mr #ne #om #ps #fa #pcm #pt #pa #ru #gd #sr #si #so #es #sw #ta #te #th #ti #tr #uk #ur #uz #vi #cy #yo #dataset-csebuetnlp/xlsum #model-index #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n" ]
[ -0.08646338433027267, -0.02580511011183262, -0.006402633152902126, 0.014841850847005844, 0.11805697530508041, 0.03955939784646034, 0.07664599269628525, 0.1215146854519844, 0.04115590453147888, 0.06827494502067566, 0.1432306319475174, 0.13022325932979584, 0.066834457218647, 0.09859861433506012, -0.020829707384109497, -0.2796865999698639, 0.042838387191295624, 0.01805652119219303, -0.06560946255922318, 0.14316698908805847, 0.07420980930328369, -0.07293959707021713, 0.12529557943344116, -0.059878136962652206, -0.03137265890836716, -0.009355505928397179, -0.03475627303123474, -0.03452621027827263, 0.10409579426050186, 0.08351270854473114, 0.0011220359010621905, 0.10217955708503723, -0.009927472099661827, -0.19220183789730072, 0.05530533567070961, 0.010962228290736675, -0.06121176853775978, 0.06295283883810043, 0.06624240428209305, -0.05702592432498932, 0.20684295892715454, 0.010274659842252731, -0.10116490721702576, 0.07062696665525436, -0.13991835713386536, -0.10073749721050262, -0.06109984964132309, 0.09174918383359909, 0.022892916575074196, 0.0860346108675003, -0.041013505309820175, 0.12542006373405457, -0.08066155761480331, 0.07819471508264542, 0.25837138295173645, -0.2753373384475708, -0.013708444312214851, 0.10153517872095108, 0.06425516307353973, 0.11544177681207657, -0.0648888573050499, 0.07164723426103592, 0.03990350291132927, 0.011139418929815292, 0.014654522761702538, -0.05404828488826752, 0.09234413504600525, 0.06270287185907364, -0.12023982405662537, -0.02208303101360798, 0.2882609963417053, 0.00890812836587429, 0.043102510273456573, 0.054322902113199234, -0.10071641206741333, -0.11903024464845657, -0.03480162099003792, 0.0030002507846802473, -0.032737456262111664, 0.035928044468164444, -0.016664890572428703, -0.007063908036798239, -0.11757451295852661, -0.012095363810658455, -0.12029002606868744, 0.08956138789653778, 0.020348967984318733, 0.009425998665392399, -0.05437120050191879, 0.08388996869325638, -0.0407414436340332, -0.1163187250494957, 0.01163626927882433, -0.05937284231185913, 0.03361080586910248, 0.0175373163074255, 0.0023424476385116577, -0.02653774805366993, 0.08220241963863373, 0.06497935205698013, -0.09780006855726242, 0.045716144144535065, 0.05472186952829361, 0.09946230053901672, 0.04137740656733513, 0.03110693208873272, -0.13020409643650055, -0.12313950061798096, -0.06357461214065552, -0.004433367867022753, -0.023852603510022163, -0.04243079200387001, -0.10859720408916473, -0.02107400819659233, 0.0749984160065651, 0.02045246958732605, 0.03208538889884949, 0.11141199618577957, -0.03804052993655205, -0.015583659522235394, 0.059466972947120667, -0.03752642124891281, -0.01465922873467207, 0.054823342710733414, -0.015965940430760384, 0.13291332125663757, -0.01358629111200571, 0.005811266601085663, -0.07705327868461609, 0.0746295377612114, -0.09481575340032578, 0.011774844489991665, -0.02097959630191326, -0.07440358400344849, 0.044464945793151855, -0.08328255265951157, 0.002461118856444955, -0.15154998004436493, -0.13312439620494843, -0.011392655782401562, -0.016019552946090698, -0.0510411262512207, 0.006991948001086712, -0.04892702028155327, -0.08705591410398483, 0.04527730867266655, -0.04525529593229294, 0.06288579851388931, -0.09293904155492783, 0.11157049238681793, -0.037971239537000656, 0.09139513224363327, -0.06133711710572243, 0.05996314808726311, -0.06494206935167313, 0.015841085463762283, -0.04956357553601265, 0.02019020915031433, -0.03552056849002838, 0.016650451347231865, -0.08050727099180222, -0.07340698689222336, -0.10303705930709839, 0.042004916816949844, -0.010168171487748623, 0.20999889075756073, -0.12815506756305695, -0.1229950413107872, 0.1728687435388565, -0.07732796669006348, -0.11980258673429489, 0.13947376608848572, 0.04672984033823013, 0.012982234358787537, 0.04674148187041283, 0.12775863707065582, 0.017619166523218155, -0.056185781955718994, -0.08064185827970505, 0.08863220363855362, -0.0008219906012527645, -0.021924324333667755, 0.12521758675575256, 0.04606008902192116, -0.029642032459378242, 0.04374675825238228, 0.025907909497618675, 0.0524730421602726, -0.09625004976987839, -0.0479864627122879, 0.02365446835756302, -0.025242431089282036, 0.06894630938768387, 0.05423135682940483, 0.04871244728565216, -0.04982810840010643, -0.05945693701505661, -0.06991522014141083, 0.0721476674079895, 0.01388605684041977, 0.035613901913166046, -0.05702692270278931, 0.1623762547969818, -0.048149287700653076, 0.03330444544553757, -0.16488759219646454, 0.0654362216591835, -0.05291084200143814, 0.08847109973430634, 0.04705169051885605, 0.1513070911169052, 0.08368470519781113, 0.020035747438669205, -0.07014743238687515, -0.056721754372119904, 0.07458966970443726, -0.0036694095470011234, -0.07594125717878342, -0.2001580148935318, 0.03095705807209015, -0.05246477946639061, 0.0541391484439373, -0.12220260500907898, 0.043900344520807266, 0.1203746572136879, 0.13357628881931305, -0.013496517203748226, 0.028763482347130775, -0.008525585755705833, 0.0537622794508934, -0.07923253625631332, 0.030123241245746613, 0.09258973598480225, -0.01786765456199646, -0.09028387814760208, 0.1356632113456726, -0.13933894038200378, 0.19381265342235565, 0.18729551136493683, -0.16480375826358795, -0.0106788519769907, 0.00305515737272799, -0.04154086485505104, -0.005570478271692991, 0.05160908401012421, -0.058879874646663666, 0.021410977467894554, 0.012488342821598053, 0.13865934312343597, -0.07460424304008484, -0.04306861758232117, 0.026865415275096893, -0.05621049180626869, -0.06938046962022781, 0.17617401480674744, 0.13984617590904236, -0.184221088886261, 0.2193295657634735, 0.22216098010540009, 0.07638496905565262, 0.20972466468811035, 0.02146581932902336, -0.07523148506879807, -0.0067572640255093575, -0.03842756524682045, -0.03014184907078743, 0.05528867989778519, -0.17208194732666016, -0.004855314735323191, 0.05302481725811958, 0.028850145637989044, 0.07292649149894714, -0.09483453631401062, -0.03986866772174835, -0.020175183191895485, -0.012613067403435707, -0.005751879420131445, 0.09126702696084976, -0.020969988778233528, 0.14618010818958282, 0.015278870239853859, -0.01508206408470869, -0.021914981305599213, 0.012596981599926949, -0.11008677631616592, 0.2006336897611618, -0.1434328556060791, -0.2647164463996887, -0.06791103631258011, -0.10699635744094849, -0.03864356875419617, 0.010203001089394093, 0.06351496279239655, -0.1494152545928955, -0.005632020533084869, -0.0401131734251976, 0.1266208440065384, -0.12256360799074173, 0.02756439335644245, -0.08154728263616562, 0.04155242443084717, -0.06654123961925507, -0.05503842979669571, -0.044777773320674896, -0.0032507332507520914, -0.03223313391208649, 0.15086406469345093, -0.14564742147922516, 0.10209407657384872, 0.14454345405101776, 0.05491498485207558, 0.04454679787158966, -0.01745566725730896, 0.18424974381923676, -0.11019963026046753, 0.0546313039958477, 0.12980739772319794, 0.03108617290854454, 0.08790569007396698, 0.11257853358983994, 0.012022046372294426, -0.03635305166244507, -0.0031973563600331545, -0.025873683393001556, -0.050823986530303955, -0.19269588589668274, -0.12297598272562027, -0.09652827680110931, 0.14376114308834076, -0.013132900930941105, 0.07093209773302078, 0.06360015273094177, 0.06547761708498001, -0.0629582554101944, -0.061056386679410934, -0.012075182981789112, 0.03522107005119324, 0.07370967417955399, -0.0680118203163147, 0.10450724512338638, -0.04872577264904976, -0.04218374565243721, 0.10744914412498474, 0.024929512292146683, 0.031761534512043, 0.03159615397453308, 0.06261876225471497, 0.05901128798723221, 0.09176083654165268, 0.08170662075281143, 0.0828763023018837, 0.022844430059194565, -0.03887558355927467, -0.04769158363342285, -0.05327599495649338, -0.038153111934661865, 0.04842197522521019, 0.07148966938257217, -0.019605575129389763, -0.05496722459793091, -0.018697328865528107, 0.09307091683149338, 0.02155941165983677, 0.09615350514650345, -0.20223158597946167, -0.04868096485733986, 0.049459151923656464, 0.024035051465034485, -0.05029739439487457, 0.03762015700340271, 0.07877764850854874, -0.10010737180709839, 0.048669926822185516, 0.019626609981060028, 0.1006847396492958, -0.046271178871393204, 0.07306604832410812, -0.011013571172952652, 0.0003125910006929189, -0.027551395818591118, 0.08456190675497055, -0.2643141448497772, 0.2700752913951874, 0.021132059395313263, -0.04820789396762848, -0.03734063729643822, -0.05281893163919449, 0.03586646169424057, 0.17501327395439148, 0.11537172645330429, 0.054651711136102676, -0.07836266607046127, -0.14567236602306366, -0.03072880581021309, 0.011262441985309124, 0.12249168008565903, -0.0760888084769249, 0.06281032413244247, -0.026217155158519745, 0.0003677707863971591, -0.019980330020189285, 0.0870484709739685, -0.0485142283141613, -0.10698268562555313, 0.10248531401157379, -0.004980690777301788, 0.039586085826158524, -0.022746069356799126, -0.07157531380653381, -0.1603529155254364, 0.08944152295589447, -0.10459475964307785, -0.051959335803985596, -0.10517973452806473, 0.01071182545274496, 0.06875112652778625, -0.12003149837255478, -0.09473168849945068, -0.030588233843445778, 0.0022939175833016634, -0.05684424936771393, -0.09701033681631088, 0.08882134407758713, -0.08890765905380249, -0.16795499622821808, -0.029955841600894928, 0.1478097140789032, 0.016809793189167976, 0.09889959543943405, -0.06433264166116714, 0.032100699841976166, -0.07950994372367859, -0.09678886085748672, 0.08184576779603958, 0.03300520405173302, 0.05671609193086624, 0.05158074572682381, -0.05103420838713646, 0.0024466700851917267, -0.04485040903091431, -0.09979300945997238, 0.1561969667673111, 0.2654974162578583, -0.03831382840871811, 0.08190181851387024, 0.1898847222328186, -0.06698823720216751, -0.32883501052856445, -0.1020924523472786, -0.12290880084037781, 0.006095423363149166, -0.054667167365550995, -0.1428288221359253, -0.007802795618772507, 0.020302485674619675, 0.008326495997607708, 0.061280831694602966, -0.28666168451309204, -0.0804954245686531, 0.11632534861564636, 0.026943285018205643, 0.20262247323989868, -0.1612684428691864, -0.04419039189815521, -0.031638771295547485, -0.11420679092407227, -0.0013124970719218254, -0.08047515153884888, 0.1119115874171257, -0.05642223358154297, 0.03384925052523613, 0.0031363742891699076, -0.0412670336663723, 0.14393746852874756, -0.005251957569271326, -0.01978449523448944, -0.09357107430696487, -0.06671282649040222, 0.05618290603160858, 0.020172258839011192, -0.008968192152678967, -0.13067586719989777, 0.002749864710494876, -0.10936305671930313, 0.001712290570139885, -0.10428362339735031, 0.042806562036275864, -0.025254538282752037, -0.07440433651208878, -0.09491828083992004, 0.06143561005592346, 0.015914084389805794, -0.015384151600301266, 0.15805964171886444, -0.09795594960451126, 0.17690154910087585, 0.14932550489902496, 0.12163546681404114, -0.10477377474308014, -0.0029790252447128296, -0.056381579488515854, -0.04294315725564957, 0.03388737142086029, -0.15141144394874573, -0.006584265735000372, 0.15575098991394043, -0.0026862001977860928, 0.08538544178009033, 0.0842437893152237, -0.08821997046470642, -0.0017864579567685723, 0.12189710140228271, -0.18256902694702148, -0.1587209552526474, -0.07199584692716599, -0.04493510350584984, 0.025237757712602615, 0.032388217747211456, 0.13786470890045166, -0.06252606213092804, 0.007686957251280546, -0.0044171069748699665, 0.03348146751523018, -0.0261247456073761, 0.14706666767597198, 0.054604947566986084, 0.04910893738269806, -0.10807984322309494, 0.0745367780327797, -0.018365714699029922, -0.09974934160709381, -0.019174844026565552, 0.14350084960460663, -0.13603082299232483, -0.1390450894832611, -0.04728773981332779, 0.07022230327129364, -0.05800842121243477, -0.04889875277876854, -0.06902402639389038, -0.11054613441228867, 0.10943786799907684, 0.18772326409816742, 0.06845293194055557, 0.03541647642850876, -0.006301398854702711, -0.005879333708435297, -0.0032616930548101664, 0.07122199237346649, 0.047307852655649185, 0.008813084103167057, -0.11481264978647232, 0.06905263662338257, -0.01946692354977131, 0.14809025824069977, -0.06259697675704956, -0.038330283015966415, -0.13862833380699158, 0.013722794130444527, -0.11192870885133743, -0.07710595428943634, -0.12475071102380753, -0.08276855945587158, -0.034661807119846344, -0.10541734099388123, -0.060688551515340805, -0.05759627744555473, -0.08924674242734909, -0.0077257161028683186, -0.015023157000541687, 0.07556517422199249, -0.05175592750310898, 0.011267132125794888, 0.103173166513443, -0.05046331137418747, 0.08374372124671936, 0.10673636198043823, -0.0431801974773407, 0.09389308094978333, -0.13563120365142822, -0.01780153624713421, 0.040358442813158035, 0.043687961995601654, 0.0316753163933754, 0.02403990365564823, -0.01372868474572897, -0.020762581378221512, 0.027986908331513405, 0.06853371858596802, 0.04971487075090408, -0.05410198122262955, 0.15225590765476227, -0.09667443484067917, -0.056180987507104874, -0.07200424373149872, 0.0443526990711689, 0.06957680732011795, 0.04593314230442047, 0.08024001121520996, -0.0779152438044548, 0.05859170854091644, -0.13844361901283264, 0.06522088497877121, 0.008075354620814323, -0.15024586021900177, -0.008301224559545517, -0.061293747276067734, 0.05687815696001053, -0.053940434008836746, 0.13230113685131073, -0.0344119518995285, -0.07035045325756073, 0.041370514780282974, 0.031971968710422516, -0.026774905622005463, -0.007100245915353298, 0.10602365434169769, 0.08839527517557144, -0.04227568954229355, -0.0873025432229042, 0.09140364825725555, 0.02627820521593094, 0.019871899858117104, 0.15128456056118011, 0.09062705188989639, 0.0976504534482956, 0.08410143107175827, -0.046059153974056244, -0.03251180425286293, -0.0656374990940094, -0.052535828202962875, -0.03929354250431061, -0.0025122733786702156, -0.024486053735017776, 0.20401577651500702, 0.27737337350845337, -0.0894511267542839, 0.028562651947140694, -0.04784608259797096, -0.03032585047185421, -0.14754462242126465, -0.14468172192573547, -0.07479477673768997, -0.11383165419101715, -0.004045257344841957, -0.11564740538597107, -0.011477400548756123, 0.07924777269363403, 0.07451639324426651, -0.004670888185501099, 0.06655564159154892, 0.05333368107676506, -0.08459985256195068, 0.04941840097308159, -0.010902351699769497, -0.004488688427954912, -0.0811287984251976, 0.011388746090233326, 0.010560373775660992, -0.07143492996692657, 0.0036565193440765142, 0.04611961171030998, -0.08140251040458679, 0.007694265339523554, -0.09253548830747604, -0.12355495989322662, -0.048570416867733, 0.04097408428788185, -0.01347232423722744, 0.10258661210536957, 0.0002143421588698402, -0.015863237902522087, 0.017246780917048454, 0.18903088569641113, -0.06997811794281006, -0.028809864073991776, -0.029772967100143433, 0.15632179379463196, 0.015792034566402435, 0.06093810871243477, 0.007155471947044134, -0.033846162259578705, -0.04435848817229271, 0.28692737221717834, 0.3263632357120514, -0.09259019047021866, 0.040054187178611755, 0.04673274606466293, 0.04663088545203209, 0.05650675669312477, 0.019446365535259247, 0.11920265853404999, 0.20709876716136932, -0.10886888951063156, 0.04138452932238579, -0.09229446947574615, -0.0092853344976902, -0.01768067106604576, 0.05967481806874275, 0.07244032621383667, -0.02558945305645466, -0.07466726005077362, 0.07927776873111725, -0.19803692400455475, -0.04248494654893875, -0.054057929664850235, -0.16209806501865387, -0.05589992552995682, -0.010377940721809864, 0.05967050418257713, 0.08711781352758408, 0.076105497777462, -0.019122811034321785, -0.0351237989962101, -0.03232823312282562, 0.05595700815320015, -0.17337502539157867, 0.03204787150025368, 0.07457692921161652, -0.04495998099446297, 0.03231082484126091, -0.0657830536365509, -0.015107393264770508, 0.12348698824644089, 0.004296677187085152, -0.03410175442695618, 0.06970909982919693, 0.06582240015268326, -0.03800734132528305, -0.033503949642181396, 0.05947599187493324, 0.05240248143672943, 0.005586260464042425, 0.11067622154951096, -0.09385061264038086, 0.05889048054814339, 0.02572837471961975, -0.05974463373422623, 0.003003775840625167, 0.0928146168589592, -0.034844622015953064, 0.11481953412294388, 0.09349633008241653, -0.002403327263891697, -0.030757077038288116, -0.06704383343458176, -0.028497202321887016, -0.013408532366156578, -0.0070684547536075115, -0.008555013686418533, -0.12491217255592346, -0.06465089321136475, -0.00555462297052145, 0.048126619309186935, -0.18978245556354523, -0.02049051970243454, -0.04568195715546608, 0.023115165531635284, -0.09293416142463684, 0.10014216601848602, 0.07051248103380203, 0.0012292563915252686, -0.02616814337670803, -0.10490471869707108, 0.029085248708724976, 0.11885254830121994, -0.13522009551525116, -0.05303233116865158 ]
null
null
transformers
# FrALBERT Base Cased Pretrained model on French language using a masked language modeling (MLM) objective. It was introduced in [this paper](https://arxiv.org/abs/1909.11942) and first released in [this repository](https://github.com/google-research/albert). This model, unlike other ALBERT models, is cased: it does make a difference between french and French. ## Model description FrALBERT is a transformers model pretrained on 16Go of French Wikipedia in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was pretrained with two objectives: - Masked language modeling (MLM): taking a sentence, the model randomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict the masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the sentence. - Sentence Ordering Prediction (SOP): FrALBERT uses a pretraining loss based on predicting the ordering of two consecutive segments of text. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the FrALBERT model as inputs. FrALBERT is particular in that it shares its layers across its Transformer. Therefore, all layers have the same weights. Using repeating layers results in a small memory footprint, however, the computational cost remains similar to a BERT-like architecture with the same number of hidden layers as it has to iterate through the same number of (repeating) layers. This is the second version of the base model. This model has the following configuration: - 12 repeating layers - 128 embedding dimension - 768 hidden dimension - 12 attention heads - 11M parameters ## Intended uses & limitations You can use the raw model for either masked language modeling or next sentence prediction, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=fralbert-base-cased) to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use You can use this model directly with a pipeline for masked language modeling: ```python >>> from transformers import pipeline >>> unmasker = pipeline('fill-mask', model='cservan/fralbert-base-cased') >>> unmasker("Paris est la capitale de la [MASK] .") [ { "sequence": "paris est la capitale de la france.", "score": 0.6231236457824707, "token": 3043, "token_str": "france" }, { "sequence": "paris est la capitale de la region.", "score": 0.2993471622467041, "token": 10531, "token_str": "region" }, { "sequence": "paris est la capitale de la societe.", "score": 0.02028230018913746, "token": 24622, "token_str": "societe" }, { "sequence": "paris est la capitale de la bretagne.", "score": 0.012089950032532215, "token": 24987, "token_str": "bretagne" }, { "sequence": "paris est la capitale de la chine.", "score": 0.010002839379012585, "token": 14860, "token_str": "chine" } ] ``` Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import AlbertTokenizer, AlbertModel tokenizer = AlbertTokenizer.from_pretrained('cservan/fralbert-base-cased') model = AlbertModel.from_pretrained("cservan/fralbert-base-cased") text = "Remplacez-moi par le texte en français que vous souhaitez." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import AlbertTokenizer, TFAlbertModel tokenizer = AlbertTokenizer.from_pretrained('cservan/fralbert-base-cased') model = TFAlbertModel.from_pretrained("cservan/fralbert-base-cased") text = "Remplacez-moi par le texte en français que vous souhaitez." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` ## Training data The FrALBERT model was pretrained on 4go of [French Wikipedia](https://fr.wikipedia.org/wiki/French_Wikipedia) (excluding lists, tables and headers). ## Training procedure ### Preprocessing The texts are lowercased and tokenized using SentencePiece and a vocabulary size of 32,000. The inputs of the model are then of the form: ``` [CLS] Sentence A [SEP] Sentence B [SEP] ``` ### Training The FrALBERT procedure follows the BERT setup. The details of the masking procedure for each sentence are the following: - 15% of the tokens are masked. - In 80% of the cases, the masked tokens are replaced by `[MASK]`. - In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace. - In the 10% remaining cases, the masked tokens are left as is. ## Evaluation results When fine-tuned on downstream tasks, the ALBERT models achieve the following results: Slot-filling: | | FrALBERT-base | FrALBERT-base-cased |----------------|---------------|-------------------- | MEDIA | 81.76 (0.59) | 85.09 (0.14) | ### BibTeX entry and citation info ```bibtex @inproceedings{cattan2021fralbert, author = {Oralie Cattan and Christophe Servan and Sophie Rosset}, booktitle = {Recent Advances in Natural Language Processing, RANLP 2021}, title = {{On the Usability of Transformers-based models for a French Question-Answering task}}, year = {2021}, address = {Online}, month = sep, } ``` Link to the paper: [PDF](https://hal.archives-ouvertes.fr/hal-03336060)
{"language": "fr", "license": "apache-2.0", "datasets": ["wikipedia"]}
fill-mask
cservan/fralbert-base-cased
[ "transformers", "pytorch", "albert", "fill-mask", "fr", "dataset:wikipedia", "arxiv:1909.11942", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "1909.11942" ]
[ "fr" ]
TAGS #transformers #pytorch #albert #fill-mask #fr #dataset-wikipedia #arxiv-1909.11942 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
FrALBERT Base Cased =================== Pretrained model on French language using a masked language modeling (MLM) objective. It was introduced in this paper and first released in this repository. This model, unlike other ALBERT models, is cased: it does make a difference between french and French. Model description ----------------- FrALBERT is a transformers model pretrained on 16Go of French Wikipedia in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was pretrained with two objectives: * Masked language modeling (MLM): taking a sentence, the model randomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict the masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the sentence. * Sentence Ordering Prediction (SOP): FrALBERT uses a pretraining loss based on predicting the ordering of two consecutive segments of text. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the FrALBERT model as inputs. FrALBERT is particular in that it shares its layers across its Transformer. Therefore, all layers have the same weights. Using repeating layers results in a small memory footprint, however, the computational cost remains similar to a BERT-like architecture with the same number of hidden layers as it has to iterate through the same number of (repeating) layers. This is the second version of the base model. This model has the following configuration: * 12 repeating layers * 128 embedding dimension * 768 hidden dimension * 12 attention heads * 11M parameters Intended uses & limitations --------------------------- You can use the raw model for either masked language modeling or next sentence prediction, but it's mostly intended to be fine-tuned on a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use You can use this model directly with a pipeline for masked language modeling: Here is how to use this model to get the features of a given text in PyTorch: and in TensorFlow: Training data ------------- The FrALBERT model was pretrained on 4go of French Wikipedia (excluding lists, tables and headers). Training procedure ------------------ ### Preprocessing The texts are lowercased and tokenized using SentencePiece and a vocabulary size of 32,000. The inputs of the model are then of the form: ### Training The FrALBERT procedure follows the BERT setup. The details of the masking procedure for each sentence are the following: * 15% of the tokens are masked. * In 80% of the cases, the masked tokens are replaced by '[MASK]'. * In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace. * In the 10% remaining cases, the masked tokens are left as is. Evaluation results ------------------ When fine-tuned on downstream tasks, the ALBERT models achieve the following results: Slot-filling: FrALBERT-base: MEDIA, FrALBERT-base-cased: 81.76 (0.59) FrALBERT-base: , FrALBERT-base-cased: ### BibTeX entry and citation info Link to the paper: PDF
[ "### How to use\n\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\nand in TensorFlow:\n\n\nTraining data\n-------------\n\n\nThe FrALBERT model was pretrained on 4go of French Wikipedia (excluding lists, tables and\nheaders).\n\n\nTraining procedure\n------------------", "### Preprocessing\n\n\nThe texts are lowercased and tokenized using SentencePiece and a vocabulary size of 32,000. The inputs of the model are\nthen of the form:", "### Training\n\n\nThe FrALBERT procedure follows the BERT setup.\n\n\nThe details of the masking procedure for each sentence are the following:\n\n\n* 15% of the tokens are masked.\n* In 80% of the cases, the masked tokens are replaced by '[MASK]'.\n* In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace.\n* In the 10% remaining cases, the masked tokens are left as is.\n\n\nEvaluation results\n------------------\n\n\nWhen fine-tuned on downstream tasks, the ALBERT models achieve the following results:\n\n\nSlot-filling:\n\n\nFrALBERT-base: MEDIA, FrALBERT-base-cased: 81.76 (0.59)\nFrALBERT-base: , FrALBERT-base-cased:", "### BibTeX entry and citation info\n\n\nLink to the paper: PDF" ]
[ "TAGS\n#transformers #pytorch #albert #fill-mask #fr #dataset-wikipedia #arxiv-1909.11942 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### How to use\n\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\nand in TensorFlow:\n\n\nTraining data\n-------------\n\n\nThe FrALBERT model was pretrained on 4go of French Wikipedia (excluding lists, tables and\nheaders).\n\n\nTraining procedure\n------------------", "### Preprocessing\n\n\nThe texts are lowercased and tokenized using SentencePiece and a vocabulary size of 32,000. The inputs of the model are\nthen of the form:", "### Training\n\n\nThe FrALBERT procedure follows the BERT setup.\n\n\nThe details of the masking procedure for each sentence are the following:\n\n\n* 15% of the tokens are masked.\n* In 80% of the cases, the masked tokens are replaced by '[MASK]'.\n* In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace.\n* In the 10% remaining cases, the masked tokens are left as is.\n\n\nEvaluation results\n------------------\n\n\nWhen fine-tuned on downstream tasks, the ALBERT models achieve the following results:\n\n\nSlot-filling:\n\n\nFrALBERT-base: MEDIA, FrALBERT-base-cased: 81.76 (0.59)\nFrALBERT-base: , FrALBERT-base-cased:", "### BibTeX entry and citation info\n\n\nLink to the paper: PDF" ]
[ 60, 86, 42, 185, 17 ]
[ "passage: TAGS\n#transformers #pytorch #albert #fill-mask #fr #dataset-wikipedia #arxiv-1909.11942 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### How to use\n\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\nand in TensorFlow:\n\n\nTraining data\n-------------\n\n\nThe FrALBERT model was pretrained on 4go of French Wikipedia (excluding lists, tables and\nheaders).\n\n\nTraining procedure\n------------------### Preprocessing\n\n\nThe texts are lowercased and tokenized using SentencePiece and a vocabulary size of 32,000. The inputs of the model are\nthen of the form:### Training\n\n\nThe FrALBERT procedure follows the BERT setup.\n\n\nThe details of the masking procedure for each sentence are the following:\n\n\n* 15% of the tokens are masked.\n* In 80% of the cases, the masked tokens are replaced by '[MASK]'.\n* In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace.\n* In the 10% remaining cases, the masked tokens are left as is.\n\n\nEvaluation results\n------------------\n\n\nWhen fine-tuned on downstream tasks, the ALBERT models achieve the following results:\n\n\nSlot-filling:\n\n\nFrALBERT-base: MEDIA, FrALBERT-base-cased: 81.76 (0.59)\nFrALBERT-base: , FrALBERT-base-cased:### BibTeX entry and citation info\n\n\nLink to the paper: PDF" ]
[ -0.03126373142004013, 0.015831228345632553, -0.005797833204269409, 0.09988299757242203, 0.02111339382827282, 0.0010972425807267427, 0.0966000109910965, 0.0784265398979187, -0.07355821877717972, 0.09975865483283997, 0.033671341836452484, 0.05030170455574989, 0.0803571566939354, 0.14475573599338531, -0.0023422869853675365, -0.25827556848526, 0.08825349807739258, -0.007973266765475273, 0.05110989138484001, 0.07126928865909576, 0.08964456617832184, -0.07734489440917969, -0.014853608794510365, 0.03814234212040901, -0.009291918016970158, -0.052537646144628525, -0.00396701181307435, -0.05471326783299446, 0.08350078761577606, 0.03464047610759735, 0.0712599903345108, 0.04687013104557991, 0.002947910688817501, -0.1370858997106552, 0.029356304556131363, 0.06817523390054703, -0.036716002970933914, 0.07374241203069687, 0.08421848714351654, -0.040386080741882324, 0.09577756375074387, -0.01874254085123539, 0.056160181760787964, 0.04625200480222702, -0.13657629489898682, -0.24023714661598206, -0.12885290384292603, 0.13538125157356262, 0.053403254598379135, 0.0030310465954244137, -0.01399570144712925, 0.1315125674009323, -0.021797573193907738, 0.1024475023150444, 0.2739783227443695, -0.25322529673576355, -0.010178772732615471, -0.05497116968035698, 0.025915829464793205, -0.05491338297724724, -0.07798278331756592, -0.03055965155363083, 0.0011877523502334952, 0.03656814619898796, 0.10937156528234482, -0.027761679142713547, -0.011256376281380653, -0.06585794687271118, -0.16706448793411255, -0.05667265132069588, 0.009395605884492397, 0.006422834470868111, -0.11433397978544235, -0.13469940423965454, -0.0777079313993454, -0.028535936027765274, -0.07338167726993561, 0.0164161529392004, 0.008116135373711586, 0.0033872511703521013, 0.07742544263601303, -0.06011240556836128, -0.05353892967104912, -0.005297193303704262, -0.05240635573863983, 0.17509743571281433, 0.0547267384827137, 0.010174150578677654, -0.05038491263985634, 0.05312125012278557, -0.16328538954257965, -0.09796810150146484, -0.05222005397081375, -0.07414095103740692, -0.07239633798599243, -0.020173851400613785, -0.026419268921017647, -0.2143239974975586, -0.03875771909952164, 0.21598637104034424, -0.10344083607196808, 0.06714069098234177, -0.1527629792690277, 0.0386815071105957, 0.01422026939690113, 0.10706017166376114, -0.09526859223842621, 0.09090418368577957, 0.026210976764559746, 0.02120141126215458, 0.0038376757875084877, -0.04634750261902809, 0.009663492441177368, 0.08582809567451477, 0.023847628384828568, 0.03734150528907776, 0.045383092015981674, 0.08064950257539749, -0.06275366246700287, -0.042710427194833755, 0.16599518060684204, -0.15146586298942566, 0.019146475940942764, 0.01870676316320896, -0.024260155856609344, -0.0371098667383194, 0.07111024856567383, -0.046655360609292984, -0.11241098493337631, 0.10352036356925964, -0.04689731076359749, -0.019825445488095284, -0.03816835954785347, -0.17033204436302185, 0.008827226236462593, -0.06556632369756699, -0.09246677905321121, -0.011509446427226067, -0.08031734079122543, -0.009881095960736275, 0.07152464985847473, -0.005408904515206814, 0.05139679089188576, -0.011237211525440216, -0.05628899112343788, 0.025723814964294434, 0.04505803436040878, -0.012099724262952805, 0.0007605130667798221, 0.025670524686574936, -0.010379797779023647, 0.08783318847417831, 0.027804555371403694, 0.014454703778028488, -0.051468752324581146, 0.05378050357103348, -0.28227928280830383, 0.06671958416700363, -0.10687236487865448, -0.09950369596481323, -0.0776059478521347, -0.06238848716020584, -0.025997288525104523, -0.008059604093432426, 0.03582329675555229, 0.14051593840122223, -0.1596592664718628, -0.05518638715147972, 0.2829727232456207, -0.1303163766860962, 0.028972167521715164, 0.13350529968738556, -0.053812846541404724, -0.040512632578611374, 0.02787865325808525, 0.15400102734565735, -0.023335132747888565, -0.17560267448425293, -0.01134188286960125, -0.059657201170921326, -0.015845229849219322, 0.1756686568260193, 0.06165381893515587, -0.05888708308339119, 0.06034056842327118, 0.0048727234825491905, -0.04295628145337105, -0.08135946094989777, -0.024344725534319878, -0.03407374396920204, 0.00859373714774847, 0.00977044552564621, 0.07994694262742996, -0.0541708804666996, -0.013235509395599365, -0.05524790287017822, -0.09228874742984772, -0.06386640667915344, 0.04496641084551811, -0.07454599440097809, 0.08350690454244614, -0.07815075665712357, 0.1015094667673111, -0.012317189015448093, 0.023776959627866745, -0.21155209839344025, -0.0191126000136137, 0.08792262524366379, -0.11591916531324387, 0.037296511232852936, -0.028142565861344337, 0.04210653156042099, 0.0815669372677803, -0.052031632512807846, -0.001386136980727315, -0.03363979235291481, -0.04485993832349777, -0.0627434030175209, -0.16748927533626556, -0.06270036101341248, -0.06299711763858795, 0.010145383886992931, -0.046862512826919556, 0.03603888303041458, 0.02627391368150711, 0.08711986243724823, 0.03514697030186653, -0.051629625260829926, 0.026504218578338623, -0.00046325798030011356, -0.027337193489074707, -0.031706538051366806, -0.013317739591002464, 0.02287021093070507, -0.002746132668107748, 0.13962770998477936, -0.15864698588848114, -0.10874614864587784, 0.06298030912876129, 0.06440801918506622, -0.08542723953723907, 0.029274536296725273, -0.060880549252033234, 0.01563912257552147, -0.049893055111169815, 0.0069137816317379475, 0.17443732917308807, 0.0478915199637413, 0.14362718164920807, -0.06681738048791885, -0.01114121824502945, 0.0418156273663044, -0.004176169168204069, -0.09554029256105423, 0.055709727108478546, -0.024143161252141, -0.10109300166368484, 0.050064072012901306, -0.04925931245088577, 0.06828182190656662, 0.09725053608417511, -0.008003286086022854, -0.0964212417602539, -0.06349979341030121, 0.05415915325284004, 0.07235915958881378, 0.08316905051469803, 0.022997582331299782, 0.010099820792675018, 0.08177686482667923, 0.04069485515356064, 0.0271091777831316, -0.07987174391746521, 0.059414952993392944, 0.06271633505821228, -0.027571868151426315, -0.08267057687044144, -0.018607409670948982, 0.027455154806375504, 0.099863700568676, 0.021624170243740082, 0.026943758130073547, -0.03885194659233093, -0.027828309684991837, -0.08013005554676056, 0.1597585529088974, -0.12858393788337708, -0.23613305389881134, -0.15396441519260406, -0.06609634310007095, -0.043687913566827774, 0.04208178445696831, 0.06474273651838303, -0.01542885135859251, -0.08142628520727158, -0.1012234315276146, 0.03915518894791603, -0.004087105393409729, -0.0032715261913836002, 0.0385480672121048, -0.06947419047355652, 0.03645993024110794, -0.14376838505268097, -0.025903897359967232, -0.009863142855465412, 0.039599575102329254, 0.04916936531662941, 0.021375827491283417, 0.10668128728866577, 0.08709923923015594, -0.00806447397917509, 0.00024632542044855654, -0.0361468531191349, 0.18356619775295258, -0.060598164796829224, 0.07967719435691833, 0.016597798094153404, -0.07831062376499176, 0.12429367750883102, 0.06585169583559036, 0.0024800309911370277, -0.06359836459159851, -0.004548979923129082, 0.06313393265008926, -0.035094697028398514, -0.2279394418001175, -0.015937382355332375, -0.04802478849887848, 0.03575640544295311, 0.09041113406419754, 0.03848110884428024, 0.030289916321635246, -0.03732447326183319, -0.09795568138360977, -0.006218831054866314, 0.06720910966396332, 0.1031751036643982, -0.04279840737581253, -0.013690667226910591, 0.08086207509040833, -0.04891398921608925, -0.015445621684193611, 0.08085103332996368, -0.05766119807958603, 0.13784247636795044, -0.0480380542576313, 0.19922897219657898, 0.1077343225479126, 0.040160927921533585, 0.02431841753423214, 0.15070828795433044, -0.0660848319530487, 0.03725374490022659, -0.029555590823292732, -0.07264284044504166, 0.012659692205488682, 0.007637558039277792, 0.007564081810414791, -0.007597476709634066, -0.09747279435396194, -0.024163594469428062, 0.06660313159227371, 0.2880840301513672, 0.034353725612163544, -0.1150619313120842, -0.08102244883775711, -0.005385955795645714, -0.03608572855591774, -0.061926908791065216, -0.0487096793949604, 0.03818075731396675, -0.10015642642974854, 0.09023429453372955, -0.06536151468753815, 0.0875980332493782, -0.02604423649609089, 0.012455393560230732, -0.07726864516735077, 0.026901274919509888, -0.03213633596897125, 0.03691813349723816, -0.18734893202781677, 0.19763745367527008, 0.0597897469997406, 0.040348559617996216, -0.0948561429977417, 0.044490646570920944, -0.005151356104761362, 0.0005061870906502008, 0.18142080307006836, 0.00040908955270424485, -0.07156725972890854, -0.063368059694767, -0.05056830123066902, -0.037200357764959335, 0.158786803483963, -0.007862322963774204, 0.11508572101593018, 0.03774850815534592, -0.025114746764302254, 0.011703078635036945, 0.10541470348834991, -0.07836329936981201, -0.1547905057668686, 0.09183057397603989, -0.10034766793251038, -0.1299218386411667, -0.06824725866317749, -0.07030791789293289, -0.13805794715881348, 0.18633458018302917, -0.074905164539814, -0.009385260753333569, -0.14310462772846222, -0.027071423828601837, 0.09443897753953934, -0.10559144616127014, 0.03905096650123596, -0.06568122655153275, 0.15881872177124023, -0.13868418335914612, -0.0992833822965622, 0.08819664269685745, -0.06285540014505386, -0.1249719187617302, -0.059734851121902466, 0.11206851899623871, 0.12209577858448029, 0.04852120205760002, -0.001197535777464509, 0.05442117899656296, 0.07410480082035065, -0.08006606996059418, 0.017028870061039925, 0.017288556322455406, 0.10815107822418213, 0.02366064116358757, -0.09814003109931946, -0.007213374134153128, -0.0882011204957962, 0.06391803175210953, 0.08310677856206894, 0.3353902995586395, -0.07252423465251923, 0.16271494328975677, 0.14252664148807526, -0.07668792456388474, -0.20947521924972534, -0.021347226575016975, 0.026913313195109367, 0.02796291373670101, 0.04092103987932205, -0.17783817648887634, -0.06355960667133331, 0.0013000473845750093, -0.026130441576242447, -0.0126236816868186, -0.1777590960264206, -0.10271882265806198, 0.08665873855352402, 0.043620526790618896, 0.0065283640287816525, -0.09256807714700699, -0.046878669410943985, -0.00457672169432044, -0.10361000895500183, 0.056024886667728424, -0.013085621409118176, 0.12116934359073639, 0.004780254792422056, -0.10930193215608597, 0.02956816554069519, -0.07464984059333801, 0.1123441830277443, -0.01977194845676422, 0.08985209465026855, -0.05692250281572342, 0.013511055149137974, 0.14900988340377808, -0.054388049989938736, 0.14121153950691223, 0.044394392520189285, 0.04109340161085129, -0.048373330384492874, -0.05279785767197609, -0.06245533004403114, 0.05107474699616432, -0.07461639493703842, -0.027469297870993614, -0.05310722440481186, 0.04848920553922653, 0.12685120105743408, -0.019933098927140236, 0.08295408636331558, 0.010572533123195171, 0.10640636831521988, 0.22146284580230713, -0.01502812746912241, 0.03494073823094368, -0.06962869316339493, 0.008946234360337257, -0.006378259044140577, 0.036226239055395126, -0.08927065879106522, 0.06682566553354263, 0.08172203600406647, 0.042319245636463165, 0.16358520090579987, 0.028316184878349304, -0.17607273161411285, -0.0473959855735302, 0.02111664041876793, -0.14700847864151, -0.18069210648536682, 0.020382406190037727, -0.04149399697780609, -0.0992322564125061, 0.0017833077581599355, 0.08241505175828934, -0.0011721700429916382, -0.046027421951293945, -0.012518768198788166, 0.046917106956243515, 0.01707732118666172, 0.12415369600057602, 0.016802750527858734, 0.06322890520095825, -0.07572177797555923, 0.08828515559434891, 0.10153067857027054, -0.13533169031143188, 0.05686407908797264, 0.05569439381361008, -0.07820814102888107, -0.030651092529296875, 0.08646082878112793, 0.1007210984826088, 0.12700505554676056, 0.026836631819605827, -0.10670231282711029, -0.051309455186128616, 0.10241896659135818, 0.1838139146566391, 0.022932013496756554, 0.01699824072420597, -0.0009611759451217949, -0.005334113724529743, -0.07186871767044067, 0.05726674571633339, 0.0448339618742466, 0.03151725232601166, 0.060971491038799286, 0.12230922281742096, -0.019359102472662926, 0.05889289826154709, -0.012284096330404282, -0.04598497971892357, -0.0842597559094429, -0.02883070707321167, -0.04671144485473633, 0.004115395713597536, -0.08828680962324142, -0.028980402275919914, -0.02805166132748127, 0.03278094902634621, 0.0392017662525177, 0.05443492904305458, -0.08304021507501602, -0.036961816251277924, -0.041627123951911926, 0.0011705616489052773, -0.15885372459888458, 0.008719800040125847, 0.04749401658773422, -0.07611341774463654, 0.0813336968421936, 0.039800092577934265, -0.03396344557404518, 0.046178970485925674, -0.06241088733077049, -0.01585695706307888, -0.02105962485074997, -0.010883974842727184, 0.017934100702404976, -0.1186569407582283, 0.016981730237603188, -0.05976860225200653, -0.05016003176569939, -0.012023200280964375, 0.07823647558689117, -0.10828373581171036, -0.0005995439132675529, 0.038201238960027695, -0.050452299416065216, -0.028795145452022552, 0.11744943261146545, 0.04125773534178734, -0.01021589059382677, 0.1024302989244461, -0.04549725353717804, 0.07335717231035233, -0.10074403136968613, -0.02576710470020771, 0.001072047045454383, -0.05448366701602936, 0.11990317702293396, -0.04699785262346268, 0.03629770874977112, -0.036475714296102524, 0.10681664198637009, -0.05085902661085129, -0.03252386301755905, 0.013116220943629742, -0.047490306198596954, -0.08789572864770889, 0.027507638558745384, 0.022560378536581993, -0.05539262294769287, -0.06943154335021973, 0.03815314173698425, 0.036775846034288406, 0.017720356583595276, 0.15225407481193542, 0.18536700308322906, 0.12625649571418762, 0.13579820096492767, 0.022001221776008606, 0.02848350629210472, -0.05781369283795357, -0.09623599052429199, -0.043217457830905914, 0.02232157625257969, 0.08444730192422867, -0.00278739002533257, -0.030395152047276497, 0.09834554046392441, -0.13139191269874573, 0.19244931638240814, 0.03308645635843277, -0.06233701854944229, -0.08219818770885468, -0.2261313945055008, -0.020105749368667603, 0.0671517625451088, -0.004894794896245003, -0.10086239129304886, 0.05543721839785576, -0.03630228713154793, 0.04245966672897339, 0.015983201563358307, 0.12757064402103424, -0.12569718062877655, -0.060717325657606125, 0.08646707981824875, -0.002724743215367198, -0.06538991630077362, 0.09519529342651367, -0.007444037590175867, 0.037833407521247864, 0.011956132017076015, 0.06479208916425705, 0.051378607749938965, 0.11107685416936874, 0.03836570680141449, -0.06732919067144394, -0.09800158441066742, 0.043532006442546844, -0.022174913436174393, 0.0569736547768116, 0.22106319665908813, 0.02599242515861988, -0.040562357753515244, -0.01581711508333683, 0.10791146010160446, -0.0187229011207819, -0.07740899175405502, -0.15278717875480652, 0.18583831191062927, 0.0731990709900856, 0.00641772523522377, 0.0076959701254963875, -0.10091305524110794, -0.022714365273714066, 0.19675381481647491, 0.17682857811450958, 0.07894260436296463, 0.06011941283941269, 0.035742055624723434, 0.007595642935484648, 0.08625811338424683, 0.02564554661512375, 0.039802368730306625, 0.07414482533931732, -0.00877761747688055, 0.041036609560251236, -0.05290122702717781, -0.03725820034742355, -0.04803069680929184, 0.17605628073215485, 0.007108672987669706, -0.013499867171049118, -0.05459071695804596, 0.012450768612325191, 0.003111298428848386, -0.2592938244342804, -0.024055704474449158, -0.09063160419464111, -0.11535383015871048, -0.0035498221404850483, -0.0695076733827591, 0.038260165601968765, 0.08520223200321198, 0.06321877241134644, -0.012289637699723244, 0.20652809739112854, 0.03389843553304672, -0.012417060323059559, -0.07257089763879776, 0.03928513452410698, -0.15838338434696198, 0.1955021768808365, 0.07029462605714798, 0.0062413448467850685, 0.11987299472093582, 0.005583424586802721, -0.06181890144944191, 0.045237645506858826, 0.027626272290945053, 0.018370099365711212, -0.02937506139278412, 0.19683299958705902, -0.012648436240851879, -0.03161856532096863, -0.003585538361221552, -0.1006641834974289, 0.06310900300741196, -0.11803949624300003, -0.052048247307538986, -0.060140449553728104, 0.08627867698669434, -0.07969646155834198, 0.08128221333026886, 0.20717179775238037, 0.019680267199873924, 0.005136874038726091, -0.0767279788851738, -0.014154604636132717, 0.007332661189138889, 0.04697529599070549, -0.03806417062878609, -0.19388775527477264, 0.04737045243382454, -0.04374242573976517, 0.06101393699645996, -0.24061523377895355, -0.07307584583759308, 0.056562915444374084, -0.07962829619646072, 0.009198267944157124, 0.07023456692695618, 0.0343620739877224, 0.024302758276462555, -0.04297609254717827, -0.05238189175724983, 0.012070182710886002, 0.08260146528482437, -0.07417739927768707, 0.017056994140148163 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-large-xls-r-1b-bemba-fds This model is a fine-tuned version of [facebook/wav2vec2-xls-r-1b](https://huggingface.co/facebook/wav2vec2-xls-r-1b) on the [BembaSpeech](https://github.com/csikasote/BembaSpeech) dataset. It achieves the following results on the evaluation set: - Loss: 0.2898 - Wer: 0.3435 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 8 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 15 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | 1.7986 | 0.34 | 500 | 0.4549 | 0.7292 | | 0.5358 | 0.67 | 1000 | 0.3325 | 0.4491 | | 0.4559 | 1.01 | 1500 | 0.3090 | 0.3954 | | 0.3983 | 1.35 | 2000 | 0.3067 | 0.4105 | | 0.4067 | 1.68 | 2500 | 0.2838 | 0.3678 | | 0.3722 | 2.02 | 3000 | 0.2824 | 0.3762 | | 0.3286 | 2.36 | 3500 | 0.2810 | 0.3670 | | 0.3239 | 2.69 | 4000 | 0.2643 | 0.3501 | | 0.3187 | 3.03 | 4500 | 0.2838 | 0.3754 | | 0.2801 | 3.36 | 5000 | 0.2815 | 0.3507 | | 0.2806 | 3.7 | 5500 | 0.2725 | 0.3486 | | 0.2714 | 4.04 | 6000 | 0.2898 | 0.3435 | ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.0
{"license": "apache-2.0", "tags": ["generated_from_trainer", "bem", "robust-speech-event"], "model-index": [{"name": "wav2vec2-large-xls-r-1b-bemba-fds", "results": []}]}
automatic-speech-recognition
csikasote/wav2vec2-large-xls-r-1b-bemba-fds
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "bem", "robust-speech-event", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #bem #robust-speech-event #license-apache-2.0 #endpoints_compatible #region-us
wav2vec2-large-xls-r-1b-bemba-fds ================================= This model is a fine-tuned version of facebook/wav2vec2-xls-r-1b on the BembaSpeech dataset. It achieves the following results on the evaluation set: * Loss: 0.2898 * Wer: 0.3435 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 5e-05 * train\_batch\_size: 4 * eval\_batch\_size: 8 * seed: 42 * gradient\_accumulation\_steps: 2 * total\_train\_batch\_size: 8 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 500 * num\_epochs: 15 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.16.2 * Pytorch 1.10.0+cu111 * Datasets 1.18.3 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 8\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 15\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #bem #robust-speech-event #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 8\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 15\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ 66, 159, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #bem #robust-speech-event #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 8\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 15\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ -0.11465804278850555, 0.0729159489274025, -0.003411562880501151, 0.05645596608519554, 0.1167273223400116, 0.006166545208543539, 0.08970609307289124, 0.14165210723876953, -0.07306689023971558, 0.08457065373659134, 0.10702508687973022, 0.09272751957178116, 0.05682758986949921, 0.12470467388629913, -0.02086191624403, -0.31695669889450073, 0.008644022047519684, 0.013749250210821629, -0.1080833449959755, 0.1188637763261795, 0.09670364111661911, -0.1104830801486969, 0.01604842022061348, 0.017682787030935287, -0.1203666478395462, 0.010089853778481483, -0.028642775490880013, -0.07182403653860092, 0.11872098594903946, 0.03271980211138725, 0.0941983237862587, 0.01801241934299469, 0.08845213055610657, -0.26146501302719116, 0.017468051984906197, 0.060490284115076065, 0.04079675301909447, 0.07159990072250366, 0.11077789217233658, -0.01215266901999712, 0.11997175216674805, -0.07529519498348236, 0.07582072168588638, 0.03586238995194435, -0.10904458165168762, -0.30871808528900146, -0.079629085958004, 0.040390171110630035, 0.12419858574867249, 0.0947251170873642, -0.030139686539769173, 0.07717125117778778, -0.08391585201025009, 0.08233310282230377, 0.2338806837797165, -0.2553621232509613, -0.07997356355190277, -0.042304519563913345, 0.04617169499397278, 0.0484105683863163, -0.12355860322713852, -0.03512994945049286, 0.02445291168987751, 0.03886450082063675, 0.11416621506214142, 0.004651391878724098, -0.03317520394921303, 0.018279917538166046, -0.1564047634601593, -0.04874109476804733, 0.09482134133577347, 0.07333564013242722, -0.021545208990573883, -0.09844724088907242, -0.019445078447461128, -0.18673738837242126, -0.055682770907878876, 0.005414553452283144, 0.03029453381896019, -0.04023255035281181, -0.10070443898439407, 0.013943028636276722, -0.08755800127983093, -0.08377258479595184, 0.014507779851555824, 0.13889020681381226, 0.049072638154029846, -0.025331629440188408, -0.0004927952541038394, 0.10312613099813461, 0.039224572479724884, -0.1331842839717865, 0.003712362376973033, 0.041197482496500015, -0.10960282385349274, -0.01897326484322548, -0.03130091354250908, -0.02863425575196743, -0.0017270103562623262, 0.11698354780673981, -0.030381467193365097, 0.08707468956708908, 0.024512337520718575, 0.03189198672771454, -0.08301780372858047, 0.1762590855360031, -0.06963635981082916, -0.028689783066511154, -0.057863641530275345, 0.09910757094621658, -0.022618943825364113, -0.008321649394929409, -0.06389568746089935, 0.032693710178136826, 0.10189776122570038, 0.04133057966828346, -0.01868143118917942, 0.023220181465148926, -0.06539179384708405, -0.02705351449549198, -0.035207122564315796, -0.10217826068401337, 0.042340755462646484, 0.03196457028388977, -0.08098205178976059, 0.005847196560353041, 0.008040736429393291, 0.017546167597174644, -0.03202008083462715, 0.1363699734210968, -0.060650669038295746, 0.02123217098414898, -0.08559124171733856, -0.10204990208148956, 0.03434909135103226, -0.048630062490701675, -0.001579316915012896, -0.07215074449777603, -0.10872471332550049, -0.046992696821689606, 0.06687621027231216, -0.052045729011297226, -0.07044627517461777, -0.06542174518108368, -0.06382743269205093, 0.048201221972703934, -0.03310520574450493, 0.17923149466514587, -0.06275186687707901, 0.10930201411247253, 0.009898796677589417, 0.04393261671066284, 0.045221272855997086, 0.07575681805610657, -0.03808322921395302, 0.035179175436496735, -0.13919302821159363, 0.0759599581360817, -0.0879511684179306, 0.05317996069788933, -0.15111590921878815, -0.12070140987634659, -0.0273148063570261, 0.007198154926300049, 0.09825922548770905, 0.0946950688958168, -0.1697043627500534, -0.09949410706758499, 0.1888808012008667, -0.0636448785662651, -0.08538898825645447, 0.13942573964595795, -0.03193500265479088, -0.0033397749066352844, 0.0357743538916111, 0.17291919887065887, 0.07795273512601852, -0.09831254929304123, 0.028188295662403107, -0.04748594015836716, 0.10749498009681702, 0.01837144047021866, 0.09269227832555771, -0.045231323689222336, 0.040062423795461655, -0.006286522839218378, -0.03188760578632355, 0.07762469351291656, -0.08819938451051712, -0.08365539461374283, -0.014384817332029343, -0.07356658577919006, 0.02287701517343521, 0.057874131947755814, 0.04168354347348213, -0.08838514983654022, -0.1314988136291504, 0.03825211897492409, 0.09806649386882782, -0.09878745675086975, 0.034033410251140594, -0.07617934793233871, 0.0376780666410923, -0.03052746318280697, -0.00965869054198265, -0.17085058987140656, 0.0009372279746457934, 0.02661212719976902, -0.04143603518605232, 0.030777109786868095, -0.0032901172526180744, 0.08567831665277481, 0.02942909114062786, -0.04472601041197777, -0.07264291495084763, -0.06535214930772781, -0.006914857774972916, -0.08942613750696182, -0.2281254678964615, -0.07109202444553375, -0.03306642919778824, 0.15681514143943787, -0.22735461592674255, 0.006822066381573677, 0.014779395423829556, 0.10274270921945572, 0.03748010843992233, -0.04477597400546074, -0.010423821397125721, 0.08078636974096298, -0.00904101599007845, -0.06199902668595314, 0.045869309455156326, -0.002089110901579261, -0.11663604527711868, 0.03044520691037178, -0.11194563657045364, 0.0929076299071312, 0.10881432145833969, -0.024921398609876633, -0.08674421161413193, -0.06010991334915161, -0.0667123794555664, -0.06847027689218521, -0.03197035565972328, 0.0055082389153540134, 0.22282421588897705, 0.039874766021966934, 0.12354154884815216, -0.07152245938777924, -0.057848576456308365, 0.029067683964967728, 0.004046874586492777, -0.01600673981010914, 0.14721687138080597, 0.07891273498535156, -0.03861532732844353, 0.09765312820672989, 0.10653368383646011, -0.08523266017436981, 0.14924311637878418, -0.0682525485754013, -0.12872377038002014, -0.01713109388947487, 0.030145512893795967, 0.02617504447698593, 0.11897778511047363, -0.15225085616111755, -0.01231789868324995, 0.012990618124604225, 0.026569994166493416, 0.029713785275816917, -0.21502336859703064, -0.011365440674126148, 0.04339442774653435, -0.06994078308343887, -0.02636057883501053, -0.021857235580682755, 0.007816189900040627, 0.09266335517168045, 0.009805914014577866, -0.07301336526870728, -0.008978160098195076, -0.02535031922161579, -0.08181121200323105, 0.18968629837036133, -0.09315365552902222, -0.14080514013767242, -0.11690033227205276, -0.038552042096853256, 0.012376666069030762, -0.02406039834022522, 0.04863140732049942, -0.11400895565748215, -0.0323059968650341, -0.06206660345196724, 0.04691091552376747, -0.0588332824409008, 0.030682072043418884, -0.0022099833004176617, 0.005909142550081015, 0.07713494449853897, -0.09788522124290466, 0.02251456491649151, -0.03687961772084236, -0.03416372835636139, 0.03535033017396927, 0.037076640874147415, 0.088966004550457, 0.17458751797676086, 0.037667158991098404, 0.022147314622998238, -0.04370063543319702, 0.1360781341791153, -0.10045673698186874, -0.03533150255680084, 0.10981252044439316, -0.012900268658995628, 0.04497615247964859, 0.11990101635456085, 0.05808635801076889, -0.07652527093887329, 0.017612114548683167, 0.049648355692625046, -0.012775846756994724, -0.24490858614444733, -0.033738065510988235, -0.06024482473731041, -0.013661086559295654, 0.12493278086185455, 0.0319879949092865, -0.030148280784487724, 0.024877365678548813, 0.00038008380215615034, -0.010481288656592369, -0.0018099669832736254, 0.059515345841646194, 0.06250002235174179, 0.034609466791152954, 0.11636824160814285, -0.015246170572936535, -0.05010418966412544, 0.025503262877464294, 0.004829812794923782, 0.25913628935813904, -0.00002113589653163217, 0.16799314320087433, 0.05143214017152786, 0.1585153490304947, 0.013110645115375519, 0.0614105723798275, 0.008158217184245586, -0.027765445411205292, 0.013531479984521866, -0.04886045306921005, -0.015871815383434296, 0.0477352999150753, 0.10790809243917465, 0.030619872733950615, -0.11673513799905777, -0.03547490015625954, 0.025883955880999565, 0.35830268263816833, 0.06122482195496559, -0.28662410378456116, -0.08033651113510132, 0.0027683842927217484, -0.09504377096891403, -0.038235217332839966, 0.03147856518626213, 0.11149973422288895, -0.09588140994310379, 0.04095664620399475, -0.08359957486391068, 0.10061703622341156, -0.05957798287272453, 0.005650090053677559, 0.09092158079147339, 0.07738037407398224, 0.0030934421811252832, 0.05793919786810875, -0.2666623294353485, 0.30875128507614136, -0.01928752474486828, 0.07001787424087524, -0.04718998819589615, 0.03159574419260025, 0.03347394987940788, -0.0680202841758728, 0.0763864740729332, -0.01799360290169716, -0.0964474007487297, -0.18794690072536469, -0.08248554915189743, 0.022296791896224022, 0.12869687378406525, -0.04282720014452934, 0.12166975438594818, -0.021652724593877792, -0.009814586490392685, 0.06052037328481674, -0.07758308947086334, -0.09337455779314041, -0.10886475443840027, 0.025225773453712463, 0.0161589328199625, 0.05695087090134621, -0.10088509321212769, -0.11603646725416183, -0.08493371307849884, 0.1581026017665863, -0.09952891618013382, -0.010108173824846745, -0.12803015112876892, 0.09493040293455124, 0.15948888659477234, -0.06869146227836609, 0.052216824144124985, 0.028950160369277, 0.10829783231019974, 0.0258466973900795, -0.013217565603554249, 0.12339247763156891, -0.08515296876430511, -0.1805020421743393, -0.057040926069021225, 0.1582033485174179, 0.04713145270943642, 0.06842582672834396, -0.024611515924334526, 0.021077850833535194, -0.02122957818210125, -0.07866830378770828, 0.06269138306379318, 0.0070550814270973206, 0.011770983226597309, 0.05388942360877991, -0.03803795203566551, -0.012445102445781231, -0.08088935911655426, -0.06406118720769882, 0.1680067926645279, 0.2763645052909851, -0.08723480999469757, 0.039395254105329514, 0.03939918056130409, -0.04854800924658775, -0.15163478255271912, 0.025840414687991142, 0.14123649895191193, 0.03279050439596176, 0.01629791408777237, -0.2261388748884201, 0.05704843997955322, 0.0900326743721962, -0.02135155349969864, 0.07912681251764297, -0.34417444467544556, -0.13178205490112305, 0.12466740608215332, 0.10615707188844681, -0.02286899834871292, -0.15823805332183838, -0.05385438725352287, 0.007116110064089298, -0.08417138457298279, 0.0668724998831749, -0.041439007967710495, 0.12175412476062775, -0.0004911621217615902, 0.06833121180534363, 0.018702805042266846, -0.05559452250599861, 0.14005762338638306, -0.015934228897094727, 0.054680489003658295, -0.006386719178408384, 0.043917857110500336, -0.012733899988234043, -0.04636942222714424, 0.007204148918390274, -0.06961990892887115, 0.00434862170368433, -0.13828659057617188, -0.028931856155395508, -0.08625278621912003, 0.023634405806660652, -0.037344470620155334, -0.042113956063985825, 0.004308102186769247, 0.041049446910619736, 0.06180603429675102, 0.0057421294040977955, 0.11786676198244095, -0.05680074915289879, 0.1428832709789276, 0.07794640213251114, 0.09179182350635529, -0.0038341328036040068, -0.11457566916942596, -0.021404245868325233, -0.00882038939744234, 0.05385178327560425, -0.11863362789154053, 0.02721124328672886, 0.1454533040523529, 0.04411330074071884, 0.153669536113739, 0.06116523593664169, -0.07936936616897583, 0.02028798684477806, 0.063605397939682, -0.08400379121303558, -0.11589139699935913, -0.0022862947080284357, 0.06656406074762344, -0.13528025150299072, -0.001793139846995473, 0.10693079233169556, -0.04954594001173973, -0.012932416051626205, 0.013385909609496593, 0.01976807788014412, -0.05273105949163437, 0.23116420209407806, 0.026781344786286354, 0.07347119599580765, -0.09077958762645721, 0.079373799264431, 0.06021811440587044, -0.18612957000732422, 0.027225831523537636, 0.0896759033203125, -0.03062223084270954, -0.021516527980566025, 0.030065637081861496, 0.07806956022977829, 0.020655840635299683, -0.054434556514024734, -0.11046270281076431, -0.14972352981567383, 0.09341984987258911, 0.10901831090450287, 0.03405376523733139, 0.02205805294215679, -0.045820947736501694, 0.04814862087368965, -0.10193610191345215, 0.08900731801986694, 0.10672987252473831, 0.06962910294532776, -0.12947487831115723, 0.15112201869487762, 0.012357527390122414, 0.013198789209127426, 0.0013188612647354603, -0.004249102436006069, -0.10612910240888596, 0.03373146057128906, -0.10775202512741089, -0.0212356299161911, -0.04826769232749939, -0.0034540712367743254, 0.010564284399151802, -0.05062839388847351, -0.047148656100034714, 0.015634896233677864, -0.1240227222442627, -0.039637576788663864, -0.006906122900545597, 0.06637855619192123, -0.10561075806617737, -0.028990864753723145, 0.0438205748796463, -0.10511544346809387, 0.10308638960123062, 0.06234126538038254, 0.020992159843444824, 0.03801307454705238, -0.1357487142086029, 0.00665896013379097, 0.04002389684319496, -0.0027673854492604733, 0.0239156112074852, -0.15673336386680603, -0.012788317166268826, -0.02935931645333767, 0.038234539330005646, -0.005211932584643364, 0.028872564435005188, -0.13593384623527527, -0.0394141711294651, -0.03293919563293457, -0.07213085889816284, -0.05868282541632652, 0.04416143521666527, 0.0432087779045105, 0.04572809487581253, 0.15897540748119354, -0.09119901806116104, 0.05427839607000351, -0.21388612687587738, 0.015857785940170288, -0.03424825519323349, -0.06355158239603043, -0.05466584116220474, -0.030466757714748383, 0.08793677389621735, -0.0689416229724884, 0.0870811715722084, -0.038802195340394974, 0.03561116382479668, 0.03726715222001076, -0.121780626475811, 0.02228458970785141, 0.04998272657394409, 0.21091605722904205, 0.04472191631793976, -0.02356491982936859, 0.055778998881578445, 0.012624809518456459, 0.05729154869914055, 0.15730193257331848, 0.1587958037853241, 0.19344058632850647, 0.061144519597291946, 0.09419524669647217, 0.06006782129406929, -0.11463050544261932, -0.11116614192724228, 0.1278000921010971, -0.03475121408700943, 0.12978215515613556, -0.019207848235964775, 0.2534855306148529, 0.09744298458099365, -0.19280439615249634, 0.047736089676618576, -0.0359317809343338, -0.0852498859167099, -0.09182455390691757, -0.04217104613780975, -0.06380519270896912, -0.17535658180713654, 0.010174648836255074, -0.10154163092374802, 0.057794034481048584, 0.0592498853802681, 0.046883925795555115, 0.022670958191156387, 0.12295301258563995, 0.0526496097445488, -0.006831868086010218, 0.1301276683807373, 0.016083097085356712, -0.009887893684208393, -0.055895786732435226, -0.08547303080558777, 0.030557259917259216, -0.03102712333202362, 0.044698312878608704, -0.05051182582974434, -0.1089639887213707, 0.0627068504691124, 0.00767051987349987, -0.10593350976705551, 0.01965685747563839, -0.013170073740184307, 0.07301969826221466, 0.0961654856801033, 0.03366152569651604, -0.007710068020969629, -0.02927478775382042, 0.2534811496734619, -0.10421669483184814, -0.053058747202157974, -0.12722285091876984, 0.2409396916627884, 0.010610807687044144, -0.015738368034362793, 0.010018600150942802, -0.06611084938049316, -0.0013128221035003662, 0.16565552353858948, 0.13441741466522217, -0.0418873094022274, -0.011236431077122688, 0.01822277531027794, -0.009532976895570755, -0.05019773915410042, 0.08721230179071426, 0.13398852944374084, 0.07951795309782028, -0.08063901960849762, -0.04520833492279053, -0.04388820752501488, -0.04458744451403618, -0.03970989212393761, 0.05586434528231621, 0.03372360020875931, -0.00788616668432951, -0.030714042484760284, 0.10977508127689362, -0.06908083707094193, -0.11668902635574341, 0.011216970160603523, -0.18445949256420135, -0.19061565399169922, -0.0358768068253994, 0.07944762706756592, 0.03496045991778374, 0.051622726023197174, -0.008183971047401428, -0.023791372776031494, 0.08853431046009064, 0.00301368348300457, -0.042341940104961395, -0.10826966166496277, 0.08794477581977844, -0.0967019572854042, 0.1787789762020111, -0.04385281354188919, 0.032416652888059616, 0.11710511893033981, 0.07866038382053375, -0.06696678698062897, 0.046546898782253265, 0.06976951658725739, -0.1333034634590149, 0.04785196855664253, 0.18845361471176147, -0.03645295277237892, 0.13342595100402832, 0.043881434947252274, -0.1124177873134613, 0.027876203879714012, -0.10619135200977325, -0.059690993279218674, -0.05666695907711983, -0.00679200142621994, -0.03924848884344101, 0.13430576026439667, 0.22071141004562378, -0.06213665008544922, -0.01892172545194626, -0.06094592809677124, 0.0014529037289321423, 0.03220866620540619, 0.12829981744289398, -0.05180111527442932, -0.26491057872772217, 0.017862115055322647, -0.004963963758200407, 0.019309109076857567, -0.24263381958007812, -0.09641294926404953, 0.040871307253837585, -0.05532550811767578, -0.05992060527205467, 0.11141109466552734, 0.07717228680849075, 0.05313652381300926, -0.053489673882722855, -0.0904872864484787, -0.0340319499373436, 0.18386074900627136, -0.1734144389629364, -0.054290201514959335 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-large-xls-r-300m-bemba-fds This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the [BembaSpeech](https://github.com/csikasote/BembaSpeech) dataset. It achieves the following results on the evaluation set: - Loss: 0.3594 - Wer: 0.3838 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | 2.9961 | 0.67 | 500 | 0.5157 | 0.7133 | | 0.5903 | 1.34 | 1000 | 0.3663 | 0.4989 | | 0.4804 | 2.02 | 1500 | 0.3547 | 0.4653 | | 0.4146 | 2.69 | 2000 | 0.3274 | 0.4345 | | 0.3792 | 3.36 | 2500 | 0.3586 | 0.4640 | | 0.3509 | 4.03 | 3000 | 0.3360 | 0.4316 | | 0.3114 | 4.7 | 3500 | 0.3382 | 0.4303 | | 0.2935 | 5.38 | 4000 | 0.3263 | 0.4091 | | 0.2723 | 6.05 | 4500 | 0.3348 | 0.4175 | | 0.2502 | 6.72 | 5000 | 0.3317 | 0.4147 | | 0.2334 | 7.39 | 5500 | 0.3542 | 0.4030 | | 0.2287 | 8.06 | 6000 | 0.3594 | 0.4067 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer", "bem", "robust-speech-event"], "model-index": [{"name": "wav2vec2-large-xls-r-300m-bemba-fds", "results": []}]}
automatic-speech-recognition
csikasote/wav2vec2-large-xls-r-300m-bemba-fds
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "bem", "robust-speech-event", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #bem #robust-speech-event #license-apache-2.0 #endpoints_compatible #region-us
wav2vec2-large-xls-r-300m-bemba-fds =================================== This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the BembaSpeech dataset. It achieves the following results on the evaluation set: * Loss: 0.3594 * Wer: 0.3838 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.0003 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * gradient\_accumulation\_steps: 2 * total\_train\_batch\_size: 16 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 500 * num\_epochs: 30 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.11.3 * Pytorch 1.10.0+cu111 * Datasets 1.13.3 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.13.3\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #bem #robust-speech-event #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.13.3\n* Tokenizers 0.10.3" ]
[ 66, 158, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #bem #robust-speech-event #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.13.3\n* Tokenizers 0.10.3" ]
[ -0.11642204970121384, 0.07105162739753723, -0.0032327650114893913, 0.05935138836503029, 0.10804925858974457, 0.008824776858091354, 0.09769383817911148, 0.1461581140756607, -0.08202971518039703, 0.0781208798289299, 0.10415594279766083, 0.09022080153226852, 0.057156842201948166, 0.11077716201543808, -0.01727232336997986, -0.32496529817581177, 0.00855289027094841, 0.018603701144456863, -0.10595814883708954, 0.11788330227136612, 0.0965743213891983, -0.11080584675073624, 0.01853482984006405, 0.02176075428724289, -0.10881074517965317, 0.008794841356575489, -0.03086256980895996, -0.07456693798303604, 0.12404712289571762, 0.02672526054084301, 0.09635687619447708, 0.01481279544532299, 0.08993694186210632, -0.2617224156856537, 0.01805497705936432, 0.0601360909640789, 0.0452447384595871, 0.06894588470458984, 0.11001991480588913, -0.01918674446642399, 0.1439713090658188, -0.06951609253883362, 0.07789056748151779, 0.0381220206618309, -0.11263908445835114, -0.3121357560157776, -0.07556230574846268, 0.039194945245981216, 0.11500761657953262, 0.09932972490787506, -0.02624616026878357, 0.07412296533584595, -0.08596387505531311, 0.08044642955064774, 0.23574751615524292, -0.24767117202281952, -0.08367420732975006, -0.03843418508768082, 0.044787317514419556, 0.0508071668446064, -0.12937705218791962, -0.03697773441672325, 0.022230105474591255, 0.039575207978487015, 0.11633675545454025, 0.003035113448277116, -0.025166595354676247, 0.020055586472153664, -0.14748865365982056, -0.04856153205037117, 0.09924551099538803, 0.06699060648679733, -0.024013152346014977, -0.1031462624669075, -0.016670098528265953, -0.19324225187301636, -0.05854259058833122, 0.012616897001862526, 0.027831731364130974, -0.0390564501285553, -0.1074926033616066, 0.013527637347579002, -0.08634815365076065, -0.09000428766012192, 0.02399410307407379, 0.13273650407791138, 0.047846537083387375, -0.027514217421412468, -0.0033466094173491, 0.10992901027202606, 0.03789927065372467, -0.13441012799739838, -0.0015383580466732383, 0.03848718851804733, -0.10261216014623642, -0.016479594632983208, -0.029522741213440895, -0.022602232173085213, -0.005432269535958767, 0.11180628836154938, -0.03467527776956558, 0.08259224146604538, 0.022198209539055824, 0.031548719853162766, -0.08750500530004501, 0.18048010766506195, -0.06367941945791245, -0.035749245434999466, -0.0621141642332077, 0.10002798587083817, -0.022160494700074196, -0.009863988496363163, -0.06187991425395012, 0.028657488524913788, 0.10093221068382263, 0.04218939319252968, -0.025469662621617317, 0.02639133669435978, -0.06383515894412994, -0.026714803650975227, -0.042377445846796036, -0.0998840406537056, 0.04405368119478226, 0.03523106500506401, -0.08237608522176743, 0.007194072008132935, 0.007452397607266903, 0.016093892976641655, -0.031514331698417664, 0.1380993276834488, -0.060353729873895645, 0.020023714751005173, -0.09330327063798904, -0.09817984700202942, 0.03174858167767525, -0.05332774668931961, 0.0014473290648311377, -0.0659838318824768, -0.11715225875377655, -0.037345681339502335, 0.07214713841676712, -0.053171996027231216, -0.06172272935509682, -0.06468234956264496, -0.0643346980214119, 0.0473281666636467, -0.02786576934158802, 0.17190037667751312, -0.0649404525756836, 0.11146839708089828, 0.010632720775902271, 0.042202942073345184, 0.041975025087594986, 0.07583526521921158, -0.0471305213868618, 0.03618790954351425, -0.1367097645998001, 0.07246523350477219, -0.08542721718549728, 0.05528459697961807, -0.14423692226409912, -0.12663748860359192, -0.035410746932029724, 0.005450117401778698, 0.09583024680614471, 0.08968889713287354, -0.17589305341243744, -0.09818606823682785, 0.18748538196086884, -0.0658225268125534, -0.09061046689748764, 0.13710302114486694, -0.03514309599995613, 0.001434636884368956, 0.03586574271321297, 0.16498295962810516, 0.08187981694936752, -0.09542520344257355, 0.030670829117298126, -0.04751589521765709, 0.1085071936249733, 0.019410010427236557, 0.09359801560640335, -0.04591483622789383, 0.040867824107408524, -0.008653232827782631, -0.024437393993139267, 0.06962781399488449, -0.08650228381156921, -0.0918339267373085, -0.016018547117710114, -0.07248746603727341, 0.01806606538593769, 0.05936158075928688, 0.03623482957482338, -0.09356402605772018, -0.13542614877223969, 0.020977597683668137, 0.09724149107933044, -0.0999893769621849, 0.03318372368812561, -0.07781395316123962, 0.03291673958301544, -0.0234936885535717, -0.006928673014044762, -0.16472488641738892, 0.006022842135280371, 0.027724526822566986, -0.03251330181956291, 0.02942669577896595, 0.007198340259492397, 0.09034429490566254, 0.03366789594292641, -0.04564465954899788, -0.07492038607597351, -0.07198088616132736, -0.009423617273569107, -0.08505536615848541, -0.2303803712129593, -0.0728880986571312, -0.03313658386468887, 0.14923390746116638, -0.22927798330783844, 0.007641419768333435, 0.018822025507688522, 0.09565579146146774, 0.03894369304180145, -0.045168861746788025, -0.013378158211708069, 0.08079294860363007, -0.014222204685211182, -0.06478890776634216, 0.045129623264074326, 0.0003432764206081629, -0.11647335439920425, 0.03198867291212082, -0.11463712155818939, 0.09648185223340988, 0.10946719348430634, -0.029050694778561592, -0.08319709450006485, -0.058812499046325684, -0.07038350403308868, -0.06764727830886841, -0.03204961493611336, 0.007347938138991594, 0.22847974300384521, 0.039383381605148315, 0.12236560881137848, -0.07488413900136948, -0.057170454412698746, 0.027011068537831306, 0.0036384763661772013, -0.015733839944005013, 0.14567887783050537, 0.06663595139980316, -0.02865545265376568, 0.09468844532966614, 0.10481377691030502, -0.08937918394804001, 0.16136722266674042, -0.06713955849409103, -0.13388969004154205, -0.014283857308328152, 0.026248082518577576, 0.0318816676735878, 0.13149094581604004, -0.16994845867156982, -0.011627382598817348, 0.012051718309521675, 0.02743375673890114, 0.028924981132149696, -0.21799778938293457, -0.014090554788708687, 0.048563748598098755, -0.07135152816772461, -0.030067671090364456, -0.023785635828971863, 0.009864375926554203, 0.0921010747551918, 0.003547406056895852, -0.07470991462469101, -0.009074850007891655, -0.026401594281196594, -0.07883517444133759, 0.19178971648216248, -0.085688516497612, -0.1335299015045166, -0.12261681258678436, -0.041579440236091614, 0.010261958464980125, -0.01946091651916504, 0.04814016446471214, -0.11200560629367828, -0.0299681406468153, -0.0651434138417244, 0.053100213408470154, -0.06704670935869217, 0.03087404929101467, -0.004574017599225044, 0.009628809057176113, 0.07936310768127441, -0.10203389078378677, 0.01904400624334812, -0.03219730034470558, -0.04056921973824501, 0.038246069103479385, 0.03636636957526207, 0.09174990653991699, 0.1730836182832718, 0.03168663755059242, 0.02193032205104828, -0.04221764951944351, 0.13141675293445587, -0.09798125177621841, -0.03518134728074074, 0.11046680063009262, -0.007410194259136915, 0.04529428854584694, 0.11518555879592896, 0.06299767643213272, -0.07526032626628876, 0.01587788760662079, 0.052214063704013824, -0.017150668427348137, -0.23942255973815918, -0.02804531157016754, -0.062119197100400925, -0.009484624490141869, 0.12589356303215027, 0.030341779813170433, -0.019592707976698875, 0.028553687036037445, -0.0007235237862914801, -0.0033322246745228767, -0.0012167568784207106, 0.0631004050374031, 0.05352884158492088, 0.03585305064916611, 0.11341435462236404, -0.015800822526216507, -0.054018791764974594, 0.02248472347855568, 0.006871303077787161, 0.2613057494163513, -0.002978159347549081, 0.17276810109615326, 0.052130475640296936, 0.15718844532966614, 0.014976056292653084, 0.0653093233704567, 0.003019133349880576, -0.02921583503484726, 0.01114977989345789, -0.055232226848602295, -0.02408219687640667, 0.04849093407392502, 0.09847261756658554, 0.04298056289553642, -0.11722464859485626, -0.03057108260691166, 0.02557172067463398, 0.35804688930511475, 0.05881059169769287, -0.2943343222141266, -0.07927178591489792, 0.004134088754653931, -0.08846427500247955, -0.04031017795205116, 0.032678406685590744, 0.1088128313422203, -0.09653764218091965, 0.034130487591028214, -0.08836890757083893, 0.1008419618010521, -0.06662992388010025, 0.007587817031890154, 0.09294367581605911, 0.0750386118888855, -0.00037260568933561444, 0.06314448267221451, -0.2689807713031769, 0.3096538186073303, -0.02033594809472561, 0.07992976903915405, -0.05180973559617996, 0.03139610216021538, 0.03785758465528488, -0.06360232084989548, 0.07726127654314041, -0.0199615266174078, -0.09144127368927002, -0.19417549669742584, -0.07940112799406052, 0.021342435851693153, 0.1294575333595276, -0.043080635368824005, 0.12498253583908081, -0.0238997470587492, -0.005502240266650915, 0.06190606206655502, -0.07609972357749939, -0.0963478535413742, -0.11330961436033249, 0.025220856070518494, 0.018669266253709793, 0.05978403985500336, -0.10657833516597748, -0.11162808537483215, -0.0737803727388382, 0.1587395966053009, -0.0933208018541336, -0.0075712488032877445, -0.12869760394096375, 0.08382958918809891, 0.1607457399368286, -0.06697754561901093, 0.046112995594739914, 0.027778232470154762, 0.11211419105529785, 0.02475745789706707, -0.01023709774017334, 0.11652958393096924, -0.08930914103984833, -0.17883701622486115, -0.059259310364723206, 0.16129136085510254, 0.04708833247423172, 0.0722474604845047, -0.022078804671764374, 0.017122970893979073, -0.02367626689374447, -0.07888272404670715, 0.059478726238012314, 0.0019942910876125097, 0.013575609773397446, 0.04687574878334999, -0.038936953991651535, 0.004059562459588051, -0.08018998801708221, -0.05716080963611603, 0.16430659592151642, 0.2663310766220093, -0.0837044045329094, 0.033973969519138336, 0.038778144866228104, -0.048922717571258545, -0.1503930687904358, 0.02971569448709488, 0.1277240365743637, 0.030548470094799995, 0.005215685814619064, -0.23457874357700348, 0.05807854235172272, 0.08690203726291656, -0.01924259215593338, 0.08075001090765, -0.34997645020484924, -0.12842272222042084, 0.12790314853191376, 0.11160246282815933, -0.028205951675772667, -0.1571047008037567, -0.05232829973101616, 0.006147623527795076, -0.08941793441772461, 0.060492176562547684, -0.04468868672847748, 0.12465638667345047, -0.008387075737118721, 0.07196033746004105, 0.018478073179721832, -0.059567514806985855, 0.14030463993549347, -0.018645944073796272, 0.06053952872753143, -0.008291688747704029, 0.053259074687957764, 0.0009981045732274652, -0.04881560429930687, 0.012655270285904408, -0.07762975990772247, 0.003297524293884635, -0.1392897516489029, -0.03173503279685974, -0.08175375312566757, 0.02543814666569233, -0.03829570114612579, -0.05272878333926201, 0.006541144102811813, 0.04404396936297417, 0.060537129640579224, 0.00399318290874362, 0.11765255779027939, -0.059166572988033295, 0.1423650085926056, 0.06166262924671173, 0.10472243279218674, -0.013692797161638737, -0.11587861180305481, -0.018775708973407745, -0.0031067815143615007, 0.058151811361312866, -0.11414626985788345, 0.026379374787211418, 0.14308024942874908, 0.04463285207748413, 0.15381598472595215, 0.06504663079977036, -0.07863503694534302, 0.019573718309402466, 0.06480123847723007, -0.08416147530078888, -0.11178526282310486, -0.006011219695210457, 0.0755656287074089, -0.1414736807346344, 0.0045037842355668545, 0.10113336890935898, -0.05399082601070404, -0.007004304323345423, 0.013653994537889957, 0.01372851338237524, -0.05382975563406944, 0.22513410449028015, 0.023262914270162582, 0.07283252477645874, -0.08913706988096237, 0.0782613679766655, 0.05121917650103569, -0.18275173008441925, 0.024669094011187553, 0.0900406688451767, -0.02616008184850216, -0.020464003086090088, 0.029390821233391762, 0.08891602605581284, 0.018866000697016716, -0.055180590599775314, -0.11484213173389435, -0.15117982029914856, 0.09277691692113876, 0.10148010402917862, 0.03244103118777275, 0.02657465636730194, -0.059381451457738876, 0.04984970763325691, -0.10537686944007874, 0.08840467035770416, 0.10146014392375946, 0.07159218192100525, -0.12910374999046326, 0.14832879602909088, 0.009839209727942944, 0.006337940227240324, 0.004112366586923599, -0.008717585355043411, -0.10116711258888245, 0.037157054990530014, -0.11039368808269501, -0.02392484061419964, -0.054976850748062134, -0.0016911604907363653, 0.006276225205510855, -0.04820113256573677, -0.04745686799287796, 0.014724994078278542, -0.11911686509847641, -0.04186311364173889, -0.0065686083398759365, 0.06895825266838074, -0.106974296271801, -0.02320483885705471, 0.046767715364694595, -0.10935318470001221, 0.09944076836109161, 0.05987754464149475, 0.0168056171387434, 0.03753725066781044, -0.13740654289722443, 0.01219850778579712, 0.035885993391275406, -0.003142037196084857, 0.02874649316072464, -0.15354806184768677, -0.012834065593779087, -0.026474183425307274, 0.038067519664764404, -0.004230188671499491, 0.020253852009773254, -0.1395770162343979, -0.03791797161102295, -0.025256024673581123, -0.0799713060259819, -0.05594110116362572, 0.05011704936623573, 0.053903527557849884, 0.03896036744117737, 0.1659637987613678, -0.09072128683328629, 0.05486588925123215, -0.21362453699111938, 0.01662796549499035, -0.02860872820019722, -0.06160346046090126, -0.05748927965760231, -0.03425170108675957, 0.08527806401252747, -0.0718810111284256, 0.09161219000816345, -0.033815860748291016, 0.03035951964557171, 0.035811588168144226, -0.12240120768547058, 0.0337725505232811, 0.04660840705037117, 0.22958451509475708, 0.04829967021942139, -0.02561136707663536, 0.06149125099182129, 0.009579875506460667, 0.057599637657403946, 0.16682027280330658, 0.1645125299692154, 0.19963935017585754, 0.07329437881708145, 0.09309066832065582, 0.06433558464050293, -0.11104727536439896, -0.11860987544059753, 0.12292242795228958, -0.02486320771276951, 0.1247721016407013, -0.01674807071685791, 0.2606149911880493, 0.10573980957269669, -0.19518589973449707, 0.04914055019617081, -0.0347851924598217, -0.08653022348880768, -0.08988930284976959, -0.0518781803548336, -0.06154841557145119, -0.17124047875404358, 0.011974550783634186, -0.10315363854169846, 0.05867791920900345, 0.06041288375854492, 0.04719462990760803, 0.018299778923392296, 0.13236099481582642, 0.05613773688673973, -0.00013302125444170088, 0.1280936896800995, 0.01984352618455887, -0.0070845335721969604, -0.05445149168372154, -0.08836081624031067, 0.026036176830530167, -0.03955768421292305, 0.043172646313905716, -0.053678177297115326, -0.10786096751689911, 0.05654522404074669, 0.00924211461097002, -0.1051776111125946, 0.02060002088546753, -0.021230138838291168, 0.07467207312583923, 0.09668188542127609, 0.033540233969688416, -0.008884420618414879, -0.028330523520708084, 0.2552233636379242, -0.1063055619597435, -0.05909974128007889, -0.12636949121952057, 0.24111811816692352, 0.010326492600142956, -0.01855415850877762, 0.010307423770427704, -0.06437242776155472, 0.007066073827445507, 0.167831152677536, 0.14516139030456543, -0.04295959696173668, -0.012839007191359997, 0.015348630025982857, -0.01036688219755888, -0.05425727367401123, 0.08574160188436508, 0.13324017822742462, 0.07465913891792297, -0.07852044701576233, -0.04272184148430824, -0.04796802997589111, -0.045292407274246216, -0.042928364127874374, 0.06502865999937057, 0.038791585713624954, -0.008285760879516602, -0.03336536884307861, 0.10913986712694168, -0.07068584859371185, -0.10683175176382065, 0.014461176469922066, -0.17723605036735535, -0.18719615042209625, -0.03738768398761749, 0.08035410195589066, 0.030862746760249138, 0.05375700443983078, -0.0074777379631996155, -0.02240670472383499, 0.08420631289482117, 0.005653626751154661, -0.044513002038002014, -0.10858403891324997, 0.08933018893003464, -0.08807437866926193, 0.19239313900470734, -0.04636577144265175, 0.02029094286262989, 0.12016241997480392, 0.0779624730348587, -0.066758893430233, 0.05166080966591835, 0.07169485837221146, -0.133012592792511, 0.04257615655660629, 0.19021637737751007, -0.03350203111767769, 0.14098648726940155, 0.044737521559000015, -0.11258760094642639, 0.030386250466108322, -0.10244165360927582, -0.05987321585416794, -0.05715717375278473, -0.010790384374558926, -0.03547893464565277, 0.13327933847904205, 0.22671107947826385, -0.0602031834423542, -0.021650169044733047, -0.0603790245950222, 0.002122508594766259, 0.03706914186477661, 0.1202511340379715, -0.05504283308982849, -0.26333460211753845, 0.010316315107047558, 0.000727058679331094, 0.015040190890431404, -0.25290733575820923, -0.10330981761217117, 0.04995787516236305, -0.05252068117260933, -0.06006643921136856, 0.1072792336344719, 0.07924424856901169, 0.0580766536295414, -0.05177401751279831, -0.08006090670824051, -0.0351671501994133, 0.18152499198913574, -0.1717243492603302, -0.05104629695415497 ]
null
null
transformers
# Wav2Vec2-Large-XLSR-53-Bemba Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on Bemba language of Zambia using the [BembaSpeech](https://csikasote.github.io/BembaSpeech). When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ```python import torch import torchaudio from datasets import load_dataset from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor test_dataset = load_dataset("csv", data_files={"test": "/content/test.csv"}, delimiter="\t")["test"] # Adapt the path to test.csv processor = Wav2Vec2Processor.from_pretrained("csikasote/wav2vec2-large-xlsr-bemba") model = Wav2Vec2ForCTC.from_pretrained("csikasote/wav2vec2-large-xlsr-bemba") #BembaSpeech is sample at 16kHz so we you do not need to resample #resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): speech_array, sampling_rate = torchaudio.load(batch["path"]) batch["speech"] = speech_array.squeeze().numpy() return batch test_dataset = test_dataset.map(speech_file_to_array_fn) inputs = processor(test_dataset["speech"][:2], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits predicted_ids = torch.argmax(logits, dim=-1) print("Prediction:", processor.batch_decode(predicted_ids)) print("Reference:", test_dataset["sentence"][:2]) ``` ## Evaluation The model can be evaluated as follows on the Bemba test data of BembaSpeech. ```python import torch import torchaudio from datasets import load_dataset, load_metric from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor import re test_dataset = load_dataset("csv", data_files={"test": "/content/test.csv"}, delimiter="\\t")["test"] wer = load_metric("wer") processor = Wav2Vec2Processor.from_pretrained("csikasote/wav2vec2-large-xlsr-bemba") model = Wav2Vec2ForCTC.from_pretrained("csikasote/wav2vec2-large-xlsr-bemba") model.to("cuda") chars_to_ignore_regex = '[\,\_\?\.\!\;\:\"\“]' #resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): batch["sentence"] = re.sub(chars_to_ignore_regex, '', batch["sentence"]).lower() speech_array, sampling_rate = torchaudio.load(batch["path"]) batch["speech"] = speech_array.squeeze().numpy() return batch test_dataset = test_dataset.map(speech_file_to_array_fn) # Preprocessing the datasets. # We need to read the aduio files as arrays def evaluate(batch): inputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits pred_ids = torch.argmax(logits, dim=-1) batch["pred_strings"] = processor.batch_decode(pred_ids) return batch result = test_dataset.map(evaluate, batched=True, batch_size=8) print("WER: {:2f}".format(100 * wer.compute(predictions=result["pred_strings"], references=result["sentence"]))) ``` **Test Result**: 42.17 % ## Training The BembaSpeech `train`, `dev` and `test` datasets were used for training, development and evaluation respectively. The script used for evaluating the model on the test dataset can be found [here](https://colab.research.google.com/drive/1aplFHfaXE68HGDwBYV2KqUWPasrk7bXv?usp=sharing).
{"language": "bem", "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech", "xlsr-fine-tuning-week"], "datasets": ["BembaSpeech"], "metrics": ["wer"], "model-index": [{"name": "XLSR Wav2Vec2 Bemba by Claytone Sikasote", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Speech Recognition"}, "dataset": {"name": "BembaSpeech bem", "type": "bembaspeech", "args": "bem"}, "metrics": [{"type": "wer", "value": 42.17, "name": "Test WER"}]}]}]}
automatic-speech-recognition
csikasote/wav2vec2-large-xlsr-bemba
[ "transformers", "pytorch", "jax", "wav2vec2", "automatic-speech-recognition", "audio", "speech", "xlsr-fine-tuning-week", "bem", "dataset:BembaSpeech", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "bem" ]
TAGS #transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #bem #dataset-BembaSpeech #license-apache-2.0 #model-index #endpoints_compatible #region-us
# Wav2Vec2-Large-XLSR-53-Bemba Fine-tuned facebook/wav2vec2-large-xlsr-53 on Bemba language of Zambia using the BembaSpeech. When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ## Evaluation The model can be evaluated as follows on the Bemba test data of BembaSpeech. Test Result: 42.17 % ## Training The BembaSpeech 'train', 'dev' and 'test' datasets were used for training, development and evaluation respectively. The script used for evaluating the model on the test dataset can be found here.
[ "# Wav2Vec2-Large-XLSR-53-Bemba\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Bemba language of Zambia using the BembaSpeech. When using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:", "## Evaluation\n\nThe model can be evaluated as follows on the Bemba test data of BembaSpeech. \n\n\n\n\nTest Result: 42.17 %", "## Training\n\nThe BembaSpeech 'train', 'dev' and 'test' datasets were used for training, development and evaluation respectively. The script used for evaluating the model on the test dataset can be found here." ]
[ "TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #bem #dataset-BembaSpeech #license-apache-2.0 #model-index #endpoints_compatible #region-us \n", "# Wav2Vec2-Large-XLSR-53-Bemba\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Bemba language of Zambia using the BembaSpeech. When using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:", "## Evaluation\n\nThe model can be evaluated as follows on the Bemba test data of BembaSpeech. \n\n\n\n\nTest Result: 42.17 %", "## Training\n\nThe BembaSpeech 'train', 'dev' and 'test' datasets were used for training, development and evaluation respectively. The script used for evaluating the model on the test dataset can be found here." ]
[ 80, 70, 20, 32, 52 ]
[ "passage: TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #bem #dataset-BembaSpeech #license-apache-2.0 #model-index #endpoints_compatible #region-us \n# Wav2Vec2-Large-XLSR-53-Bemba\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Bemba language of Zambia using the BembaSpeech. When using this model, make sure that your speech input is sampled at 16kHz.## Usage\n\nThe model can be used directly (without a language model) as follows:## Evaluation\n\nThe model can be evaluated as follows on the Bemba test data of BembaSpeech. \n\n\n\n\nTest Result: 42.17 %## Training\n\nThe BembaSpeech 'train', 'dev' and 'test' datasets were used for training, development and evaluation respectively. The script used for evaluating the model on the test dataset can be found here." ]
[ -0.11402676999568939, 0.05119815841317177, -0.002567385323345661, 0.032641369849443436, 0.02792268805205822, -0.007559408899396658, 0.1468786597251892, 0.09148366749286652, 0.04825704172253609, 0.0037539098411798477, 0.041110049933195114, -0.01816781610250473, 0.019245078787207603, 0.13581177592277527, -0.031562089920043945, -0.15989825129508972, 0.01000458188354969, 0.04719199240207672, 0.011200756765902042, 0.13568225502967834, 0.09662226587533951, -0.05515754595398903, 0.011758833192288876, 0.053429264575242996, -0.12224258482456207, 0.0031478942837566137, 0.009304547682404518, -0.10648516565561295, 0.050544317811727524, 0.06362245976924896, 0.06863436847925186, 0.032085105776786804, 0.05212550610303879, -0.1724461019039154, 0.018365079537034035, 0.038605883717536926, 0.04521987587213516, -0.010347997769713402, 0.09459565579891205, 0.049519505351781845, 0.10866860300302505, 0.07134600728750229, -0.024740854278206825, 0.06101509928703308, -0.05834802985191345, -0.17510481178760529, -0.048470787703990936, -0.02722611464560032, 0.09817183017730713, 0.1095299944281578, -0.047259457409381866, 0.21910113096237183, -0.1298525184392929, 0.03394460678100586, 0.057418908923864365, -0.12852613627910614, -0.0030671358108520508, 0.032356955111026764, -0.005085684824734926, 0.10380692034959793, -0.03927881270647049, -0.02348422445356846, 0.05534601956605911, -0.00011405583063606173, 0.019994694739580154, -0.040368854999542236, -0.07918297499418259, -0.07066065818071365, -0.17165187001228333, -0.0385994054377079, 0.2746437191963196, -0.03183497488498688, -0.08278901875019073, -0.168972909450531, -0.029163489118218422, 0.12276619672775269, 0.0026826246175915003, -0.03797613084316254, -0.01996498927474022, 0.034884482622146606, -0.00927745457738638, -0.022918803617358208, -0.08447576314210892, -0.09975846856832504, 0.01561116985976696, 0.06140368431806564, 0.0043815672397613525, 0.02420131303369999, -0.15158821642398834, 0.09410829097032547, -0.036046840250492096, -0.11366578191518784, -0.09445445984601974, -0.050803110003471375, -0.0589204765856266, -0.04888574406504631, -0.041607268154621124, -0.1221025213599205, 0.013835631310939789, 0.08382418751716614, 0.0627203956246376, 0.019495561718940735, -0.05641818046569824, -0.022904550656676292, 0.07554906606674194, 0.15911853313446045, -0.009862028062343597, -0.12195843458175659, -0.004000307992100716, -0.014971519820392132, -0.04096590727567673, -0.010266456753015518, -0.01742146909236908, -0.13591083884239197, 0.029558679088950157, 0.03982339799404144, 0.014316733926534653, -0.007207429502159357, -0.09192502498626709, -0.09773623943328857, 0.044380296021699905, -0.10779235512018204, 0.01029337476938963, 0.0954606831073761, 0.026097120717167854, 0.061006635427474976, 0.06529638171195984, 0.033245258033275604, -0.07873187214136124, 0.00894546415656805, -0.009607864543795586, 0.043795835226774216, -0.06110897660255432, 0.0011063746642321348, 0.03484431654214859, 0.03041890449821949, -0.009663070552051067, -0.05551423132419586, -0.1877417415380478, -0.021451566368341446, 0.01582864671945572, -0.026279568672180176, 0.04641132801771164, -0.01708274334669113, 0.011126321740448475, -0.01919720135629177, -0.07026433944702148, 0.048882585018873215, -0.05165340378880501, 0.04389283433556557, 0.10050851851701736, 0.017603067681193352, 0.06058492511510849, 0.0639244094491005, -0.051123231649398804, -0.017634257674217224, -0.0659731924533844, 0.16305950284004211, -0.047074273228645325, -0.08798190951347351, -0.08302280306816101, -0.0628179982304573, -0.09615723788738251, 0.012734967283904552, 0.0009433082304894924, 0.14242075383663177, -0.19441474974155426, -0.10037834197282791, 0.29011282324790955, -0.1533249467611313, -0.03706808760762215, 0.2002486139535904, -0.06116309016942978, 0.12458428740501404, 0.0732731968164444, 0.09326325356960297, 0.25242769718170166, -0.24607202410697937, 0.0595317967236042, -0.0028976909816265106, -0.015646815299987793, -0.028511714190244675, 0.10774178057909012, -0.07178083807229996, 0.034416262060403824, 0.022609593346714973, -0.11185327917337418, 0.04741741716861725, 0.012579974718391895, -0.04390370100736618, -0.06353918462991714, -0.043154027312994, 0.0898401290178299, -0.007273537106812, -0.032990243285894394, -0.017568524926900864, -0.032375603914260864, 0.11288773268461227, 0.11997359991073608, -0.09233096987009048, 0.02224144898355007, -0.12038425356149673, 0.08138405531644821, -0.12840905785560608, -0.04613185301423073, -0.1431664526462555, 0.1678144484758377, 0.04481362923979759, 0.05265092849731445, 0.06366736441850662, 0.10593932867050171, 0.01963040977716446, -0.02659965120255947, 0.00006569059769390151, -0.00092760578263551, -0.0036655620206147432, -0.047775913029909134, -0.023103656247258186, -0.049400102347135544, -0.033814121037721634, -0.043489232659339905, 0.1380770355463028, -0.16648966073989868, -0.024239934980869293, 0.06385377049446106, 0.043203435838222504, 0.03233013302087784, -0.02036481909453869, 0.07500220835208893, 0.05005182325839996, -0.0022660617250949144, -0.0036004353314638138, 0.011058742180466652, 0.023394852876663208, 0.007206321228295565, 0.08963929861783981, -0.07012096047401428, -0.07655948400497437, 0.10755623131990433, 0.05983589217066765, 0.03671487793326378, 0.03412112221121788, 0.006359199061989784, -0.0194699764251709, -0.09747394174337387, -0.03364308923482895, 0.2849104404449463, -0.0002595496771391481, 0.12454771995544434, -0.13422122597694397, -0.11305107921361923, 0.004002481698989868, -0.09278848767280579, 0.0023345891386270523, -0.007501774933189154, -0.025169292464852333, 0.06136335805058479, 0.03525887802243233, 0.09008271247148514, -0.05469901114702225, 0.37327003479003906, -0.07062440365552902, -0.15040352940559387, -0.03721587359905243, -0.01442366000264883, -0.03502550721168518, 0.06296265870332718, -0.21201345324516296, 0.0013183228438720107, 0.022777849808335304, 0.08300813287496567, 0.06733010709285736, -0.10727293789386749, 0.05026540532708168, 0.07194896787405014, -0.12010014057159424, -0.08505628257989883, 0.06597176194190979, -0.004756307229399681, 0.027366047725081444, -0.10573527961969376, 0.03062012977898121, 0.008399078622460365, -0.060784853994846344, -0.18085569143295288, 0.12640699744224548, -0.07644478976726532, -0.12343776971101761, -0.17758555710315704, 0.02291776053607464, 0.009832089766860008, 0.01341453567147255, 0.04111046716570854, -0.10137458890676498, -0.01079022977501154, -0.07074816524982452, 0.0749952495098114, 0.003562260651960969, 0.0011466802097856998, 0.08747672289609909, 0.04891251400113106, 0.07298069447278976, -0.14551319181919098, 0.030319849029183388, -0.025761524215340614, 0.0289551243185997, 0.08175504207611084, -0.06016882136464119, -0.023810306563973427, 0.1665678322315216, 0.0023630065843462944, 0.015042845159769058, 0.029394501820206642, 0.1878678798675537, -0.08391479402780533, -0.009103856980800629, 0.24412864446640015, -0.08055933564901352, -0.024373240768909454, 0.030973365530371666, -0.0415944866836071, -0.07421433925628662, 0.018266255035996437, -0.027148820459842682, -0.0476236455142498, -0.2726791501045227, -0.10725116729736328, -0.07684226334095001, -0.0397978313267231, -0.020245671272277832, -0.0624493770301342, -0.017107617110013962, 0.02362840808928013, -0.0230970848351717, -0.09212023764848709, 0.03189076855778694, 0.010160288773477077, 0.06616007536649704, 0.019883714616298676, 0.07480582594871521, -0.052167050540447235, 0.0016543969977647066, 0.06360036879777908, 0.04046715423464775, 0.15992900729179382, 0.06493605673313141, 0.1402808129787445, 0.044693246483802795, 0.05988204479217529, 0.1088060736656189, 0.050519488751888275, -0.07994088530540466, 0.013482474721968174, 0.012778524309396744, -0.06213885918259621, -0.10586826503276825, 0.04130791872739792, 0.09067783504724503, -0.048366330564022064, 0.004199671093374491, -0.014635931700468063, 0.03456130996346474, 0.1383078545331955, 0.09571094810962677, -0.1588689237833023, -0.0734986737370491, 0.021597711369395256, -0.07000913470983505, -0.03145158663392067, 0.01377718336880207, 0.14760246872901917, -0.08939633518457413, 0.002753143198788166, -0.05190659686923027, 0.07609601318836212, -0.004292440600693226, -0.022738557308912277, -0.02656693384051323, -0.0002134158567059785, -0.0016307021724060178, 0.03233605623245239, -0.21354453265666962, 0.11329182982444763, 0.005600000731647015, 0.1148214042186737, -0.0709080770611763, 0.05218496173620224, 0.003848955500870943, 0.014175275340676308, 0.08534117043018341, 0.02938135899603367, -0.055840786546468735, -0.06571288406848907, -0.044679850339889526, 0.03460681810975075, 0.04495524615049362, 0.11052785813808441, 0.13031268119812012, -0.03636139631271362, 0.03810178115963936, 0.026198802515864372, -0.014020930044353008, -0.14199389517307281, -0.06736090779304504, 0.05907151475548744, 0.08022863417863846, 0.07071982324123383, -0.03623070567846298, -0.03135858476161957, -0.030399955809116364, -0.029320264235138893, -0.30610525608062744, -0.09821724146604538, -0.09336109459400177, -0.08102483302354813, 0.16308315098285675, -0.056585896760225296, 0.0067940219305455685, 0.04960436746478081, 0.18473686277866364, -0.021877571940422058, -0.053759459406137466, -0.01611659675836563, -0.11844930052757263, -0.16067031025886536, -0.06691883504390717, 0.14397098124027252, 0.12278534471988678, 0.05514903739094734, 0.13213564455509186, -0.007941613905131817, 0.016118640080094337, -0.0777367427945137, -0.013303716666996479, -0.008064281195402145, -0.00722178490832448, -0.013772809877991676, -0.020284265279769897, -0.16410236060619354, -0.11972325295209885, -0.0932646244764328, 0.09428597241640091, 0.0773499608039856, -0.036065611988306046, 0.13466379046440125, 0.131798654794693, -0.12319918721914291, -0.18839648365974426, -0.005505163222551346, 0.15768221020698547, 0.09725981950759888, -0.03528701141476631, -0.2258608192205429, 0.07612390071153641, -0.004147874657064676, -0.03167828917503357, 0.0010643138084560633, -0.3411238193511963, -0.1571797877550125, 0.15210069715976715, -0.05897950381040573, 0.13372956216335297, -0.031187715008854866, -0.014116554521024227, 0.008616571314632893, -0.06502488255500793, 0.06197519972920418, -0.17603731155395508, 0.06579435616731644, 0.006767685525119305, 0.14619959890842438, 0.023661985993385315, -0.031049374490976334, 0.10495965927839279, 0.11981704086065292, 0.010064659640192986, 0.011250914074480534, 0.1281462013721466, 0.002638953970745206, -0.013070867396891117, 0.12075597792863846, 0.004360131919384003, 0.04252878576517105, -0.19425877928733826, -0.1115245670080185, -0.05914901942014694, 0.06670618057250977, -0.005736277438700199, -0.07461367547512054, 0.08949266374111176, -0.001211735769174993, 0.015025469474494457, -0.02263505570590496, -0.09886185079813004, -0.1328015774488449, -0.10228448361158371, 0.1783398985862732, 0.19953781366348267, -0.04518771171569824, -0.05807019770145416, -0.022539330646395683, 0.02340712770819664, 0.13001057505607605, -0.06613221764564514, 0.09600427746772766, -0.005525417160242796, 0.05094051733613014, 0.08875790983438492, -0.008935141377151012, -0.1000806912779808, 0.061726756393909454, 0.012642735615372658, -0.03448742628097534, 0.007508476264774799, 0.03973696753382683, -0.023532560095191002, -0.07782895117998123, -0.029690144583582878, 0.13595251739025116, -0.03479207307100296, -0.029690463095903397, -0.015380775555968285, -0.04193135350942612, -0.09177077561616898, 0.21457697451114655, -0.008167346939444542, 0.052927032113075256, -0.06374526768922806, 0.04346054047346115, -0.034701451659202576, 0.04031415656208992, 0.03317999094724655, -0.06606732308864594, -0.09464219957590103, -0.031064288690686226, -0.04751535877585411, 0.06856422871351242, 0.02863999456167221, -0.11312317848205566, -0.01259578112512827, -0.059983618557453156, 0.020391426980495453, 0.17140142619609833, 0.08305946737527847, 0.048882149159908295, -0.08487977832555771, -0.05058855935931206, -0.10209020227193832, 0.0842592790722847, 0.12565277516841888, -0.004672727547585964, -0.0964738130569458, 0.10078300535678864, 0.0010211686603724957, -0.013930203393101692, -0.08237078785896301, -0.049396514892578125, -0.008221421390771866, 0.05862531065940857, -0.15691453218460083, 0.01930248737335205, -0.034537363797426224, -0.0016348998760804534, -0.015143400058150291, -0.09842921793460846, -0.0047645606100559235, 0.08801037818193436, -0.04367654398083687, 0.05770228058099747, -0.03978825733065605, 0.10553249716758728, -0.06316021084785461, 0.006277130916714668, 0.12002372741699219, -0.10096951574087143, 0.09244203567504883, 0.12858596444129944, -0.07513650506734848, 0.07524171471595764, -0.18421579897403717, -0.05984693765640259, -0.0063031925819814205, 0.07542303204536438, -0.03346122056245804, -0.17882612347602844, 0.05773866921663284, 0.06747439503669739, 0.08202527463436127, 0.0021239868365228176, 0.16336041688919067, -0.07043690234422684, 0.00866671558469534, -0.037976332008838654, -0.06292599439620972, 0.009330593049526215, 0.03590119630098343, 0.03899691626429558, 0.08188971132040024, 0.20004884898662567, -0.10805771499872208, 0.062675341963768, -0.15348194539546967, 0.001283116522245109, 0.010913359932601452, 0.023781538009643555, -0.08169926702976227, -0.07565008848905563, 0.03164950758218765, -0.01775483600795269, 0.1698802262544632, 0.018968505784869194, -0.019118165597319603, 0.011408350430428982, -0.11767680197954178, 0.013037394732236862, -0.01622205413877964, 0.32799655199050903, 0.07933727651834488, 0.0526682510972023, 0.019118532538414, -0.03185566887259483, 0.031174685806035995, 0.021244777366518974, 0.027955137193202972, 0.23655539751052856, -0.03584771975874901, 0.055941395461559296, 0.026248235255479813, -0.03956519812345505, -0.04291485622525215, -0.06063557416200638, -0.14225977659225464, 0.0072090644389390945, -0.005212206393480301, 0.13297975063323975, 0.1491418182849884, -0.1566811352968216, 0.038653019815683365, 0.033466484397649765, -0.09304343163967133, -0.14850623905658722, -0.04349571466445923, -0.0751299187541008, -0.12249922752380371, 0.05198162421584129, -0.11120176315307617, 0.04862847924232483, 0.013661680743098259, 0.054477509111166, -0.03944515064358711, 0.1572827249765396, 0.034164804965257645, -0.15979093313217163, 0.08589629083871841, -0.04941673204302788, -0.002540275687351823, 0.022945189848542213, 0.018409019336104393, 0.09687808156013489, 0.019978446885943413, 0.059022605419158936, 0.027956482023000717, -0.06983789801597595, -0.018371352925896645, -0.021409625187516212, -0.08530139923095703, -0.05306343361735344, -0.007301513105630875, 0.030104996636509895, 0.15259894728660583, 0.08321331441402435, -0.032554350793361664, -0.004840237554162741, 0.025256795808672905, -0.02610851265490055, -0.1591571867465973, -0.17115171253681183, 0.06047922745347023, -0.020567791536450386, 0.03853590041399002, -0.009661458432674408, -0.04909299314022064, 0.008531340397894382, 0.2331211119890213, 0.0982486680150032, -0.011599550023674965, 0.01166444830596447, -0.015591875649988651, 0.010415596887469292, -0.022088050842285156, 0.11301323771476746, 0.03368637338280678, 0.22746892273426056, -0.000146206803037785, 0.021272625774145126, -0.05836465582251549, -0.07232623547315598, -0.03929532319307327, 0.009007580578327179, -0.11145032942295074, -0.08763757348060608, 0.005727680400013924, 0.08912855386734009, -0.05028896406292915, -0.10550379008054733, -0.041181594133377075, -0.008882356807589531, -0.08305074274539948, 0.014646690338850021, -0.0023573474027216434, 0.02787480689585209, 0.007349254097789526, -0.04862155765295029, -0.02827600948512554, 0.20125311613082886, -0.02425425872206688, -0.033055298030376434, 0.018468398600816727, 0.03013998456299305, -0.07238476723432541, 0.06435370445251465, 0.0243894774466753, 0.21409352123737335, 0.018085859715938568, 0.05105030536651611, -0.06675748527050018, 0.11822754144668579, 0.031270138919353485, 0.017909174785017967, -0.003152647754177451, 0.09750186651945114, -0.023254379630088806, 0.15412300825119019, 0.05374908447265625, -0.055436231195926666, 0.018474822863936424, -0.056974444538354874, -0.07902318984270096, -0.14260268211364746, 0.06208081543445587, -0.04710603877902031, 0.10597787797451019, 0.10424135625362396, -0.102015420794487, -0.06680101156234741, -0.06005467101931572, 0.09355487674474716, 0.013736818917095661, 0.05787495896220207, -0.0724249929189682, -0.20198598504066467, -0.007800431922078133, 0.017532341182231903, 0.0069350916892290115, -0.21665997803211212, 0.026965105906128883, 0.042214978486299515, -0.062201905995607376, 0.04779614135622978, 0.07420890033245087, 0.03364920616149902, 0.049475278705358505, 0.024295317009091377, -0.006542949005961418, 0.035829219967126846, 0.06735040247440338, -0.14909765124320984, -0.06513954699039459 ]
null
null
transformers
### marianmt-th-zh_cn * source languages: th * target languages: zh_cn * dataset: * model: transformer-align * pre-processing: normalization + SentencePiece * test set translations: * test set scores: ## Training Training scripts from [LalitaDeelert/NLP-ZH_TH-Project](https://github.com/LalitaDeelert/NLP-ZH_TH-Project). Experiments tracked at [cstorm125/marianmt-th-zh_cn](https://wandb.ai/cstorm125/marianmt-th-zh_cn). ``` export WANDB_PROJECT=marianmt-th-zh_cn python train_model.py --input_fname ../data/v1/Train.csv \ --output_dir ../models/marianmt-th-zh_cn \ --source_lang th --target_lang zh \ --metric_tokenize zh --fp16 ``` ## Usage ``` from transformers import AutoTokenizer, AutoModelForSeq2SeqLM tokenizer = AutoTokenizer.from_pretrained("cstorm125/marianmt-zh_cn-th") model = AutoModelForSeq2SeqLM.from_pretrained("cstorm125/marianmt-zh_cn-th").cpu() src_text = [ 'ฉันรักคุณ', 'ฉันอยากกินข้าว', ] translated = model.generate(**tokenizer(src_text, return_tensors="pt", padding=True)) print([tokenizer.decode(t, skip_special_tokens=True) for t in translated]) > ['我爱你', '我想吃饭。'] ``` ## Requirements ``` transformers==4.6.0 torch==1.8.0 ```
{"tags": ["translation", "torch==1.8.0"], "widget": [{"text": "Inference Unavailable"}]}
translation
cstorm125/marianmt-th-zh_cn
[ "transformers", "pytorch", "marian", "text2text-generation", "translation", "torch==1.8.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #marian #text2text-generation #translation #torch==1.8.0 #autotrain_compatible #endpoints_compatible #region-us
### marianmt-th-zh_cn * source languages: th * target languages: zh_cn * dataset: * model: transformer-align * pre-processing: normalization + SentencePiece * test set translations: * test set scores: ## Training Training scripts from LalitaDeelert/NLP-ZH_TH-Project. Experiments tracked at cstorm125/marianmt-th-zh_cn. ## Usage ## Requirements
[ "### marianmt-th-zh_cn\n* source languages: th\n* target languages: zh_cn\n* dataset: \n* model: transformer-align\n* pre-processing: normalization + SentencePiece\n* test set translations: \n* test set scores:", "## Training\n\nTraining scripts from LalitaDeelert/NLP-ZH_TH-Project. Experiments tracked at cstorm125/marianmt-th-zh_cn.", "## Usage", "## Requirements" ]
[ "TAGS\n#transformers #pytorch #marian #text2text-generation #translation #torch==1.8.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### marianmt-th-zh_cn\n* source languages: th\n* target languages: zh_cn\n* dataset: \n* model: transformer-align\n* pre-processing: normalization + SentencePiece\n* test set translations: \n* test set scores:", "## Training\n\nTraining scripts from LalitaDeelert/NLP-ZH_TH-Project. Experiments tracked at cstorm125/marianmt-th-zh_cn.", "## Usage", "## Requirements" ]
[ 49, 63, 41, 3, 5 ]
[ "passage: TAGS\n#transformers #pytorch #marian #text2text-generation #translation #torch==1.8.0 #autotrain_compatible #endpoints_compatible #region-us \n### marianmt-th-zh_cn\n* source languages: th\n* target languages: zh_cn\n* dataset: \n* model: transformer-align\n* pre-processing: normalization + SentencePiece\n* test set translations: \n* test set scores:## Training\n\nTraining scripts from LalitaDeelert/NLP-ZH_TH-Project. Experiments tracked at cstorm125/marianmt-th-zh_cn.## Usage## Requirements" ]
[ -0.17309942841529846, -0.02402368374168873, -0.0025235656648874283, 0.04286745563149452, 0.08233149349689484, -0.004156056325882673, 0.0526319295167923, 0.08169356733560562, -0.023285694420337677, 0.006210209336131811, 0.12547250092029572, 0.08747147768735886, 0.005561168305575848, 0.030811857432127, -0.053131479769945145, -0.17585749924182892, 0.025161365047097206, 0.040346842259168625, -0.1667453944683075, 0.13419905304908752, 0.11687688529491425, -0.049648385494947433, 0.1047242134809494, 0.03446459770202637, -0.04025081545114517, 0.02850279025733471, -0.022890605032444, -0.11238297075033188, 0.13362814486026764, 0.060239315032958984, 0.0926518589258194, 0.10185325890779495, 0.07365555316209793, -0.1769273728132248, 0.01952873356640339, -0.0172830019146204, -0.0032912064343690872, 0.02604632079601288, 0.09201551228761673, -0.05003215745091438, 0.2323170006275177, -0.09471257776021957, -0.026848401874303818, 0.07056120783090591, -0.05314893275499344, -0.0943462997674942, -0.06700225174427032, -0.03450635075569153, 0.024889755994081497, 0.1406448781490326, -0.012420358136296272, 0.19792751967906952, -0.169630765914917, 0.09684810042381287, 0.19849856197834015, -0.2584282159805298, -0.04796166718006134, 0.07071937620639801, 0.11320475488901138, -0.012481819838285446, 0.013054423034191132, 0.09186478704214096, 0.09372807294130325, 0.0055234432220458984, -0.09208069741725922, -0.09758717566728592, -0.020047591999173164, 0.04602647200226784, -0.15966439247131348, -0.007107562851160765, 0.2540832757949829, -0.014960519969463348, -0.04581663757562637, -0.00582644110545516, -0.03126008063554764, -0.018019165843725204, -0.010455912910401821, -0.007116223219782114, -0.04998723790049553, -0.0371963307261467, -0.04042372107505798, -0.026547111570835114, -0.10129772126674652, -0.09563121199607849, -0.04754158854484558, 0.16776001453399658, 0.0699511170387268, -0.005491908639669418, -0.0990767851471901, 0.11782071739435196, 0.14426645636558533, -0.12656310200691223, -0.05872180312871933, -0.06933476030826569, -0.03659417852759361, -0.006343928165733814, -0.04225388169288635, -0.11000965535640717, 0.11715041100978851, 0.08201926201581955, -0.021041065454483032, 0.10778217017650604, 0.001935038948431611, 0.019849035888910294, -0.05578485131263733, 0.14707127213478088, -0.05746309086680412, -0.06542734056711197, 0.023519758135080338, -0.028671463951468468, -0.04796788841485977, -0.01634110137820244, -0.09203075617551804, -0.12278519570827484, 0.046650223433971405, 0.11805443465709686, -0.04782428219914436, 0.06854087859392166, 0.022504117339849472, -0.029098495841026306, -0.0323047861456871, -0.10502062737941742, -0.0392073318362236, 0.06343823671340942, -0.05906146019697189, 0.08269798010587692, 0.04426981508731842, 0.0035111564211547375, -0.0659901425242424, 0.015080384910106659, -0.03539261221885681, 0.0008680910686962306, -0.06180749088525772, -0.06929893791675568, 0.0389714241027832, 0.04387116804718971, -0.008799808099865913, -0.15968622267246246, -0.12808287143707275, -0.030135702341794968, 0.030380362644791603, -0.02957158349454403, 0.0541713647544384, -0.10342281311750412, -0.039654407650232315, 0.024808695539832115, -0.0017189773498103023, 0.0968913584947586, -0.07448779046535492, 0.07540237158536911, 0.05002184212207794, 0.018965231254696846, 0.009443058632314205, -0.005015185568481684, -0.07232912629842758, 0.03970874100923538, 0.008217696100473404, 0.09840381890535355, -0.0866256132721901, 0.07900069653987885, -0.12384587526321411, -0.15830077230930328, -0.021173574030399323, 0.0738418698310852, 0.04902176558971405, 0.2069670855998993, -0.19652488827705383, -0.0033183456398546696, 0.19358505308628082, -0.03489898517727852, -0.11817865073680878, 0.10936225950717926, -0.06359244883060455, 0.0815722793340683, 0.04769326001405716, 0.11103356629610062, 0.056711990386247635, -0.11193760484457016, 0.12159236520528793, -0.0028071526903659105, 0.021005742251873016, -0.006069888826459646, 0.07641284167766571, -0.028389275074005127, -0.16928140819072723, 0.04283297806978226, -0.12158514559268951, 0.05412532016634941, -0.06922708451747894, -0.08070934563875198, -0.012272477149963379, -0.06023050472140312, 0.08354542404413223, 0.0158791933208704, 0.11044079810380936, -0.10404433310031891, -0.06327320635318756, -0.07115773111581802, 0.11394862085580826, -0.07092136144638062, 0.037652768194675446, -0.10532156378030777, 0.07905411720275879, 0.05001009255647659, 0.012283749878406525, -0.11332421004772186, 0.047407716512680054, 0.0023041635286062956, 0.04077019542455673, 0.10494048148393631, 0.06898488849401474, 0.04754549637436867, 0.03771291673183441, -0.02898671291768551, -0.007534065283834934, 0.03452224284410477, 0.0017252754187211394, -0.026350637897849083, -0.158772811293602, 0.02678435668349266, -0.005190092138946056, 0.13702809810638428, -0.26260218024253845, -0.0030683435034006834, 0.09647921472787857, 0.0755094662308693, -0.06580416113138199, 0.06516500562429428, 0.002857421524822712, 0.04417876899242401, -0.05408836528658867, -0.009079898707568645, 0.04844564571976662, -0.03333953395485878, -0.13650773465633392, 0.11324939131736755, -0.04970283806324005, -0.005108973011374474, 0.0745224729180336, -0.14704445004463196, -0.029950158670544624, -0.03673172742128372, 0.0034908857196569443, -0.008188499137759209, 0.021215111017227173, -0.00013446417870000005, 0.15426190197467804, -0.019916292279958725, 0.140868678689003, -0.11148179322481155, -0.004799454007297754, -0.008153132162988186, -0.08285219967365265, 0.014448937959969044, 0.18609996140003204, 0.0275820754468441, -0.2075110375881195, 0.022409066557884216, 0.07184971868991852, -0.09625409543514252, 0.2416023313999176, -0.03930353373289108, -0.04577493295073509, 0.011831541545689106, 0.04939308762550354, 0.004171349573880434, 0.0750691294670105, -0.1275360882282257, -0.008584984578192234, 0.047494884580373764, 0.04751748591661453, 0.0732058435678482, -0.1384473741054535, -0.030592158436775208, 0.02623257413506508, -0.042806774377822876, -0.0424383170902729, 0.11348497122526169, 0.009912990033626556, 0.07502260059118271, -0.04508271440863609, -0.09459389001131058, -0.007362199015915394, -0.030892860144376755, -0.15273548662662506, 0.25472819805145264, -0.09272509813308716, -0.1703944206237793, -0.1568310558795929, 0.005125884432345629, -0.08741629868745804, -0.02388095296919346, 0.06397917121648788, -0.12564383447170258, -0.012833958491683006, -0.02189537137746811, 0.0726134181022644, -0.06364124268293381, -0.03750508278608322, -0.05572205036878586, 0.05365900322794914, -0.044352851808071136, -0.13338373601436615, -0.013231364078819752, -0.029469404369592667, -0.07872219383716583, 0.06789351254701614, -0.14816582202911377, 0.11905292421579361, 0.14961297810077667, -0.0026049634907394648, 0.06268330663442612, -0.03279966861009598, 0.12500081956386566, -0.1050175130367279, -0.02488104999065399, 0.1680292785167694, 0.023929297924041748, -0.007530268281698227, 0.04846853390336037, 0.007486885413527489, -0.053075261414051056, 0.022217897698283195, -0.047446079552173615, -0.07162372022867203, -0.30478140711784363, -0.14570069313049316, -0.09227485209703445, -0.01726452261209488, 0.014626799151301384, 0.024795731529593468, 0.12410324066877365, 0.07270807027816772, 0.002623711945489049, -0.05200747027993202, 0.001263757236301899, 0.052593428641557693, 0.1368359476327896, 0.05086134746670723, 0.08163908123970032, -0.06845174729824066, -0.07385435700416565, 0.028603944927453995, -0.029064487665891647, 0.22408849000930786, 0.00978158414363861, 0.12217734009027481, 0.07615545392036438, 0.17717459797859192, 0.05047805979847908, 0.09153976291418076, 0.05800618603825569, -0.015570526011288166, 0.016323670744895935, -0.07876899093389511, -0.008104098029434681, 0.04277665540575981, -0.011060495860874653, -0.052835095673799515, -0.09553717076778412, 0.0205510463565588, 0.06569597870111465, 0.12301753461360931, 0.018621299415826797, -0.18077747523784637, -0.016367556527256966, -0.007439970970153809, 0.03519948199391365, -0.04336400702595711, 0.08164379745721817, 0.03435178101062775, -0.15441283583641052, 0.09560991823673248, -0.047310929745435715, 0.11614561080932617, 0.021589204668998718, 0.021296149119734764, -0.06581444293260574, 0.028698353096842766, 0.004840277601033449, 0.17152561247348785, -0.3451928198337555, 0.2990512251853943, 0.008542178198695183, 0.06149239093065262, -0.10734815895557404, -0.050877850502729416, 0.03178733214735985, 0.10086182504892349, 0.10845470428466797, 0.00913910660892725, -0.17183426022529602, -0.08004745095968246, -0.013421379029750824, 0.03317137807607651, 0.10091470181941986, 0.0491049699485302, 0.06769727170467377, 0.0012100661406293511, -0.02534019574522972, -0.01058289222419262, 0.04236100614070892, -0.20138375461101532, -0.07912564277648926, 0.035814303904771805, -0.00665730657055974, -0.07957152277231216, -0.09094364196062088, -0.08208860456943512, -0.06409646570682526, 0.05529634281992912, -0.046510253101587296, -0.01754147745668888, -0.1009678989648819, -0.044222839176654816, 0.1489141434431076, -0.11145270615816116, 0.024399876594543457, -0.0005426584393717349, -0.011931357905268669, -0.019802285358309746, -0.06847517192363739, 0.08033864945173264, -0.09157039225101471, -0.13536757230758667, -0.023677386343479156, 0.1740117073059082, 0.04028617590665817, 0.06963475048542023, 0.05438428744673729, -0.026667211204767227, -0.024019284173846245, -0.11095430701971054, -0.10125245153903961, -0.05135587975382805, -0.024294529110193253, 0.03024164028465748, -0.08634275197982788, -0.028709005564451218, -0.09596384316682816, -0.13025537133216858, 0.21809548139572144, 0.17642022669315338, -0.06547719240188599, 0.022165028378367424, 0.07140405476093292, -0.07467879354953766, -0.2617526650428772, 0.03567102551460266, 0.009737713262438774, 0.08301573991775513, -0.05593804642558098, -0.14389941096305847, 0.03337109461426735, -0.005115840118378401, -0.002010377123951912, 0.091753751039505, -0.30239415168762207, -0.15985319018363953, 0.17575116455554962, 0.04408952593803406, 0.16842348873615265, -0.07745727896690369, -0.04626963660120964, -0.08862372487783432, -0.1613837033510208, 0.01745631732046604, -0.13710619509220123, 0.08443039655685425, -0.004860365763306618, 0.044515855610370636, 0.03236960619688034, -0.04765687137842178, 0.16282692551612854, 0.032018501311540604, 0.0022555955220013857, -0.05253055691719055, -0.0284102950245142, 0.015865053981542587, 0.010205673985183239, 0.06991399824619293, -0.09129000455141068, 0.05740198493003845, -0.11072210967540741, -0.03575708717107773, -0.07650195062160492, 0.09432979673147202, -0.02949357032775879, -0.023772917687892914, -0.0936819463968277, 0.015516435727477074, 0.042089544236660004, -0.005614600144326687, 0.09954240173101425, -0.09891607612371445, -0.009508956223726273, 0.06970719248056412, 0.21415680646896362, -0.022291401401162148, 0.11474169045686722, 0.006859428249299526, 0.002399260411038995, 0.07840695977210999, -0.10240977257490158, 0.029813023284077644, 0.1600416600704193, 0.030940214172005653, 0.048370134085416794, 0.03478633612394333, -0.0847863256931305, -0.01618010364472866, 0.09740892052650452, -0.05667397007346153, -0.06620978564023972, -0.11278241127729416, 0.01553039439022541, 0.04860310256481171, 0.05132000148296356, 0.14488773047924042, -0.04196063429117203, -0.00999775342643261, -0.028079308569431305, -0.057266440242528915, -0.08481968939304352, 0.22570106387138367, 0.08142334967851639, 0.06723113358020782, -0.09747568517923355, 0.046900518238544464, -0.004282205365598202, 0.052145328372716904, 0.015221801586449146, 0.1500464677810669, -0.11685601621866226, -0.1065080314874649, 0.05865577235817909, 0.20680458843708038, -0.16253674030303955, -0.1027313619852066, -0.13419878482818604, -0.12574055790901184, 0.00001887303005787544, 0.18706728518009186, 0.09727148711681366, -0.031144948676228523, -0.04324769973754883, -0.07612881809473038, -0.018254395574331284, 0.03545002266764641, 0.14522315561771393, 0.03071916662156582, -0.06846729665994644, 0.13008908927440643, -0.017566977068781853, 0.13429653644561768, -0.05956333503127098, -0.01742156408727169, -0.08573991060256958, 0.09530481696128845, -0.2410985827445984, -0.0036721033975481987, -0.04739873856306076, -0.024494251236319542, -0.01440828014165163, -0.06755571067333221, -0.04911162704229355, 0.03924141451716423, -0.1115669384598732, 0.032847125083208084, -0.06913909316062927, 0.07981332391500473, -0.00066593405790627, -0.009565822780132294, 0.0608842559158802, -0.07456718385219574, 0.02720438875257969, 0.07621970772743225, -0.06784428656101227, 0.1712464541196823, -0.16704878211021423, 0.00987914763391018, -0.005955812521278858, 0.06324853003025055, 0.03447266295552254, -0.053403157740831375, 0.01752512902021408, 0.05431919917464256, 0.10847380012273788, 0.019280223175883293, -0.04035598039627075, -0.0533885657787323, -0.061516981571912766, -0.0662013441324234, -0.08295472711324692, -0.01662747748196125, 0.0699634850025177, 0.04369480535387993, 0.06998354941606522, 0.09057410806417465, -0.09036428481340408, 0.05520870164036751, -0.06796623766422272, -0.014291825704276562, -0.0010600771056488156, -0.07726658880710602, -0.06337477266788483, -0.09290598332881927, 0.0908130407333374, -0.01759311743080616, 0.13196249306201935, -0.0513492077589035, 0.022470448166131973, -0.022075414657592773, 0.04379909858107567, 0.08818678557872772, 0.026418952271342278, 0.2282741814851761, 0.017376871779561043, 0.0482858307659626, -0.05074381083250046, 0.01502111367881298, 0.016096075996756554, 0.1301531344652176, 0.14554902911186218, 0.15665075182914734, 0.01771457865834236, 0.13063283264636993, -0.004922846332192421, 0.003859135787934065, 0.026925865560770035, -0.07870876789093018, 0.010444995015859604, 0.021940559148788452, -0.025761500000953674, 0.12158038467168808, 0.1719643622636795, -0.13841122388839722, 0.03378140553832054, -0.05607050284743309, -0.10122702270746231, -0.13785967230796814, -0.1561506986618042, -0.10358734428882599, -0.10271783173084259, 0.01642220839858055, -0.07444076985120773, -0.019282791763544083, 0.017019499093294144, 0.10242869704961777, -0.04732147976756096, 0.2053244411945343, 0.03686041384935379, -0.11324216425418854, 0.09450255334377289, -0.04309156909584999, 0.046816423535346985, 0.055454373359680176, -0.012533345259726048, -0.003562218975275755, -0.08385278284549713, 0.01518509816378355, -0.002535533858463168, -0.09232174605131149, 0.01724018156528473, -0.03499394282698631, -0.06580178439617157, -0.017819073051214218, 0.02497720718383789, 0.09400846809148788, 0.18169663846492767, 0.04489925876259804, -0.019087864086031914, -0.002948748879134655, 0.10735724121332169, -0.008198583498597145, -0.1732204556465149, -0.11524084955453873, 0.18152636289596558, 0.06977560371160507, 0.039565376937389374, 0.015861190855503082, -0.008571803569793701, 0.03156546875834465, 0.3101840913295746, 0.24211104214191437, -0.06820940971374512, -0.01604117639362812, 0.00863416027277708, 0.02128016948699951, 0.007528241258114576, 0.09413257986307144, 0.040861744433641434, 0.1572205275297165, -0.06085661053657532, -0.08381308615207672, -0.06388851255178452, -0.024455612525343895, -0.09744902700185776, 0.11584045737981796, 0.046661462634801865, -0.09468964487314224, -0.0020476195495575666, 0.13607625663280487, -0.10675752907991409, -0.006493146065622568, -0.039259810000658035, -0.08940247446298599, -0.11121157556772232, -0.049552373588085175, 0.017957204952836037, 0.038400933146476746, 0.040942948311567307, -0.07127345353364944, -0.03116253763437271, -0.0031364497262984514, 0.024196932092308998, -0.10156998038291931, -0.06509953737258911, 0.12446589767932892, -0.06011210009455681, 0.13325874507427216, -0.013444619253277779, 0.15759721398353577, 0.11073623597621918, 0.058475129306316376, -0.009872660040855408, 0.10305201262235641, 0.07947872579097748, -0.04193814471364021, 0.06109168380498886, 0.06091545149683952, -0.02382410503923893, 0.000053315288823796436, 0.03365492448210716, -0.11576233804225922, 0.09029227495193481, 0.07355352491140366, 0.01118296105414629, -0.09939298033714294, 0.08047472685575485, -0.10547266900539398, 0.11455691605806351, 0.13963451981544495, -0.031615231186151505, 0.012306365184485912, -0.08323150873184204, 0.05889449268579483, 0.013147806748747826, -0.007723946124315262, -0.044504135847091675, -0.19014431536197662, -0.048822421580553055, -0.0698038712143898, -0.0065048690885305405, -0.1704726219177246, -0.025378098711371422, -0.07050401717424393, -0.047361403703689575, -0.024475332349538803, 0.14638826251029968, 0.0927625298500061, 0.019667061045765877, 0.02854015678167343, -0.12102469056844711, -0.0026769954711198807, 0.0945979505777359, -0.17445798218250275, -0.07904449850320816 ]
null
null
transformers
### marianmt-zh_cn-th * source languages: zh_cn * target languages: th * dataset: * model: transformer-align * pre-processing: normalization + SentencePiece * test set translations: * test set scores: ## Training Training scripts from [LalitaDeelert/NLP-ZH_TH-Project](https://github.com/LalitaDeelert/NLP-ZH_TH-Project). Experiments tracked at [cstorm125/marianmt-zh_cn-th](https://wandb.ai/cstorm125/marianmt-zh_cn-th). ``` export WANDB_PROJECT=marianmt-zh_cn-th python train_model.py --input_fname ../data/v1/Train.csv \ \\t--output_dir ../models/marianmt-zh_cn-th \ \\t--source_lang zh --target_lang th \ \\t--metric_tokenize th_syllable --fp16 ``` ## Usage ``` from transformers import AutoTokenizer, AutoModelForSeq2SeqLM tokenizer = AutoTokenizer.from_pretrained("cstorm125/marianmt-zh_cn-th") model = AutoModelForSeq2SeqLM.from_pretrained("cstorm125/marianmt-zh_cn-th").cpu() src_text = [ '我爱你', '我想吃米饭', ] translated = model.generate(**tokenizer(src_text, return_tensors="pt", padding=True)) print([tokenizer.decode(t, skip_special_tokens=True) for t in translated]) > ['ผมรักคุณนะ', 'ฉันอยากกินข้าว'] ``` ## Requirements ``` transformers==4.6.0 torch==1.8.0 ```
{"tags": ["translation", "torch==1.8.0"], "widget": [{"text": "Inference Unavailable"}]}
translation
cstorm125/marianmt-zh_cn-th
[ "transformers", "pytorch", "marian", "text2text-generation", "translation", "torch==1.8.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #marian #text2text-generation #translation #torch==1.8.0 #autotrain_compatible #endpoints_compatible #region-us
### marianmt-zh_cn-th * source languages: zh_cn * target languages: th * dataset: * model: transformer-align * pre-processing: normalization + SentencePiece * test set translations: * test set scores: ## Training Training scripts from LalitaDeelert/NLP-ZH_TH-Project. Experiments tracked at cstorm125/marianmt-zh_cn-th. ## Usage ## Requirements
[ "### marianmt-zh_cn-th\n* source languages: zh_cn\n* target languages: th\n* dataset: \n* model: transformer-align\n* pre-processing: normalization + SentencePiece\n* test set translations: \n* test set scores:", "## Training\n\nTraining scripts from LalitaDeelert/NLP-ZH_TH-Project. Experiments tracked at cstorm125/marianmt-zh_cn-th.", "## Usage", "## Requirements" ]
[ "TAGS\n#transformers #pytorch #marian #text2text-generation #translation #torch==1.8.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### marianmt-zh_cn-th\n* source languages: zh_cn\n* target languages: th\n* dataset: \n* model: transformer-align\n* pre-processing: normalization + SentencePiece\n* test set translations: \n* test set scores:", "## Training\n\nTraining scripts from LalitaDeelert/NLP-ZH_TH-Project. Experiments tracked at cstorm125/marianmt-zh_cn-th.", "## Usage", "## Requirements" ]
[ 49, 63, 41, 3, 5 ]
[ "passage: TAGS\n#transformers #pytorch #marian #text2text-generation #translation #torch==1.8.0 #autotrain_compatible #endpoints_compatible #region-us \n### marianmt-zh_cn-th\n* source languages: zh_cn\n* target languages: th\n* dataset: \n* model: transformer-align\n* pre-processing: normalization + SentencePiece\n* test set translations: \n* test set scores:## Training\n\nTraining scripts from LalitaDeelert/NLP-ZH_TH-Project. Experiments tracked at cstorm125/marianmt-zh_cn-th.## Usage## Requirements" ]
[ -0.16600242257118225, -0.027812577784061432, -0.002706715138629079, 0.04715290293097496, 0.0873231440782547, -0.005283234175294638, 0.050016067922115326, 0.07865055650472641, -0.025380514562129974, 0.008193920366466045, 0.12634031474590302, 0.09176836162805557, 0.004395538941025734, 0.04229692742228508, -0.05487890541553497, -0.18166756629943848, 0.02715873345732689, 0.03612595424056053, -0.15910914540290833, 0.13827678561210632, 0.11968296766281128, -0.0454656258225441, 0.10597463697195053, 0.02751600183546543, -0.04340993985533714, 0.02637927047908306, -0.01568298041820526, -0.11511150747537613, 0.13452479243278503, 0.06041799113154411, 0.09637763351202011, 0.09812972694635391, 0.06990458816289902, -0.1823704093694687, 0.019417934119701385, -0.015727100893855095, -0.0027074534446001053, 0.026586124673485756, 0.09203238785266876, -0.05236917361617088, 0.23085398972034454, -0.0931372120976448, -0.02185753732919693, 0.06848341971635818, -0.06224217265844345, -0.08235563337802887, -0.06336579471826553, -0.033810101449489594, 0.023781949654221535, 0.1438513845205307, -0.012827564030885696, 0.20232518017292023, -0.16352961957454681, 0.10443394631147385, 0.19104807078838348, -0.2612217962741852, -0.05128660053014755, 0.0704120546579361, 0.11383768171072006, -0.008452331647276878, 0.011386653408408165, 0.09152668714523315, 0.09218355268239975, 0.010467529296875, -0.08692561089992523, -0.09838569909334183, -0.026912959292531013, 0.037766166031360626, -0.16374622285366058, -0.012138135731220245, 0.25750306248664856, -0.016383042559027672, -0.048943303525447845, -0.0032462929375469685, -0.03288828581571579, -0.017772000283002853, -0.012550790794193745, -0.014735860750079155, -0.05290371552109718, -0.032794516533613205, -0.04475047066807747, -0.028066884726285934, -0.10192220658063889, -0.09086021780967712, -0.05372299998998642, 0.1610981971025467, 0.06459642946720123, -0.0057967123575508595, -0.10393962264060974, 0.1128072589635849, 0.12517982721328735, -0.12105956673622131, -0.05493324249982834, -0.07226864248514175, -0.03490560129284859, -0.012327487580478191, -0.038940753787755966, -0.11001955717802048, 0.11101428419351578, 0.09088649600744247, -0.02764444425702095, 0.10269812494516373, -0.008176573552191257, 0.020563943311572075, -0.05590398609638214, 0.1417982578277588, -0.050612710416316986, -0.0661921501159668, 0.020641490817070007, -0.035388071089982986, -0.04853130131959915, -0.017048489302396774, -0.09110485017299652, -0.1212567463517189, 0.04320840910077095, 0.11824239790439606, -0.03739992901682854, 0.07652276009321213, 0.01705155335366726, -0.029689468443393707, -0.027813391759991646, -0.1035325676202774, -0.03750718757510185, 0.06006040424108505, -0.05108531937003136, 0.08169142156839371, 0.044794365763664246, 0.002572472672909498, -0.06938183307647705, 0.020949555560946465, -0.03465719893574715, -0.00320239644497633, -0.06106717139482498, -0.07386945188045502, 0.042201295495033264, 0.03693694621324539, -0.007780165411531925, -0.16202138364315033, -0.12623170018196106, -0.030258784070611, 0.030196428298950195, -0.030354760587215424, 0.047904904931783676, -0.10270994156599045, -0.04460665583610535, 0.030744319781661034, -0.0032439175993204117, 0.09237226098775864, -0.07521375268697739, 0.07761859893798828, 0.051780588924884796, 0.02371702715754509, 0.007212596479803324, -0.005828367546200752, -0.07800965011119843, 0.03814557194709778, 0.0036834687925875187, 0.09365436434745789, -0.08777495473623276, 0.0759875625371933, -0.12740223109722137, -0.15419414639472961, -0.027834242209792137, 0.07481652498245239, 0.048086751252412796, 0.20854295790195465, -0.1960277110338211, -0.0016304492019116879, 0.2061179280281067, -0.039377741515636444, -0.11675256490707397, 0.11152754724025726, -0.05863349139690399, 0.08647800981998444, 0.050932589918375015, 0.1141979917883873, 0.05660087242722511, -0.11092456430196762, 0.11810749024152756, 0.00032358794123865664, 0.019335387274622917, -0.003408460645005107, 0.07471564412117004, -0.026273809373378754, -0.17271658778190613, 0.04261317849159241, -0.12311361730098724, 0.054921168833971024, -0.06896557658910751, -0.08009664714336395, -0.012438623234629631, -0.052473004907369614, 0.08658628165721893, 0.012518600560724735, 0.10329394042491913, -0.10541462898254395, -0.06359580159187317, -0.07361330091953278, 0.10716305673122406, -0.06200804188847542, 0.03794100880622864, -0.10600699484348297, 0.07538964599370956, 0.04501007869839668, 0.016266493126749992, -0.1196778267621994, 0.040778595954179764, 0.0030583448242396116, 0.050615690648555756, 0.10592199862003326, 0.06795231252908707, 0.05045321583747864, 0.031600471585989, -0.03011801466345787, -0.005340940784662962, 0.03659328818321228, -0.0014882220420986414, -0.027680950239300728, -0.16059237718582153, 0.03289502114057541, -0.010936220176517963, 0.1258634328842163, -0.2624475061893463, -0.0027807841543108225, 0.09944988787174225, 0.0718110203742981, -0.05926883965730667, 0.06324192136526108, 0.001985258189961314, 0.04677413031458855, -0.05641396343708038, -0.008025256916880608, 0.05014386028051376, -0.031540222465991974, -0.12830029428005219, 0.12023009359836578, -0.0548446848988533, -0.006136483978480101, 0.07933022081851959, -0.1498950868844986, -0.03918980434536934, -0.03444627299904823, 0.0024539725854992867, -0.006251748651266098, 0.020438551902770996, 0.0008608216303400695, 0.14227432012557983, -0.015983257442712784, 0.13813892006874084, -0.10802624374628067, -0.0019361082231625915, -0.006177693605422974, -0.08398593217134476, 0.009107859805226326, 0.18368418514728546, 0.03625183552503586, -0.203501358628273, 0.030389169231057167, 0.07886873185634613, -0.08925731480121613, 0.23722721636295319, -0.03560017794370651, -0.0458141528069973, 0.014596851542592049, 0.051900047808885574, 0.008299252949655056, 0.06714285910129547, -0.12686368823051453, -0.006035390309989452, 0.0486767403781414, 0.05088723450899124, 0.06917206197977066, -0.1350783109664917, -0.031377945095300674, 0.027296368032693863, -0.043602585792541504, -0.04272513836622238, 0.11083832383155823, 0.008693534880876541, 0.07852812111377716, -0.0383252389729023, -0.09249567240476608, -0.004940792918205261, -0.02683286927640438, -0.15330855548381805, 0.25412553548812866, -0.09317144751548767, -0.18210142850875854, -0.1621117740869522, -0.0001317799324169755, -0.08391600102186203, -0.01865096390247345, 0.06586698442697525, -0.12591174244880676, -0.016631541773676872, -0.02159576676785946, 0.07674067467451096, -0.06989327818155289, -0.03806963562965393, -0.057541631162166595, 0.05623767897486687, -0.04329857975244522, -0.12918917834758759, -0.013490560464560986, -0.02207852713763714, -0.07758182287216187, 0.07212129980325699, -0.14578720927238464, 0.12565648555755615, 0.1452740728855133, -0.0006020978908054531, 0.05618072673678398, -0.0343228280544281, 0.13339866697788239, -0.10851246863603592, -0.02129307948052883, 0.1627371609210968, 0.019851956516504288, -0.004910178482532501, 0.057189635932445526, 0.0037521633785218, -0.05638308823108673, 0.026800164952874184, -0.044402457773685455, -0.07179545611143112, -0.3084999620914459, -0.1458289474248886, -0.0931360200047493, -0.005453138146549463, 0.016010740771889687, 0.02685544081032276, 0.12059278786182404, 0.07292500883340836, -0.0010048558469861746, -0.04861126095056534, 0.0034431512467563152, 0.05761023238301277, 0.142024427652359, 0.04527931660413742, 0.08428727090358734, -0.06972901523113251, -0.07721636444330215, 0.03432466834783554, -0.02352486364543438, 0.2227739542722702, 0.018063466995954514, 0.125370055437088, 0.07492278516292572, 0.17070885002613068, 0.05151119828224182, 0.09465314447879791, 0.05807521194219589, -0.013927099294960499, 0.013535981066524982, -0.07713786512613297, -0.00875845830887556, 0.04069213569164276, -0.016141366213560104, -0.05759594589471817, -0.09132090210914612, 0.020606085658073425, 0.06786098331212997, 0.12438427656888962, 0.016406018286943436, -0.18401233851909637, -0.021796321496367455, -0.004772582091391087, 0.03269033879041672, -0.04830925166606903, 0.08209218084812164, 0.0335112139582634, -0.15679356455802917, 0.09498900175094604, -0.04613291844725609, 0.11839796602725983, 0.0187680646777153, 0.022331731393933296, -0.06551235169172287, 0.02165280655026436, 0.004399498458951712, 0.1713276505470276, -0.3509644567966461, 0.3038150668144226, 0.00868775974959135, 0.05817576125264168, -0.10433991998434067, -0.048698414117097855, 0.022091854363679886, 0.10782203078269958, 0.11770376563072205, 0.009901098906993866, -0.1770709604024887, -0.08797946572303772, -0.012727437540888786, 0.02802317962050438, 0.10863403975963593, 0.04824315756559372, 0.06945089995861053, -0.002446170197799802, -0.024690330028533936, -0.009284021332859993, 0.034253936260938644, -0.1990301012992859, -0.0837135910987854, 0.03520051762461662, -0.00629215594381094, -0.07812930643558502, -0.08408542722463608, -0.07853744179010391, -0.057220056653022766, 0.068489208817482, -0.054186683148145676, -0.014424082823097706, -0.10529156774282455, -0.0462985597550869, 0.14372289180755615, -0.11302772164344788, 0.02424662746489048, -0.005315183196216822, -0.009964118711650372, -0.021679624915122986, -0.06904258579015732, 0.08661238104104996, -0.09566308557987213, -0.1328759491443634, -0.02206914685666561, 0.16943557560443878, 0.03566956892609596, 0.07135098427534103, 0.053163591772317886, -0.019357791170477867, -0.02102123573422432, -0.1105794757604599, -0.09787429869174957, -0.04967039078474045, -0.029093975201249123, 0.026381075382232666, -0.09089897572994232, -0.03279408439993858, -0.09718990325927734, -0.13266201317310333, 0.2263089120388031, 0.17884261906147003, -0.06458063423633575, 0.024409234523773193, 0.07130778580904007, -0.07880304008722305, -0.26867708563804626, 0.037747375667095184, 0.007822367362678051, 0.07625961303710938, -0.051242806017398834, -0.14312531054019928, 0.029443323612213135, -0.012245007790625095, -0.0007672050851397216, 0.09235537797212601, -0.29802948236465454, -0.15575692057609558, 0.17060205340385437, 0.043730489909648895, 0.1686781495809555, -0.08069398254156113, -0.05032600089907646, -0.08706816285848618, -0.158794566988945, 0.025325428694486618, -0.13051554560661316, 0.0806502103805542, 0.0006698024226352572, 0.049134500324726105, 0.03194265067577362, -0.045712992548942566, 0.1579827517271042, 0.030016692355275154, -0.0005587919731624424, -0.05735376477241516, -0.026291532441973686, 0.01596114970743656, 0.013952843844890594, 0.07258599996566772, -0.08012593537569046, 0.05515005439519882, -0.110883928835392, -0.03982483595609665, -0.07660875469446182, 0.09744945168495178, -0.03101804107427597, -0.026918288320302963, -0.08824873715639114, 0.014109198935329914, 0.03811926394701004, -0.00908446591347456, 0.0962422788143158, -0.09888793528079987, -0.0015149010578170419, 0.0730455294251442, 0.2172602266073227, -0.008591579273343086, 0.11385484039783478, 0.010044820606708527, -0.0023165473248809576, 0.07641846686601639, -0.10259316861629486, 0.028711779043078423, 0.15646803379058838, 0.026797190308570862, 0.04815082997083664, 0.03338273987174034, -0.08310088515281677, -0.018814269453287125, 0.1001296266913414, -0.06542666256427765, -0.0711185559630394, -0.10911905020475388, 0.010754607617855072, 0.04661419242620468, 0.04682145640254021, 0.1488005816936493, -0.04327710345387459, -0.008818818256258965, -0.028316006064414978, -0.05072356015443802, -0.0828469768166542, 0.22147905826568604, 0.07815995812416077, 0.06286395341157913, -0.09437797963619232, 0.04521525651216507, 0.0011203160975128412, 0.03209218382835388, 0.020366793498396873, 0.15299633145332336, -0.11939382553100586, -0.10868803411722183, 0.06397537887096405, 0.1939624845981598, -0.16425278782844543, -0.09930428117513657, -0.12582385540008545, -0.11866503953933716, 0.004289319273084402, 0.18659508228302002, 0.09276051074266434, -0.02953184023499489, -0.04292576014995575, -0.07236737012863159, -0.015460293740034103, 0.035987626761198044, 0.14336061477661133, 0.03393687680363655, -0.06965192407369614, 0.12923552095890045, -0.01902848854660988, 0.13263966143131256, -0.0588928647339344, -0.01582559011876583, -0.08991747349500656, 0.09431697428226471, -0.22629013657569885, -0.0028061505872756243, -0.04602424055337906, -0.02380671724677086, -0.01618875004351139, -0.06845083832740784, -0.049387045204639435, 0.032949719578027725, -0.11157773435115814, 0.03023529052734375, -0.0697985291481018, 0.07955125719308853, -0.006573433522135019, -0.006925638299435377, 0.0663638561964035, -0.07090011984109879, 0.026212746277451515, 0.07666855305433273, -0.06712788343429565, 0.17251069843769073, -0.16754990816116333, 0.013837786391377449, -0.002336397534236312, 0.062187910079956055, 0.03356492891907692, -0.04491019994020462, 0.017494244500994682, 0.05624759942293167, 0.11172628402709961, 0.02285376563668251, -0.032507799565792084, -0.06222148612141609, -0.058056384325027466, -0.061434514820575714, -0.0922132134437561, -0.018379483371973038, 0.06736551225185394, 0.046705763787031174, 0.06289549171924591, 0.09685657173395157, -0.09362667798995972, 0.05784488841891289, -0.06787732988595963, -0.010172479785978794, -0.0012354189530014992, -0.07692597061395645, -0.06505681574344635, -0.09087013453245163, 0.09123918414115906, -0.020442213863134384, 0.12641768157482147, -0.04637434706091881, 0.031214188784360886, -0.017408521845936775, 0.03366268053650856, 0.07865840196609497, 0.023569956421852112, 0.2205260545015335, 0.02470816858112812, 0.04164375364780426, -0.05750175192952156, 0.015269921161234379, 0.0156038086861372, 0.12529154121875763, 0.14662835001945496, 0.16087184846401215, 0.015528787858784199, 0.12855572998523712, -0.0013865503715351224, 0.007953155785799026, 0.019481051713228226, -0.08353973180055618, 0.007047333754599094, 0.02232562191784382, -0.02714814990758896, 0.12401817739009857, 0.17382900416851044, -0.13781563937664032, 0.03483574092388153, -0.061411626636981964, -0.10050879418849945, -0.141856849193573, -0.1615498661994934, -0.106770358979702, -0.09695596992969513, 0.019070366397500038, -0.08032909035682678, -0.016655895859003067, 0.015577757731080055, 0.10164356231689453, -0.04961265251040459, 0.206912562251091, 0.032367367297410965, -0.11533848196268082, 0.0952874943614006, -0.043212004005908966, 0.050738900899887085, 0.056018680334091187, -0.010894078761339188, -0.007717194966971874, -0.07637061178684235, 0.016146177425980568, 0.0034932508133351803, -0.09210208803415298, 0.015182727947831154, -0.03978579491376877, -0.06738632917404175, -0.02116398513317108, 0.029746050015091896, 0.08788377046585083, 0.1830219030380249, 0.04262876510620117, -0.01955346204340458, 0.0006461511948145926, 0.11416105926036835, -0.012963689863681793, -0.17403925955295563, -0.11950952559709549, 0.1871212273836136, 0.06487806141376495, 0.04618682339787483, 0.008408553898334503, -0.009127442725002766, 0.02706526592373848, 0.3117702901363373, 0.2465764880180359, -0.06424082070589066, -0.014098159037530422, 0.011757227592170238, 0.02521596848964691, 0.01346118375658989, 0.09998629242181778, 0.042985327541828156, 0.16384638845920563, -0.06529894471168518, -0.08434753119945526, -0.06327579915523529, -0.020851388573646545, -0.09378155320882797, 0.11766384541988373, 0.04687928408384323, -0.08843761682510376, -0.005371768027544022, 0.14156338572502136, -0.10860739648342133, -0.006172310560941696, -0.049296800047159195, -0.09114862978458405, -0.10960453003644943, -0.04970785602927208, 0.016255877912044525, 0.03765437752008438, 0.042112771421670914, -0.06716272979974747, -0.03251567855477333, -0.004731037188321352, 0.025191999971866608, -0.10520124435424805, -0.06153574958443642, 0.12089362740516663, -0.047358401119709015, 0.1266084611415863, -0.011368351057171822, 0.15603423118591309, 0.1102476492524147, 0.05746765434741974, -0.011550636030733585, 0.1020030602812767, 0.0783248320221901, -0.03264625743031502, 0.062273088842630386, 0.06223654747009277, -0.02739381231367588, 0.001606744364835322, 0.037881962954998016, -0.11702331900596619, 0.09138418734073639, 0.07537859678268433, 0.012805039063096046, -0.10182074457406998, 0.08361755311489105, -0.11058178544044495, 0.11374874413013458, 0.1463797241449356, -0.028194086626172066, 0.010148935951292515, -0.08713217824697495, 0.05788881331682205, 0.010874228551983833, -0.010617641732096672, -0.046551238745450974, -0.19032250344753265, -0.05414820834994316, -0.06366172432899475, 0.0024869453627616167, -0.18242672085762024, -0.02467728778719902, -0.06765307486057281, -0.03724600747227669, -0.03213454410433769, 0.14669546484947205, 0.09171243011951447, 0.016895977780222893, 0.02629663608968258, -0.11891276389360428, 0.0018964578630402684, 0.0965452566742897, -0.17318075895309448, -0.07812336832284927 ]
null
null
transformers
# wangchan-deberta_v1-base-wiki-20210520-news-spm-finetune-qa Finetuning `airesearch/wangchan-deberta_v1-base-wiki-20210520-news-spm` with the training set of `iapp_wiki_qa_squad`, `thaiqa_squad`, and `nsc_qa` (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 `newmm` words). Benchmarks shared on [wandb](https://wandb.ai/cstorm125/wangchanberta-qa) using validation and test sets of `iapp_wiki_qa_squad`. Trained with [thai2transformers](https://github.com/vistec-AI/thai2transformers/blob/dev/scripts/downstream/train_question_answering_lm_finetuning.py). Run with: ``` export MODEL_NAME=wangchan-deberta_v1-base-wiki-20210520-news-spm CUDA_LAUNCH_BLOCKING=1 python train_question_answering_lm_finetuning.py \ --model_name $MODEL_NAME \ --dataset_name chimera_qa \ --revision mlm@ckp-41100 \ --output_dir $MODEL_NAME-finetune-chimera_qa-model \ --log_dir $MODEL_NAME-finetune-chimera_qa-log \ --model_max_length 400 \ --pad_on_right \ --fp16 \ --use_auth_token ```
{"widget": [{"text": "\u0e2a\u0e27\u0e19\u0e01\u0e38\u0e2b\u0e25\u0e32\u0e1a\u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e2d\u0e30\u0e44\u0e23", "context": "\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e2a\u0e27\u0e19\u0e01\u0e38\u0e2b\u0e25\u0e32\u0e1a\u0e27\u0e34\u0e17\u0e22\u0e32\u0e25\u0e31\u0e22 (Suankularb Wittayalai School) (\u0e2d\u0e31\u0e01\u0e29\u0e23\u0e22\u0e48\u0e2d : \u0e2a.\u0e01. / S.K.) \u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e0a\u0e32\u0e22\u0e25\u0e49\u0e27\u0e19 \u0e23\u0e30\u0e14\u0e31\u0e1a\u0e0a\u0e31\u0e49\u0e19\u0e21\u0e31\u0e18\u0e22\u0e21\u0e28\u0e36\u0e01\u0e29\u0e32\u0e02\u0e19\u0e32\u0e14\u0e43\u0e2b\u0e0d\u0e48\u0e1e\u0e34\u0e40\u0e28\u0e29 \u0e2a\u0e31\u0e07\u0e01\u0e31\u0e14\u0e2a\u0e33\u0e19\u0e31\u0e01\u0e07\u0e32\u0e19\u0e40\u0e02\u0e15\u0e1e\u0e37\u0e49\u0e19\u0e17\u0e35\u0e48\u0e01\u0e32\u0e23\u0e28\u0e36\u0e01\u0e29\u0e32\u0e21\u0e31\u0e18\u0e22\u0e21\u0e28\u0e36\u0e01\u0e29\u0e32\u0e40\u0e02\u0e15 1 \u0e2a\u0e33\u0e19\u0e31\u0e01\u0e07\u0e32\u0e19\u0e04\u0e13\u0e30\u0e01\u0e23\u0e23\u0e21\u0e01\u0e32\u0e23\u0e01\u0e32\u0e23\u0e28\u0e36\u0e01\u0e29\u0e32\u0e02\u0e31\u0e49\u0e19\u0e1e\u0e37\u0e49\u0e19\u0e10\u0e32\u0e19 (\u0e0a\u0e37\u0e48\u0e2d\u0e40\u0e14\u0e34\u0e21: \u0e01\u0e23\u0e21\u0e2a\u0e32\u0e21\u0e31\u0e0d\u0e28\u0e36\u0e01\u0e29\u0e32) \u0e01\u0e23\u0e30\u0e17\u0e23\u0e27\u0e07\u0e28\u0e36\u0e01\u0e29\u0e32\u0e18\u0e34\u0e01\u0e32\u0e23 \u0e01\u0e48\u0e2d\u0e15\u0e31\u0e49\u0e07\u0e42\u0e14\u0e22 \u0e1e\u0e23\u0e30\u0e1a\u0e32\u0e17\u0e2a\u0e21\u0e40\u0e14\u0e47\u0e08\u0e1e\u0e23\u0e30\u0e08\u0e38\u0e25\u0e08\u0e2d\u0e21\u0e40\u0e01\u0e25\u0e49\u0e32\u0e40\u0e08\u0e49\u0e32\u0e2d\u0e22\u0e39\u0e48\u0e2b\u0e31\u0e27 \u0e44\u0e14\u0e49\u0e23\u0e31\u0e1a\u0e01\u0e32\u0e23\u0e2a\u0e16\u0e32\u0e1b\u0e19\u0e32\u0e02\u0e36\u0e49\u0e19\u0e43\u0e19\u0e27\u0e31\u0e19\u0e17\u0e35\u0e48 8 \u0e21\u0e35\u0e19\u0e32\u0e04\u0e21 \u0e1e.\u0e28. 2424 (\u0e02\u0e13\u0e30\u0e19\u0e31\u0e49\u0e19\u0e19\u0e31\u0e1a\u0e27\u0e31\u0e19\u0e17\u0e35\u0e48 1 \u0e40\u0e21\u0e29\u0e32\u0e22\u0e19 \u0e40\u0e1b\u0e47\u0e19\u0e27\u0e31\u0e19\u0e02\u0e36\u0e49\u0e19\u0e1b\u0e35\u0e43\u0e2b\u0e21\u0e48 \u0e40\u0e21\u0e37\u0e48\u0e2d\u0e19\u0e31\u0e1a\u0e2d\u0e22\u0e48\u0e32\u0e07\u0e2a\u0e32\u0e01\u0e25\u0e16\u0e37\u0e2d\u0e40\u0e1b\u0e47\u0e19 \u0e1e.\u0e28. 2425) \u0e42\u0e14\u0e22\u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e23\u0e31\u0e10\u0e1a\u0e32\u0e25\u0e41\u0e2b\u0e48\u0e07\u0e41\u0e23\u0e01\u0e02\u0e2d\u0e07\u0e1b\u0e23\u0e30\u0e40\u0e17\u0e28\u0e44\u0e17\u0e22"}]}
question-answering
cstorm125/wangchan-deberta_v1-base-wiki-20210520-news-spm-finetune-qa
[ "transformers", "pytorch", "deberta", "question-answering", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #deberta #question-answering #endpoints_compatible #region-us
# wangchan-deberta_v1-base-wiki-20210520-news-spm-finetune-qa Finetuning 'airesearch/wangchan-deberta_v1-base-wiki-20210520-news-spm' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'. Trained with thai2transformers. Run with:
[ "# wangchan-deberta_v1-base-wiki-20210520-news-spm-finetune-qa\n\nFinetuning 'airesearch/wangchan-deberta_v1-base-wiki-20210520-news-spm' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:" ]
[ "TAGS\n#transformers #pytorch #deberta #question-answering #endpoints_compatible #region-us \n", "# wangchan-deberta_v1-base-wiki-20210520-news-spm-finetune-qa\n\nFinetuning 'airesearch/wangchan-deberta_v1-base-wiki-20210520-news-spm' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:" ]
[ 31, 166 ]
[ "passage: TAGS\n#transformers #pytorch #deberta #question-answering #endpoints_compatible #region-us \n# wangchan-deberta_v1-base-wiki-20210520-news-spm-finetune-qa\n\nFinetuning 'airesearch/wangchan-deberta_v1-base-wiki-20210520-news-spm' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:" ]
[ -0.03964553028345108, -0.13523806631565094, -0.0036429017782211304, 0.0469239205121994, 0.041265372186899185, -0.001220006844960153, 0.09441833198070526, 0.08874376863241196, -0.010469491593539715, 0.013886729255318642, 0.06614906340837479, 0.037891749292612076, 0.050557609647512436, 0.05073443427681923, -0.07806675136089325, -0.2249918282032013, 0.09185003489255905, 0.11173123121261597, -0.04190700128674507, 0.14337725937366486, 0.13323982059955597, -0.06521935015916824, 0.07074622064828873, 0.06181171163916588, -0.09304570406675339, 0.005544273182749748, 0.02328803390264511, -0.09098009020090103, 0.14152809977531433, 0.005325653590261936, 0.14552246034145355, 0.0884888768196106, -0.03917224705219269, -0.19798417389392853, 0.04782523587346077, -0.004278290085494518, 0.037588175386190414, -0.005781333427876234, -0.026305461302399635, 0.05793054774403572, -0.027301188558340073, -0.027283474802970886, 0.036811623722314835, 0.00043573984294198453, -0.0988268181681633, -0.07832182198762894, -0.04809620976448059, 0.01812036707997322, 0.14755919575691223, 0.09371072053909302, -0.03486338257789612, 0.18911856412887573, -0.19289329648017883, 0.04097588360309601, 0.11081963777542114, -0.39115962386131287, -0.013333627954125404, 0.04986076056957245, 0.024028578773140907, 0.047075774520635605, -0.03275908902287483, 0.015083156526088715, 0.05320301651954651, 0.0334831178188324, -0.05875849351286888, -0.06755463778972626, -0.027838924899697304, -0.02737538330256939, -0.09628593921661377, 0.016434159129858017, 0.16484282910823822, 0.05211958289146423, -0.06287345290184021, -0.10004466027021408, -0.04829955846071243, 0.033349934965372086, -0.003967587370425463, -0.12086115032434464, -0.029394039884209633, 0.017244262620806694, 0.00012411287752911448, 0.015140073373913765, -0.10678932815790176, -0.07268576323986053, -0.1480409801006317, 0.07979077845811844, 0.07017774134874344, 0.04277492314577103, -0.1441257745027542, 0.02068425342440605, -0.020301802083849907, -0.13658525049686432, -0.07542064785957336, -0.05524318665266037, 0.0034836085978895426, 0.0201125331223011, -0.08767865598201752, -0.0466904379427433, 0.08556964993476868, 0.08771400153636932, -0.15118153393268585, 0.04142029955983162, -0.08731068670749664, 0.051801420748233795, -0.027307147160172462, 0.15556924045085907, -0.0225241519510746, 0.06899938732385635, 0.035446371883153915, 0.016630591824650764, 0.0004519057983998209, 0.012227696366608143, -0.000016570536899962462, -0.08900296688079834, 0.03326911851763725, 0.028358573094010353, -0.08285384625196457, 0.08981844037771225, -0.05114540830254555, -0.027130307629704475, 0.12821324169635773, -0.11152497678995132, -0.09489023685455322, 0.011423435993492603, -0.028599195182323456, 0.08693736791610718, 0.0006118241581134498, 0.09738261997699738, -0.021307772025465965, 0.09444485604763031, -0.020560843870043755, -0.013734548352658749, 0.03463394194841385, -0.04074323549866676, 0.021270081400871277, -0.08508365601301193, 0.0050997124053537846, -0.15369150042533875, -0.12586164474487305, -0.00020341953495517373, -0.027092838659882545, 0.027918575331568718, -0.01926829293370247, -0.03554418310523033, 0.009732340462505817, -0.06910140812397003, -0.011426771059632301, -0.02034059911966324, -0.03312075883150101, 0.10256940871477127, 0.07174728810787201, 0.12190507352352142, -0.10256006568670273, 0.04119809344410896, -0.14950823783874512, 0.05795414000749588, -0.10341285914182663, 0.0869172215461731, -0.045411527156829834, -0.013585643842816353, -0.06106827035546303, -0.023377010598778725, -0.09593096375465393, 0.025166692212224007, 0.0063321953639388084, 0.10588479787111282, -0.11816687881946564, -0.0661119893193245, 0.20574630796909332, -0.12208522856235504, -0.14152272045612335, 0.17930607497692108, -0.057388875633478165, 0.0964154452085495, 0.10660958290100098, 0.2176109254360199, 0.031135106459259987, -0.13595151901245117, 0.06215289980173111, -0.035483431071043015, -0.025775430724024773, 0.06995284557342529, 0.06545154750347137, 0.023516010493040085, -0.03247624635696411, 0.07479282468557358, -0.1184515431523323, 0.03395149111747742, -0.049229834228754044, -0.07727234065532684, -0.032693151384592056, -0.08685877174139023, 0.09819796681404114, -0.038094040006399155, 0.11750875413417816, -0.0475754514336586, -0.05192602053284645, 0.0036638176534324884, 0.06261371821165085, -0.002463730750605464, 0.014301792718470097, -0.15599124133586884, 0.051133282482624054, -0.042097847908735275, -0.0025282977148890495, -0.09984544664621353, 0.0026534132193773985, 0.04040621221065521, 0.001586543396115303, 0.03158341720700264, 0.10859646648168564, -0.009829345159232616, -0.056467317044734955, -0.05927613377571106, 0.04705396667122841, -0.03321235254406929, -0.001267920946702361, -0.03152746334671974, -0.00716171320527792, 0.09088297933340073, -0.024986356496810913, 0.10676365345716476, -0.0850147157907486, 0.03000103309750557, 0.035515494644641876, 0.11061518639326096, 0.007873614318668842, 0.06687530130147934, 0.034992579370737076, 0.0825955793261528, 0.03209371492266655, 0.033357515931129456, 0.03680216893553734, 0.008967868983745575, -0.13732998073101044, 0.10755743086338043, -0.05534199997782707, 0.14929375052452087, 0.13464659452438354, -0.11067931354045868, 0.015888415277004242, 0.015085071325302124, -0.0430341437458992, -0.029752029106020927, -0.09956617653369904, 0.0459517277777195, 0.1586785465478897, -0.015766171738505363, 0.12873856723308563, -0.1174500435590744, -0.025319507345557213, -0.018858464434742928, -0.0122576504945755, -0.011127769947052002, 0.03594670072197914, 0.017452795058488846, -0.17317788302898407, 0.059294480830430984, 0.1913202553987503, -0.047849081456661224, 0.13561101257801056, -0.06919945776462555, -0.0709945484995842, 0.005010632798075676, 0.07735864818096161, -0.06159668043255806, 0.028194820508360863, -0.11914651095867157, 0.04332182928919792, 0.054001010954380035, 0.06397905945777893, 0.058258961886167526, -0.1263829618692398, -0.0708836242556572, 0.009215966798365116, -0.0748254582285881, -0.1285163015127182, 0.10426085442304611, 0.05070926249027252, 0.07624394446611404, 0.04690077155828476, 0.03602917492389679, 0.04494396969676018, 0.009085949510335922, -0.13419803977012634, 0.15050143003463745, -0.034831639379262924, -0.34912797808647156, -0.03409019485116005, 0.09434914588928223, -0.04539947211742401, -0.04525769129395485, 0.061148907989263535, -0.1666322946548462, 0.01588713563978672, -0.00037349952617660165, 0.038898006081581116, -0.0007825231878086925, 0.015726864337921143, 0.0822867676615715, 0.027400799095630646, 0.03714154660701752, -0.08496090769767761, -0.039934828877449036, -0.08539316803216934, -0.06339694559574127, 0.08496460318565369, -0.12961672246456146, 0.038331545889377594, 0.02689221315085888, -0.06154846027493477, 0.020044395700097084, -0.021918315440416336, 0.25879985094070435, -0.08735295385122299, 0.010247420519590378, 0.14928843080997467, -0.12924548983573914, 0.02571197785437107, 0.1469881683588028, -0.021672742441296577, -0.09273982793092728, 0.0680340975522995, 0.056351084262132645, -0.01574140600860119, -0.24390535056591034, -0.01983964629471302, -0.017649643123149872, -0.00639861635863781, -0.0904071107506752, 0.030973218381404877, 0.06429192423820496, 0.08317839354276657, -0.006789624225348234, -0.058485325425863266, -0.030097227543592453, 0.01833752728998661, 0.2222054898738861, 0.03759608417749405, 0.12540358304977417, -0.04600447043776512, -0.0639946237206459, 0.024275964125990868, 0.09815912693738937, 0.08099458366632462, 0.04964635893702507, -0.01685021072626114, 0.10912510752677917, 0.19492806494235992, 0.17979542911052704, 0.026352746412158012, -0.06892112642526627, -0.049460191279649734, -0.01391538791358471, -0.014122571796178818, -0.06931345909833908, 0.03594903647899628, -0.022734595462679863, -0.0301434975117445, -0.012731348164379597, 0.06825247406959534, 0.06559862196445465, 0.22153283655643463, 0.04618977755308151, -0.11729921400547028, -0.10630511492490768, 0.019386667758226395, -0.092210553586483, 0.007764512673020363, 0.06473853439092636, 0.0319364033639431, -0.1373714953660965, 0.07070440798997879, 0.029348159208893776, 0.12758395075798035, -0.07663482427597046, 0.07175026834011078, -0.05240427702665329, -0.18168215453624725, 0.01806592382490635, 0.03802090883255005, -0.36218082904815674, 0.1855057328939438, 0.010046476498246193, 0.026293180882930756, -0.04538657143712044, -0.04299401864409447, -0.03909990191459656, 0.05689476430416107, 0.07100257277488708, -0.020124539732933044, 0.007341781165450811, -0.09641318768262863, -0.03052595630288124, 0.09976986795663834, 0.05611196160316467, 0.038535211235284805, 0.035100311040878296, 0.01787007600069046, 0.048928212374448776, 0.018775269389152527, 0.05187604948878288, -0.2168058305978775, -0.024653209373354912, 0.0328778401017189, 0.015289828181266785, 0.03332621604204178, -0.04894150421023369, -0.05026279762387276, -0.1221374049782753, 0.11329550296068192, -0.11129748076200485, -0.11473153531551361, -0.06498581916093826, 0.06818735599517822, 0.01573241502046585, -0.07742141932249069, -0.005606579128652811, -0.019919319078326225, 0.022299811244010925, -0.04385669156908989, -0.05367192625999451, 0.07615046948194504, -0.02322353795170784, -0.1251642256975174, 0.023103075101971626, 0.09773983806371689, 0.026317337527871132, 0.05572840943932533, 0.08485281467437744, -0.026214003562927246, 0.04837996885180473, -0.0709821954369545, -0.03537744656205177, 0.001126654096879065, -0.06717360019683838, -0.036802563816308975, -0.09041166305541992, 0.008905279450118542, -0.13697311282157898, -0.07513025403022766, 0.16341543197631836, 0.15817369520664215, -0.051550570875406265, 0.09390424937009811, 0.17113140225410461, 0.013964718207716942, -0.12953655421733856, -0.008122354745864868, 0.011574974283576012, 0.039610009640455246, -0.05266452953219414, -0.05424536392092705, 0.12159707397222519, 0.11533783376216888, -0.02428716979920864, 0.003065894590690732, -0.16135235130786896, -0.11518917977809906, 0.1041150838136673, -0.017578257247805595, 0.29201704263687134, -0.12603570520877838, -0.04217913746833801, 0.00486409105360508, -0.2568013072013855, 0.07150240242481232, -0.06523620337247849, 0.058289334177970886, -0.03482051566243172, 0.07154475152492523, -0.008053096011281013, -0.001326818484812975, 0.13984255492687225, 0.044788505882024765, 0.007591368164867163, -0.060701511800289154, -0.012231211178004742, 0.06510938704013824, 0.0378466472029686, 0.10012119263410568, 0.003920756746083498, 0.054012712091207504, -0.17143192887306213, -0.06457504630088806, -0.08572402596473694, -0.03594553470611572, -0.0452277772128582, -0.06023741513490677, -0.03854816406965256, 0.034546274691820145, 0.014912388287484646, 0.011260184459388256, 0.019980430603027344, -0.08583595603704453, 0.05478881672024727, 0.003247666172683239, 0.08813011646270752, -0.07277432829141617, 0.04613110050559044, -0.041859615594148636, -0.05187537521123886, 0.12257750332355499, -0.178974911570549, 0.04606940597295761, 0.10309114307165146, -0.014022739604115486, 0.08661479502916336, 0.03414091095328331, -0.03467842936515808, 0.09165123105049133, 0.037856586277484894, -0.106887087225914, -0.10682990401983261, 0.037379730492830276, -0.13799163699150085, 0.023923518136143684, 0.01002503465861082, 0.117428719997406, -0.022164050489664078, -0.041465479880571365, 0.009220607578754425, -0.013071568682789803, -0.0800926461815834, 0.0819300189614296, 0.05743185058236122, 0.03321155160665512, -0.08499567955732346, 0.09456615895032883, 0.04912583902478218, -0.15450482070446014, 0.0023081479594111443, 0.027613256126642227, -0.17443880438804626, -0.042807597666978836, -0.11280424147844315, 0.05440044403076172, -0.02909133769571781, -0.0850832387804985, -0.10940232872962952, -0.1427372843027115, 0.04591602459549904, 0.1431533694267273, 0.07215593010187149, 0.03838910534977913, 0.03570856153964996, 0.005145614966750145, 0.0005491864285431802, 0.030433781445026398, 0.07339853048324585, 0.05918349698185921, -0.09474094957113266, -0.04255056008696556, 0.02775123529136181, 0.17242343723773956, -0.05170850083231926, -0.022351011633872986, -0.08758708089590073, 0.07108794897794724, -0.24875445663928986, 0.060475368052721024, -0.028482867404818535, -0.03546518459916115, -0.03008725307881832, -0.15927094221115112, -0.05026839300990105, 0.01764511875808239, -0.03457817807793617, 0.031150933355093002, -0.03245268017053604, -0.021383823826909065, -0.07610439509153366, -0.018650807440280914, 0.11815597116947174, -0.06765275448560715, 0.1063288152217865, 0.06456310302019119, -0.07578178495168686, 0.10065814107656479, -0.08372348546981812, -0.04496735706925392, 0.059810999780893326, 0.025451544672250748, 0.027658823877573013, 0.003491915762424469, 0.045853979885578156, 0.05110572278499603, 0.05981012433767319, 0.05928643047809601, 0.2085205316543579, -0.11118753254413605, -0.04339536651968956, -0.03814433515071869, -0.0679766833782196, -0.04707787558436394, -0.0881853923201561, 0.16501273214817047, 0.062335219234228134, 0.12267941236495972, -0.07359664887189865, 0.05564722418785095, -0.10055836290121078, -0.005725170020014048, 0.013124492950737476, -0.07603725045919418, 0.01857525110244751, -0.04866505786776543, 0.02705702930688858, -0.0535009391605854, 0.2014784961938858, -0.11974319815635681, -0.004596707411110401, 0.06004063040018082, -0.04909021034836769, -0.022361215204000473, 0.029932817444205284, 0.2606382966041565, 0.11327296495437622, -0.010474716313183308, -0.08197222650051117, -0.0530007928609848, -0.020716916769742966, -0.017303647473454475, 0.06908117234706879, 0.2986431121826172, -0.02387140318751335, 0.0765424445271492, 0.07696174085140228, 0.1154627725481987, -0.05433383211493492, -0.03465309366583824, -0.11879149824380875, -0.03074863739311695, 0.032842569053173065, -0.014248042367398739, 0.26189202070236206, -0.08219341933727264, 0.02795613743364811, -0.04316955804824829, -0.05906490236520767, -0.13355639576911926, -0.012706642970442772, -0.13744962215423584, -0.10553054511547089, 0.05745919421315193, -0.07193827629089355, -0.06217426434159279, 0.1376897245645523, 0.07039018720388412, -0.044483792036771774, 0.21120195090770721, 0.08641792088747025, -0.0694768950343132, 0.08345131576061249, -0.04902343451976776, -0.007738600485026836, 0.05599581450223923, 0.011617873795330524, 0.04914214462041855, 0.056667134165763855, 0.004958674777299166, -0.003949407488107681, -0.05323909968137741, -0.046926774084568024, -0.10858152061700821, -0.0458916611969471, -0.008828933350741863, 0.038102224469184875, -0.0017246523639187217, 0.07838600128889084, 0.055669937282800674, 0.036806803196668625, -0.002607713220641017, 0.09152407944202423, 0.03463519364595413, -0.1870381236076355, -0.18481120467185974, 0.037760209292173386, 0.028327811509370804, 0.06573953479528427, 0.040074799209833145, -0.056129105389118195, -0.03259839490056038, 0.31550073623657227, 0.19431228935718536, -0.05566665157675743, 0.04109997674822807, 0.05412213131785393, 0.042672500014305115, -0.004334264900535345, 0.04501243680715561, 0.11548613756895065, 0.14313171803951263, -0.06790722906589508, -0.09975873678922653, -0.06202331930398941, -0.08381101489067078, -0.08324384689331055, 0.04168587177991867, 0.06578278541564941, -0.007668302860110998, -0.08741346746683121, 0.08800508826971054, -0.04951449856162071, 0.00251093041151762, -0.05109037831425667, -0.14309506118297577, -0.11874251812696457, -0.03325094282627106, -0.07695502042770386, 0.0003530853136908263, 0.014396986924111843, -0.0074133556336164474, 0.009965546429157257, 0.04610896110534668, 0.01752399280667305, -0.024230988696217537, 0.002997289877384901, 0.1286107897758484, 0.08856961876153946, -0.03553637117147446, 0.04571190103888512, 0.16436034440994263, 0.0471607968211174, 0.09231231361627579, 0.025476500391960144, 0.15932036936283112, 0.06029670685529709, 0.06884951889514923, -0.07559294998645782, 0.058610476553440094, 0.03401279076933861, 0.027434593066573143, 0.12680977582931519, -0.16601425409317017, 0.05144305154681206, -0.03349217772483826, -0.06154589354991913, -0.1470811516046524, 0.11567503213882446, -0.04427654296159744, 0.08556554466485977, 0.1527124047279358, -0.03687634691596031, 0.053801365196704865, -0.040937405079603195, 0.046814195811748505, 0.022740036249160767, -0.075853630900383, -0.06855916976928711, -0.1666407734155655, 0.027590613812208176, 0.019091598689556122, 0.0025507472455501556, -0.2284461408853531, -0.03560779243707657, -0.02970220521092415, 0.003730073105543852, 0.003377749351784587, 0.06043329834938049, 0.06482624262571335, 0.034312035888433456, -0.0032150449696928263, -0.2682606279850006, 0.02524673007428646, 0.06748870760202408, -0.13957887887954712, -0.09159059822559357 ]
null
null
transformers
# airesearch/wangchanberta-base-att-spm-uncased Finetuning `airesearch/wangchanberta-base-att-spm-uncased` with the training set of `iapp_wiki_qa_squad`, `thaiqa_squad`, and `nsc_qa` (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 `newmm` words). Benchmarks shared on [wandb](https://wandb.ai/cstorm125/wangchanberta-qa) using validation and test sets of `iapp_wiki_qa_squad`. Trained with [thai2transformers](https://github.com/vistec-AI/thai2transformers/blob/dev/scripts/downstream/train_question_answering_lm_finetuning.py). Run with: ``` export MODEL_NAME=airesearch/wangchanberta-base-att-spm-uncased python train_question_answering_lm_finetuning.py \ --model_name $MODEL_NAME \ --dataset_name chimera_qa \ --output_dir $MODEL_NAME-finetune-chimera_qa-model \ --log_dir $MODEL_NAME-finetune-chimera_qa-log \ --lowercase \ --pad_on_right \ --fp16 ```
{"widget": [{"text": "\u0e2a\u0e27\u0e19\u0e01\u0e38\u0e2b\u0e25\u0e32\u0e1a\u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e2d\u0e30\u0e44\u0e23", "context": "\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e2a\u0e27\u0e19\u0e01\u0e38\u0e2b\u0e25\u0e32\u0e1a\u0e27\u0e34\u0e17\u0e22\u0e32\u0e25\u0e31\u0e22 (Suankularb Wittayalai School) (\u0e2d\u0e31\u0e01\u0e29\u0e23\u0e22\u0e48\u0e2d : \u0e2a.\u0e01. / S.K.) \u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e0a\u0e32\u0e22\u0e25\u0e49\u0e27\u0e19 \u0e23\u0e30\u0e14\u0e31\u0e1a\u0e0a\u0e31\u0e49\u0e19\u0e21\u0e31\u0e18\u0e22\u0e21\u0e28\u0e36\u0e01\u0e29\u0e32\u0e02\u0e19\u0e32\u0e14\u0e43\u0e2b\u0e0d\u0e48\u0e1e\u0e34\u0e40\u0e28\u0e29 \u0e2a\u0e31\u0e07\u0e01\u0e31\u0e14\u0e2a\u0e33\u0e19\u0e31\u0e01\u0e07\u0e32\u0e19\u0e40\u0e02\u0e15\u0e1e\u0e37\u0e49\u0e19\u0e17\u0e35\u0e48\u0e01\u0e32\u0e23\u0e28\u0e36\u0e01\u0e29\u0e32\u0e21\u0e31\u0e18\u0e22\u0e21\u0e28\u0e36\u0e01\u0e29\u0e32\u0e40\u0e02\u0e15 1 \u0e2a\u0e33\u0e19\u0e31\u0e01\u0e07\u0e32\u0e19\u0e04\u0e13\u0e30\u0e01\u0e23\u0e23\u0e21\u0e01\u0e32\u0e23\u0e01\u0e32\u0e23\u0e28\u0e36\u0e01\u0e29\u0e32\u0e02\u0e31\u0e49\u0e19\u0e1e\u0e37\u0e49\u0e19\u0e10\u0e32\u0e19 (\u0e0a\u0e37\u0e48\u0e2d\u0e40\u0e14\u0e34\u0e21: \u0e01\u0e23\u0e21\u0e2a\u0e32\u0e21\u0e31\u0e0d\u0e28\u0e36\u0e01\u0e29\u0e32) \u0e01\u0e23\u0e30\u0e17\u0e23\u0e27\u0e07\u0e28\u0e36\u0e01\u0e29\u0e32\u0e18\u0e34\u0e01\u0e32\u0e23 \u0e01\u0e48\u0e2d\u0e15\u0e31\u0e49\u0e07\u0e42\u0e14\u0e22 \u0e1e\u0e23\u0e30\u0e1a\u0e32\u0e17\u0e2a\u0e21\u0e40\u0e14\u0e47\u0e08\u0e1e\u0e23\u0e30\u0e08\u0e38\u0e25\u0e08\u0e2d\u0e21\u0e40\u0e01\u0e25\u0e49\u0e32\u0e40\u0e08\u0e49\u0e32\u0e2d\u0e22\u0e39\u0e48\u0e2b\u0e31\u0e27 \u0e44\u0e14\u0e49\u0e23\u0e31\u0e1a\u0e01\u0e32\u0e23\u0e2a\u0e16\u0e32\u0e1b\u0e19\u0e32\u0e02\u0e36\u0e49\u0e19\u0e43\u0e19\u0e27\u0e31\u0e19\u0e17\u0e35\u0e48 8 \u0e21\u0e35\u0e19\u0e32\u0e04\u0e21 \u0e1e.\u0e28. 2424 (\u0e02\u0e13\u0e30\u0e19\u0e31\u0e49\u0e19\u0e19\u0e31\u0e1a\u0e27\u0e31\u0e19\u0e17\u0e35\u0e48 1 \u0e40\u0e21\u0e29\u0e32\u0e22\u0e19 \u0e40\u0e1b\u0e47\u0e19\u0e27\u0e31\u0e19\u0e02\u0e36\u0e49\u0e19\u0e1b\u0e35\u0e43\u0e2b\u0e21\u0e48 \u0e40\u0e21\u0e37\u0e48\u0e2d\u0e19\u0e31\u0e1a\u0e2d\u0e22\u0e48\u0e32\u0e07\u0e2a\u0e32\u0e01\u0e25\u0e16\u0e37\u0e2d\u0e40\u0e1b\u0e47\u0e19 \u0e1e.\u0e28. 2425) \u0e42\u0e14\u0e22\u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e23\u0e31\u0e10\u0e1a\u0e32\u0e25\u0e41\u0e2b\u0e48\u0e07\u0e41\u0e23\u0e01\u0e02\u0e2d\u0e07\u0e1b\u0e23\u0e30\u0e40\u0e17\u0e28\u0e44\u0e17\u0e22"}]}
question-answering
cstorm125/wangchanberta-base-att-spm-uncased-finetune-qa
[ "transformers", "pytorch", "camembert", "question-answering", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us
# airesearch/wangchanberta-base-att-spm-uncased Finetuning 'airesearch/wangchanberta-base-att-spm-uncased' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'. Trained with thai2transformers. Run with:
[ "# airesearch/wangchanberta-base-att-spm-uncased\n\nFinetuning 'airesearch/wangchanberta-base-att-spm-uncased' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:" ]
[ "TAGS\n#transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us \n", "# airesearch/wangchanberta-base-att-spm-uncased\n\nFinetuning 'airesearch/wangchanberta-base-att-spm-uncased' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:" ]
[ 31, 154 ]
[ "passage: TAGS\n#transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us \n# airesearch/wangchanberta-base-att-spm-uncased\n\nFinetuning 'airesearch/wangchanberta-base-att-spm-uncased' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:" ]
[ -0.06169361621141434, -0.10002221912145615, -0.004205552861094475, 0.07189788669347763, 0.053869858384132385, 0.00043351310887373984, 0.07556740939617157, 0.09258310496807098, 0.0015330344904214144, 0.018894976004958153, 0.056050099432468414, 0.03379638120532036, 0.06882385164499283, 0.02088298462331295, -0.07881523668766022, -0.14780306816101074, 0.06325415521860123, 0.07154237478971481, -0.039426371455192566, 0.1499026119709015, 0.08045346289873123, -0.05171479284763336, 0.08370348066091537, 0.06119164824485779, -0.10609216243028641, -0.015914801508188248, 0.013856706209480762, -0.10142084211111069, 0.1307935267686844, -0.029467513784766197, 0.18172860145568848, 0.05570966750383377, -0.07346896827220917, -0.17744092643260956, 0.05200200900435448, 0.0028096248861402273, 0.06438665091991425, 0.004760744050145149, -0.013892127200961113, 0.07254758477210999, -0.027095070108771324, 0.055308498442173004, 0.04892750829458237, 0.015893742442131042, -0.10955753177404404, -0.022485611960291862, -0.06075170263648033, -0.014564082957804203, 0.12420249730348587, 0.05757579207420349, -0.019274257123470306, 0.2233375608921051, -0.2180650681257248, 0.04848141968250275, 0.03335811570286751, -0.37762749195098877, -0.01927879825234413, 0.03250092640519142, -0.00508640892803669, 0.03554008528590202, -0.07425438612699509, -0.018375463783740997, 0.02229929529130459, 0.03983553498983383, -0.04115912690758705, -0.08321043103933334, -0.13996732234954834, -0.02841419354081154, -0.07779299467802048, 0.01802234724164009, 0.15968483686447144, 0.062126852571964264, -0.08258774131536484, -0.07678709924221039, -0.02983308769762516, 0.06028047204017639, 0.0077493153512477875, -0.08891028165817261, -0.03643720969557762, 0.007560596335679293, -0.015414334833621979, 0.020807800814509392, -0.10042005777359009, -0.06494557112455368, -0.11268854141235352, 0.03826236352324486, 0.06577445566654205, 0.041410256177186966, -0.09102389216423035, 0.009106609970331192, -0.030024448409676552, -0.11758098006248474, -0.0799388661980629, -0.074864961206913, -0.042097192257642746, 0.0007490995340049267, -0.07643626630306244, -0.046362247318029404, 0.12150228768587112, 0.0990038588643074, -0.06157189980149269, 0.04946226254105568, -0.08012381196022034, 0.05913054570555687, -0.05930085480213165, 0.1384335309267044, 0.013986639678478241, 0.10091322660446167, 0.026065785437822342, 0.010957028716802597, -0.039922911673784256, 0.004855569452047348, 0.030675673857331276, -0.047453414648771286, 0.03447260335087776, 0.025958500802516937, -0.09219276905059814, 0.08182504028081894, -0.07938089966773987, -0.037510983645915985, 0.15957383811473846, -0.09615491330623627, -0.07862246781587601, 0.018472490832209587, -0.007903626188635826, 0.06077247112989426, 0.021369000896811485, 0.10068856924772263, -0.024465274065732956, 0.04754210263490677, -0.008798479102551937, 0.010557956993579865, 0.04452326521277428, -0.03458027169108391, 0.012442853301763535, -0.11534678936004639, 0.0009347318555228412, -0.14984340965747833, -0.1437915414571762, -0.0070157768204808235, 0.006285584066063166, 0.0446736179292202, -0.006384925451129675, -0.028707843273878098, 0.016606729477643967, -0.08181065320968628, -0.008981768041849136, -0.02999919466674328, -0.03934856876730919, 0.08228738605976105, 0.047177936881780624, 0.12361925095319748, -0.125436931848526, 0.04347721114754677, -0.18220451474189758, 0.045577261596918106, -0.1444200724363327, 0.054821815341711044, -0.05909958481788635, 0.02893843501806259, -0.04693695902824402, -0.015292626805603504, -0.0862935408949852, -0.0027207338716834784, 0.015572995878756046, 0.12878727912902832, -0.1123671680688858, -0.06375480443239212, 0.1725902259349823, -0.1311982125043869, -0.18728207051753998, 0.19393785297870636, -0.04058343544602394, 0.07396263629198074, 0.06410014629364014, 0.2733392119407654, 0.02717333287000656, -0.08730173110961914, 0.046496644616127014, -0.04424462467432022, -0.025206154212355614, 0.0005662143230438232, 0.04915955290198326, 0.00440028915181756, -0.04320681467652321, 0.08445252478122711, -0.08673129230737686, 0.05889536440372467, -0.046149078756570816, -0.09840618818998337, -0.0703902542591095, -0.11425910145044327, 0.06548812240362167, -0.027742184698581696, 0.12231721729040146, -0.05326024815440178, 0.006483196746557951, -0.0592370368540287, 0.04442058131098747, 0.012288780882954597, 0.0022907035890966654, -0.11051249504089355, 0.08227857202291489, -0.04858686402440071, -0.023635245859622955, -0.11662784963846207, -0.020959611982107162, 0.0523233562707901, 0.009824077598750591, 0.04187772423028946, 0.037129007279872894, 0.004784129094332457, -0.01726248860359192, -0.041229620575904846, 0.02251034416258335, 0.0057913619093596935, -0.005497370380908251, -0.031967002898454666, -0.02054222859442234, 0.06335511803627014, -0.03324764594435692, 0.08697859197854996, -0.10404743999242783, 0.027222832664847374, -0.015048091299831867, 0.12252667546272278, 0.02222803421318531, 0.011955438181757927, 0.08058122545480728, 0.10088168829679489, 0.030368493869900703, 0.023635977879166603, 0.061980314552783966, 0.0252189002931118, -0.13756632804870605, 0.06765339523553848, 0.0053394390270113945, 0.14501088857650757, 0.1414213329553604, -0.12246610224246979, -0.01764488033950329, -0.017051201313734055, -0.04647308960556984, -0.04518350586295128, -0.08538533002138138, 0.10162870585918427, 0.200114905834198, 0.013411130756139755, 0.1528967171907425, -0.10508503764867783, -0.004123337101191282, -0.011907669715583324, -0.018922701478004456, 0.014597728848457336, 0.049043554812669754, 0.05241959169507027, -0.14901019632816315, 0.06038863956928253, 0.22179122269153595, -0.05515649542212486, 0.141051784157753, -0.08156585693359375, -0.057337693870067596, 0.02244514413177967, 0.12324333935976028, -0.04422465339303017, 0.0336410328745842, -0.14592836797237396, 0.014853546395897865, 0.06479912996292114, 0.0414838008582592, 0.022096576169133186, -0.13086377084255219, -0.057797327637672424, -0.0037631848827004433, -0.052512235939502716, -0.1335785835981369, 0.08514571934938431, 0.05557866394519806, 0.07111447304487228, 0.005498175974935293, -0.0034974655136466026, 0.02184721827507019, 0.007310779765248299, -0.10709719359874725, 0.14943477511405945, -0.027205105870962143, -0.2909665107727051, -0.031847432255744934, 0.13552425801753998, -0.09623327851295471, -0.049343179911375046, 0.04693479835987091, -0.14313484728336334, 0.03745276480913162, 0.004610451404005289, -0.017236188054084778, 0.04192980378866196, 0.01729697361588478, 0.017933903262019157, 0.011490057222545147, 0.041749339550733566, -0.06797100603580475, -0.03891343995928764, -0.07998894155025482, -0.07915297895669937, 0.10557576268911362, -0.09593117982149124, 0.043032750487327576, 0.017215635627508163, -0.07243600487709045, 0.021132053807377815, -0.004550217650830746, 0.24405474960803986, -0.05479152128100395, 0.012433988973498344, 0.10716351866722107, -0.1253141164779663, 0.02880443073809147, 0.10679998993873596, -0.014378624968230724, -0.09109317511320114, 0.04782459884881973, 0.03865228220820427, -0.03956577926874161, -0.23330718278884888, -0.021192627027630806, -0.03961390256881714, 0.0038678369019180536, -0.05782794579863548, 0.021200846880674362, 0.024960583075881004, 0.10769889503717422, 0.02499094232916832, -0.08388269692659378, -0.09722240269184113, 0.00567483389750123, 0.11865472793579102, 0.047723475843667984, 0.1299983710050583, -0.04046669229865074, -0.04323042184114456, 0.02878458984196186, 0.12913280725479126, 0.1296088546514511, 0.03773890808224678, -0.029365822672843933, 0.10636596381664276, 0.2556578814983368, 0.15492616593837738, 0.052548978477716446, -0.07707565277814865, -0.06477530300617218, -0.006206491030752659, -0.022761518135666847, -0.10691992193460464, 0.0416632741689682, 0.012434408068656921, -0.005094287917017937, -0.03379865735769272, 0.04600908234715462, 0.07021338492631912, 0.25410622358322144, 0.02924523502588272, -0.09427227824926376, -0.08052118122577667, -0.008222493343055248, -0.09003116190433502, -0.008498786017298698, 0.06659353524446487, 0.037630561739206314, -0.15649546682834625, 0.04546568542718887, 0.009362280368804932, 0.12366414815187454, -0.07129386812448502, 0.05941012129187584, -0.0680471882224083, -0.15080858767032623, 0.01797897182404995, 0.01667250692844391, -0.3056921660900116, 0.19709573686122894, 0.008104043081402779, 0.02410631813108921, -0.025889290496706963, -0.04305456206202507, 0.007793188560754061, 0.022794902324676514, 0.09057972580194473, -0.006996447220444679, -0.018983198329806328, -0.11567658931016922, -0.031826868653297424, 0.12249711155891418, 0.06111735850572586, 0.0763811469078064, 0.017325384542346, 0.03975439816713333, 0.059196632355451584, 0.0039195953868329525, 0.05778883025050163, -0.20031100511550903, -0.009421936236321926, 0.011063692159950733, 0.029392186552286148, -0.008061188273131847, -0.032133717089891434, -0.031081847846508026, -0.17016977071762085, 0.17616112530231476, -0.06621050089597702, -0.10067577660083771, -0.07693468034267426, 0.003513741074129939, 0.03810020908713341, -0.06551644206047058, 0.029722420498728752, -0.06026227027177811, -0.04726657271385193, -0.02681538462638855, -0.048768751323223114, 0.07808596640825272, -0.03581555560231209, -0.07505828887224197, 0.005627943202853203, 0.06264296174049377, -0.004832218401134014, 0.05743733048439026, 0.07679804414510727, -0.011898175813257694, 0.017700601369142532, -0.08890075981616974, -0.014457624405622482, -0.0444558821618557, -0.052028998732566833, 0.040820639580488205, -0.10606168210506439, 0.04327971488237381, -0.11109162122011185, -0.08143185079097748, 0.19476056098937988, 0.16286590695381165, -0.03830110281705856, 0.11106874793767929, 0.17610420286655426, -0.0002615197154227644, -0.11108866333961487, -0.004705268424004316, 0.01022527739405632, 0.052097249776124954, -0.059899818152189255, -0.06701533496379852, 0.16358929872512817, 0.11553844809532166, -0.008818134665489197, -0.013109995983541012, -0.08157234638929367, -0.10417088866233826, 0.1317957043647766, 0.03267379477620125, 0.2772156596183777, -0.11232812702655792, -0.05089246481657028, 0.038667865097522736, -0.1873830407857895, 0.03267952799797058, -0.06122628599405289, 0.08475341647863388, -0.05026750639081001, 0.04026272892951965, -0.00038236932596191764, -0.02417699806392193, 0.13777956366539001, 0.07060838490724564, 0.03893033042550087, -0.033857300877571106, 0.030088260769844055, 0.01789572462439537, 0.04270599037408829, 0.11460387706756592, 0.0022367697674781084, 0.06534150242805481, -0.1819716989994049, -0.01358972117304802, -0.10792557895183563, -0.00410406943410635, -0.005451282951980829, -0.05070118606090546, -0.027830207720398903, 0.04011179506778717, 0.005807036999613047, 0.032246701419353485, 0.0017646305495873094, -0.06203322485089302, 0.1301763951778412, 0.08305943757295609, 0.08800799399614334, -0.11015021055936813, 0.04797990992665291, -0.05137540400028229, -0.04442062973976135, 0.1126914769411087, -0.14321351051330566, 0.03404460474848747, 0.12610577046871185, 0.017370332032442093, 0.07634872198104858, 0.036785539239645004, 0.009356587193906307, 0.08525122702121735, 0.047443728893995285, -0.10432633757591248, -0.059030722826719284, 0.032023631036281586, -0.14456824958324432, -0.007398600224405527, 0.016361868008971214, 0.1019783616065979, -0.03853105008602142, -0.027260305359959602, -0.00385312270373106, -0.03982248529791832, -0.08822116255760193, 0.09679339081048965, 0.07738488167524338, 0.03591972589492798, -0.0967877134680748, 0.08390846848487854, 0.02563735470175743, -0.12706442177295685, -0.0032031163573265076, -0.0039563109166920185, -0.1714022159576416, -0.05908457562327385, -0.07526203989982605, 0.10358908027410507, -0.018426012247800827, -0.05482950061559677, -0.10769602656364441, -0.12631656229496002, 0.04333760589361191, 0.12553802132606506, 0.08480799943208694, 0.044392045587301254, 0.019968733191490173, -0.025111857801675797, 0.0007954351603984833, 0.002413736889138818, 0.047946564853191376, 0.06741069257259369, -0.11232953518629074, -0.027234936133027077, 0.016353042796254158, 0.14100387692451477, -0.03659394383430481, -0.04088450223207474, -0.11104282736778259, 0.06190106272697449, -0.18850429356098175, 0.04973035305738449, -0.05154803395271301, -0.009293662384152412, 0.0048852404579520226, -0.15151241421699524, -0.04334920644760132, 0.01382733415812254, -0.062490709125995636, 0.024343056604266167, -0.016252733767032623, 0.02488892711699009, -0.0672866627573967, -0.025007013231515884, 0.12215842306613922, -0.058295249938964844, 0.08162006735801697, 0.05741802975535393, -0.09396599233150482, 0.09168445318937302, -0.10742981731891632, -0.05108426511287689, 0.05277048423886299, 0.03596753627061844, 0.03542184457182884, -0.0027388064190745354, 0.0464644655585289, 0.049456678330898285, 0.06271960586309433, 0.0708705484867096, 0.2423771470785141, -0.11448174715042114, -0.10107527673244476, -0.020830774679780006, -0.06300143897533417, -0.06520339846611023, -0.07181468605995178, 0.18103745579719543, 0.057584863156080246, 0.1323997527360916, -0.056769996881484985, 0.06504740566015244, -0.13220000267028809, -0.018026063218712807, 0.009753167629241943, -0.10372498631477356, 0.027559585869312286, -0.05973218008875847, 0.017645899206399918, -0.03413645550608635, 0.18920549750328064, -0.1295851171016693, 0.025772910565137863, 0.031217779964208603, 0.0029166999738663435, -0.03343433141708374, 0.029039662331342697, 0.2753395736217499, 0.11997071653604507, -0.03170079365372658, -0.028389951214194298, -0.03135001286864281, -0.035072509199380875, 0.007990779355168343, 0.036549393087625504, 0.29954794049263, -0.06481378525495529, 0.0648762509226799, 0.08465846627950668, 0.12617439031600952, -0.019791804254055023, -0.03305554389953613, -0.07437150180339813, 0.02939327247440815, 0.03638555109500885, -0.05411858111619949, 0.23659482598304749, -0.10902774333953857, 0.0519835501909256, -0.043244630098342896, -0.07163432985544205, -0.09900432825088501, 0.00798000954091549, -0.10727941989898682, -0.08851034194231033, 0.06724265962839127, -0.0633554682135582, -0.06808631122112274, 0.16642814874649048, 0.06810217350721359, -0.007940524257719517, 0.19051098823547363, 0.05629001557826996, -0.040219634771347046, 0.10916294902563095, -0.06081356108188629, -0.02990959770977497, 0.019412606954574585, -0.008496788330376148, 0.032619066536426544, 0.04330091550946236, -0.00928207952529192, -0.0000866249029058963, -0.04226485639810562, -0.039829641580581665, -0.12233851850032806, -0.05587826669216156, -0.012137824669480324, -0.002866057213395834, -0.003344410564750433, 0.07090970873832703, 0.03338225558400154, 0.034170836210250854, 0.00014552322681993246, 0.09995167702436447, 0.005156655330210924, -0.14912573993206024, -0.23349186778068542, 0.05662763863801956, -0.0038171224296092987, 0.014516431838274002, 0.02510402910411358, -0.029946263879537582, 0.00011264016211498529, 0.3191787600517273, 0.17379260063171387, -0.06751825660467148, 0.04835344851016998, 0.0514242947101593, 0.02553579956293106, -0.03010503575205803, 0.0723547711968422, 0.10658762603998184, 0.04303156957030296, -0.07991938292980194, -0.10427272319793701, -0.08924830704927444, -0.11073092371225357, -0.09611870348453522, 0.04778620973229408, 0.08633558452129364, -0.009221549145877361, -0.105065256357193, 0.07006194442510605, -0.05468941852450371, 0.06834042817354202, -0.05919628217816353, -0.14584235846996307, -0.14787869155406952, -0.07504476606845856, -0.0999111607670784, 0.0444989874958992, 0.0006293095648288727, -0.022529518231749535, 0.006499325856566429, 0.026564303785562515, 0.006221753545105457, -0.027467550709843636, -0.04456228390336037, 0.11044169217348099, 0.04708419740200043, -0.04304299131035805, 0.06578782200813293, 0.13966000080108643, 0.04367126524448395, 0.09135967493057251, 0.023108504712581635, 0.1198631152510643, 0.04138518497347832, 0.07996600866317749, -0.06453801691532135, 0.07830676436424255, 0.04432335123419762, 0.057898759841918945, 0.07512499392032623, -0.12583927810192108, 0.05939020961523056, -0.017162257805466652, -0.04777595400810242, -0.11666560173034668, 0.11338791996240616, -0.027847182005643845, 0.1042686253786087, 0.15346252918243408, -0.02051912620663643, 0.05718081444501877, -0.04712582752108574, 0.05831754952669144, 0.023700280115008354, -0.1251724660396576, -0.1089579239487648, -0.15888993442058563, 0.02765781618654728, 0.0028320623096078634, -0.02728802151978016, -0.21526964008808136, -0.06561625003814697, -0.02503202296793461, -0.016123821958899498, -0.014764483086764812, 0.07415689527988434, 0.0719742402434349, 0.04055998846888542, 0.0011646741768345237, -0.22894616425037384, 0.04626140743494034, 0.0538916252553463, -0.0918298065662384, -0.09028176218271255 ]
null
null
transformers
# wangchanberta-base-wiki-20210520-news-spm-finetune-qa Finetuning `airesearchth/wangchanberta-base-wiki-20210520-news-spm` with the training set of `iapp_wiki_qa_squad`, `thaiqa_squad`, and `nsc_qa` (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 `newmm` words). Benchmarks shared on [wandb](https://wandb.ai/cstorm125/wangchanberta-qa) using validation and test sets of `iapp_wiki_qa_squad`. Trained with [thai2transformers](https://github.com/vistec-AI/thai2transformers/blob/dev/scripts/downstream/train_question_answering_lm_finetuning.py). Run with: ``` export MODEL_NAME=airesearchth/wangchanberta-base-wiki-20210520-news-spm CUDA_LAUNCH_BLOCKING=1 python train_question_answering_lm_finetuning.py \ --model_name $MODEL_NAME \ --dataset_name chimera_qa \ --output_dir $MODEL_NAME-finetune-chimera_qa-model \ --log_dir $MODEL_NAME-finetune-chimera_qa-log \ --model_max_length 400 \ --pad_on_right \ --fp16 ```
{"widget": [{"text": "\u0e2a\u0e27\u0e19\u0e01\u0e38\u0e2b\u0e25\u0e32\u0e1a\u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e2d\u0e30\u0e44\u0e23", "context": "\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e2a\u0e27\u0e19\u0e01\u0e38\u0e2b\u0e25\u0e32\u0e1a\u0e27\u0e34\u0e17\u0e22\u0e32\u0e25\u0e31\u0e22 (Suankularb Wittayalai School) (\u0e2d\u0e31\u0e01\u0e29\u0e23\u0e22\u0e48\u0e2d : \u0e2a.\u0e01. / S.K.) \u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e0a\u0e32\u0e22\u0e25\u0e49\u0e27\u0e19 \u0e23\u0e30\u0e14\u0e31\u0e1a\u0e0a\u0e31\u0e49\u0e19\u0e21\u0e31\u0e18\u0e22\u0e21\u0e28\u0e36\u0e01\u0e29\u0e32\u0e02\u0e19\u0e32\u0e14\u0e43\u0e2b\u0e0d\u0e48\u0e1e\u0e34\u0e40\u0e28\u0e29 \u0e2a\u0e31\u0e07\u0e01\u0e31\u0e14\u0e2a\u0e33\u0e19\u0e31\u0e01\u0e07\u0e32\u0e19\u0e40\u0e02\u0e15\u0e1e\u0e37\u0e49\u0e19\u0e17\u0e35\u0e48\u0e01\u0e32\u0e23\u0e28\u0e36\u0e01\u0e29\u0e32\u0e21\u0e31\u0e18\u0e22\u0e21\u0e28\u0e36\u0e01\u0e29\u0e32\u0e40\u0e02\u0e15 1 \u0e2a\u0e33\u0e19\u0e31\u0e01\u0e07\u0e32\u0e19\u0e04\u0e13\u0e30\u0e01\u0e23\u0e23\u0e21\u0e01\u0e32\u0e23\u0e01\u0e32\u0e23\u0e28\u0e36\u0e01\u0e29\u0e32\u0e02\u0e31\u0e49\u0e19\u0e1e\u0e37\u0e49\u0e19\u0e10\u0e32\u0e19 (\u0e0a\u0e37\u0e48\u0e2d\u0e40\u0e14\u0e34\u0e21: \u0e01\u0e23\u0e21\u0e2a\u0e32\u0e21\u0e31\u0e0d\u0e28\u0e36\u0e01\u0e29\u0e32) \u0e01\u0e23\u0e30\u0e17\u0e23\u0e27\u0e07\u0e28\u0e36\u0e01\u0e29\u0e32\u0e18\u0e34\u0e01\u0e32\u0e23 \u0e01\u0e48\u0e2d\u0e15\u0e31\u0e49\u0e07\u0e42\u0e14\u0e22 \u0e1e\u0e23\u0e30\u0e1a\u0e32\u0e17\u0e2a\u0e21\u0e40\u0e14\u0e47\u0e08\u0e1e\u0e23\u0e30\u0e08\u0e38\u0e25\u0e08\u0e2d\u0e21\u0e40\u0e01\u0e25\u0e49\u0e32\u0e40\u0e08\u0e49\u0e32\u0e2d\u0e22\u0e39\u0e48\u0e2b\u0e31\u0e27 \u0e44\u0e14\u0e49\u0e23\u0e31\u0e1a\u0e01\u0e32\u0e23\u0e2a\u0e16\u0e32\u0e1b\u0e19\u0e32\u0e02\u0e36\u0e49\u0e19\u0e43\u0e19\u0e27\u0e31\u0e19\u0e17\u0e35\u0e48 8 \u0e21\u0e35\u0e19\u0e32\u0e04\u0e21 \u0e1e.\u0e28. 2424 (\u0e02\u0e13\u0e30\u0e19\u0e31\u0e49\u0e19\u0e19\u0e31\u0e1a\u0e27\u0e31\u0e19\u0e17\u0e35\u0e48 1 \u0e40\u0e21\u0e29\u0e32\u0e22\u0e19 \u0e40\u0e1b\u0e47\u0e19\u0e27\u0e31\u0e19\u0e02\u0e36\u0e49\u0e19\u0e1b\u0e35\u0e43\u0e2b\u0e21\u0e48 \u0e40\u0e21\u0e37\u0e48\u0e2d\u0e19\u0e31\u0e1a\u0e2d\u0e22\u0e48\u0e32\u0e07\u0e2a\u0e32\u0e01\u0e25\u0e16\u0e37\u0e2d\u0e40\u0e1b\u0e47\u0e19 \u0e1e.\u0e28. 2425) \u0e42\u0e14\u0e22\u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e23\u0e31\u0e10\u0e1a\u0e32\u0e25\u0e41\u0e2b\u0e48\u0e07\u0e41\u0e23\u0e01\u0e02\u0e2d\u0e07\u0e1b\u0e23\u0e30\u0e40\u0e17\u0e28\u0e44\u0e17\u0e22"}]}
question-answering
cstorm125/wangchanberta-base-wiki-20210520-news-spm-finetune-qa
[ "transformers", "pytorch", "camembert", "question-answering", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us
# wangchanberta-base-wiki-20210520-news-spm-finetune-qa Finetuning 'airesearchth/wangchanberta-base-wiki-20210520-news-spm' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'. Trained with thai2transformers. Run with:
[ "# wangchanberta-base-wiki-20210520-news-spm-finetune-qa\n\nFinetuning 'airesearchth/wangchanberta-base-wiki-20210520-news-spm' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:" ]
[ "TAGS\n#transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us \n", "# wangchanberta-base-wiki-20210520-news-spm-finetune-qa\n\nFinetuning 'airesearchth/wangchanberta-base-wiki-20210520-news-spm' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:" ]
[ 31, 159 ]
[ "passage: TAGS\n#transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us \n# wangchanberta-base-wiki-20210520-news-spm-finetune-qa\n\nFinetuning 'airesearchth/wangchanberta-base-wiki-20210520-news-spm' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:" ]
[ -0.05998677387833595, -0.11627022922039032, -0.003607646794989705, 0.047817789018154144, 0.026026777923107147, 0.0034189617726951838, 0.07305090874433517, 0.08659139275550842, -0.0019146837294101715, 0.0176298338919878, 0.06823267787694931, 0.005805567372590303, 0.0698540061712265, 0.02400720678269863, -0.08120398968458176, -0.19051747024059296, 0.073115274310112, 0.09388849884271622, -0.06575071066617966, 0.14523082971572876, 0.10974657535552979, -0.06184523552656174, 0.080986388027668, 0.08090835064649582, -0.1015891507267952, -0.013785232789814472, 0.026258286088705063, -0.09847381711006165, 0.14216487109661102, 0.0077484166249632835, 0.15683990716934204, 0.06991502642631531, -0.05883506312966347, -0.2115025818347931, 0.05869748443365097, -0.010367019101977348, 0.04850764945149422, 0.005671354476362467, -0.003966958727687597, 0.06872488558292389, -0.005680215544998646, 0.006985425483435392, 0.024920504540205002, 0.013415164314210415, -0.10346293449401855, -0.015129061415791512, -0.05383995547890663, 0.02034040167927742, 0.14747627079486847, 0.08754530549049377, -0.021264584735035896, 0.20202967524528503, -0.20548678934574127, 0.04154345393180847, 0.10880858451128006, -0.4099293053150177, -0.03323443606495857, 0.02706247940659523, 0.0024889043997973204, 0.0353328213095665, -0.05015769600868225, 0.014098210260272026, 0.028940493240952492, 0.0359792523086071, -0.08083879202604294, -0.0603475384414196, -0.07095978409051895, -0.023585917428135872, -0.09508562833070755, 0.010909431613981724, 0.17128241062164307, 0.051367223262786865, -0.05666772276163101, -0.09266793727874756, -0.026610128581523895, 0.03722511976957321, -0.0026004831306636333, -0.08817540854215622, -0.06186773255467415, 0.010040314868092537, -0.005798691883683205, 0.008502845652401447, -0.11521129310131073, -0.07405633479356766, -0.12667782604694366, 0.08711286634206772, 0.07361230254173279, 0.05218959227204323, -0.12901048362255096, 0.013267606496810913, -0.006514499895274639, -0.13073286414146423, -0.07734176516532898, -0.06050126627087593, -0.007644480559974909, 0.013665328733623028, -0.08424083143472672, -0.03474488481879234, 0.09256439656019211, 0.036221083253622055, -0.13586024940013885, 0.04196052625775337, -0.0655650720000267, 0.058089353144168854, -0.04812217131257057, 0.14354924857616425, 0.02502429112792015, 0.09696486592292786, 0.018876347690820694, 0.02874775230884552, -0.03350912779569626, 0.015201115980744362, 0.018791284412145615, -0.06562302261590958, 0.0007265793974511325, 0.025031480938196182, -0.0861712172627449, 0.07147690653800964, -0.05354004353284836, -0.021903200075030327, 0.12694492936134338, -0.10511471331119537, -0.09168524295091629, 0.01941707730293274, -0.0173184871673584, 0.07617122679948807, 0.011001328006386757, 0.10739396512508392, -0.03328050673007965, 0.06726278364658356, -0.013346186839044094, -0.0019135072361677885, 0.04832175374031067, 0.0002830341982189566, 0.018117375671863556, -0.08613725751638412, -0.003364939009770751, -0.14503854513168335, -0.11391928046941757, 0.013430907391011715, -0.0041270204819738865, 0.041052620857954025, -0.030090106651186943, -0.025343716144561768, 0.009870430454611778, -0.06396173685789108, 0.008846846409142017, 0.00934659969061613, -0.029680361971259117, 0.099256232380867, 0.048019591718912125, 0.11215152591466904, -0.09519494324922562, 0.03242276981472969, -0.1421627700328827, 0.05968758836388588, -0.09636510163545609, 0.06828965991735458, -0.050707198679447174, -0.013686919584870338, -0.05853891745209694, -0.0249848123639822, -0.10386598855257034, 0.008634260855615139, -0.008142286911606789, 0.13131672143936157, -0.11295925080776215, -0.07453663647174835, 0.1861482411623001, -0.12715598940849304, -0.15154238045215607, 0.188584566116333, -0.0492490790784359, 0.07511419802904129, 0.08338732272386551, 0.2285599410533905, 0.02193031646311283, -0.11407876759767532, 0.036245010793209076, -0.034521620720624924, -0.004931426141411066, 0.04894110932946205, 0.0696987584233284, 0.027594206854701042, -0.05499277263879776, 0.07585965842008591, -0.09218164533376694, 0.04294548183679581, -0.05192537233233452, -0.1046672984957695, -0.05558261275291443, -0.1113847866654396, 0.08209778368473053, -0.043060820549726486, 0.12381788343191147, -0.045256972312927246, -0.03992076590657234, -0.07498989254236221, 0.06477459520101547, 0.012280493974685669, 0.005949456710368395, -0.15539875626564026, 0.06921307742595673, -0.042045336216688156, -0.01527697779238224, -0.09236979484558105, -0.0014133088989183307, 0.03924965485930443, 0.0031116134487092495, 0.03959512710571289, 0.07103856652975082, -0.001514190691523254, -0.03565216436982155, -0.0558171309530735, 0.051967959851026535, -0.02307497337460518, -0.01788361929357052, -0.04781492426991463, -0.019967708736658096, 0.0824524313211441, -0.023761626332998276, 0.11796523630619049, -0.12608154118061066, 0.026503846049308777, 0.030350370332598686, 0.11845937371253967, 0.018997617065906525, 0.03520263358950615, 0.05705440044403076, 0.0787416324019432, 0.03369918838143349, 0.04560023546218872, 0.04032342880964279, 0.012306369841098785, -0.14395660161972046, 0.09543802589178085, -0.008855076506733894, 0.13177305459976196, 0.13109134137630463, -0.09498357027769089, 0.03510291129350662, -0.019703427329659462, -0.0535796582698822, -0.03002641163766384, -0.08879055827856064, 0.0902794823050499, 0.18155179917812347, 0.00586409866809845, 0.13442453742027283, -0.11845973879098892, -0.026207629591226578, -0.010776662267744541, -0.008617518469691277, -0.009033379144966602, 0.05705016851425171, 0.04754915460944176, -0.17212574183940887, 0.036987822502851486, 0.21095480024814606, -0.03795456513762474, 0.143288254737854, -0.07381082326173782, -0.06091918796300888, 0.022797975689172745, 0.09733492136001587, -0.04849855229258537, 0.037956010550260544, -0.12058445066213608, 0.045068662613630295, 0.06597510725259781, 0.039790280163288116, 0.051910217851400375, -0.1406411975622177, -0.0841236338019371, 0.008663168177008629, -0.06085124611854553, -0.156227245926857, 0.09724122285842896, 0.05246598273515701, 0.07250657677650452, 0.0327121801674366, 0.03767922893166542, 0.02638568915426731, 0.0034181135706603527, -0.12206479907035828, 0.14171043038368225, -0.01979479193687439, -0.31973567605018616, -0.05301983281970024, 0.09529411792755127, -0.055073801428079605, -0.06265915930271149, 0.04594616964459419, -0.18548797070980072, 0.03484101966023445, 0.0005574728129431605, 0.014795361086726189, 0.017753805965185165, 0.02234623022377491, 0.07616691291332245, 0.01950986683368683, 0.02790115214884281, -0.06627893447875977, -0.03109339252114296, -0.09040703624486923, -0.07687472552061081, 0.08244894444942474, -0.13455946743488312, 0.037059977650642395, 0.026631463319063187, -0.052343063056468964, 0.036270588636398315, -0.009029584005475044, 0.27541863918304443, -0.08937236666679382, 0.012951698154211044, 0.1425587236881256, -0.1382226049900055, 0.031114917248487473, 0.11329995840787888, -0.007124569732695818, -0.08432348072528839, 0.04594920203089714, 0.041338007897138596, -0.023339400067925453, -0.25878116488456726, -0.025960346683859825, -0.018996620550751686, -0.01347443275153637, -0.0801951140165329, 0.03662737086415291, 0.027114450931549072, 0.10499394685029984, -0.0019254089565947652, -0.0792451798915863, -0.0718548595905304, 0.00654886057600379, 0.12957876920700073, 0.056117717176675797, 0.11434309184551239, -0.047412533313035965, -0.04137048497796059, 0.02861369587481022, 0.1084737703204155, 0.10065539181232452, 0.03980393707752228, -0.023310648277401924, 0.10510362684726715, 0.25631898641586304, 0.19542300701141357, 0.03714650869369507, -0.08448204398155212, -0.07295016944408417, 0.0008509985636919737, -0.01901201903820038, -0.08075105398893356, 0.05210105702280998, -0.024804340675473213, 0.002590315882116556, -0.012421180494129658, 0.09413672983646393, 0.051052920520305634, 0.2727530896663666, 0.03070620633661747, -0.07661200314760208, -0.10693885385990143, 0.011560567654669285, -0.11977285146713257, 0.001187590998597443, 0.07540040463209152, 0.06606664508581161, -0.15002067387104034, 0.04686130955815315, 0.017017148435115814, 0.13418740034103394, -0.05813944712281227, 0.06168919429183006, -0.06995542347431183, -0.18959668278694153, 0.013998341746628284, 0.0410933792591095, -0.2963142395019531, 0.20991797745227814, 0.013450887985527515, 0.03601989150047302, -0.05275307968258858, -0.0480181984603405, -0.016603443771600723, 0.06089313328266144, 0.07905038446187973, -0.014426850713789463, 0.0006508055957965553, -0.10448193550109863, -0.047985486686229706, 0.11609014868736267, 0.07263720035552979, 0.0307297483086586, 0.033748771995306015, 0.024013133719563484, 0.05345909297466278, 0.005401741713285446, 0.04360348358750343, -0.19607169926166534, -0.040530361235141754, 0.033217694610357285, 0.03578021004796028, 0.02007213793694973, -0.05675423517823219, -0.04586167261004448, -0.17265909910202026, 0.1194014698266983, -0.12125591188669205, -0.10105548053979874, -0.06607739627361298, 0.03425133600831032, 0.02642533741891384, -0.07550635933876038, 0.006112344097346067, -0.02472393587231636, 0.021129924803972244, -0.048270780593156815, -0.043363768607378006, 0.05512123927474022, -0.04243355244398117, -0.12590958178043365, 0.02178768813610077, 0.11736694723367691, 0.027989903464913368, 0.07399840652942657, 0.08684127777814865, -0.01891312561929226, 0.01863466016948223, -0.07731688022613525, -0.021218545734882355, -0.03483981266617775, -0.059023622423410416, -0.005724424961954355, -0.0896216332912445, 0.028614968061447144, -0.14648400247097015, -0.09555945545434952, 0.1862419992685318, 0.14917904138565063, -0.05205561965703964, 0.09610815346240997, 0.14812906086444855, 0.023759106174111366, -0.10670413076877594, -0.014553889632225037, 0.028176939114928246, 0.045083943754434586, -0.06282171607017517, -0.060084328055381775, 0.15254093706607819, 0.11028483510017395, -0.029182342812418938, -0.008228513412177563, -0.13771402835845947, -0.10536394268274307, 0.08013562113046646, -0.0032225081231445074, 0.2698207199573517, -0.12010712921619415, -0.04342171549797058, 0.031208621338009834, -0.24128150939941406, 0.0580989308655262, -0.08066021651029587, 0.05707031860947609, -0.04218451678752899, 0.08004917204380035, -0.007990825921297073, -0.01268966868519783, 0.14632229506969452, 0.03721940889954567, 0.0003658049099612981, -0.04197666421532631, -0.016157690435647964, 0.007578840013593435, 0.04460960999131203, 0.07370810955762863, -0.022231237962841988, 0.06233539432287216, -0.19562667608261108, -0.045763835310935974, -0.10244742035865784, -0.026644781231880188, -0.040239498019218445, -0.04016419127583504, -0.028338788077235222, 0.03314116969704628, 0.008546906523406506, 0.03357600420713425, 0.024830717593431473, -0.09351760894060135, 0.10233718156814575, 0.02491975575685501, 0.11228886991739273, -0.07992538064718246, 0.040910426527261734, -0.060340188443660736, -0.050025053322315216, 0.11687002331018448, -0.1875673532485962, 0.03317240998148918, 0.09684830158948898, 0.006125830579549074, 0.06255091726779938, 0.02666187472641468, -0.009551660157740116, 0.09539449214935303, 0.051177944988012314, -0.11091668903827667, -0.12445065379142761, 0.014568577520549297, -0.054898012429475784, -0.00044072410673834383, 0.006758151575922966, 0.10075923800468445, -0.01588866487145424, -0.03663083538413048, -0.0010708653135225177, -0.031225528568029404, -0.0762796625494957, 0.07573699206113815, 0.07252006232738495, 0.04070081561803818, -0.10375235974788666, 0.07122351229190826, 0.0398109145462513, -0.13759852945804596, -0.0005435931961983442, 0.04584469646215439, -0.17321044206619263, -0.05296416953206062, -0.10671980679035187, 0.06103084981441498, -0.018618693575263023, -0.07593514025211334, -0.08974909037351608, -0.14864614605903625, 0.05389104411005974, 0.16787207126617432, 0.05826585739850998, 0.016978327184915543, 0.023017289116978645, -0.01562504656612873, 0.005454711616039276, 0.02216821350157261, 0.05903239548206329, 0.05132446065545082, -0.10151340812444687, -0.03191627562046051, 0.011674672365188599, 0.18397732079029083, -0.04176994785666466, -0.0238204225897789, -0.06769641488790512, 0.07774517685174942, -0.22090879082679749, 0.06979905813932419, -0.03590122610330582, -0.016994567587971687, -0.0289923083037138, -0.17664021253585815, -0.059354059398174286, 0.010113739408552647, -0.05746536701917648, 0.04008884355425835, -0.03620194271206856, -0.0111400056630373, -0.047727398574352264, -0.0007748480420559645, 0.14051760733127594, -0.06595069169998169, 0.09950956702232361, 0.05964333564043045, -0.08085672557353973, 0.12623770534992218, -0.07133127003908157, -0.05040488392114639, 0.03311267867684364, 0.03304349631071091, 0.022721657529473305, 0.024826617911458015, 0.04753779247403145, 0.04059547185897827, 0.05642355978488922, 0.06854487210512161, 0.21480756998062134, -0.11187073588371277, -0.05771850794553757, -0.0036319270730018616, -0.06279648095369339, -0.04307873547077179, -0.08058926463127136, 0.17455732822418213, 0.03269685059785843, 0.1346166878938675, -0.05678367614746094, 0.060748208314180374, -0.11418306827545166, 0.004021932370960712, 0.018021471798419952, -0.08916378021240234, 0.004342753440141678, -0.057323090732097626, 0.02286280132830143, -0.05184945836663246, 0.17615842819213867, -0.15037038922309875, -0.010270455852150917, 0.04696404188871384, -0.02051754854619503, -0.03902424871921539, 0.020506327971816063, 0.27192240953445435, 0.15053458511829376, -0.011173578910529613, -0.05573581904172897, -0.07319046556949615, -0.033770062029361725, 0.0059178913943469524, 0.06528232991695404, 0.29661568999290466, -0.009653138928115368, 0.0831252932548523, 0.07925058156251907, 0.11808599531650543, -0.009784885682165623, -0.031774770468473434, -0.1380470246076584, -0.018467923626303673, 0.030947979539632797, -0.027171071618795395, 0.2718485891819, -0.09488452970981598, 0.0218422319740057, -0.04000731557607651, -0.05533602461218834, -0.12487273663282394, -0.012654115445911884, -0.12676464021205902, -0.08280695974826813, 0.07238540053367615, -0.06295689195394516, -0.0512939877808094, 0.15865522623062134, 0.07207449525594711, -0.021424241364002228, 0.21425561606884003, 0.08011376857757568, -0.05023388937115669, 0.10213205963373184, -0.05106305703520775, -0.002883884822949767, 0.024087127298116684, 0.010742886923253536, 0.05424150079488754, 0.021989846602082253, 0.00632523512467742, -0.0009779547108337283, -0.056659918278455734, -0.04546131566166878, -0.11286120116710663, -0.055469296872615814, -0.008094334974884987, 0.012495160102844238, 0.02212664484977722, 0.051780443638563156, 0.03127003088593483, 0.06127306446433067, -0.013768339529633522, 0.09228281676769257, 0.02290290966629982, -0.17281222343444824, -0.20026624202728271, 0.004288277123123407, 0.0005470117903314531, 0.0381028912961483, 0.028483837842941284, -0.037359971553087234, -0.017819635570049286, 0.362479031085968, 0.17092789709568024, -0.06823226809501648, 0.04575801640748978, 0.05250988155603409, 0.045429181307554245, -0.03461220860481262, 0.03886113315820694, 0.10650050640106201, 0.10244264453649521, -0.08905567973852158, -0.11330749094486237, -0.06687948107719421, -0.08176212757825851, -0.0903545543551445, 0.04693090170621872, 0.07576346397399902, -0.014754929579794407, -0.0940021425485611, 0.07614202797412872, -0.051483504474163055, 0.05495968461036682, -0.07858002185821533, -0.1499720960855484, -0.1367143988609314, -0.05900723487138748, -0.07792842388153076, 0.02581060864031315, 0.006975416559726, -0.00787192489951849, -0.009087282232940197, 0.010646896436810493, 0.015641668811440468, -0.023387962952256203, -0.022182593122124672, 0.15200182795524597, 0.06244107708334923, -0.03503631427884102, 0.06585492938756943, 0.15055938065052032, 0.038191474974155426, 0.07604879885911942, 0.03273149952292442, 0.12362753599882126, 0.06322918832302094, 0.06080253794789314, -0.0908081904053688, 0.08971697837114334, 0.04993118345737457, 0.04043011739850044, 0.11748308688402176, -0.1207563653588295, 0.06113775074481964, -0.037966467440128326, -0.04998144507408142, -0.11828859895467758, 0.12855030596256256, -0.031691063195466995, 0.10061809420585632, 0.16853441298007965, -0.03329012170433998, 0.05353747680783272, -0.03274111822247505, 0.07262856513261795, 0.024150948971509933, -0.07946788519620895, -0.07369479537010193, -0.17624782025814056, 0.022387376055121422, -0.02043765038251877, -0.014464063569903374, -0.23627908527851105, -0.04915015771985054, -0.03033972531557083, 0.0019466548692435026, -0.0006079393206164241, 0.054665446281433105, 0.11854720115661621, 0.03636569529771805, 0.006296330597251654, -0.23919245600700378, 0.043719738721847534, 0.04323435202240944, -0.11095013469457626, -0.07466676831245422 ]
null
null
transformers
# wangchanberta-base-wiki-20210520-news-spm_span-mask-finetune-qa Finetuning `airesearch/wangchanberta-base-wiki-20210520-news-spm_span-mask` with the training set of `iapp_wiki_qa_squad`, `thaiqa_squad`, and `nsc_qa` (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 `newmm` words). Benchmarks shared on [wandb](https://wandb.ai/cstorm125/wangchanberta-qa) using validation and test sets of `iapp_wiki_qa_squad`. Trained with [thai2transformers](https://github.com/vistec-AI/thai2transformers/blob/dev/scripts/downstream/train_question_answering_lm_finetuning.py). Run with: ``` export MODEL_NAME=airesearch/wangchanberta-base-wiki-20210520-news-spm_span-mask CUDA_LAUNCH_BLOCKING=1 python train_question_answering_lm_finetuning.py \ --model_name $MODEL_NAME \ --dataset_name chimera_qa \ --output_dir $MODEL_NAME-finetune-chimera_qa-model \ --log_dir $MODEL_NAME-finetune-chimera_qa-log \ --model_max_length 400 \ --pad_on_right \ --fp16 \ --use_auth_token ```
{"widget": [{"text": "\u0e2a\u0e27\u0e19\u0e01\u0e38\u0e2b\u0e25\u0e32\u0e1a\u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e2d\u0e30\u0e44\u0e23", "context": "\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e2a\u0e27\u0e19\u0e01\u0e38\u0e2b\u0e25\u0e32\u0e1a\u0e27\u0e34\u0e17\u0e22\u0e32\u0e25\u0e31\u0e22 (Suankularb Wittayalai School) (\u0e2d\u0e31\u0e01\u0e29\u0e23\u0e22\u0e48\u0e2d : \u0e2a.\u0e01. / S.K.) \u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e0a\u0e32\u0e22\u0e25\u0e49\u0e27\u0e19 \u0e23\u0e30\u0e14\u0e31\u0e1a\u0e0a\u0e31\u0e49\u0e19\u0e21\u0e31\u0e18\u0e22\u0e21\u0e28\u0e36\u0e01\u0e29\u0e32\u0e02\u0e19\u0e32\u0e14\u0e43\u0e2b\u0e0d\u0e48\u0e1e\u0e34\u0e40\u0e28\u0e29 \u0e2a\u0e31\u0e07\u0e01\u0e31\u0e14\u0e2a\u0e33\u0e19\u0e31\u0e01\u0e07\u0e32\u0e19\u0e40\u0e02\u0e15\u0e1e\u0e37\u0e49\u0e19\u0e17\u0e35\u0e48\u0e01\u0e32\u0e23\u0e28\u0e36\u0e01\u0e29\u0e32\u0e21\u0e31\u0e18\u0e22\u0e21\u0e28\u0e36\u0e01\u0e29\u0e32\u0e40\u0e02\u0e15 1 \u0e2a\u0e33\u0e19\u0e31\u0e01\u0e07\u0e32\u0e19\u0e04\u0e13\u0e30\u0e01\u0e23\u0e23\u0e21\u0e01\u0e32\u0e23\u0e01\u0e32\u0e23\u0e28\u0e36\u0e01\u0e29\u0e32\u0e02\u0e31\u0e49\u0e19\u0e1e\u0e37\u0e49\u0e19\u0e10\u0e32\u0e19 (\u0e0a\u0e37\u0e48\u0e2d\u0e40\u0e14\u0e34\u0e21: \u0e01\u0e23\u0e21\u0e2a\u0e32\u0e21\u0e31\u0e0d\u0e28\u0e36\u0e01\u0e29\u0e32) \u0e01\u0e23\u0e30\u0e17\u0e23\u0e27\u0e07\u0e28\u0e36\u0e01\u0e29\u0e32\u0e18\u0e34\u0e01\u0e32\u0e23 \u0e01\u0e48\u0e2d\u0e15\u0e31\u0e49\u0e07\u0e42\u0e14\u0e22 \u0e1e\u0e23\u0e30\u0e1a\u0e32\u0e17\u0e2a\u0e21\u0e40\u0e14\u0e47\u0e08\u0e1e\u0e23\u0e30\u0e08\u0e38\u0e25\u0e08\u0e2d\u0e21\u0e40\u0e01\u0e25\u0e49\u0e32\u0e40\u0e08\u0e49\u0e32\u0e2d\u0e22\u0e39\u0e48\u0e2b\u0e31\u0e27 \u0e44\u0e14\u0e49\u0e23\u0e31\u0e1a\u0e01\u0e32\u0e23\u0e2a\u0e16\u0e32\u0e1b\u0e19\u0e32\u0e02\u0e36\u0e49\u0e19\u0e43\u0e19\u0e27\u0e31\u0e19\u0e17\u0e35\u0e48 8 \u0e21\u0e35\u0e19\u0e32\u0e04\u0e21 \u0e1e.\u0e28. 2424 (\u0e02\u0e13\u0e30\u0e19\u0e31\u0e49\u0e19\u0e19\u0e31\u0e1a\u0e27\u0e31\u0e19\u0e17\u0e35\u0e48 1 \u0e40\u0e21\u0e29\u0e32\u0e22\u0e19 \u0e40\u0e1b\u0e47\u0e19\u0e27\u0e31\u0e19\u0e02\u0e36\u0e49\u0e19\u0e1b\u0e35\u0e43\u0e2b\u0e21\u0e48 \u0e40\u0e21\u0e37\u0e48\u0e2d\u0e19\u0e31\u0e1a\u0e2d\u0e22\u0e48\u0e32\u0e07\u0e2a\u0e32\u0e01\u0e25\u0e16\u0e37\u0e2d\u0e40\u0e1b\u0e47\u0e19 \u0e1e.\u0e28. 2425) \u0e42\u0e14\u0e22\u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e23\u0e31\u0e10\u0e1a\u0e32\u0e25\u0e41\u0e2b\u0e48\u0e07\u0e41\u0e23\u0e01\u0e02\u0e2d\u0e07\u0e1b\u0e23\u0e30\u0e40\u0e17\u0e28\u0e44\u0e17\u0e22"}]}
question-answering
cstorm125/wangchanberta-base-wiki-20210520-news-spm_span-mask-finetune-qa
[ "transformers", "pytorch", "camembert", "question-answering", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us
# wangchanberta-base-wiki-20210520-news-spm_span-mask-finetune-qa Finetuning 'airesearch/wangchanberta-base-wiki-20210520-news-spm_span-mask' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'. Trained with thai2transformers. Run with:
[ "# wangchanberta-base-wiki-20210520-news-spm_span-mask-finetune-qa\n\nFinetuning 'airesearch/wangchanberta-base-wiki-20210520-news-spm_span-mask' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:" ]
[ "TAGS\n#transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us \n", "# wangchanberta-base-wiki-20210520-news-spm_span-mask-finetune-qa\n\nFinetuning 'airesearch/wangchanberta-base-wiki-20210520-news-spm_span-mask' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:" ]
[ 31, 168 ]
[ "passage: TAGS\n#transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us \n# wangchanberta-base-wiki-20210520-news-spm_span-mask-finetune-qa\n\nFinetuning 'airesearch/wangchanberta-base-wiki-20210520-news-spm_span-mask' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:" ]
[ -0.05645165219902992, -0.15578137338161469, -0.003975315950810909, 0.06371725350618362, 0.028402362018823624, -0.00334802339784801, 0.10726091265678406, 0.09047532081604004, -0.009096259251236916, 0.02531369775533676, 0.06127292662858963, -0.0026147738099098206, 0.07015769183635712, 0.09560193121433258, -0.048190489411354065, -0.24415038526058197, 0.08125502616167068, 0.0861651599407196, -0.05217559263110161, 0.1618712991476059, 0.11556685715913773, -0.07393321394920349, 0.06784527748823166, 0.0834026113152504, -0.10574711859226227, -0.01679716818034649, 0.03422338888049126, -0.10367725044488907, 0.1333068460226059, -0.004372229799628258, 0.17342953383922577, 0.0700506865978241, -0.03057202883064747, -0.17775164544582367, 0.05375754460692406, -0.004029496107250452, 0.04726012051105499, 0.015807438641786575, -0.006195432506501675, 0.06686002761125565, -0.030359333381056786, -0.024774329736828804, 0.02383538708090782, 0.015091247856616974, -0.09852325916290283, -0.066645547747612, -0.04286876320838928, 0.025934411212801933, 0.11792894452810287, 0.07881458103656769, -0.03761140629649162, 0.18603329360485077, -0.19750891625881195, 0.05044373497366905, 0.13119077682495117, -0.4157165288925171, -0.02768372744321823, 0.02105094864964485, 0.033281054347753525, 0.028933078050613403, -0.04602588713169098, 0.01290867943316698, 0.03625987470149994, 0.02726447768509388, -0.026248546317219734, -0.06289022415876389, -0.06225086376070976, -0.048625219613313675, -0.1089450940489769, 0.02139902673661709, 0.11461251229047775, 0.03860693424940109, -0.06399373710155487, -0.10389754921197891, -0.06502378731966019, 0.036381665617227554, -0.009629866108298302, -0.07338462024927139, -0.0364442877471447, 0.020629502832889557, -0.02692311443388462, 0.008887811563909054, -0.10264462232589722, -0.06737029552459717, -0.1558917611837387, 0.11004480719566345, 0.0732748806476593, 0.05009687691926956, -0.13868533074855804, -0.00897529348731041, -0.03203938528895378, -0.14262373745441437, -0.06516434252262115, -0.059015028178691864, -0.012103626504540443, 0.04680560901761055, -0.07336738705635071, -0.08792723715305328, 0.091684490442276, 0.0738968625664711, -0.16564004123210907, 0.053737737238407135, -0.09410662949085236, 0.07177819311618805, -0.055066462606191635, 0.14851242303848267, -0.02236366458237171, 0.10206802189350128, 0.03132730722427368, -0.0076262252405285835, 0.021591030061244965, -0.0034246535506099463, 0.00953826867043972, -0.08524952083826065, -0.004357611760497093, 0.01668895035982132, -0.06778482347726822, 0.09788382798433304, -0.05657027289271355, -0.020758964121341705, 0.08207695186138153, -0.12052488327026367, -0.08154972642660141, 0.004460074007511139, -0.004799318965524435, 0.03615960851311684, 0.024416524916887283, 0.08840680867433548, -0.019984766840934753, 0.10322345048189163, -0.021340323612093925, 0.007727453950792551, 0.03279855102300644, -0.052327536046504974, 0.016964823007583618, -0.09368434548377991, -0.02776949480175972, -0.13772724568843842, -0.09753260761499405, 0.02911374159157276, -0.015969393774867058, 0.04605834186077118, -0.01214260421693325, -0.014105312526226044, 0.0061781806871294975, -0.06043658405542374, -0.0018248173873871565, 0.021213388070464134, -0.03478660807013512, 0.10106168687343597, 0.06477238237857819, 0.14710521697998047, -0.08211812376976013, 0.02419007197022438, -0.15100987255573273, 0.06369543820619583, -0.13181306421756744, 0.07371138781309128, -0.04689563810825348, -0.017399175092577934, -0.05628116801381111, -0.04840604588389397, -0.11448999494314194, 0.02137947641313076, 0.019602999091148376, 0.10521119832992554, -0.11747425049543381, -0.0731862410902977, 0.2813548147678375, -0.13584426045417786, -0.1443224549293518, 0.1942521631717682, -0.05075673758983612, 0.06908208876848221, 0.06882669776678085, 0.23642200231552124, 0.032922569662332535, -0.1572342813014984, 0.05905246362090111, -0.024904360994696617, -0.03180669620633125, 0.0466734804213047, 0.0809616893529892, 0.010784859769046307, -0.03827410936355591, 0.07427258789539337, -0.10888773202896118, 0.05550534278154373, -0.05131752789020538, -0.0932859405875206, -0.052332282066345215, -0.09054175764322281, 0.1292944997549057, -0.041898056864738464, 0.12166837602853775, -0.05191865563392639, -0.05994275584816933, -0.06824152171611786, 0.03887249901890755, -0.005055443849414587, 0.03493788093328476, -0.17229850590229034, 0.07678613066673279, -0.03695080801844597, -0.009906494058668613, -0.1222047507762909, 0.010653662495315075, 0.03956134244799614, 0.019036397337913513, 0.04009195789694786, 0.03871827945113182, 0.004358310718089342, -0.03245937451720238, -0.06898299604654312, 0.03870321437716484, -0.01142857689410448, -0.02049609273672104, -0.042482778429985046, -0.016320331022143364, 0.06578545272350311, -0.030722135677933693, 0.10358788073062897, -0.054744310677051544, 0.02723276987671852, 0.013491441495716572, 0.09330595284700394, 0.007189548574388027, 0.04844487085938454, 0.054848477244377136, 0.07434631884098053, 0.03050103969871998, 0.030394168570637703, 0.048753827810287476, 0.015145153738558292, -0.17928318679332733, 0.10594724118709564, -0.05684894695878029, 0.12418027967214584, 0.12928356230258942, -0.13985568284988403, 0.025113634765148163, 0.013372362591326237, -0.06022104248404503, -0.02567988820374012, -0.11626686900854111, 0.07160074263811111, 0.15131038427352905, -0.002860398031771183, 0.12099212408065796, -0.11867613345384598, 0.0030787510331720114, -0.007004202343523502, -0.02378193289041519, -0.004700901452451944, 0.03332814946770668, 0.012058869004249573, -0.1314307153224945, 0.04624590277671814, 0.20292189717292786, -0.038550712168216705, 0.13664980232715607, -0.06490408629179001, -0.054673902690410614, 0.007971933111548424, 0.09531109780073166, -0.044260118156671524, 0.017422642558813095, -0.10744903981685638, 0.031825121492147446, 0.05897724628448486, 0.04162351414561272, 0.04420775547623634, -0.14425188302993774, -0.06665658950805664, 0.0011216654675081372, -0.0744219645857811, -0.15799081325531006, 0.09589588642120361, 0.051240473985672, 0.07317370921373367, 0.0456082709133625, 0.0340593196451664, 0.03998574614524841, 0.0068251946941018105, -0.12711025774478912, 0.15868569910526276, -0.040327273309230804, -0.3563890755176544, -0.046893712133169174, 0.061563100665807724, -0.045807842165231705, -0.04441053047776222, 0.04723352566361427, -0.16923876106739044, 0.023387938737869263, -0.0022869673557579517, 0.025988757610321045, 0.01587662659585476, 0.04229474812746048, 0.060286957770586014, 0.01624622941017151, 0.059984732419252396, -0.07098153978586197, -0.03893434256315231, -0.0876036286354065, -0.0726843774318695, 0.10448531061410904, -0.12638235092163086, 0.0618516206741333, 0.035028740763664246, -0.059566840529441833, 0.025313934311270714, -0.033398669213056564, 0.23763255774974823, -0.0788927972316742, 0.004982548765838146, 0.1569685935974121, -0.1251358538866043, 0.04411948844790459, 0.1573708951473236, -0.01271434873342514, -0.0709698274731636, 0.06387582421302795, 0.04476165398955345, -0.04996141418814659, -0.2243911623954773, -0.02413898892700672, -0.03477882966399193, -0.011948734521865845, -0.07173380255699158, 0.02517900988459587, 0.027421507984399796, 0.09205605834722519, 0.0001933203311637044, -0.07960685342550278, -0.032453037798404694, 0.019291533157229424, 0.1323135495185852, 0.0453004315495491, 0.13742691278457642, -0.03815677389502525, -0.05465081334114075, 0.02154124341905117, 0.06903959065675735, 0.1262403279542923, 0.030950428918004036, -0.02074783481657505, 0.1303718090057373, 0.2521170675754547, 0.1931368112564087, 0.04449344053864479, -0.06666159629821777, -0.045453865081071854, -0.002646553795784712, -0.03264681249856949, -0.07596889138221741, 0.025024618953466415, -0.009739906527101994, 0.005826374515891075, -0.02140367403626442, 0.0904291495680809, 0.0687747374176979, 0.23931416869163513, 0.04921963810920715, -0.10510202497243881, -0.10872305184602737, 0.010597066022455692, -0.07664304971694946, 0.013129840604960918, 0.06450960040092468, 0.053273558616638184, -0.1424940824508667, 0.09238754212856293, -0.006918950472027063, 0.12409234791994095, -0.018174726516008377, 0.0694262683391571, -0.08389951288700104, -0.14666932821273804, 0.013733753003180027, 0.03447822108864784, -0.33849450945854187, 0.25234460830688477, 0.00713298050686717, 0.034632839262485504, -0.050640009343624115, -0.03494016081094742, 0.0028525409288704395, 0.036094311624765396, 0.10685926675796509, -0.01118082832545042, -0.013135116547346115, -0.09911953657865524, -0.03177577629685402, 0.10172165185213089, 0.08495207130908966, 0.0704447478055954, 0.028829388320446014, 0.02387024648487568, 0.03336026892066002, 0.02035253494977951, 0.08619460463523865, -0.21086567640304565, -0.014499422162771225, 0.03288273140788078, 0.033718809485435486, -0.010357825085520744, -0.06234145909547806, -0.058893557637929916, -0.15615105628967285, 0.14148657023906708, -0.11031268537044525, -0.11262644827365875, -0.07787282764911652, 0.0885225236415863, 0.03802981972694397, -0.09562625735998154, 0.00989309512078762, -0.03763686865568161, 0.04117213562130928, -0.04490004852414131, -0.052452750504016876, 0.0821690782904625, -0.02067621424794197, -0.1188993901014328, 0.030677499249577522, 0.09560912847518921, 0.016302907839417458, 0.07696668803691864, 0.08186209946870804, -0.003333958564326167, 0.03153949975967407, -0.08143407851457596, -0.02790510654449463, -0.04256097972393036, -0.06286454200744629, 0.011274293065071106, -0.09618093073368073, -0.026925988495349884, -0.12939485907554626, -0.058172211050987244, 0.19188842177391052, 0.1516365110874176, -0.06966100633144379, 0.11370330303907394, 0.1601092368364334, 0.012129274196922779, -0.16243842244148254, -0.017171340063214302, 0.015334893018007278, 0.05995607003569603, -0.030482754111289978, -0.050392504781484604, 0.12976886332035065, 0.08510090410709381, -0.029459398239850998, -0.04841260612010956, -0.14722906053066254, -0.1233307421207428, 0.09791356325149536, 0.00761268800124526, 0.26061221957206726, -0.11959756910800934, -0.05337437987327576, 0.0010068112751469016, -0.21429729461669922, 0.04421420767903328, -0.08569993823766708, 0.08155646175146103, -0.03101450577378273, 0.05889633670449257, -0.00683694938197732, -0.020256081596016884, 0.12315419316291809, 0.012659359723329544, -0.007534236181527376, -0.05974123254418373, -0.00598837761208415, 0.07186516374349594, 0.03798792138695717, 0.07320533692836761, 0.003622655989602208, 0.042057473212480545, -0.14047503471374512, -0.03874216228723526, -0.09903033077716827, -0.022127274423837662, -0.03819279372692108, -0.05495528504252434, -0.038980767130851746, 0.04265791177749634, 0.04265185818076134, 0.03039495274424553, 0.0528106726706028, -0.08112549781799316, 0.10376928001642227, -0.0020388534758239985, 0.09085554629564285, -0.09089762717485428, 0.03489893302321434, -0.04310543090105057, -0.03939225152134895, 0.10782921314239502, -0.1849430650472641, 0.04541826993227005, 0.10224716365337372, 0.0053063202649354935, 0.07715350389480591, 0.05674015358090401, -0.02221371978521347, 0.10788600891828537, 0.06146606057882309, -0.09638608992099762, -0.11751127988100052, 0.04755084216594696, -0.13107463717460632, -0.02532198093831539, 0.000405985425459221, 0.10311058163642883, -0.013571041636168957, -0.03571176156401634, -0.00037755814264528453, 0.000621594546828419, -0.09030323475599289, 0.08523041009902954, 0.07044848799705505, 0.03537596017122269, -0.0855194553732872, 0.07624754309654236, 0.04448126628994942, -0.11791151016950607, -0.01402331329882145, 0.044152386486530304, -0.17679978907108307, -0.040157996118068695, -0.07682318985462189, 0.025651317089796066, 0.004485803190618753, -0.07336962968111038, -0.10682182013988495, -0.15395870804786682, 0.03279250115156174, 0.15345053374767303, 0.06887253373861313, 0.009461878798902035, 0.019313979893922806, 0.0020494575146585703, 0.008692781440913677, 0.02367996796965599, 0.07320357859134674, 0.05315642058849335, -0.07568085193634033, 0.029136957600712776, 0.03456750512123108, 0.1862347275018692, -0.05804112181067467, -0.020328717306256294, -0.09454434365034103, 0.0799933448433876, -0.20637862384319305, 0.06910024583339691, -0.030562210828065872, -0.04060971364378929, -0.045557886362075806, -0.15620961785316467, -0.054748643189668655, 0.025947153568267822, -0.038972653448581696, 0.028549320995807648, -0.034839265048503876, -0.02635052800178528, -0.0800986960530281, -0.014757287688553333, 0.12921614944934845, -0.05508222430944443, 0.09601178765296936, 0.058791376650333405, -0.09166315197944641, 0.09740674495697021, -0.13584481179714203, -0.0753355473279953, 0.055561210960149765, 0.021988825872540474, 0.018057290464639664, 0.009738794527947903, 0.04446424916386604, 0.04967714846134186, 0.05413934960961342, 0.06503071635961533, 0.2236274778842926, -0.1047842726111412, -0.029805220663547516, -0.04008534923195839, -0.05438491702079773, -0.04365277290344238, -0.06317665427923203, 0.14987410604953766, 0.0645315945148468, 0.11730875074863434, -0.08226016908884048, 0.06199562922120094, -0.11232834309339523, -0.007497240789234638, 0.0049714744091033936, -0.08312064409255981, 0.006797037087380886, -0.05524333566427231, 0.03332368656992912, -0.0449766144156456, 0.17435061931610107, -0.10551266372203827, 0.008485675789415836, 0.04954545199871063, 0.005280391313135624, -0.028978366404771805, 0.013038692064583302, 0.2741159200668335, 0.09899080544710159, -0.020456206053495407, 0.0067034270614385605, -0.05657985433936119, -0.03087596781551838, 0.018583107739686966, 0.10116090625524521, 0.2798546552658081, 0.012569633312523365, 0.07931564003229141, 0.07070605456829071, 0.14550256729125977, -0.05268070474267006, -0.06165361776947975, -0.11771334707736969, -0.017518222332000732, 0.0337199866771698, -0.03332908824086189, 0.2294331192970276, -0.08726039528846741, 0.04942779988050461, -0.03250725567340851, -0.07401353865861893, -0.14127188920974731, -0.020210707560181618, -0.13061438500881195, -0.05716592073440552, 0.06348954141139984, -0.06945988535881042, -0.06099657714366913, 0.15722978115081787, 0.06486516445875168, -0.018841980025172234, 0.23413832485675812, 0.03509972244501114, -0.05381190404295921, 0.09144502133131027, -0.04897899925708771, 0.007790836971253157, 0.06953942030668259, 0.003064451040700078, 0.04981398954987526, 0.06989816576242447, 0.014387210831046104, -0.00792248360812664, -0.04984894022345543, -0.027987414970993996, -0.12140931934118271, -0.059778448194265366, -0.016094928607344627, 0.01943817548453808, 0.027810780331492424, 0.06886626034975052, 0.04837511479854584, 0.03828977420926094, -0.02126351185142994, 0.10680917650461197, 0.015447737649083138, -0.18841877579689026, -0.188200443983078, 0.029639899730682373, 0.0005057085654698312, 0.025602994486689568, 0.023490356281399727, -0.04816146939992905, -0.029371798038482666, 0.34897422790527344, 0.2282310277223587, -0.06503435969352722, 0.04994513466954231, 0.05080592259764671, 0.03726101666688919, -0.0010308146011084318, 0.06878799945116043, 0.10701823979616165, 0.1320493370294571, -0.0748148187994957, -0.07790536433458328, -0.06437265127897263, -0.09954023361206055, -0.1299128234386444, 0.015754355117678642, 0.05160360410809517, 0.001727293711155653, -0.0874411016702652, 0.08896085619926453, -0.05334225669503212, 0.017634548246860504, -0.046170689165592194, -0.15938560664653778, -0.12217520922422409, -0.035359252244234085, -0.06232749670743942, 0.03437512740492821, 0.0079569760710001, -0.01440426055341959, 0.03553403913974762, 0.004970451816916466, 0.03173965588212013, -0.008316252380609512, -0.003199238795787096, 0.11449633538722992, 0.03580276295542717, 0.010280794464051723, 0.06237341836094856, 0.13421103358268738, 0.0513298325240612, 0.08147658407688141, 0.026739811524748802, 0.14336702227592468, 0.05537400022149086, 0.06243152171373367, -0.07312624156475067, 0.07120147347450256, 0.02347569167613983, -0.006148068234324455, 0.1134309321641922, -0.14451684057712555, 0.06678295880556107, -0.04926222562789917, -0.041110560297966, -0.12792548537254333, 0.11433670669794083, -0.05923744663596153, 0.08394601941108704, 0.16661304235458374, -0.03404637426137924, 0.031223097816109657, -0.040925946086645126, 0.02898281067609787, 0.008738464675843716, -0.06407328695058823, -0.085347980260849, -0.17400729656219482, 0.01590942218899727, 0.020673677325248718, -0.008004342205822468, -0.2563948631286621, -0.056728608906269073, -0.04525475949048996, 0.0005499205435626209, -0.014929153956472874, 0.05711062252521515, 0.09244736284017563, 0.02335730753839016, 0.0036072642542421818, -0.2897733449935913, 0.045558493584394455, 0.06405538320541382, -0.13738404214382172, -0.0778832659125328 ]
null
null
k2
# Introduction This repo contains pre-trained model using <https://github.com/k2-fsa/icefall/pull/219>. It is trained on [AIShell](https://www.openslr.org/33/) dataset using modified transducer from [optimized_transducer](https://github.com/csukuangfj/optimized_transducer). Also, it uses [aidatatang_200zh](http://www.openslr.org/62/) as extra training data. ## How to clone this repo ``` sudo apt-get install git-lfs git clone https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2-2022-03-01 cd icefall-aishell-transducer-stateless-modified-2-2022-03-01 git lfs pull ``` **Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later. The model in this repo is trained using the commit `TODO`. You can use ``` git clone https://github.com/k2-fsa/icefall cd icefall git checkout TODO ``` to download `icefall`. You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/TODO/egs/aishell/ASR/transducer_stateless_modified-2/train.py#L232>. In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 512-dim embedding layer and a Conv1d with kernel size 2. The decoder architecture is modified from [Rnn-Transducer with Stateless Prediction Network](https://ieeexplore.ieee.org/document/9054419). A Conv1d layer is placed right after the input embedding layer. ----- ## Description This repo provides pre-trained transducer Conformer model for the AIShell dataset using [icefall][icefall]. There are no RNNs in the decoder. The decoder is stateless and contains only an embedding layer and a Conv1d. The commands for training are: ```bash cd egs/aishell/ASR ./prepare.sh --stop-stage 6 ./prepare_aidatatang_200zh.sh export CUDA_VISIBLE_DEVICES="0,1,2" ./transducer_stateless_modified-2/train.py \ --world-size 3 \ --num-epochs 90 \ --start-epoch 0 \ --exp-dir transducer_stateless_modified-2/exp-2 \ --max-duration 250 \ --lr-factor 2.0 \ --context-size 2 \ --modified-transducer-prob 0.25 \ --datatang-prob 0.2 ``` The tensorboard training log can be found at <https://tensorboard.dev/experiment/oG72ZlWaSGua6fXkcGRRjA/> The commands for decoding are ```bash # greedy search for epoch in 89; do for avg in 38; do ./transducer_stateless_modified-2/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer_stateless_modified-2/exp-2 \ --max-duration 100 \ --context-size 2 \ --decoding-method greedy_search \ --max-sym-per-frame 1 done done # modified beam search for epoch in 89; do for avg in 38; do ./transducer_stateless_modified-2/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer_stateless_modified-2/exp-2 \ --max-duration 100 \ --context-size 2 \ --decoding-method modified_beam_search \ --beam-size 4 done done ``` You can find the decoding log for the above command in this repo (in the folder [log][log]). The WER for the test dataset is | | test |comment | |------------------------|------|----------------------------------------------------------------| | greedy search | 4.94 |--epoch 89, --avg 38, --max-duration 100, --max-sym-per-frame 1 | | modified beam search | 4.68 |--epoch 89, --avg 38, --max-duration 100 --beam-size 4 | # File description - [log][log], this directory contains the decoding log and decoding results - [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model - [data][data], this directory contains files generated by [prepare.sh][prepare] - [exp][exp], this directory contains only one file: `preprained.pt` `exp/pretrained.pt` is generated by the following command: ```bash epoch=89 avg=38 ./transducer_stateless_modified-2/export.py \ --exp-dir ./transducer_stateless_modified-2/exp-2 \ --lang-dir ./data/lang_char \ --epoch $epoch \ --avg $avg ``` **HINT**: To use `pretrained.pt` to compute the WER for the `test` dataset, just do the following: ```bash cp icefall-aishell-transducer-stateless-modified-2-2022-03-01/exp/pretrained.pt \ /path/to/icefall/egs/aishell/ASR/transducer_stateless_modified-2/exp/epoch-999.pt ``` and pass `--epoch 999 --avg 1` to `transducer_stateless_modified-2/decode.py`. [icefall]: https://github.com/k2-fsa/icefall [prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/aishell/ASR/prepare.sh [exp]: https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2-2022-03-01/tree/main/exp [data]: https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2-2022-03-01/tree/main/data [test_wavs]: https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2-2022-03-01/tree/main/test_wavs [log]: https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2-2022-03-01/tree/main/log [icefall]: https://github.com/k2-fsa/icefall
{"language": "en", "license": "apache-2.0", "tags": ["icefall", "k2", "transducer", "aishell", "ASR", "stateless transducer", "PyTorch"], "datasets": ["aishell", "aidatatang_200zh"], "metrics": ["WER"]}
null
csukuangfj/icefall-aishell-transducer-stateless-modified-2-2022-03-01
[ "k2", "icefall", "transducer", "aishell", "ASR", "stateless transducer", "PyTorch", "en", "dataset:aishell", "dataset:aidatatang_200zh", "license:apache-2.0", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #k2 #icefall #transducer #aishell #ASR #stateless transducer #PyTorch #en #dataset-aishell #dataset-aidatatang_200zh #license-apache-2.0 #region-us
Introduction ============ This repo contains pre-trained model using <URL It is trained on AIShell dataset using modified transducer from optimized\_transducer. Also, it uses aidatatang\_200zh as extra training data. How to clone this repo ---------------------- Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later. The model in this repo is trained using the commit 'TODO'. You can use to download 'icefall'. You can find the model information by visiting <URL In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 512-dim embedding layer and a Conv1d with kernel size 2. The decoder architecture is modified from Rnn-Transducer with Stateless Prediction Network. A Conv1d layer is placed right after the input embedding layer. --- Description ----------- This repo provides pre-trained transducer Conformer model for the AIShell dataset using [icefall](URL). There are no RNNs in the decoder. The decoder is stateless and contains only an embedding layer and a Conv1d. The commands for training are: The tensorboard training log can be found at <URL The commands for decoding are You can find the decoding log for the above command in this repo (in the folder [log](URL)). The WER for the test dataset is test: greedy search, comment: 4.94 test: modified beam search, comment: 4.68 File description ================ * [log](URL), this directory contains the decoding log and decoding results * [test\_wavs](URL), this directory contains wave files for testing the pre-trained model * [data](URL), this directory contains files generated by <URL> * [exp](URL), this directory contains only one file: 'URL' 'exp/URL' is generated by the following command: HINT: To use 'URL' to compute the WER for the 'test' dataset, just do the following: and pass '--epoch 999 --avg 1' to 'transducer\_stateless\_modified-2/URL'.
[]
[ "TAGS\n#k2 #icefall #transducer #aishell #ASR #stateless transducer #PyTorch #en #dataset-aishell #dataset-aidatatang_200zh #license-apache-2.0 #region-us \n" ]
[ 59 ]
[ "passage: TAGS\n#k2 #icefall #transducer #aishell #ASR #stateless transducer #PyTorch #en #dataset-aishell #dataset-aidatatang_200zh #license-apache-2.0 #region-us \n" ]
[ -0.1148015484213829, 0.10091722756624222, -0.005711867939680815, 0.07724861800670624, 0.011987430043518543, -0.029239220544695854, 0.20202097296714783, 0.05115976929664612, 0.18732497096061707, -0.08075223863124847, 0.19758202135562897, 0.11648067086935043, 0.03819235414266586, 0.15277323126792908, 0.013063925318419933, -0.23367811739444733, 0.029090605676174164, -0.05850335583090782, -0.06745374202728271, 0.08112197369337082, 0.07438448816537857, -0.02390199527144432, 0.022162649780511856, 0.0009555229917168617, 0.05282209813594818, -0.000254948710789904, 0.02319534309208393, -0.0818755030632019, 0.106153205037117, -0.051216769963502884, 0.029070869088172913, 0.05592772364616394, 0.026166610419750214, -0.10673113912343979, 0.023400070145726204, -0.0324745699763298, -0.050855886191129684, 0.06551963835954666, 0.018584974110126495, 0.025359569117426872, -0.034253790974617004, -0.07581587135791779, -0.05255710333585739, 0.03154519572854042, -0.053218912333250046, -0.1781885325908661, -0.08833780884742737, 0.0696774423122406, 0.029151100665330887, 0.06046759709715843, 0.044264234602451324, 0.1484770029783249, -0.10634864866733551, 0.08207681775093079, 0.09348854422569275, -0.38248297572135925, 0.03998624160885811, -0.0035798142198473215, -0.08166542649269104, 0.07526382803916931, 0.024904796853661537, 0.018727976828813553, 0.005931975319981575, -0.02555565908551216, 0.017922701314091682, -0.004578035324811935, -0.13343076407909393, 0.06466837227344513, -0.09468796104192734, -0.020967308431863785, 0.24798037111759186, 0.001363372546620667, 0.03105809912085533, 0.07612738758325577, -0.05686473473906517, -0.06683357805013657, 0.06773005425930023, 0.06384042650461197, 0.007082329131662846, 0.08480021357536316, 0.024228207767009735, -0.0003360282862558961, -0.1182275041937828, -0.01695646531879902, -0.16147975623607635, 0.11133717000484467, -0.005225215572863817, 0.12750962376594543, -0.15582223236560822, 0.023647334426641464, 0.027938229963183403, -0.1396714746952057, 0.04334425553679466, -0.06869920343160629, 0.03529244288802147, 0.10652870684862137, -0.017306316643953323, 0.023390330374240875, 0.12033690512180328, 0.17717312276363373, 0.07021744549274445, -0.0026351886335760355, 0.03152914345264435, 0.1294643133878708, 0.07635761797428131, 0.031011853367090225, 0.027817318215966225, -0.025541765615344048, 0.033667512238025665, -0.016377722844481468, 0.10481766611337662, -0.06630479544401169, -0.048740487545728683, -0.027710415422916412, 0.007423692848533392, 0.08481381088495255, 0.10501531511545181, -0.06651268899440765, -0.09562108665704727, 0.014560188166797161, 0.15716271102428436, -0.06698399037122726, -0.0008345293463207781, 0.026797259226441383, -0.010839767754077911, 0.06782073527574539, -0.013921457342803478, 0.07382947951555252, 0.03167569637298584, 0.04981767758727074, -0.07938603311777115, -0.014168706722557545, 0.030012192204594612, 0.03652556613087654, 0.08975958079099655, -0.14743782579898834, 0.04161372035741806, -0.13468100130558014, -0.18266251683235168, 0.0567227303981781, 0.04342533275485039, 0.0025394605472683907, -0.06508456915616989, 0.04062739387154579, -0.01978158950805664, 0.03579029440879822, -0.07235369831323624, -0.10035927593708038, -0.05030660703778267, 0.008216163143515587, -0.12174719572067261, 0.06130810081958771, -0.13179577887058258, -0.007826377637684345, -0.16415561735630035, 0.00982065312564373, 0.07549134641885757, -0.012398339807987213, -0.1588166356086731, 0.09383723139762878, -0.0494166761636734, -0.0022445477079600096, -0.05397187918424606, -0.06588722765445709, 0.05580703541636467, 0.1483178287744522, -0.2645792067050934, -0.06368101388216019, 0.07377907633781433, -0.13664370775222778, -0.09225593507289886, 0.09673742204904556, 0.05494162440299988, -0.06880464404821396, 0.038225430995225906, 0.18308226764202118, -0.04957277327775955, 0.00995603110641241, -0.09152165055274963, 0.13488750159740448, -0.14360177516937256, -0.17740345001220703, 0.12049251794815063, -0.028780797496438026, 0.027948463335633278, 0.023580100387334824, 0.0673627257347107, 0.11822225898504257, -0.023646125569939613, -0.15073691308498383, -0.057993389666080475, -0.07341615110635757, 0.052298784255981445, -0.012518909759819508, 0.017202099785208702, -0.012901557609438896, 0.005458363331854343, -0.06565352529287338, 0.03329962491989136, 0.09046793729066849, 0.047037381678819656, -0.051521461457014084, -0.0030687961261719465, 0.01338664535433054, -0.0037942028138786554, -0.06672512739896774, -0.066722072660923, -0.017620760947465897, -0.04774579405784607, -0.017878739163279533, 0.06663636863231659, 0.07280325144529343, -0.029358793050050735, 0.012992985546588898, -0.00006046418275218457, 0.027517156675457954, 0.06472375243902206, -0.04667041078209877, -0.20747056603431702, 0.030041834339499474, -0.03561898693442345, 0.04022201523184776, 0.021193675696849823, 0.02493033930659294, 0.13669008016586304, 0.05899077653884888, -0.014099531807005405, 0.051306117326021194, -0.03099757991731167, -0.021341243758797646, 0.018016474321484566, 0.0062652211636304855, 0.09037521481513977, 0.015094535425305367, -0.16625846922397614, 0.1532743275165558, 0.010977949015796185, 0.08053315430879593, 0.11544673889875412, 0.002508155070245266, 0.13293395936489105, 0.013532248325645924, -0.06271301209926605, -0.08457291126251221, 0.07565964013338089, 0.059033505618572235, 0.027567381039261818, 0.058439452201128006, 0.02992931753396988, 0.032459720969200134, -0.008283737115561962, -0.008900943212211132, -0.07277532666921616, -0.04095272719860077, 0.06842021644115448, 0.031034037470817566, -0.18014082312583923, 0.10507648438215256, 0.3310094177722931, -0.009461190551519394, 0.14952638745307922, -0.09993211925029755, -0.06344986706972122, -0.04271498695015907, -0.09525483846664429, -0.044503483921289444, 0.16427797079086304, 0.023340653628110886, 0.039982568472623825, 0.08674288541078568, 0.009699800983071327, -0.0018702357774600387, -0.095360167324543, -0.06608012318611145, -0.021184531971812248, 0.03721543028950691, -0.20433136820793152, 0.06325513124465942, -0.09995322674512863, 0.03270009532570839, -0.021619463339447975, -0.119318388402462, 0.1257971227169037, -0.011602917686104774, -0.026071697473526, 0.12296675145626068, -0.1650463044643402, -0.13475733995437622, -0.06287362426519394, 0.12879055738449097, -0.020791806280612946, -0.08198481053113937, 0.056459713727235794, -0.12323685735464096, -0.02755378745496273, 0.009753264486789703, -0.06083650887012482, 0.029620200395584106, 0.02392806112766266, 0.08798269182443619, -0.02565748430788517, -0.01103246957063675, -0.10330172628164291, -0.03757186233997345, -0.028709353879094124, -0.002745653036981821, 0.1473575234413147, 0.02740565501153469, 0.02039276249706745, 0.08389554172754288, 0.07063369452953339, 0.00495782308280468, 0.011808724142611027, 0.09281165152788162, -0.02293284423649311, 0.03077532723546028, 0.13363592326641083, -0.03932245820760727, 0.08806796371936798, 0.07557173073291779, 0.0631294772028923, 0.010427074506878853, -0.025052839890122414, -0.02001882530748844, -0.10672782361507416, -0.22966593503952026, -0.07061851769685745, -0.05051672086119652, 0.18572653830051422, 0.00008583541784901172, 0.06081751361489296, 0.009941480122506618, 0.09022683650255203, 0.07332286983728409, -0.08076893538236618, -0.06572859734296799, -0.009571438655257225, 0.08542108535766602, -0.0028580473735928535, 0.08933636546134949, -0.067864328622818, 0.036803800612688065, 0.13075418770313263, 0.10097254067659378, 0.17850039899349213, 0.07061534374952316, 0.054705314338207245, 0.0720491111278534, 0.23296688497066498, 0.044114693999290466, 0.11057999730110168, 0.06678697466850281, -0.013932537287473679, -0.0076210591942071915, -0.02170269377529621, -0.057605963200330734, 0.05974460020661354, -0.05721434950828552, -0.13072127103805542, 0.09600239992141724, 0.13247254490852356, 0.082707479596138, 0.2932901680469513, 0.0664379894733429, -0.08956611156463623, -0.03453537076711655, 0.04506199061870575, 0.01009838841855526, 0.034305404871702194, 0.09125459939241409, -0.022582322359085083, -0.06447640061378479, 0.07140086591243744, -0.06434786319732666, 0.02617541514337063, 0.009931505657732487, 0.03295358642935753, -0.0072774216532707214, -0.08536293357610703, 0.05464601516723633, 0.07279948145151138, -0.21003960072994232, 0.22071564197540283, -0.012300503440201283, -0.02240607887506485, -0.06691130995750427, -0.03293239325284958, 0.07363315671682358, 0.08371198922395706, 0.0707201361656189, 0.09919799119234085, -0.0989479348063469, -0.0015371923800557852, -0.20447106659412384, 0.08307737112045288, -0.0029684058390557766, 0.012418256141245365, -0.10903327167034149, -0.02186156064271927, 0.04366733878850937, 0.041408970952034, -0.029706353321671486, -0.18491800129413605, -0.04468965902924538, 0.08639267832040787, 0.14922116696834564, 0.03894747048616409, -0.05360452085733414, -0.13112182915210724, -0.0036891307681798935, 0.041440997272729874, -0.12044662982225418, -0.09576204419136047, -0.07608950883150101, -0.05073711648583412, 0.12482964247465134, -0.08441577106714249, 0.02391030266880989, -0.00690739369019866, -0.06632465124130249, -0.022457744926214218, -0.16547898948192596, 0.09471742808818817, -0.1201910600066185, -0.14242209494113922, 0.02904740534722805, 0.03980318084359169, 0.07152464985847473, 0.04241269826889038, -0.05331417918205261, 0.08339832723140717, -0.09783415496349335, -0.0906013697385788, 0.10238237679004669, -0.018963336944580078, -0.013795220293104649, 0.03320249170064926, 0.055919039994478226, -0.05139300227165222, 0.008331102319061756, -0.08144025504589081, 0.19805340468883514, 0.2969817519187927, -0.09824652969837189, 0.2299693524837494, 0.24374297261238098, -0.011523820459842682, -0.23339492082595825, -0.1753263920545578, -0.1291065663099289, -0.05536515265703201, 0.12574587762355804, -0.1965564638376236, 0.11102968454360962, 0.1354658454656601, -0.1131785586476326, 0.09919629991054535, -0.23892073333263397, -0.06148752570152283, 0.19973769783973694, -0.09727280586957932, 0.3090699017047882, -0.10259290784597397, -0.04503165930509567, -0.03563323989510536, -0.09421144425868988, 0.12061432749032974, -0.12152884900569916, 0.029258938506245613, -0.05064014345407486, 0.038011323660612106, -0.022474752739071846, -0.0666390061378479, 0.12864261865615845, 0.034910403192043304, -0.015314378775656223, -0.04856010898947716, -0.049115560948848724, 0.15301190316677094, -0.017435722053050995, 0.04057178646326065, -0.15213564038276672, 0.07323353737592697, -0.04989440739154816, 0.04412394016981125, -0.10939613729715347, 0.08143379539251328, -0.00045677594607695937, -0.08974894136190414, 0.017379555851221085, 0.015520214103162289, 0.0069947452284395695, -0.017514511942863464, 0.21325655281543732, 0.10157203674316406, -0.09806722402572632, 0.0690135657787323, 0.036232803016901016, -0.09059890359640121, 0.007226115558296442, -0.12738728523254395, -0.054685965180397034, 0.0643787607550621, -0.10978461056947708, 0.0510978065431118, 0.11754311621189117, 0.021787559613585472, -0.04200707748532295, 0.0677449181675911, -0.047744929790496826, 0.007186018396168947, 0.10690263658761978, -0.16860921680927277, -0.030130213126540184, 0.024418221786618233, 0.14354895055294037, 0.07778661698102951, 0.16595713794231415, 0.16886548697948456, -0.0014274085406213999, -0.01944117806851864, -0.00520746735855937, 0.11769451946020126, -0.0671636238694191, 0.007200663909316063, 0.10181494802236557, 0.006079449318349361, -0.10545367002487183, 0.16792921721935272, -0.05657822638750076, -0.007519190199673176, 0.05405334383249283, 0.05909876525402069, -0.09745743125677109, -0.10224730521440506, -0.06640386581420898, -0.052639931440353394, -0.1781580001115799, -0.10955797135829926, -0.04319985955953598, -0.042737022042274475, -0.005648382939398289, 0.0017959390534088016, 0.05505835637450218, 0.07234556972980499, -0.010789436288177967, -0.06746252626180649, 0.06552114337682724, 0.048114702105522156, -0.07925665378570557, -0.021262967959046364, -0.129445418715477, -0.0810498297214508, -0.029748426750302315, 0.10815152525901794, -0.0420716293156147, -0.01422110479325056, -0.03993930295109749, 0.012782135978341103, -0.0544874481856823, -0.08381707966327667, -0.09103242307901382, -0.008020336739718914, 0.007274485658854246, -0.08861040323972702, -0.02808341756463051, 0.06212720274925232, -0.09209234267473221, 0.008680742233991623, -0.00958373211324215, 0.03371232748031616, -0.1680988371372223, -0.06492803990840912, 0.08960868418216705, 0.025081193074584007, 0.13699910044670105, 0.1099754124879837, -0.006110748741775751, 0.07390075922012329, -0.024433569982647896, 0.010918797925114632, 0.042463190853595734, 0.03341212123632431, 0.032477930188179016, 0.04447389394044876, -0.08023177087306976, 0.03975401073694229, 0.0018375942017883062, 0.05786708742380142, 0.12794379889965057, -0.07527396827936172, -0.019702745601534843, 0.0350460410118103, -0.10218365490436554, -0.07671384513378143, -0.10002622753381729, 0.07336261868476868, 0.06702952831983566, 0.1783405840396881, -0.009416835382580757, 0.011883482336997986, -0.01511877216398716, 0.026678331196308136, 0.036386363208293915, -0.07881330698728561, -0.043482810258865356, 0.01830647885799408, -0.023587306961417198, -0.062283869832754135, 0.16358621418476105, -0.0335439071059227, -0.1494579017162323, 0.0484217070043087, -0.006114899646490812, -0.06856246292591095, 0.01217444147914648, 0.15546225011348724, 0.01707691140472889, -0.0010288683697581291, -0.09997974336147308, 0.026580221951007843, -0.06798923015594482, -0.08481443673372269, 0.13246016204357147, 0.0024109340738505125, 0.07813962548971176, 0.039905447512865067, 0.06216132268309593, -0.09243014454841614, -0.13563430309295654, -0.12757523357868195, -0.0933513194322586, 0.07063347846269608, 0.06336767226457596, 0.1578318178653717, 0.1248278096318245, 0.018958615139126778, -0.019824815914034843, -0.06728646159172058, -0.04965100809931755, -0.12956605851650238, -0.14456047117710114, 0.00004444519800017588, -0.09188754856586456, 0.03445559740066528, -0.021311776712536812, 0.036771468818187714, 0.1971183717250824, 0.03747646510601044, -0.010775414295494556, -0.016659341752529144, -0.012138805352151394, 0.05873779207468033, 0.007611739914864302, -0.020882828161120415, -0.07346170395612717, -0.03684861212968826, -0.02391767129302025, -0.04635409638285637, 0.012399713508784771, -0.0715688094496727, -0.03235851600766182, -0.07524725049734116, 0.021394232288002968, -0.0269028190523386, -0.07859896123409271, -0.06683474034070969, -0.006011278834193945, 0.020925350487232208, 0.07303860038518906, 0.06478817015886307, 0.07325267791748047, 0.04698638617992401, 0.12860339879989624, -0.023450205102562904, -0.10800128430128098, -0.052480410784482956, 0.044411856681108475, -0.0652337446808815, 0.03148108348250389, -0.018023915588855743, -0.05711107328534126, -0.04742569848895073, 0.029931195080280304, 0.19975987076759338, -0.06587405502796173, 0.016222313046455383, -0.06352584809064865, 0.00852664653211832, -0.04027804732322693, 0.06103311851620674, 0.08890693634748459, 0.1376587599515915, -0.12213917821645737, -0.006943514570593834, -0.07608677446842194, -0.020005587488412857, -0.13620126247406006, 0.0064850496128201485, -0.05119314789772034, -0.05175105482339859, -0.0734827071428299, 0.11291849613189697, -0.0721370056271553, 0.07872288674116135, -0.043865978717803955, -0.09845808148384094, -0.03448650613427162, -0.024293003603816032, 0.1709374338388443, 0.035199593752622604, 0.08750098198652267, -0.059315990656614304, -0.010157641023397446, 0.01288573071360588, 0.04547644406557083, -0.182626411318779, 0.020508253946900368, 0.07102642208337784, -0.10771642625331879, 0.1531551629304886, -0.016674213111400604, 0.14644810557365417, 0.10287383943796158, 0.03800848498940468, -0.1295185387134552, 0.023571547120809555, 0.02512074075639248, -0.019764738157391548, -0.11407361179590225, -0.07629604637622833, 0.028142137452960014, -0.08464132249355316, 0.06089257448911667, 0.05766482651233673, -0.005139557644724846, 0.15443135797977448, 0.03270568698644638, 0.00476139597594738, 0.07068683952093124, -0.022975077852606773, 0.07021715492010117, -0.027968736365437508, -0.04925323650240898, -0.06437640637159348, -0.04676100239157677, -0.03053719736635685, 0.01083003357052803, -0.19969512522220612, -0.057057905942201614, 0.08547097444534302, 0.022587521001696587, 0.05868929997086525, 0.0220847949385643, 0.05767853930592537, -0.058477647602558136, -0.167682945728302, -0.022224415093660355, -0.09989747405052185, -0.010763054713606834, 0.023328151553869247, -0.02353227697312832, 0.028539419174194336, -0.02381320856511593, 0.018801046535372734, -0.026001261547207832, -0.13209518790245056, -0.07569220662117004 ]
null
null
k2
# Introduction This repo contains pre-trained model using <https://github.com/k2-fsa/icefall/pull/219>. It is trained on [AIShell](https://www.openslr.org/33/) dataset using modified transducer from [optimized_transducer](https://github.com/csukuangfj/optimized_transducer). ## How to clone this repo ``` sudo apt-get install git-lfs git clone https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2022-03-01 cd icefall-aishell-transducer-stateless-modified-2022-03-01 git lfs pull ``` **Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later. The model in this repo is trained using the commit `TODO`. You can use ``` git clone https://github.com/k2-fsa/icefall cd icefall git checkout TODO ``` to download `icefall`. You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/TODO/egs/aishell/ASR/transducer_stateless_modified/train.py#L232>. In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 512-dim embedding layer and a Conv1d with kernel size 2. The decoder architecture is modified from [Rnn-Transducer with Stateless Prediction Network](https://ieeexplore.ieee.org/document/9054419). A Conv1d layer is placed right after the input embedding layer. ----- ## Description This repo provides pre-trained transducer Conformer model for the AIShell dataset using [icefall][icefall]. There are no RNNs in the decoder. The decoder is stateless and contains only an embedding layer and a Conv1d. The commands for training are: ```bash cd egs/aishell/ASR ./prepare.sh --stop-stage 6 export CUDA_VISIBLE_DEVICES="0,1,2" ./transducer_stateless_modified/train.py \ --world-size 3 \ --num-epochs 90 \ --start-epoch 0 \ --exp-dir transducer_stateless_modified/exp-4 \ --max-duration 250 \ --lr-factor 2.0 \ --context-size 2 \ --modified-transducer-prob 0.25 ``` The tensorboard training log can be found at <https://tensorboard.dev/experiment/C27M8YxRQCa1t2XglTqlWg> The commands for decoding are ```bash # greedy search for epoch in 64; do for avg in 33; do ./transducer_stateless_modified-2/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer_stateless_modified/exp-4 \ --max-duration 100 \ --context-size 2 \ --decoding-method greedy_search \ --max-sym-per-frame 1 done done # modified beam search for epoch in 64; do for avg in 33; do ./transducer_stateless_modified/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer_stateless_modified/exp-4 \ --max-duration 100 \ --context-size 2 \ --decoding-method modified_beam_search \ --beam-size 4 done done ``` You can find the decoding log for the above command in this repo (in the folder [log][log]). The WER for the test dataset is | | test |comment | |------------------------|------|----------------------------------------------------------------| | greedy search | 5.22 |--epoch 64, --avg 33, --max-duration 100, --max-sym-per-frame 1 | | modified beam search | 5.02 |--epoch 64, --avg 33, --max-duration 100 --beam-size 4 | # File description - [log][log], this directory contains the decoding log and decoding results - [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model - [data][data], this directory contains files generated by [prepare.sh][prepare] - [exp][exp], this directory contains only one file: `preprained.pt` `exp/pretrained.pt` is generated by the following command: ```bash epoch=64 avg=33 ./transducer_stateless_modified/export.py \ --exp-dir ./transducer_stateless_modified/exp-4 \ --lang-dir ./data/lang_char \ --epoch $epoch \ --avg $avg ``` **HINT**: To use `pretrained.pt` to compute the WER for the `test` dataset, just do the following: ```bash cp icefall-aishell-transducer-stateless-modified-2022-03-01/exp/pretrained.pt \ /path/to/icefall/egs/aishell/ASR/transducer_stateless_modified/exp/epoch-999.pt ``` and pass `--epoch 999 --avg 1` to `transducer_stateless_modified/decode.py`. [icefall]: https://github.com/k2-fsa/icefall [prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/aishell/ASR/prepare.sh [exp]: https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2022-03-01/tree/main/exp [data]: https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2022-03-01/tree/main/data [test_wavs]: https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2022-03-01/tree/main/test_wavs [log]: https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2022-03-01/tree/main/log [icefall]: https://github.com/k2-fsa/icefall
{"language": "en", "license": "apache-2.0", "tags": ["icefall", "k2", "transducer", "aishell", "ASR", "stateless transducer", "PyTorch"], "datasets": ["aishell"], "metrics": ["WER"]}
null
csukuangfj/icefall-aishell-transducer-stateless-modified-2022-03-01
[ "k2", "icefall", "transducer", "aishell", "ASR", "stateless transducer", "PyTorch", "en", "dataset:aishell", "license:apache-2.0", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #k2 #icefall #transducer #aishell #ASR #stateless transducer #PyTorch #en #dataset-aishell #license-apache-2.0 #region-us
Introduction ============ This repo contains pre-trained model using <URL It is trained on AIShell dataset using modified transducer from optimized\_transducer. How to clone this repo ---------------------- Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later. The model in this repo is trained using the commit 'TODO'. You can use to download 'icefall'. You can find the model information by visiting <URL In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 512-dim embedding layer and a Conv1d with kernel size 2. The decoder architecture is modified from Rnn-Transducer with Stateless Prediction Network. A Conv1d layer is placed right after the input embedding layer. --- Description ----------- This repo provides pre-trained transducer Conformer model for the AIShell dataset using [icefall](URL). There are no RNNs in the decoder. The decoder is stateless and contains only an embedding layer and a Conv1d. The commands for training are: The tensorboard training log can be found at <URL The commands for decoding are You can find the decoding log for the above command in this repo (in the folder [log](URL)). The WER for the test dataset is test: greedy search, comment: 5.22 test: modified beam search, comment: 5.02 File description ================ * [log](URL), this directory contains the decoding log and decoding results * [test\_wavs](URL), this directory contains wave files for testing the pre-trained model * [data](URL), this directory contains files generated by <URL> * [exp](URL), this directory contains only one file: 'URL' 'exp/URL' is generated by the following command: HINT: To use 'URL' to compute the WER for the 'test' dataset, just do the following: and pass '--epoch 999 --avg 1' to 'transducer\_stateless\_modified/URL'.
[]
[ "TAGS\n#k2 #icefall #transducer #aishell #ASR #stateless transducer #PyTorch #en #dataset-aishell #license-apache-2.0 #region-us \n" ]
[ 49 ]
[ "passage: TAGS\n#k2 #icefall #transducer #aishell #ASR #stateless transducer #PyTorch #en #dataset-aishell #license-apache-2.0 #region-us \n" ]
[ -0.0912378653883934, 0.07988960295915604, -0.006669369526207447, 0.07851304858922958, 0.008801380172371864, -0.03460881486535072, 0.1909128725528717, 0.05409783869981766, 0.18148064613342285, -0.0828222706913948, 0.2081434726715088, 0.14270251989364624, 0.02130713500082493, 0.1469823122024536, 0.0018595680594444275, -0.2536882758140564, 0.03303275257349014, -0.07665423303842545, -0.06312529742717743, 0.08486305177211761, 0.06858454644680023, -0.021544598042964935, 0.013000999577343464, 0.0025744843296706676, 0.05133645981550217, -0.0026905182749032974, 0.029876604676246643, -0.08128637820482254, 0.10192607343196869, -0.053133197128772736, 0.039574362337589264, 0.06003212556242943, 0.012832049280405045, -0.12373247742652893, 0.030157431960105896, -0.03602173551917076, -0.05816726014018059, 0.06906068325042725, 0.020647751167416573, 0.0068633113987743855, -0.05441071838140488, -0.06818012148141861, -0.03350450471043587, 0.03624728322029114, -0.06317850947380066, -0.1613914966583252, -0.07557716220617294, 0.06197865679860115, 0.020470058545470238, 0.046767041087150574, 0.046882305294275284, 0.16259972751140594, -0.12858189642429352, 0.09164070338010788, 0.09944629669189453, -0.3794322907924652, 0.04365764185786247, -0.00095846236217767, -0.07714761793613434, 0.07382962852716446, 0.0221920907497406, 0.02707475796341896, -0.0040580881759524345, -0.030639678239822388, 0.017736317589879036, -0.002135912189260125, -0.14537127315998077, 0.06697852164506912, -0.09807232767343521, -0.023945223540067673, 0.22957146167755127, -0.00728967972099781, 0.034552887082099915, 0.07471691817045212, -0.07000577449798584, -0.060767434537410736, 0.07472936064004898, 0.05742533877491951, 0.015246530994772911, 0.08809284120798111, 0.030423996970057487, -0.005438430234789848, -0.11032744497060776, -0.017874307930469513, -0.16556604206562042, 0.12381387501955032, -0.012242615222930908, 0.13673777878284454, -0.15366560220718384, 0.02018345147371292, 0.031106892973184586, -0.1360904425382614, 0.04168323799967766, -0.08854102343320847, 0.042259540408849716, 0.10368205606937408, -0.021453138440847397, 0.0228680819272995, 0.12416563183069229, 0.17513138055801392, 0.05956288427114487, -0.01219430286437273, 0.03027266636490822, 0.13897959887981415, 0.08803866803646088, 0.02809186279773712, 0.05078687146306038, 0.0002833935141097754, 0.04131162539124489, -0.02649529092013836, 0.09184984117746353, -0.07380345463752747, -0.06052558869123459, -0.011226381175220013, 0.004484857432544231, 0.09055197238922119, 0.10788839310407639, -0.0635041743516922, -0.11025644838809967, 0.010831205174326897, 0.1636449545621872, -0.07003968954086304, 0.00603801105171442, 0.03614818677306175, -0.007903944700956345, 0.07127092033624649, -0.009302046149969101, 0.07987377047538757, 0.03484390676021576, 0.04085765779018402, -0.07472047954797745, -0.00532520329579711, 0.018744785338640213, 0.032718874514102936, 0.0898662656545639, -0.15068235993385315, 0.03934498876333237, -0.13456763327121735, -0.1865530014038086, 0.04674740880727768, 0.0331353098154068, 0.006312964018434286, -0.07606945931911469, 0.029336875304579735, -0.01591426320374012, 0.035490475594997406, -0.07179666310548782, -0.1183822974562645, -0.0454021655023098, 0.012591739185154438, -0.13687865436077118, 0.06004229560494423, -0.13905766606330872, -0.010852500796318054, -0.1709751933813095, 0.0038335579447448254, 0.06900954991579056, 0.009904221631586552, -0.15955688059329987, 0.09217049181461334, -0.04582381993532181, -0.0018194542499259114, -0.06471173465251923, -0.06852979958057404, 0.0274965837597847, 0.16167975962162018, -0.2606445848941803, -0.06605077534914017, 0.08944328874349594, -0.1412372887134552, -0.07851171493530273, 0.09584653377532959, 0.04632703214883804, -0.08110708743333817, 0.03424641862511635, 0.1972573846578598, -0.062363509088754654, 0.026740623638033867, -0.06983506679534912, 0.1493120789527893, -0.16207154095172882, -0.15197616815567017, 0.12212496250867844, -0.03475996106863022, 0.022455710917711258, 0.01562957465648651, 0.06247011199593544, 0.10661417245864868, -0.02652014046907425, -0.14012092351913452, -0.05510948225855827, -0.06013490632176399, 0.07225550711154938, -0.004663246683776379, 0.026414921507239342, -0.013939259573817253, 0.0076739112846553326, -0.019916612654924393, 0.012247229926288128, 0.09415968507528305, 0.05433037504553795, -0.048170190304517746, -0.0018609627149999142, 0.019639911130070686, -0.012817144393920898, -0.07991112768650055, -0.1020396426320076, -0.01813538931310177, -0.0333963967859745, -0.0067701926454901695, 0.08474551141262054, 0.08417068421840668, -0.028222449123859406, 0.021036790683865547, 0.000030475124731310643, 0.02479027770459652, 0.06133440136909485, -0.044664401561021805, -0.21472348272800446, 0.030629323795437813, -0.043714508414268494, 0.0173184797167778, 0.031560298055410385, 0.03290265426039696, 0.14808349311351776, 0.05829736962914467, -0.016148971393704414, 0.042777638882398605, -0.03166559338569641, -0.01655816286802292, 0.021473772823810577, 0.007769769057631493, 0.10401847213506699, 0.008981847204267979, -0.18405045568943024, 0.1666388362646103, -0.009725176729261875, 0.09301548451185226, 0.11492563039064407, -0.03621193766593933, 0.13533613085746765, 0.00034020503517240286, -0.07012379169464111, -0.07224441319704056, 0.0763489380478859, 0.06403760612010956, 0.0668460801243782, 0.05717892199754715, 0.03476601094007492, 0.03281814977526665, -0.024287691339850426, -0.014253119938075542, -0.07533284276723862, -0.03367980942130089, 0.06819384545087814, 0.01745135895907879, -0.17125874757766724, 0.10776758939027786, 0.3296218812465668, -0.004476088099181652, 0.160936176776886, -0.0984550341963768, -0.05269044265151024, -0.03202863410115242, -0.09679681062698364, -0.04375945404171944, 0.12936779856681824, 0.013030963018536568, 0.0383952371776104, 0.08960234373807907, 0.008655755780637264, -0.007258477155119181, -0.09566189348697662, -0.07011207938194275, -0.012708052061498165, 0.043217629194259644, -0.22362090647220612, 0.06542042642831802, -0.08990563452243805, 0.02482079155743122, -0.017326360568404198, -0.11404162645339966, 0.12900184094905853, -0.011971475556492805, -0.02978554368019104, 0.1255762130022049, -0.16270938515663147, -0.13144958019256592, -0.07185633480548859, 0.13855727016925812, -0.015364228747785091, -0.07895947992801666, 0.06630139797925949, -0.11734949052333832, -0.024616481736302376, 0.011017793789505959, -0.05074223875999451, 0.02330404706299305, 0.018224254250526428, 0.07981763035058975, -0.03156455606222153, -0.02406851015985012, -0.0973866656422615, -0.040477506816387177, -0.034757256507873535, -0.02620820328593254, 0.15048518776893616, 0.023970117792487144, 0.016635555773973465, 0.09501165896654129, 0.06400292366743088, 0.001875722547993064, 0.0042271665297448635, 0.09301481395959854, -0.024582967162132263, 0.021343937143683434, 0.1320800632238388, -0.06217304244637489, 0.09606315195560455, 0.06780920177698135, 0.0670260339975357, 0.016589229926466942, -0.024558909237384796, -0.019342387095093727, -0.11219567060470581, -0.2339407056570053, -0.0689534991979599, -0.04289085045456886, 0.18920962512493134, -0.012022891081869602, 0.061017945408821106, 0.022438189014792442, 0.0818771943449974, 0.08090834319591522, -0.08446662127971649, -0.04491174593567848, -0.005343109834939241, 0.0920477882027626, -0.006565549876540899, 0.09325553476810455, -0.059098981320858, 0.031006043776869774, 0.13661040365695953, 0.09189024567604065, 0.15743738412857056, 0.07265490293502808, 0.06164751574397087, 0.08071920275688171, 0.2106257677078247, 0.06398405879735947, 0.11643870919942856, 0.06173894926905632, 0.0003481822204776108, -0.010527604259550571, -0.019198084250092506, -0.0627213716506958, 0.05561863258481026, -0.03361527994275093, -0.1359178125858307, 0.09359019994735718, 0.122604601085186, 0.09443344175815582, 0.30849936604499817, 0.052944984287023544, -0.09825537353754044, -0.030246753245592117, 0.05059781298041344, 0.004838303197175264, 0.03711705282330513, 0.09356563538312912, -0.032572999596595764, -0.0751810222864151, 0.07253054529428482, -0.06378952413797379, 0.02651267871260643, 0.014406802132725716, 0.035934653133153915, -0.00044957027421332896, -0.10247434675693512, 0.0653790608048439, 0.07156770676374435, -0.19068898260593414, 0.21089856326580048, -0.017426539212465286, -0.03855733573436737, -0.051503077149391174, -0.03231717273592949, 0.07364355027675629, 0.09814643114805222, 0.07624343782663345, 0.1003316193819046, -0.10205621272325516, -0.0030949125066399574, -0.2048426866531372, 0.07306124269962311, 0.01624487154185772, -0.007220481988042593, -0.11370912939310074, -0.028865238651633263, 0.048266924917697906, 0.040572185069322586, -0.04346488416194916, -0.1654147356748581, -0.04027123749256134, 0.08227952569723129, 0.11921751499176025, 0.051582250744104385, -0.04684380441904068, -0.1345936506986618, -0.02395833283662796, 0.04973499849438667, -0.14804254472255707, -0.09184578061103821, -0.07757463306188583, -0.05341194570064545, 0.11481121182441711, -0.07870765775442123, 0.02731703780591488, -0.017524149268865585, -0.06935573369264603, -0.02935604751110077, -0.17036864161491394, 0.10326726734638214, -0.1116199865937233, -0.1531403809785843, 0.032947368919849396, 0.04079924523830414, 0.05331959202885628, 0.04850704222917557, -0.052280180156230927, 0.09681210666894913, -0.10598265379667282, -0.0931326299905777, 0.12043843418359756, -0.0038707368075847626, -0.009804202243685722, 0.03848946467041969, 0.04774193465709686, -0.031873807311058044, 0.01947372779250145, -0.05870721489191055, 0.22439776360988617, 0.3064878284931183, -0.10551347583532333, 0.2530943751335144, 0.232868954539299, -0.021601304411888123, -0.2518948018550873, -0.17849886417388916, -0.13449695706367493, -0.06718666851520538, 0.14296524226665497, -0.1871662437915802, 0.09939563274383545, 0.1399552822113037, -0.10915409028530121, 0.08808479458093643, -0.24154254794120789, -0.06261148303747177, 0.18601901829242706, -0.10382705181837082, 0.3112789988517761, -0.11591673642396927, -0.049192845821380615, -0.029166093096137047, -0.0726439580321312, 0.11274299770593643, -0.129652738571167, 0.033185284584760666, -0.04787520319223404, 0.0231630839407444, -0.01982324756681919, -0.06958181411027908, 0.12531019747257233, 0.023997360840439796, -0.01691919006407261, -0.05182144418358803, -0.053042761981487274, 0.1500619351863861, -0.022016003727912903, 0.025026241317391396, -0.1263105571269989, 0.07398365437984467, -0.044837817549705505, 0.04168490320444107, -0.10527857393026352, 0.08941194415092468, 0.005248651374131441, -0.08044013381004333, 0.026511162519454956, 0.0012417531106621027, 0.015155581757426262, -0.00895716529339552, 0.23401859402656555, 0.07795356959104538, -0.08085978031158447, 0.09375496953725815, 0.05962701514363289, -0.06345726549625397, -0.011563708074390888, -0.12432615458965302, -0.05835427716374397, 0.05946287512779236, -0.10460516810417175, 0.060526810586452484, 0.11194875836372375, 0.021192869171500206, -0.04262573644518852, 0.08463691920042038, -0.030982602387666702, 0.018708202987909317, 0.09947454929351807, -0.1750207394361496, -0.0526166632771492, 0.027329187840223312, 0.12590566277503967, 0.08218211680650711, 0.16406187415122986, 0.16672024130821228, 0.008738663978874683, -0.023008527234196663, -0.004100600723177195, 0.11766185611486435, -0.06703564524650574, -0.016708016395568848, 0.09397455304861069, 0.012226168066263199, -0.10942722111940384, 0.15835663676261902, -0.06255161017179489, -0.021798506379127502, 0.050581805408000946, 0.06369170546531677, -0.09594488888978958, -0.11283961683511734, -0.049256909638643265, -0.05734429880976677, -0.1745336949825287, -0.09350460022687912, -0.042272135615348816, -0.05557214096188545, -0.0027147396467626095, 0.04348263144493103, 0.05519729107618332, 0.08399542421102524, -0.0030440788250416517, -0.04849390685558319, 0.07509474456310272, 0.0517427921295166, -0.101190946996212, -0.02502637729048729, -0.12566472589969635, -0.07464843988418579, -0.035303350538015366, 0.1101202443242073, -0.05189479514956474, -0.02026687189936638, -0.04881826043128967, 0.00905164610594511, -0.017723189666867256, -0.07829699665307999, -0.07852557301521301, -0.002520667389035225, 0.014191958121955395, -0.07859302312135696, -0.016515860334038734, 0.06147865206003189, -0.09526724368333817, 0.01253076083958149, -0.013051867485046387, 0.024914147332310677, -0.17246019840240479, -0.06659027934074402, 0.0976024940609932, 0.033472683280706406, 0.14777353405952454, 0.11196228861808777, -0.021210787817835808, 0.0676933079957962, -0.04208976402878761, 0.013453011400997639, 0.054835572838783264, 0.026889638975262642, 0.0271427221596241, 0.06818908452987671, -0.08139735460281372, 0.030318889766931534, 0.000246904935920611, 0.0626215785741806, 0.12740010023117065, -0.07643812149763107, -0.00841483660042286, 0.0471816323697567, -0.12746913731098175, -0.07372988760471344, -0.11302640289068222, 0.05542151257395744, 0.06523116677999496, 0.18119804561138153, -0.009731446392834187, 0.026415813714265823, -0.011357779614627361, 0.02826249785721302, 0.0509478896856308, -0.0767626166343689, -0.05829155072569847, 0.01954500935971737, -0.023083651438355446, -0.05946911498904228, 0.18500490486621857, -0.007898562587797642, -0.14597561955451965, 0.055813390761613846, -0.002432363573461771, -0.04977600648999214, 0.022805001586675644, 0.14182807505130768, 0.014843160286545753, -0.008572817780077457, -0.10036437958478928, 0.029652049764990807, -0.07150762528181076, -0.11461387574672699, 0.14716260135173798, 0.004678407218307257, 0.06345683336257935, 0.034838367253541946, 0.06739576160907745, -0.06207776442170143, -0.13529038429260254, -0.14786063134670258, -0.07946500182151794, 0.06876452267169952, 0.05247573181986809, 0.13741017878055573, 0.1240936666727066, 0.02286904864013195, -0.008667477406561375, -0.0590241476893425, -0.04896598309278488, -0.16247017681598663, -0.13699871301651, -0.006556198466569185, -0.09643697738647461, 0.03727190941572189, -0.024525433778762817, 0.04021124914288521, 0.1975860744714737, 0.04075571149587631, 0.007448310498148203, -0.02112867310643196, -0.0286518856883049, 0.07285694032907486, 0.006339365616440773, -0.032252609729766846, -0.06877052038908005, -0.04114345461130142, -0.023358801379799843, -0.04452336207032204, 0.0012511782115325332, -0.07133675366640091, -0.031438764184713364, -0.07171868532896042, 0.010047467425465584, -0.050128474831581116, -0.08779004216194153, -0.06320007890462875, 0.008802184835076332, -0.003513575065881014, 0.07676175236701965, 0.0700116977095604, 0.0721297338604927, 0.04681841656565666, 0.1285131722688675, -0.022378871217370033, -0.09976514428853989, -0.052428457885980606, 0.052817825227975845, -0.06167568266391754, 0.047914065420627594, -0.017426693812012672, -0.0557255744934082, -0.047955453395843506, 0.02605794556438923, 0.21408070623874664, -0.06597311049699783, 0.022219963371753693, -0.05037020146846771, 0.01047042477875948, -0.014693941920995712, 0.06612900644540787, 0.08440433442592621, 0.1376197189092636, -0.13136446475982666, -0.01656896434724331, -0.06262663006782532, -0.02060304582118988, -0.13870801031589508, -0.010879376903176308, -0.053852055221796036, -0.045270830392837524, -0.07633572816848755, 0.11109621822834015, -0.10148157924413681, 0.09727495908737183, -0.02727482281625271, -0.10534919798374176, -0.022639401257038116, -0.03090241178870201, 0.1807427853345871, 0.02348906733095646, 0.09999065846204758, -0.05316198244690895, -0.04720750078558922, 0.020008506253361702, 0.04680195078253746, -0.20023256540298462, 0.04067599028348923, 0.06809109449386597, -0.12123260647058487, 0.14734168350696564, -0.0183585025370121, 0.15419432520866394, 0.09288114309310913, 0.031404681503772736, -0.11442625522613525, 0.013878867030143738, 0.020791780203580856, -0.029323097318410873, -0.11869718134403229, -0.05416688323020935, 0.029434235766530037, -0.08485594391822815, 0.05812680348753929, 0.03892120346426964, 0.0016652296762913465, 0.13654734194278717, 0.004301830660551786, 0.010969337075948715, 0.06596948206424713, -0.0204599741846323, 0.05306898429989815, -0.0233455803245306, -0.043262235820293427, -0.054493606090545654, -0.05146151781082153, -0.04442114755511284, 0.0075681558810174465, -0.19835273921489716, -0.07324900478124619, 0.09942108392715454, 0.022010236978530884, 0.06941249966621399, 0.025664355605840683, 0.05531547963619232, -0.05885900929570198, -0.16657498478889465, -0.011496917344629765, -0.11001061648130417, -0.012739116325974464, 0.022808190435171127, -0.02144463174045086, 0.03070256859064102, -0.05084160715341568, 0.04164474830031395, -0.026119444519281387, -0.1252989023923874, -0.06487127393484116 ]
null
null
null
# Introduction This repo contains pre-trained model using <https://github.com/k2-fsa/icefall/pull/213>. It is trained on train-clean-100 subset of the LibriSpeech dataset. Also, it uses the `S` subset from GigaSpeech as extra training data. ## How to clone this repo ``` sudo apt-get install git-lfs git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-100h-transducer-stateless-multi-datasets-bpe-500-2022-02-21 cd icefall-asr-librispeech-100h-transducer-stateless-multi-datasets-bpe-500-2022-02-21 git lfs pull ``` **Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later. The model in this repo is trained using the commit `2332ba312d7ce72f08c7bac1e3312f7e3dd722dc`. You can use ``` git clone https://github.com/k2-fsa/icefall cd icefall git checkout 2332ba312d7ce72f08c7bac1e3312f7e3dd722dc ``` to download `icefall`. You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/2332ba312d7ce72f08c7bac1e3312f7e3dd722dc/egs/librispeech/ASR/transducer_stateless_multi_datasets/train.py#L198>. In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2. The decoder architecture is modified from [Rnn-Transducer with Stateless Prediction Network](https://ieeexplore.ieee.org/document/9054419). A Conv1d layer is placed right after the input embedding layer. ----- ## Description This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset using [icefall][icefall]. There are no RNNs in the decoder. The decoder is stateless and contains only an embedding layer and a Conv1d. The commands for training are: ``` cd egs/librispeech/ASR/ ./prepare.sh ./prepare_giga_speech.sh export CUDA_VISIBLE_DEVICES="0,1" ./transducer_stateless_multi_datasets/train.py \ --world-size 2 \ --num-epochs 60 \ --start-epoch 0 \ --exp-dir transducer_stateless_multi_datasets/exp-100-2 \ --full-libri 0 \ --max-duration 300 \ --lr-factor 1 \ --bpe-model data/lang_bpe_500/bpe.model \ --modified-transducer-prob 0.25 --giga-prob 0.2 ``` The tensorboard training log can be found at <https://tensorboard.dev/experiment/qUEKzMnrTZmOz1EXPda9RA/> The command for decoding is: ``` epoch=57 avg=17 ## greedy search for epoch in 57; do for avg in 17; do for sym in 1 2 3; do ./transducer_stateless_multi_datasets/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer_stateless_multi_datasets/exp-100-2 \ --bpe-model ./data/lang_bpe_500/bpe.model \ --max-duration 100 \ --context-size 2 \ --max-sym-per-frame $sym done done done ## modified beam search epoch=57 avg=17 ./transducer_stateless_multi_datasets/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer_stateless_multi_datasets/exp-100-2 \ --bpe-model ./data/lang_bpe_500/bpe.model \ --max-duration 100 \ --context-size 2 \ --decoding-method modified_beam_search \ --beam-size 4 ``` You can find the decoding log for the above command in this repo (in the folder `log`). The WERs for the test datasets are | | test-clean | test-other | comment | |-------------------------------------|------------|------------|------------------------------------------| | greedy search (max sym per frame 1) | 6.34 | 16.7 | --epoch 57, --avg 17, --max-duration 100 | | greedy search (max sym per frame 2) | 6.34 | 16.7 | --epoch 57, --avg 17, --max-duration 100 | | greedy search (max sym per frame 3) | 6.34 | 16.7 | --epoch 57, --avg 17, --max-duration 100 | | modified beam search (beam size 4) | 6.31 | 16.3 | --epoch 57, --avg 17, --max-duration 100 | # File description - [log][log], this directory contains the decoding log and decoding results - [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model - [data][data], this directory contains files generated by [prepare.sh][prepare] - [exp][exp], this directory contains only one file: `preprained.pt` `exp/pretrained.pt` is generated by the following command: ```bash ./transducer_stateless_multi_datasets/export.py \ --epoch 57 \ --avg 17 \ --bpe-model data/lang_bpe_500/bpe.model \ --exp-dir transducer_stateless_multi_datasets/exp-full ``` **HINT**: To use `pretrained.pt` to compute the WER for test-clean and test-other, just do the following: ``` cp icefall-asr-librispeech-100h-transducer-stateless-multi-datasets-bpe-500-2022-02-21/exp/pretrained.pt \ /path/to/icefall/egs/librispeech/ASR/transducer_stateless_multi_datasets/exp/epoch-999.pt ``` and pass `--epoch 999 --avg 1` to `transducer_stateless_multi_datasets/decode.py`. [icefall]: https://github.com/k2-fsa/icefall [prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh [exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-100h-transducer-stateless-multi-datasets-bpe-500-2022-02-21/tree/main/exp [data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-100h-transducer-stateless-multi-datasets-bpe-500-2022-02-21/tree/main/data [test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-100h-transducer-stateless-multi-datasets-bpe-500-2022-02-21/tree/main/test_wavs [log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-100h-transducer-stateless-multi-datasets-bpe-500-2022-02-21/tree/main/log [icefall]: https://github.com/k2-fsa/icefall
{}
null
csukuangfj/icefall-asr-librispeech-100h-transducer-stateless-multi-datasets-bpe-500-2022-02-21
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
Introduction ============ This repo contains pre-trained model using <URL It is trained on train-clean-100 subset of the LibriSpeech dataset. Also, it uses the 'S' subset from GigaSpeech as extra training data. How to clone this repo ---------------------- Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later. The model in this repo is trained using the commit '2332ba312d7ce72f08c7bac1e3312f7e3dd722dc'. You can use to download 'icefall'. You can find the model information by visiting <URL In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2. The decoder architecture is modified from Rnn-Transducer with Stateless Prediction Network. A Conv1d layer is placed right after the input embedding layer. --- Description ----------- This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset using [icefall](URL). There are no RNNs in the decoder. The decoder is stateless and contains only an embedding layer and a Conv1d. The commands for training are: The tensorboard training log can be found at <URL The command for decoding is: You can find the decoding log for the above command in this repo (in the folder 'log'). The WERs for the test datasets are File description ================ * [log](URL), this directory contains the decoding log and decoding results * [test\_wavs](URL), this directory contains wave files for testing the pre-trained model * [data](URL), this directory contains files generated by <URL> * [exp](URL), this directory contains only one file: 'URL' 'exp/URL' is generated by the following command: HINT: To use 'URL' to compute the WER for test-clean and test-other, just do the following: and pass '--epoch 999 --avg 1' to 'transducer\_stateless\_multi\_datasets/URL'.
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
null
# Introduction ## How to clone this repo ``` sudo apt-get install git-lfs git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09 cd icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09 git lfs pull ``` **Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later. ----- ## Description This repo provides pre-trained conformer CTC model for the librispeech dataset using [icefall][icefall]. The commands for training are: ``` cd egs/librispeech/ASR/conformer_ctc ./prepare.sh export CUDA_VISIBLE_DEVICES="0,1,2,3" ./conformer_ctc/train.py \ --exp-dir conformer_ctc/exp_500_att0.8 \ --lang-dir data/lang_bpe_500 \ --att-rate 0.8 \ --full-libri 1 \ --max-duration 200 \ --concatenate-cuts 0 \ --world-size 4 \ --bucketing-sampler 1 \ --start-epoch 0 \ --num-epochs 90 ``` The command for decoding is: ``` ./conformer_ctc/decode.py \ --exp-dir conformer_ctc/exp_500_att0.8 \ --lang-dir data/lang_bpe_500 \ --max-duration 30 \ --concatenate-cuts 0 \ --bucketing-sampler 1 \ --num-paths 1000 \ --epoch 77 \ --avg 55 \ --method attention-decoder \ --nbest-scale 0.5 ``` You can find the decoding log for the above command in this repo: [log/log-decode-2021-11-09-17-38-28](log/log-decode-2021-11-09-17-38-28). The best WER for the librispeech test dataset is: | | test-clean | test-other | |-----|------------|------------| | WER | 2.42 | 5.73 | Scale values used in n-gram LM rescoring and attention rescoring for the best WERs are: | ngram_lm_scale | attention_scale | |----------------|-----------------| | 2.0 | 2.0 | # File description - [log][log], this directory contains the decoding log - [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model - [data][data], this directory contains files generated by [prepare.sh][prepare] Note: For the `data/lm` directory, we provide only `G_4_gram.pt`. If you need other files in this directory, please run [prepare.sh][prepare]. - [exp][exp], this directory contains two files: `preprained.pt` and `cpu_jit.pt`. `exp/pretrained.pt` is generated by the following command: ``` ./conformer_ctc/export.py \ --epoch 77 \ --avg 55 \ --jit 0 \ --lang-dir data/lang_bpe_500 \ --exp-dir conformer_ctc/exp_500_att0.8 ``` **HINT**: To use `pre-trained.pt` to compute the WER for test-clean and test-other, just do the following: ``` cp icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/exp/pretrained.pt \ /path/to/icefall/egs/librispeech/ASR/conformer_ctc/exp/epoch-999.pt ``` and pass `--epoch 999 --avg 1` to `conformer_ctc/decode.py`. `exp/cpu_jit.pt` is generated by the following command: ``` ./conformer_ctc/export.py \ --epoch 77 \ --avg 55 \ --jit 1 \ --lang-dir data/lang_bpe_500 \ --exp-dir conformer_ctc/exp_500_att0.8 ``` # Deploy your model in C++ using k2 To deploy your model in C++ using k2 without depending on Python, do the following: ``` # Note: It requires torch >= 1.8.0 git clone https://github.com/k2-fsa/k2 cd k2 git checkout v2.0-pre mkdir build_release cd build_release cmake -DCMAKE_BUILD_TYPE=Release .. make -j ctc_decode hlg_decode ngram_lm_rescore attention_rescore ``` ## CTC decoding ``` cd k2/build_release ./bin/ctc_decode \ --use_gpu true \ --nn_model ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/exp/cpu_jit.pt \ --bpe_model ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lang_bpe_500/bpe.model \ ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1089-134686-0001.wav \ ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1221-135766-0001.wav \ ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1221-135766-0002.wav ``` ## HLG decoding ``` ./bin/hlg_decode \ --use_gpu true \ --nn_model ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/exp/cpu_jit.pt \ --hlg ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lang_bpe_500/HLG.pt \ --word_table ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lang_bpe_500/words.txt \ ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1089-134686-0001.wav \ ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1221-135766-0001.wav \ ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1221-135766-0002.wav ``` ## HLG decoding + n-gram LM rescoring **NOTE**: V100 GPU with 16 GB RAM is known NOT to work because of OOM. V100 GPU with 32 GB RAM is known to work. ``` ./bin/ngram_lm_rescore \ --use_gpu true \ --nn_model ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/exp/cpu_jit.pt \ --hlg ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lang_bpe_500/HLG.pt \ --g ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lm/G_4_gram.pt \ --ngram_lm_scale 1.0 \ --word_table ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lang_bpe_500/words.txt \ ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1089-134686-0001.wav \ ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1221-135766-0001.wav \ ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1221-135766-0002.wav ``` ## HLG decoding + n-gram LM rescoring + attention decoder rescoring **NOTE**: V100 GPU with 16 GB RAM is known NOT to work because of OOM. V100 GPU with 32 GB RAM is known to work. ``` ./bin/attention_rescore \ --use_gpu true \ --nn_model ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/exp/cpu_jit.pt \ --hlg ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lang_bpe_500/HLG.pt \ --g ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lm/G_4_gram.pt \ --ngram_lm_scale 2.0 \ --attention_scale 2.0 \ --num_paths 100 \ --nbest_scale 0.5 \ --word_table ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lang_bpe_500/words.txt \ --sos_id 1 \ --eos_id 1 \ ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1089-134686-0001.wav \ ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1221-135766-0001.wav \ ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1221-135766-0002.wav ``` [prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh [exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/tree/main/exp [data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/tree/main/data [test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/tree/main/test_wavs [log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/tree/main/log [icefall]: https://github.com/k2-fsa/icefall
{}
null
csukuangfj/icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
Introduction ============ How to clone this repo ---------------------- Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later. --- Description ----------- This repo provides pre-trained conformer CTC model for the librispeech dataset using [icefall](URL). The commands for training are: The command for decoding is: You can find the decoding log for the above command in this repo: log/log-decode-2021-11-09-17-38-28. The best WER for the librispeech test dataset is: test-clean: WER, test-other: 2.42 Scale values used in n-gram LM rescoring and attention rescoring for the best WERs are: File description ================ * [log](URL), this directory contains the decoding log * [test\_wavs](URL), this directory contains wave files for testing the pre-trained model * [data](URL), this directory contains files generated by <URL> Note: For the 'data/lm' directory, we provide only 'G\_4\_gram.pt'. If you need other files in this directory, please run <URL>. * [exp](URL), this directory contains two files: 'URL' and 'cpu\_jit.pt'. 'exp/URL' is generated by the following command: HINT: To use 'URL' to compute the WER for test-clean and test-other, just do the following: and pass '--epoch 999 --avg 1' to 'conformer\_ctc/URL'. 'exp/cpu\_jit.pt' is generated by the following command: Deploy your model in C++ using k2 ================================= To deploy your model in C++ using k2 without depending on Python, do the following: CTC decoding ------------ HLG decoding ------------ HLG decoding + n-gram LM rescoring ---------------------------------- NOTE: V100 GPU with 16 GB RAM is known NOT to work because of OOM. V100 GPU with 32 GB RAM is known to work. HLG decoding + n-gram LM rescoring + attention decoder rescoring ---------------------------------------------------------------- NOTE: V100 GPU with 16 GB RAM is known NOT to work because of OOM. V100 GPU with 32 GB RAM is known to work.
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
null
# Introduction ## How to clone this repo ``` sudo apt-get install git-lfs git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-17 cd icefall-asr-librispeech-transducer-bpe-500-2021-12-17 git lfs pull ``` **Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later. The model in this repo is trained using the commit `cb04c8a7509425ab45fae888b0ca71bbbd23f0de`. You can use ``` git clone https://github.com/k2-fsa/icefall cd icefall git checkout cb04c8a7509425ab45fae888b0ca71bbbd23f0de ``` to download `icefall`. You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/cb04c8a7509425ab45fae888b0ca71bbbd23f0de/egs/librispeech/ASR/transducer/train.py#L196> In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 1024-dim embedding layer, plus a 4-layer LSTM with hidden size 512. ----- ## Description This repo provides pre-trained RNN-T Conformer model for the librispeech dataset using [icefall][icefall]. The commands for training are: ``` cd egs/librispeech/ASR/ ./prepare.sh export CUDA_VISIBLE_DEVICES="0,1,2,3" ./transducer/train.py \ --world-size 4 \ --num-epochs 30 \ --start-epoch 0 \ --exp-dir transducer/exp-lr-2.5-full \ --full-libri 1 \ --max-duration 250 \ --lr-factor 2.5 ``` The command for decoding is: ``` epoch=26 avg=12 ./transducer/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer/exp-lr-2.5-full \ --bpe-model ./data/lang_bpe_500/bpe.model \ --max-duration 100 ``` You can find the decoding log for the above command in this repo: [log/log-decode-epoch-26-avg-12-2021-12-17-09-33-04](log/log-decode-epoch-26-avg-12-2021-12-17-09-33-04). The best WER using greedy search is: | | test-clean | test-other | |-----|------------|------------| | WER | 3.16 | 7.71 | # File description - [log][log], this directory contains the decoding log and decoding results - [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model - [data][data], this directory contains files generated by [prepare.sh][prepare] - [exp][exp], this directory contains only one file: `preprained.pt` `exp/pretrained.pt` is generated by the following command: ``` ./transducer/export.py \ --epoch 26 \ --avg 12 \ --bpe-model data/lang_bpe_500/bpe.model \ --exp-dir transducer/exp-lr-2.5-full ``` **HINT**: To use `pre-trained.pt` to compute the WER for test-clean and test-other, just do the following: ``` cp icefall-asr-librispeech-transducer-bpe-500-2021-12-17/exp/pretrained.pt \ /path/to/icefall/egs/librispeech/ASR/transducer/exp/epoch-999.pt ``` and pass `--epoch 999 --avg 1` to `transducer/decode.py`. [icefall]: https://github.com/k2-fsa/icefall [prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh [exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-17/tree/main/exp [data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-17/tree/main/data [test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-17/tree/main/test_wavs [log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-17/tree/main/log [icefall]: https://github.com/k2-fsa/icefall
{}
null
csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-17
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
Introduction ============ How to clone this repo ---------------------- Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later. The model in this repo is trained using the commit 'cb04c8a7509425ab45fae888b0ca71bbbd23f0de'. You can use to download 'icefall'. You can find the model information by visiting <URL In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 1024-dim embedding layer, plus a 4-layer LSTM with hidden size 512. --- Description ----------- This repo provides pre-trained RNN-T Conformer model for the librispeech dataset using [icefall](URL). The commands for training are: The command for decoding is: You can find the decoding log for the above command in this repo: log/log-decode-epoch-26-avg-12-2021-12-17-09-33-04. The best WER using greedy search is: test-clean: WER, test-other: 3.16 File description ================ * [log](URL), this directory contains the decoding log and decoding results * [test\_wavs](URL), this directory contains wave files for testing the pre-trained model * [data](URL), this directory contains files generated by <URL> * [exp](URL), this directory contains only one file: 'URL' 'exp/URL' is generated by the following command: HINT: To use 'URL' to compute the WER for test-clean and test-other, just do the following: and pass '--epoch 999 --avg 1' to 'transducer/URL'.
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
null
# Introduction ## How to clone this repo ``` sudo apt-get install git-lfs git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-23 cd icefall-asr-librispeech-transducer-bpe-500-2021-12-23 git lfs pull ``` **Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later. The model in this repo is trained using the commit `5b6699a8354b70b23b252b371c612a35ed186ec2`. You can use ``` git clone https://github.com/k2-fsa/icefall cd icefall git checkout 5b6699a8354b70b23b252b371c612a35ed186ec2 ``` to download `icefall`. You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/5b6699a8354b70b23b252b371c612a35ed186ec2/egs/librispeech/ASR/transducer/train.py#L191> In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 1024-dim embedding layer, plus a 2-layer LSTM with hidden size 512. ----- ## Description This repo provides pre-trained RNN-T Conformer model for the librispeech dataset using [icefall][icefall]. The commands for training are: ``` cd egs/librispeech/ASR/ ./prepare.sh export CUDA_VISIBLE_DEVICES="0,1,2,3" ./transducer/train.py \ --world-size 4 \ --num-epochs 35 \ --start-epoch 0 \ --exp-dir transducer/exp-lr-2.5-full \ --full-libri 1 \ --max-duration 180 \ --lr-factor 2.5 ``` The command for decoding is: ``` epoch=34 avg=11 ./transducer/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer/exp-lr-2.5-full \ --bpe-model ./data/lang_bpe_500/bpe.model \ --max-duration 100 ``` You can find the decoding log for the above command in the `log` folder of this repo. The best WER using greedy search is: | | test-clean | test-other | |-----|------------|------------| | WER | 3.07 | 7.51 | # File description - [log][log], this directory contains the decoding log and decoding results - [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model - [data][data], this directory contains files generated by [prepare.sh][prepare] - [exp][exp], this directory contains only one file: `preprained.pt` `exp/pretrained.pt` is generated by the following command: ``` ./transducer/export.py \ --epoch 34 \ --avg 11 \ --bpe-model data/lang_bpe_500/bpe.model \ --exp-dir transducer/exp-lr-2.5-full ``` **HINT**: To use `pre-trained.pt` to compute the WER for test-clean and test-other, just do the following: ``` cp icefall-asr-librispeech-transducer-bpe-500-2021-12-23/exp/pretrained.pt \ /path/to/icefall/egs/librispeech/ASR/transducer/exp/epoch-999.pt ``` and pass `--epoch 999 --avg 1` to `transducer/decode.py`. [icefall]: https://github.com/k2-fsa/icefall [prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh [exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-23/tree/main/exp [data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-23/tree/main/data [test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-23/tree/main/test_wavs [log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-23/tree/main/log [icefall]: https://github.com/k2-fsa/icefall
{}
null
csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-23
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
Introduction ============ How to clone this repo ---------------------- Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later. The model in this repo is trained using the commit '5b6699a8354b70b23b252b371c612a35ed186ec2'. You can use to download 'icefall'. You can find the model information by visiting <URL In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 1024-dim embedding layer, plus a 2-layer LSTM with hidden size 512. --- Description ----------- This repo provides pre-trained RNN-T Conformer model for the librispeech dataset using [icefall](URL). The commands for training are: The command for decoding is: You can find the decoding log for the above command in the 'log' folder of this repo. The best WER using greedy search is: test-clean: WER, test-other: 3.07 File description ================ * [log](URL), this directory contains the decoding log and decoding results * [test\_wavs](URL), this directory contains wave files for testing the pre-trained model * [data](URL), this directory contains files generated by <URL> * [exp](URL), this directory contains only one file: 'URL' 'exp/URL' is generated by the following command: HINT: To use 'URL' to compute the WER for test-clean and test-other, just do the following: and pass '--epoch 999 --avg 1' to 'transducer/URL'.
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
null
# Introduction ## How to clone this repo ``` sudo apt-get install git-lfs git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-22 cd icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-22 git lfs pull ``` **Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later. The model in this repo is trained using the commit `fb6a57e9e01dd8aae2af2a6b4568daad8bc8ab32`. You can use ``` git clone https://github.com/k2-fsa/icefall cd icefall git checkout fb6a57e9e01dd8aae2af2a6b4568daad8bc8ab32 ``` to download `icefall`. You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/fb6a57e9e01dd8aae2af2a6b4568daad8bc8ab32/egs/librispeech/ASR/transducer_stateless/train.py#L195>. In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2. The decoder architecture is modified from [Rnn-Transducer with Stateless Prediction Network](https://ieeexplore.ieee.org/document/9054419). A Conv1d layer is placed right after the input embedding layer. ----- ## Description This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset using [icefall][icefall]. There are no RNNs in the decoder. The decoder is stateless and contains only an embedding layer and a Conv1d. The commands for training are: ``` cd egs/librispeech/ASR/ ./prepare.sh export CUDA_VISIBLE_DEVICES="0,1,2,3" ./transducer_stateless/train.py \ --world-size 4 \ --num-epochs 30 \ --start-epoch 0 \ --exp-dir transducer_stateless/exp-full \ --full-libri 1 \ --max-duration 250 \ --lr-factor 3 ``` The tensorboard training log can be found at <https://tensorboard.dev/experiment/PsJ3LgkEQfOmzedAlYfVeg/#scalars&_smoothingWeight=0> The command for decoding is: ``` epoch=20 avg=10 ## greedy search ./transducer_stateless/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer_stateless/exp-full \ --bpe-model ./data/lang_bpe_500/bpe.model \ --max-duration 100 ## beam search ./transducer_stateless/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer_stateless/exp-full \ --bpe-model ./data/lang_bpe_500/bpe.model \ --max-duration 100 \ --decoding-method beam_search \ --beam-size 4 ``` You can find the decoding log for the above command in this repo (in the folder `log`). The WERs for the test datasets are | | test-clean | test-other | comment | |---------------------------|------------|------------|------------------------------------------| | greedy search | 2.99 | 7.52 | --epoch 20, --avg 10, --max-duration 100 | | beam search (beam size 2) | 2.95 | 7.43 | | | beam search (beam size 3) | 2.94 | 7.37 | | | beam search (beam size 4) | 2.92 | 7.37 | | | beam search (beam size 5) | 2.93 | 7.38 | | | beam search (beam size 8) | 2.92 | 7.38 | | # File description - [log][log], this directory contains the decoding log and decoding results - [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model - [data][data], this directory contains files generated by [prepare.sh][prepare] - [exp][exp], this directory contains only one file: `preprained.pt` `exp/pretrained.pt` is generated by the following command: ``` ./transducer_stateless/export.py \ --epoch 20 \ --avg 10 \ --bpe-model data/lang_bpe_500/bpe.model \ --exp-dir transducer_stateless/exp-full ``` **HINT**: To use `pre-trained.pt` to compute the WER for test-clean and test-other, just do the following: ``` cp icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-22/exp/pretrained.pt \ /path/to/icefall/egs/librispeech/ASR/transducer_stateless/exp/epoch-999.pt ``` and pass `--epoch 999 --avg 1` to `transducer_stateless/decode.py`. [icefall]: https://github.com/k2-fsa/icefall [prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh [exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-22/tree/main/exp [data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-22/tree/main/data [test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-22/tree/main/test_wavs [log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-22/tree/main/log [icefall]: https://github.com/k2-fsa/icefall
{}
null
csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-22
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
Introduction ============ How to clone this repo ---------------------- Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later. The model in this repo is trained using the commit 'fb6a57e9e01dd8aae2af2a6b4568daad8bc8ab32'. You can use to download 'icefall'. You can find the model information by visiting <URL In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2. The decoder architecture is modified from Rnn-Transducer with Stateless Prediction Network. A Conv1d layer is placed right after the input embedding layer. --- Description ----------- This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset using [icefall](URL). There are no RNNs in the decoder. The decoder is stateless and contains only an embedding layer and a Conv1d. The commands for training are: The tensorboard training log can be found at <URL The command for decoding is: You can find the decoding log for the above command in this repo (in the folder 'log'). The WERs for the test datasets are File description ================ * [log](URL), this directory contains the decoding log and decoding results * [test\_wavs](URL), this directory contains wave files for testing the pre-trained model * [data](URL), this directory contains files generated by <URL> * [exp](URL), this directory contains only one file: 'URL' 'exp/URL' is generated by the following command: HINT: To use 'URL' to compute the WER for test-clean and test-other, just do the following: and pass '--epoch 999 --avg 1' to 'transducer\_stateless/URL'.
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
null
# Introduction ## How to clone this repo ``` sudo apt-get install git-lfs git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-27 cd icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-27 git lfs pull ``` **Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later. The model in this repo is trained using the commit `14c93add507982306f5a478cd144e0e32e0f970d`. You can use ``` git clone https://github.com/k2-fsa/icefall cd icefall git checkout 14c93add507982306f5a478cd144e0e32e0f970d ``` to download `icefall`. You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/14c93add507982306f5a478cd144e0e32e0f970d/egs/librispeech/ASR/transducer_stateless/train.py#L198>. In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2. The decoder architecture is modified from [Rnn-Transducer with Stateless Prediction Network](https://ieeexplore.ieee.org/document/9054419). A Conv1d layer is placed right after the input embedding layer. ----- ## Description This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset using [icefall][icefall]. There are no RNNs in the decoder. The decoder is stateless and contains only an embedding layer and a Conv1d. The commands for training are: ``` cd egs/librispeech/ASR/ ./prepare.sh export CUDA_VISIBLE_DEVICES="0,1,2,3" ./transducer_stateless/train.py \ --world-size 4 \ --num-epochs 30 \ --start-epoch 0 \ --exp-dir transducer_stateless/exp-full \ --full-libri 1 \ --max-duration 250 \ --lr-factor 3 ``` The tensorboard training log can be found at <https://tensorboard.dev/experiment/Mjx7MeTgR3Oyr1yBCwjozw/> The command for decoding is: ``` epoch=29 avg=13 ## greedy search ./transducer_stateless/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer_stateless/exp-full \ --bpe-model ./data/lang_bpe_500/bpe.model \ --max-duration 100 ## beam search ./transducer_stateless/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer_stateless/exp-full \ --bpe-model ./data/lang_bpe_500/bpe.model \ --max-duration 100 \ --decoding-method beam_search \ --beam-size 4 ``` You can find the decoding log for the above command in this repo (in the folder `log`). The WERs for the test datasets are | | test-clean | test-other | comment | |---------------------------|------------|------------|------------------------------------------| | greedy search | 2.85 | 7.30 | --epoch 29, --avg 13, --max-duration 100 | | beam search (beam size 4) | 2.83 | 7.19 | | # File description - [log][log], this directory contains the decoding log and decoding results - [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model - [data][data], this directory contains files generated by [prepare.sh][prepare] - [exp][exp], this directory contains only one file: `preprained.pt` `exp/pretrained.pt` is generated by the following command: ``` ./transducer_stateless/export.py \ --epoch 29 \ --avg 13 \ --bpe-model data/lang_bpe_500/bpe.model \ --exp-dir transducer_stateless/exp-full ``` **HINT**: To use `pre-trained.pt` to compute the WER for test-clean and test-other, just do the following: ``` cp icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-27/exp/pretrained.pt \ /path/to/icefall/egs/librispeech/ASR/transducer_stateless/exp/epoch-999.pt ``` and pass `--epoch 999 --avg 1` to `transducer_stateless/decode.py`. [icefall]: https://github.com/k2-fsa/icefall [prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh [exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-27/tree/main/exp [data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-27/tree/main/data [test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-27/tree/main/test_wavs [log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-27/tree/main/log [icefall]: https://github.com/k2-fsa/icefall
{}
null
csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-27
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
Introduction ============ How to clone this repo ---------------------- Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later. The model in this repo is trained using the commit '14c93add507982306f5a478cd144e0e32e0f970d'. You can use to download 'icefall'. You can find the model information by visiting <URL In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2. The decoder architecture is modified from Rnn-Transducer with Stateless Prediction Network. A Conv1d layer is placed right after the input embedding layer. --- Description ----------- This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset using [icefall](URL). There are no RNNs in the decoder. The decoder is stateless and contains only an embedding layer and a Conv1d. The commands for training are: The tensorboard training log can be found at <URL The command for decoding is: You can find the decoding log for the above command in this repo (in the folder 'log'). The WERs for the test datasets are File description ================ * [log](URL), this directory contains the decoding log and decoding results * [test\_wavs](URL), this directory contains wave files for testing the pre-trained model * [data](URL), this directory contains files generated by <URL> * [exp](URL), this directory contains only one file: 'URL' 'exp/URL' is generated by the following command: HINT: To use 'URL' to compute the WER for test-clean and test-other, just do the following: and pass '--epoch 999 --avg 1' to 'transducer\_stateless/URL'.
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
null
# Introduction ## How to clone this repo ``` sudo apt-get install git-lfs git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-01-10 cd icefall-asr-librispeech-transducer-stateless-bpe-500-2022-01-10 git lfs pull ``` **Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later. The model in this repo is trained using the commit `4c1b3665ee6efb935f4dd93a80ff0e154b13efb6`. You can use ``` git clone https://github.com/k2-fsa/icefall cd icefall git checkout 4c1b3665ee6efb935f4dd93a80ff0e154b13efb6 ``` to download `icefall`. You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/273e5fb2f3ac2620bafdffe2689b8b3ee10173d3/egs/librispeech/ASR/transducer_stateless/train.py#L198>. In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2. The decoder architecture is modified from [Rnn-Transducer with Stateless Prediction Network](https://ieeexplore.ieee.org/document/9054419). A Conv1d layer is placed right after the input embedding layer. ----- ## Description This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset using [icefall][icefall]. There are no RNNs in the decoder. The decoder is stateless and contains only an embedding layer and a Conv1d. The commands for training are: ``` cd egs/librispeech/ASR/ ./prepare.sh export CUDA_VISIBLE_DEVICES="0,1,2,3" ./transducer_stateless/train.py \ --world-size 4 \ --num-epochs 76 \ --start-epoch 0 \ --exp-dir transducer_stateless/exp-full \ --full-libri 1 \ --max-duration 250 \ --lr-factor 3 ``` The tensorboard training log can be found at <https://tensorboard.dev/experiment/qGdqzHnxS0WJ695OXfZDzA/> The command for decoding is: ``` epoch=71 avg=15 ## greedy search ./transducer_stateless/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer_stateless/exp-full \ --bpe-model ./data/lang_bpe_500/bpe.model \ --max-duration 100 ## beam search ./transducer_stateless/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer_stateless/exp-full \ --bpe-model ./data/lang_bpe_500/bpe.model \ --max-duration 100 \ --decoding-method beam_search \ --beam-size 4 ``` You can find the decoding log for the above command in this repo (in the folder `log`). The WERs for the test datasets are | | test-clean | test-other | comment | |---------------------------|------------|------------|------------------------------------------| | greedy search | 2.69 | 6.81 | --epoch 71, --avg 15, --max-duration 100 | | beam search (beam size 4) | 2.68 | 6.72 | --epoch 71, --avg 15, --max-duration 100 | # File description - [log][log], this directory contains the decoding log and decoding results - [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model - [data][data], this directory contains files generated by [prepare.sh][prepare] - [exp][exp], this directory contains only one file: `preprained.pt` `exp/pretrained.pt` is generated by the following command: ``` ./transducer_stateless/export.py \ --epoch 71 \ --avg 15 \ --bpe-model data/lang_bpe_500/bpe.model \ --exp-dir transducer_stateless/exp-full ``` **HINT**: To use `pre-trained.pt` to compute the WER for test-clean and test-other, just do the following: ``` cp icefall-asr-librispeech-transducer-stateless-bpe-500-2022-01-10/exp/pretrained.pt \ /path/to/icefall/egs/librispeech/ASR/transducer_stateless/exp/epoch-999.pt ``` and pass `--epoch 999 --avg 1` to `transducer_stateless/decode.py`. [icefall]: https://github.com/k2-fsa/icefall [prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh [exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-01-10/tree/main/exp [data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-01-10/tree/main/data [test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-01-10/tree/main/test_wavs [log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-01-10/tree/main/log [icefall]: https://github.com/k2-fsa/icefall
{}
null
csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-01-10
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
Introduction ============ How to clone this repo ---------------------- Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later. The model in this repo is trained using the commit '4c1b3665ee6efb935f4dd93a80ff0e154b13efb6'. You can use to download 'icefall'. You can find the model information by visiting <URL In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2. The decoder architecture is modified from Rnn-Transducer with Stateless Prediction Network. A Conv1d layer is placed right after the input embedding layer. --- Description ----------- This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset using [icefall](URL). There are no RNNs in the decoder. The decoder is stateless and contains only an embedding layer and a Conv1d. The commands for training are: The tensorboard training log can be found at <URL The command for decoding is: You can find the decoding log for the above command in this repo (in the folder 'log'). The WERs for the test datasets are File description ================ * [log](URL), this directory contains the decoding log and decoding results * [test\_wavs](URL), this directory contains wave files for testing the pre-trained model * [data](URL), this directory contains files generated by <URL> * [exp](URL), this directory contains only one file: 'URL' 'exp/URL' is generated by the following command: HINT: To use 'URL' to compute the WER for test-clean and test-other, just do the following: and pass '--epoch 999 --avg 1' to 'transducer\_stateless/URL'.
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
null
# Introduction ## How to clone this repo ``` sudo apt-get install git-lfs git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-02-07 cd icefall-asr-librispeech-transducer-stateless-bpe-500-2022-02-07 git lfs pull ``` **Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later. The model in this repo is trained using the commit `a8150021e01d34ecbd6198fe03a57eacf47a16f2`. You can use ``` git clone https://github.com/k2-fsa/icefall cd icefall git checkout a8150021e01d34ecbd6198fe03a57eacf47a16f2 ``` to download `icefall`. You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/a8150021e01d34ecbd6198fe03a57eacf47a16f2/egs/librispeech/ASR/transducer_stateless/train.py#L198>. In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2. The decoder architecture is modified from [Rnn-Transducer with Stateless Prediction Network](https://ieeexplore.ieee.org/document/9054419). A Conv1d layer is placed right after the input embedding layer. ----- ## Description This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset using [icefall][icefall]. There are no RNNs in the decoder. The decoder is stateless and contains only an embedding layer and a Conv1d. The commands for training are: ``` cd egs/librispeech/ASR/ ./prepare.sh export CUDA_VISIBLE_DEVICES="0,1,2,3" ./transducer_stateless/train.py \ --world-size 4 \ --num-epochs 76 \ --start-epoch 0 \ --exp-dir transducer_stateless/exp-full \ --full-libri 1 \ --max-duration 300 \ --lr-factor 5 \ --bpe-model data/lang_bpe_500/bpe.model \ --modified-transducer-prob 0.25 ``` The tensorboard training log can be found at <https://tensorboard.dev/experiment/qgvWkbF2R46FYA6ZMNmOjA/> The command for decoding is: ``` epoch=63 avg=19 ## greedy search for sym in 1 2 3; do ./transducer_stateless/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer_stateless/exp-full \ --bpe-model ./data/lang_bpe_500/bpe.model \ --max-duration 100 \ --max-sym-per-frame $sym done ## modified beam search ./transducer_stateless/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer_stateless/exp-full \ --bpe-model ./data/lang_bpe_500/bpe.model \ --max-duration 100 \ --context-size 2 \ --decoding-method modified_beam_search \ --beam-size 4 ``` You can find the decoding log for the above command in this repo (in the folder `log`). The WERs for the test datasets are | | test-clean | test-other | comment | |-------------------------------------|------------|------------|------------------------------------------| | greedy search (max sym per frame 1) | 2.67 | 6.67 | --epoch 63, --avg 19, --max-duration 100 | | greedy search (max sym per frame 2) | 2.67 | 6.67 | --epoch 63, --avg 19, --max-duration 100 | | greedy search (max sym per frame 3) | 2.67 | 6.67 | --epoch 63, --avg 19, --max-duration 100 | | modified beam search (beam size 4) | 2.67 | 6.57 | --epoch 63, --avg 19, --max-duration 100 | # File description - [log][log], this directory contains the decoding log and decoding results - [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model - [data][data], this directory contains files generated by [prepare.sh][prepare] - [exp][exp], this directory contains only one file: `preprained.pt` `exp/pretrained.pt` is generated by the following command: ``` ./transducer_stateless/export.py \ --epoch 63 \ --avg 19 \ --bpe-model data/lang_bpe_500/bpe.model \ --exp-dir transducer_stateless/exp-full ``` **HINT**: To use `pretrained.pt` to compute the WER for test-clean and test-other, just do the following: ``` cp icefall-asr-librispeech-transducer-stateless-bpe-500-2022-02-07/exp/pretrained.pt \ /path/to/icefall/egs/librispeech/ASR/transducer_stateless/exp/epoch-999.pt ``` and pass `--epoch 999 --avg 1` to `transducer_stateless/decode.py`. [icefall]: https://github.com/k2-fsa/icefall [prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh [exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-02-07/tree/main/exp [data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-02-07/tree/main/data [test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-02-07/tree/main/test_wavs [log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-02-07/tree/main/log [icefall]: https://github.com/k2-fsa/icefall
{}
null
csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-02-07
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
Introduction ============ How to clone this repo ---------------------- Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later. The model in this repo is trained using the commit 'a8150021e01d34ecbd6198fe03a57eacf47a16f2'. You can use to download 'icefall'. You can find the model information by visiting <URL In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2. The decoder architecture is modified from Rnn-Transducer with Stateless Prediction Network. A Conv1d layer is placed right after the input embedding layer. --- Description ----------- This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset using [icefall](URL). There are no RNNs in the decoder. The decoder is stateless and contains only an embedding layer and a Conv1d. The commands for training are: The tensorboard training log can be found at <URL The command for decoding is: You can find the decoding log for the above command in this repo (in the folder 'log'). The WERs for the test datasets are File description ================ * [log](URL), this directory contains the decoding log and decoding results * [test\_wavs](URL), this directory contains wave files for testing the pre-trained model * [data](URL), this directory contains files generated by <URL> * [exp](URL), this directory contains only one file: 'URL' 'exp/URL' is generated by the following command: HINT: To use 'URL' to compute the WER for test-clean and test-other, just do the following: and pass '--epoch 999 --avg 1' to 'transducer\_stateless/URL'.
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
k2
# Introduction This repo contains pre-trained model using <https://github.com/k2-fsa/icefall/pull/213>. It is trained on full LibriSpeech dataset. Also, it uses the `L` subset from [GigaSpeech](https://github.com/SpeechColab/GigaSpeech) as extra training data. ## How to clone this repo ``` sudo apt-get install git-lfs git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-multi-datasets-bpe-500-2022-03-01 cd icefall-asr-librispeech-transducer-stateless-multi-datasets-bpe-500-2022-03-01 git lfs pull ``` **Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later. The model in this repo is trained using the commit `2332ba312d7ce72f08c7bac1e3312f7e3dd722dc`. You can use ``` git clone https://github.com/k2-fsa/icefall cd icefall git checkout 2332ba312d7ce72f08c7bac1e3312f7e3dd722dc ``` to download `icefall`. You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/2332ba312d7ce72f08c7bac1e3312f7e3dd722dc/egs/librispeech/ASR/transducer_stateless_multi_datasets/train.py#L218> In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2. The decoder architecture is modified from [Rnn-Transducer with Stateless Prediction Network](https://ieeexplore.ieee.org/document/9054419). A Conv1d layer is placed right after the input embedding layer. ----- ## Description This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset using [icefall][icefall]. There are no RNNs in the decoder. The decoder is stateless and contains only an embedding layer and a Conv1d. The commands for training are: ``` cd egs/librispeech/ASR/ ./prepare.sh ./prepare_giga_speech.sh export CUDA_VISIBLE_DEVICES="0,1,2,3" ./transducer_stateless_multi_datasets/train.py \ --world-size 4 \ --num-epochs 40 \ --start-epoch 0 \ --exp-dir transducer_stateless_multi_datasets/exp-full-2 \ --full-libri 1 \ --max-duration 300 \ --lr-factor 5 \ --bpe-model data/lang_bpe_500/bpe.model \ --modified-transducer-prob 0.25 \ --giga-prob 0.2 ``` The tensorboard training log can be found at <https://tensorboard.dev/experiment/xmo5oCgrRVelH9dCeOkYBg/> The command for decoding is: ```bash epoch=39 avg=15 sym=1 # greedy search ./transducer_stateless_multi_datasets/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer_stateless_multi_datasets/exp-full-2 \ --bpe-model ./data/lang_bpe_500/bpe.model \ --max-duration 100 \ --context-size 2 \ --max-sym-per-frame $sym # modified beam search ./transducer_stateless_multi_datasets/decode.py \ --epoch $epoch \ --avg $avg \ --exp-dir transducer_stateless_multi_datasets/exp-full-2 \ --bpe-model ./data/lang_bpe_500/bpe.model \ --max-duration 100 \ --context-size 2 \ --decoding-method modified_beam_search \ --beam-size 4 ``` You can find the decoding log for the above command in this repo (in the folder `log`). The WERs for the test datasets are | | test-clean | test-other | comment | |-------------------------------------|------------|------------|------------------------------------------| | greedy search (max sym per frame 1) | 2.64 | 6.55 | --epoch 39, --avg 15, --max-duration 100 | | modified beam search (beam size 4) | 2.61 | 6.46 | --epoch 39, --avg 15, --max-duration 100 | # File description - [log][log], this directory contains the decoding log and decoding results - [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model - [data][data], this directory contains files generated by [prepare.sh][prepare] - [exp][exp], this directory contains only one file: `preprained.pt` `exp/pretrained.pt` is generated by the following command: ```bash ./transducer_stateless_multi_datasets/export.py \ --epoch 39 \ --avg 15 \ --bpe-model data/lang_bpe_500/bpe.model \ --exp-dir transducer_stateless_multi_datasets/exp-full-2 ``` **HINT**: To use `pretrained.pt` to compute the WER for test-clean and test-other, just do the following: ``` cp icefall-asr-librispeech-transducer-stateless-multi-datasets-bpe-500-2022-03-01/exp/pretrained.pt \ /path/to/icefall/egs/librispeech/ASR/transducer_stateless_multi_datasets/exp/epoch-999.pt ``` and pass `--epoch 999 --avg 1` to `transducer_stateless_multi_datasets/decode.py`. [icefall]: https://github.com/k2-fsa/icefall [prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh [exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-multi-datasets-bpe-500-2022-03-01/tree/main/exp [data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-multi-datasets-bpe-500-2022-03-01/tree/main/data [test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-multi-datasets-bpe-500-2022-03-01/tree/main/test_wavs [log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-multi-datasets-bpe-500-2022-03-01/tree/main/log [icefall]: https://github.com/k2-fsa/icefall
{"language": "en", "license": "apache-2.0", "tags": ["icefall", "k2", "transducer", "librispeech", "ASR", "stateless transducer", "PyTorch", "RNN-T", "speech recognition"], "datasets": ["librispeech"], "metrics": ["WER"]}
null
csukuangfj/icefall-asr-librispeech-transducer-stateless-multi-datasets-bpe-500-2022-03-01
[ "k2", "icefall", "transducer", "librispeech", "ASR", "stateless transducer", "PyTorch", "RNN-T", "speech recognition", "en", "dataset:librispeech", "license:apache-2.0", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #k2 #icefall #transducer #librispeech #ASR #stateless transducer #PyTorch #RNN-T #speech recognition #en #dataset-librispeech #license-apache-2.0 #region-us
Introduction ============ This repo contains pre-trained model using <URL It is trained on full LibriSpeech dataset. Also, it uses the 'L' subset from GigaSpeech as extra training data. How to clone this repo ---------------------- Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later. The model in this repo is trained using the commit '2332ba312d7ce72f08c7bac1e3312f7e3dd722dc'. You can use to download 'icefall'. You can find the model information by visiting <URL In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward; the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2. The decoder architecture is modified from Rnn-Transducer with Stateless Prediction Network. A Conv1d layer is placed right after the input embedding layer. --- Description ----------- This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset using [icefall](URL). There are no RNNs in the decoder. The decoder is stateless and contains only an embedding layer and a Conv1d. The commands for training are: The tensorboard training log can be found at <URL The command for decoding is: You can find the decoding log for the above command in this repo (in the folder 'log'). The WERs for the test datasets are File description ================ * [log](URL), this directory contains the decoding log and decoding results * [test\_wavs](URL), this directory contains wave files for testing the pre-trained model * [data](URL), this directory contains files generated by <URL> * [exp](URL), this directory contains only one file: 'URL' 'exp/URL' is generated by the following command: HINT: To use 'URL' to compute the WER for test-clean and test-other, just do the following: and pass '--epoch 999 --avg 1' to 'transducer\_stateless\_multi\_datasets/URL'.
[]
[ "TAGS\n#k2 #icefall #transducer #librispeech #ASR #stateless transducer #PyTorch #RNN-T #speech recognition #en #dataset-librispeech #license-apache-2.0 #region-us \n" ]
[ 60 ]
[ "passage: TAGS\n#k2 #icefall #transducer #librispeech #ASR #stateless transducer #PyTorch #RNN-T #speech recognition #en #dataset-librispeech #license-apache-2.0 #region-us \n" ]
[ -0.06383673846721649, 0.10928380489349365, -0.007234589196741581, 0.05235107243061066, -0.013617466203868389, -0.04787064343690872, 0.14397504925727844, 0.08755352348089218, 0.12339134514331818, -0.07475242018699646, 0.18447791039943695, 0.16649699211120605, 0.015006504952907562, 0.08972150087356567, -0.005499931052327156, -0.27207478880882263, 0.03771843761205673, -0.06523624062538147, 0.02768099308013916, 0.09897862374782562, 0.07782378047704697, 0.005415977910161018, -0.006837237626314163, 0.019284622743725777, 0.026077914983034134, 0.027298033237457275, 0.04625777155160904, -0.11693038046360016, 0.12307829409837723, -0.016338223591446877, -0.01370417233556509, 0.0958828404545784, 0.019184835255146027, -0.18825173377990723, 0.021731054410338402, -0.03515755757689476, -0.05755608528852463, 0.032191820442676544, -0.012311664409935474, -0.02449170872569084, 0.022781435400247574, -0.06276869773864746, -0.015734972432255745, 0.05784756317734718, -0.030432358384132385, -0.23722290992736816, -0.028044672682881355, 0.10369934141635895, -0.03447456657886505, 0.022027235478162766, 0.003631148487329483, 0.14839264750480652, -0.1168379932641983, 0.10449028015136719, 0.06641534715890884, -0.3441197872161865, 0.05054883286356926, -0.009389100596308708, -0.015874184668064117, 0.08305826783180237, 0.0012050634250044823, 0.049367960542440414, 0.004133995622396469, -0.030407389625906944, -0.03979415446519852, -0.026273034512996674, -0.1335282325744629, 0.06239426136016846, -0.11267382651567459, 0.02115130051970482, 0.18995831906795502, -0.022819358855485916, 0.03494962304830551, 0.03362118825316429, -0.04124321788549423, -0.02464749664068222, 0.06788698583841324, 0.005917070899158716, -0.0009863493032753468, 0.07599521428346634, -0.03279581665992737, -0.07071486860513687, -0.08110107481479645, -0.03992684558033943, -0.14506396651268005, 0.13959719240665436, 0.014829132705926895, 0.10182486474514008, -0.12479282915592194, -0.008670518174767494, -0.010906199924647808, -0.11360354721546173, 0.011912094429135323, -0.07094273716211319, 0.04019530490040779, 0.09798784554004669, -0.03292367234826088, 0.03114907443523407, 0.1590505838394165, 0.17247051000595093, 0.07883824408054352, 0.012803126126527786, -0.04353862255811691, 0.12587924301624298, 0.02626929059624672, 0.04395899176597595, 0.03509999066591263, -0.009383091703057289, 0.03678654134273529, -0.05695854127407074, 0.08210653066635132, -0.06491635739803314, -0.08518802374601364, -0.029652327299118042, 0.0007303226739168167, 0.12614524364471436, 0.07568629086017609, -0.00868998747318983, -0.02399742417037487, 0.025724301114678383, 0.06868362426757812, -0.08440397679805756, 0.007368502207100391, 0.04523409157991409, 0.005521790124475956, 0.08182868361473083, 0.00943367276340723, 0.0699942484498024, -0.010708951391279697, 0.013633018359541893, -0.023496506735682487, 0.0014902260154485703, 0.056734491139650345, -0.0013342555612325668, 0.07336751371622086, -0.16461902856826782, 0.027974482625722885, -0.1673319935798645, -0.14570698142051697, 0.03862711042165756, -0.015249224379658699, -0.0033039345871657133, -0.05420326441526413, -0.0502631776034832, -0.0346265584230423, 0.03338634967803955, -0.09465475380420685, -0.11548414081335068, -0.057220470160245895, 0.03745858743786812, -0.10069489479064941, 0.10636470466852188, -0.174094557762146, 0.006645421963185072, -0.1335250437259674, -0.02767690271139145, 0.0828544870018959, 0.05642374977469444, -0.15360940992832184, 0.06354877352714539, -0.05998864024877548, -0.025411663576960564, -0.07998380810022354, -0.036626193672418594, -0.00012705964036285877, 0.13209623098373413, -0.2973765730857849, -0.09720461070537567, 0.15061113238334656, -0.14727330207824707, -0.013910443522036076, 0.13978064060211182, 0.03583008423447609, -0.05290021002292633, 0.0809539258480072, 0.32908040285110474, -0.02669154480099678, -0.03706354647874832, -0.035011496394872665, 0.12676000595092773, -0.1539485901594162, -0.1114703118801117, 0.08442046493291855, -0.04972956329584122, 0.011421691626310349, 0.01475114282220602, 0.09025336056947708, 0.10513971745967865, -0.016642607748508453, -0.11816781759262085, -0.07015377283096313, -0.06325346976518631, 0.06908552348613739, -0.03085336647927761, 0.00017471704632043839, -0.07647177577018738, -0.002812117338180542, 0.029956970363855362, 0.020288199186325073, 0.03733455017209053, 0.07814943045377731, -0.08468569070100784, 0.007884971797466278, 0.07875169813632965, 0.004136864561587572, -0.07424233853816986, 0.052879251539707184, -0.03117980621755123, -0.017686888575553894, 0.10106021165847778, 0.0586705282330513, 0.0763435885310173, -0.053537845611572266, -0.02206852100789547, 0.010453728958964348, 0.08113359659910202, 0.07039126008749008, 0.010098753497004509, -0.2281980961561203, 0.07841155678033829, -0.04415570944547653, -0.012529093772172928, 0.005883242469280958, 0.01653343439102173, 0.1373090147972107, 0.06356740742921829, -0.02558819204568863, 0.04956354200839996, -0.033013422042131424, -0.006804578006267548, 0.01447339728474617, 0.013948456384241581, 0.08944689482450485, 0.007511379662901163, -0.1434481143951416, 0.23762771487236023, -0.02063487097620964, 0.05827820301055908, 0.16489127278327942, -0.05455051362514496, 0.1187986508011818, 0.061117447912693024, -0.031928837299346924, -0.056505683809518814, 0.10163718461990356, 0.031415026634931564, 0.12988026440143585, -0.0021414789371192455, 0.06356602907180786, -0.019012806937098503, -0.03869553655385971, -0.007663022726774216, -0.06751442700624466, -0.03944137692451477, 0.11198440194129944, -0.007330853492021561, -0.16084474325180054, 0.0852813795208931, 0.21986934542655945, -0.07926451414823532, 0.20620208978652954, -0.07047122716903687, -0.03434964269399643, -0.018637273460626602, -0.062436170876026154, -0.025717560201883316, 0.09926514327526093, -0.07218924909830093, -0.0042004757560789585, 0.07041069120168686, -0.00838548131287098, 0.033561401069164276, -0.10604636371135712, -0.054019127041101456, -0.033738017082214355, -0.023134242743253708, -0.1980949193239212, 0.0655890703201294, -0.07914240658283234, 0.053586144000291824, -0.048078425228595734, -0.1796693205833435, 0.09582934528589249, -0.04113204777240753, -0.04829468950629234, 0.09672339260578156, -0.18099553883075714, -0.20031633973121643, -0.08144374191761017, 0.051472458988428116, -0.0250119436532259, -0.0438435859978199, 0.10688114166259766, -0.154110848903656, 0.003085094504058361, -0.002262434922158718, 0.018510492518544197, -0.03562604635953903, 0.03346918895840645, -0.011853564530611038, -0.0015507154166698456, -0.023377999663352966, -0.12879124283790588, -0.01612759195268154, -0.05924277380108833, 0.021308278664946556, 0.1307649463415146, -0.008685196749866009, 0.011266396380960941, 0.1347532421350479, 0.0255938358604908, 0.012470548041164875, -0.025856995955109596, 0.07712766528129578, -0.0063236732967197895, 0.005406167823821306, 0.09370675683021545, -0.08147824555635452, 0.06762120872735977, 0.09689605236053467, 0.03078939951956272, -0.012301148846745491, -0.0036492827348411083, -0.0249185748398304, -0.11072329431772232, -0.25078582763671875, -0.08511092513799667, -0.0353110209107399, 0.16134227812290192, -0.015982134267687798, 0.06407089531421661, 0.07735896110534668, 0.005045383237302303, 0.07892639935016632, -0.11263807862997055, 0.01914583146572113, -0.0008859268855303526, 0.2221902310848236, -0.04734035208821297, 0.11230625957250595, -0.06481903791427612, -0.02045087330043316, 0.1211727187037468, 0.09284618496894836, 0.13620537519454956, 0.1136825829744339, 0.11566207557916641, 0.0721355676651001, 0.19752515852451324, 0.09913990646600723, 0.07909317314624786, 0.05828459560871124, 0.03441697731614113, 0.012093550525605679, -0.05020944029092789, -0.015524584800004959, 0.06317158788442612, 0.10492084920406342, -0.1031675636768341, 0.10075171291828156, 0.0346064418554306, 0.08077337592840195, 0.22613804042339325, 0.061994485557079315, -0.07376567274332047, 0.014972859062254429, 0.03424295783042908, -0.03065166249871254, 0.028094742447137833, 0.14542467892169952, 0.028966596350073814, -0.012762444093823433, 0.07594116777181625, -0.02191336452960968, 0.022266194224357605, -0.011959102004766464, 0.030802441760897636, -0.04243350774049759, -0.11124362051486969, 0.050925612449645996, 0.05514153838157654, -0.22515079379081726, 0.20457223057746887, -0.015622676350176334, -0.016076190397143364, -0.03274664282798767, -0.031808532774448395, 0.05455966293811798, 0.07981669902801514, 0.13152150809764862, 0.06594642996788025, -0.09834402799606323, 0.011795088648796082, -0.18484456837177277, 0.0789112001657486, 0.09462329000234604, 0.05038155987858772, -0.10649733245372772, -0.02351195365190506, 0.025850627571344376, 0.05798051878809929, -0.0828981101512909, -0.21324555575847626, -0.06599409878253937, 0.051851026713848114, 0.19234022498130798, 0.016267433762550354, -0.020522311329841614, -0.1450522243976593, -0.07479250431060791, 0.09856642782688141, -0.13939440250396729, -0.08292295783758163, -0.06182308495044708, -0.12789562344551086, 0.13508833944797516, -0.06088065356016159, 0.028409719467163086, 0.0019550109282135963, -0.066381074488163, -0.051733843982219696, -0.17846468091011047, 0.11098723113536835, -0.07848584651947021, -0.12493202835321426, 0.023249901831150055, 0.12574993073940277, 0.05630917102098465, 0.07449319958686829, 0.0006226514233276248, 0.09745419025421143, -0.0933104157447815, -0.08482041954994202, 0.10120019316673279, -0.02326878532767296, -0.07510068267583847, 0.0026262281462550163, 0.017197731882333755, -0.1086815893650055, -0.029801541939377785, -0.029741330072283745, 0.26513946056365967, 0.2669987380504608, -0.0909537523984909, 0.2705761194229126, 0.26193711161613464, -0.043350957334041595, -0.27131420373916626, -0.15351715683937073, -0.1365477591753006, -0.07271701097488403, 0.01663065329194069, -0.1932041049003601, 0.06609666347503662, 0.0940566286444664, -0.08955689519643784, 0.103122279047966, -0.23574478924274445, -0.10695315897464752, 0.23966282606124878, -0.11636923253536224, 0.2810322940349579, -0.11508533358573914, -0.09393299371004105, -0.06432218849658966, -0.012674026191234589, 0.16043274104595184, -0.14283804595470428, 0.06503663957118988, -0.02338833548128605, 0.01232925709336996, 0.004681346472352743, -0.03640098497271538, 0.10726048052310944, 0.0787605568766594, -0.045639291405677795, -0.05339784175157547, -0.0328964926302433, 0.0667746514081955, -0.0005917632952332497, 0.05452211946249008, -0.14563657343387604, 0.06152821704745293, -0.030606089159846306, -0.01654770039021969, -0.1172460988163948, 0.11825813353061676, 0.03410654515028, -0.09356779605150223, -0.0021695420145988464, -0.033174045383930206, 0.02261686511337757, -0.004465071484446526, 0.19595251977443695, -0.016001712530851364, -0.05549061298370361, 0.13084469735622406, 0.09578170627355576, -0.047929778695106506, 0.03782368451356888, -0.07049311697483063, -0.09529798477888107, 0.06894277036190033, -0.08143354952335358, 0.043276987969875336, 0.09312831610441208, 0.05156025290489197, -0.012010463513433933, 0.07226061075925827, -0.055778197944164276, -0.00023127463646233082, 0.08877293765544891, -0.15551337599754333, -0.052338287234306335, 0.01868550106883049, 0.032458771020174026, 0.1370583176612854, 0.16822269558906555, 0.1889338493347168, -0.02341107465326786, -0.015031300485134125, -0.01472497545182705, 0.07509530335664749, -0.09525543451309204, 0.044571653008461, 0.15734584629535675, -0.00654529919847846, -0.135562002658844, 0.13359549641609192, -0.06466805189847946, -0.016468582674860954, 0.04379289597272873, 0.05565175786614418, -0.10707290470600128, -0.1104918122291565, -0.09491173177957535, -0.05776533484458923, -0.12917909026145935, -0.11286789178848267, -0.007838161662220955, -0.0813290923833847, 0.026621365919709206, 0.1753275841474533, 0.04951454699039459, 0.07509913295507431, -0.04569308087229729, -0.06147531792521477, 0.06482944637537003, 0.058677203953266144, -0.1234220415353775, -0.021951917558908463, -0.08187119662761688, -0.020893801003694534, -0.01740695908665657, 0.08591844886541367, -0.06704383343458176, -0.030348528176546097, -0.09554983675479889, 0.034303516149520874, -0.047013744711875916, -0.032984066754579544, -0.04857245087623596, 0.022611942142248154, 0.031791508197784424, -0.08699704706668854, -0.014085481874644756, 0.0653146356344223, -0.09223891794681549, -0.003986412193626165, -0.0048046233132481575, 0.04389054328203201, -0.1976051926612854, -0.0456167533993721, 0.058134954422712326, 0.021636705845594406, 0.16555781662464142, 0.1256776750087738, -0.07660824805498123, 0.11742395162582397, -0.09869498759508133, -0.02244754135608673, 0.1117674931883812, 0.036437567323446274, 0.002997276373207569, 0.044210243970155716, -0.0440310463309288, 0.05848920717835426, 0.03338479623198509, 0.07383851706981659, 0.13101662695407867, -0.07768823951482773, 0.030128369107842445, -0.042837031185626984, -0.12950456142425537, -0.032028310000896454, -0.10896497964859009, 0.05543706566095352, 0.10185205191373825, 0.16718822717666626, -0.03868614509701729, -0.016639648005366325, 0.016077576205134392, 0.02386396750807762, 0.02848334051668644, -0.087058424949646, -0.044038332998752594, 0.03010496310889721, -0.018386367708444595, -0.035880256444215775, 0.21487922966480255, -0.026319758966565132, -0.17064960300922394, 0.044613003730773926, -0.020065510645508766, -0.12587395310401917, 0.03674069419503212, 0.2626139521598816, 0.038719795644283295, -0.012884745374321938, -0.08493516594171524, -0.01828238181769848, -0.05413510650396347, -0.09558413922786713, 0.06751565635204315, 0.06801538169384003, 0.09223853051662445, 0.06254774332046509, 0.12975148856639862, -0.025546874850988388, -0.05470895767211914, -0.05177507549524307, -0.05686228722333908, 0.06639227271080017, 0.027699220925569534, 0.209966778755188, 0.12033633142709732, 0.028490228578448296, -0.006040135398507118, -0.023606663569808006, -0.060566723346710205, -0.18969568610191345, -0.12327016890048981, -0.03732560947537422, -0.07416074723005295, 0.0384359285235405, -0.030362924560904503, 0.05626482516527176, 0.12329961359500885, 0.07700580358505249, -0.0009011514484882355, -0.06464418023824692, -0.05121458321809769, -0.0001124790869653225, 0.0348195806145668, -0.057012204080820084, -0.0755947157740593, -0.10105768591165543, -0.04394053667783737, 0.015768691897392273, 0.03133382648229599, -0.07752838730812073, -0.008578682318329811, -0.07873491197824478, -0.017232496291399002, -0.10239437222480774, -0.050892915576696396, -0.03472457453608513, 0.024232419207692146, 0.043623678386211395, 0.09894315898418427, 0.10073515772819519, -0.03718595206737518, 0.08316752314567566, 0.11457402259111404, -0.031032219529151917, -0.18016408383846283, -0.012126639485359192, 0.07565637677907944, -0.05338020995259285, 0.10025064647197723, -0.03842099756002426, -0.0558636412024498, -0.05840787664055824, 0.07573410868644714, 0.2607494592666626, -0.0174150038510561, 0.04491695761680603, -0.04660075157880783, 0.024686288088560104, 0.002875139005482197, 0.018702974542975426, 0.13356642425060272, 0.12520143389701843, -0.06304773688316345, -0.05177099257707596, -0.05929160863161087, -0.019041750580072403, -0.12145159393548965, 0.011404255405068398, -0.06015898287296295, -0.0897805467247963, -0.07275954633951187, 0.12121274322271347, -0.1598210483789444, 0.058650076389312744, -0.055437929928302765, -0.1029277965426445, -0.03295855224132538, -0.009737720713019371, 0.1659109741449356, 0.033398956060409546, 0.04384378343820572, -0.03790423646569252, -0.08001556992530823, -0.011370447464287281, 0.037753403186798096, -0.25237637758255005, 0.03721905127167702, 0.03265601396560669, -0.07988566160202026, 0.09082521498203278, -0.016132451593875885, 0.16533517837524414, 0.07192424684762955, 0.06218798831105232, -0.05425003916025162, 0.08918782323598862, 0.016535332426428795, -0.05450943857431412, -0.08191263675689697, -0.004588240757584572, 0.032828230410814285, -0.005491178948432207, 0.0703296959400177, 0.06606639176607132, 0.02495018020272255, 0.0801570937037468, -0.02230537123978138, -0.05728645995259285, 0.05666109174489975, -0.05386669188737869, 0.03400818258523941, -0.03569469600915909, -0.03584764152765274, -0.04073994234204292, -0.055036261677742004, -0.05986405164003372, 0.009979198686778545, -0.20742474496364594, -0.052534062415361404, 0.040035542100667953, -0.022007115185260773, 0.009698151610791683, 0.04001221805810928, -0.057599470019340515, -0.04128093644976616, -0.16891920566558838, 0.008959856815636158, -0.0872000902891159, 0.00420183502137661, 0.02597952075302601, -0.00715241115540266, 0.04049311578273773, -0.1003064215183258, 0.0926964282989502, -0.0023711728863418102, -0.15293952822685242, -0.06006823107600212 ]
null
null
null
## Pre-trained TDNN models for the yesno dataset with icefall. Refer to <https://github.com/k2-fsa/icefall/tree/master/egs/yesno/ASR> for more information about this pre-trained model. You can find usage instructions there. ## Sound files for testing the pre-trained model The folder `test_waves` contains test sound files. They are downloaded from <https://www.openslr.org/1/>. There are 60 files in the dataset, 30 are used for training. The remaining 30 files, contained in `test_waves` are kept for testing. The code for splitting the dataset can be found at <https://github.com/lhotse-speech/lhotse/blob/master/lhotse/recipes/yesno.py#L138> ```python wave_files = list(corpus_dir.glob("*.wav")) assert len(wave_files) == 60 wave_files.sort() train_set = wave_files[::2] test_set = wave_files[1::2] assert len(train_set) == 30 assert len(test_set) == 30 ```
{}
null
csukuangfj/icefall_asr_yesno_tdnn
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
## Pre-trained TDNN models for the yesno dataset with icefall. Refer to <URL for more information about this pre-trained model. You can find usage instructions there. ## Sound files for testing the pre-trained model The folder 'test_waves' contains test sound files. They are downloaded from <URL There are 60 files in the dataset, 30 are used for training. The remaining 30 files, contained in 'test_waves' are kept for testing. The code for splitting the dataset can be found at <URL
[ "## Pre-trained TDNN models for the yesno dataset with icefall.\n\nRefer to <URL\nfor more information about this pre-trained model.\n\nYou can find usage instructions there.", "## Sound files for testing the pre-trained model\n\nThe folder 'test_waves' contains test sound files. They\nare downloaded from <URL\n\nThere are 60 files in the dataset, 30 are used for training.\nThe remaining 30 files, contained in 'test_waves' are kept for testing.\n\nThe code for splitting the dataset can be found at\n<URL" ]
[ "TAGS\n#region-us \n", "## Pre-trained TDNN models for the yesno dataset with icefall.\n\nRefer to <URL\nfor more information about this pre-trained model.\n\nYou can find usage instructions there.", "## Sound files for testing the pre-trained model\n\nThe folder 'test_waves' contains test sound files. They\nare downloaded from <URL\n\nThere are 60 files in the dataset, 30 are used for training.\nThe remaining 30 files, contained in 'test_waves' are kept for testing.\n\nThe code for splitting the dataset can be found at\n<URL" ]
[ 6, 41, 81 ]
[ "passage: TAGS\n#region-us \n## Pre-trained TDNN models for the yesno dataset with icefall.\n\nRefer to <URL\nfor more information about this pre-trained model.\n\nYou can find usage instructions there.## Sound files for testing the pre-trained model\n\nThe folder 'test_waves' contains test sound files. They\nare downloaded from <URL\n\nThere are 60 files in the dataset, 30 are used for training.\nThe remaining 30 files, contained in 'test_waves' are kept for testing.\n\nThe code for splitting the dataset can be found at\n<URL" ]
[ -0.1880803406238556, 0.03516754135489464, 0.0037605923134833574, 0.08916329592466354, 0.09411059319972992, -0.0985790342092514, 0.13371308147907257, -0.038229405879974365, 0.045969851315021515, 0.05418291687965393, 0.04140839725732803, -0.01079460047185421, -0.0759991854429245, 0.10129813104867935, 0.051073409616947174, -0.2129470705986023, 0.025276580825448036, -0.06405456364154816, -0.12554752826690674, 0.09399820864200592, 0.016470279544591904, -0.05874006077647209, -0.08588943630456924, 0.01319935917854309, -0.03758620098233223, -0.004969341214746237, -0.1254567950963974, -0.00523633835837245, 0.045211952179670334, -0.14443345367908478, 0.10328345000743866, 0.1271306425333023, 0.06523473560810089, -0.13005033135414124, 0.049243975430727005, 0.031092120334506035, 0.030909419059753418, 0.02750231884419918, 0.09619278460741043, 0.017562050372362137, -0.06475088745355606, 0.09874220192432404, 0.02966049499809742, 0.006916831247508526, -0.04515824466943741, -0.07976207882165909, 0.019393568858504295, -0.19507046043872833, 0.10158983618021011, 0.16218900680541992, -0.08040091395378113, 0.21530915796756744, -0.17462323606014252, 0.11686546355485916, 0.10638821870088577, -0.053615741431713104, -0.03134378045797348, 0.06319116801023483, 0.06173056364059448, 0.09945931285619736, 0.022796690464019775, 0.07296309620141983, 0.1229342669248581, -0.0019071371061727405, -0.19702036678791046, 0.06646496057510376, 0.03056703694164753, -0.011435262858867645, -0.17553530633449554, 0.05971093848347664, 0.28510934114456177, 0.04556708037853241, -0.07942427694797516, 0.04922850802540779, -0.006346989888697863, -0.11432261765003204, 0.0782657116651535, 0.009105861186981201, 0.027147209271788597, -0.0238703154027462, -0.14547686278820038, -0.016239160671830177, -0.12616929411888123, -0.07100939750671387, 0.04357782378792763, 0.14047232270240784, -0.0725814551115036, 0.030603397637605667, -0.12209223955869675, 0.04829442873597145, -0.05621516704559326, -0.05323147401213646, -0.037545859813690186, -0.056509748101234436, -0.06067828834056854, -0.030005771666765213, -0.1612614244222641, -0.15554024279117584, 0.12327475100755692, 0.17358329892158508, 0.10815785825252533, 0.024381788447499275, -0.05290433391928673, 0.08120643347501755, 0.09849200397729874, -0.04396292567253113, 0.08815259486436844, -0.11450499296188354, -0.045963354408741, -0.06195472180843353, -0.01931038498878479, -0.10564753413200378, -0.1013089194893837, -0.0666094496846199, 0.06501788645982742, 0.023261724039912224, 0.11363302916288376, -0.1631072610616684, 0.004912975709885359, -0.06561914086341858, 0.03409762308001518, 0.013484282419085503, -0.0028537895996123552, 0.036633506417274475, 0.02478090114891529, -0.05373018980026245, -0.003472196636721492, 0.1531943380832672, 0.06860754638910294, 0.011965225450694561, -0.021414926275610924, 0.05217456817626953, -0.053679611533880234, -0.07200586795806885, 0.041585225611925125, -0.0997256487607956, 0.02167610079050064, -0.08497106283903122, -0.13500086963176727, -0.12684567272663116, 0.060020171105861664, 0.05123390629887581, -0.01448065135627985, -0.09673312306404114, -0.03199588879942894, -0.06993341445922852, -0.016200626268982887, 0.07634968310594559, -0.06432810425758362, -0.004888932686299086, -0.0854637399315834, 0.08704666048288345, -0.043007053434848785, -0.025861017405986786, -0.08532154560089111, -0.015161080285906792, 0.078467957675457, 0.09866147488355637, -0.11619049310684204, -0.052058856934309006, -0.018256470561027527, -0.09885408729314804, -0.17820176482200623, 0.035835590213537216, 0.16481859982013702, 0.20083700120449066, -0.36362218856811523, -0.007617845665663481, -0.06635487824678421, -0.1795448213815689, 0.07191967219114304, 0.10408735275268555, -0.05125323310494423, 0.05859067291021347, 0.04766402393579483, 0.19982051849365234, 0.01621687412261963, -0.07900933176279068, 0.10755137354135513, -0.10098730772733688, 0.025282809510827065, -0.17718638479709625, 0.07351180166006088, -0.054028116166591644, -0.20390821993350983, 0.001011505490168929, 0.023509813472628593, -0.008746406994760036, -0.10207290202379227, -0.05277483910322189, 0.0026974030770361423, -0.06470954418182373, -0.12710054218769073, 0.03374272212386131, -0.019253689795732498, -0.10906028002500534, -0.01989438384771347, -0.009530546143651009, -0.006218917667865753, -0.02464672364294529, 0.06190510839223862, 0.07517065852880478, 0.15349486470222473, -0.20044094324111938, -0.1382375806570053, -0.13257543742656708, -0.052933841943740845, -0.027448363602161407, 0.10647639632225037, 0.07040469348430634, 0.1413593739271164, 0.09919121116399765, 0.07920893281698227, -0.0024474537931382656, 0.028995700180530548, -0.007044538855552673, 0.054056670516729355, -0.00548539636656642, -0.15622587502002716, 0.061291586607694626, -0.07371541857719421, 0.10046681016683578, -0.1724107414484024, -0.053808823227882385, 0.029040802270174026, 0.0698317363858223, -0.046113040298223495, -0.09723630547523499, 0.09477979689836502, 0.034482475370168686, -0.0635899007320404, -0.004778503440320492, 0.006865119561553001, 0.034035440534353256, -0.09058579057455063, 0.026914848014712334, 0.06114351004362106, -0.13840840756893158, 0.057356785982847214, 0.059259843081235886, -0.05729283019900322, -0.06367228180170059, -0.11714624613523483, -0.03644217923283577, -0.042505886405706406, 0.18368352949619293, -0.03979511931538582, -0.023941513150930405, 0.01548993494361639, -0.03459645062685013, 0.05124691128730774, 0.06690623611211777, -0.0793534442782402, -0.028981804847717285, 0.11116449534893036, 0.08603693544864655, -0.07380107790231705, -0.0060881213285028934, -0.01457975897938013, -0.17614060640335083, 0.23533286154270172, -0.022432640194892883, -0.16989192366600037, -0.043786972761154175, 0.042748309671878815, 0.008515702560544014, 0.16809335350990295, 0.04763534665107727, 0.186717689037323, 0.011754332110285759, 0.04757480323314667, 0.12989182770252228, -0.11772117763757706, -0.0543564073741436, 0.06520412117242813, 0.08714562654495239, -0.2310389131307602, 0.016454454511404037, -0.1555488407611847, -0.022673429921269417, -0.08689798414707184, -0.10128414630889893, 0.023664414882659912, 0.010486166924238205, -0.050046712160110474, 0.1622432917356491, -0.09719298779964447, -0.091519296169281, -0.13970860838890076, 0.16134384274482727, -0.018023362383246422, -0.03349437192082405, 0.13703522086143494, -0.24961186945438385, -0.03376070410013199, -0.06807844340801239, -0.03244219347834587, -0.07880979776382446, 0.025089535862207413, -0.007128856610506773, -0.022533245384693146, -0.05777993053197861, -0.05706455931067467, 0.07653402537107468, -0.08401590585708618, 0.08929071575403214, 0.054533835500478745, -0.07880259305238724, 0.07826261222362518, 0.12880447506904602, 0.024671407416462898, 0.022014161571860313, 0.0667610689997673, 0.20953930914402008, -0.10072354972362518, 0.06827656179666519, 0.008020239882171154, -0.05179929733276367, -0.0038393139839172363, -0.014092104509472847, -0.0017098176758736372, -0.0670737773180008, 0.017380382865667343, -0.021422309800982475, -0.15908962488174438, -0.2667270600795746, -0.0895099937915802, -0.015691068023443222, -0.04498283937573433, -0.024022771045565605, 0.07066603749990463, 0.09906835854053497, 0.08853377401828766, 0.11569865792989731, -0.25246796011924744, -0.0473039485514164, 0.049667470157146454, -0.07171810418367386, 0.035767633467912674, 0.10974135994911194, -0.034038037061691284, 0.04686891660094261, 0.044859398156404495, 0.00836525484919548, 0.18976840376853943, 0.05605759099125862, 0.030783560127019882, 0.07689478248357773, -0.020121131092309952, 0.10279682278633118, 0.16104456782341003, 0.06945963203907013, 0.017502587288618088, 0.030026348307728767, -0.0638400986790657, 0.06764572113752365, 0.0670454204082489, 0.16118542850017548, -0.1364947408437729, 0.08045457303524017, 0.08626628667116165, -0.12928782403469086, 0.09292671084403992, 0.08071666955947876, -0.21918855607509613, -0.07121524959802628, -0.055910952389240265, 0.07025707513093948, -0.04964647814631462, 0.14251098036766052, 0.11859282851219177, 0.050527773797512054, 0.013012445531785488, -0.013436323963105679, 0.12219846248626709, 0.11537163704633713, -0.05677303671836853, 0.03342222794890404, 0.0616806223988533, -0.05917765572667122, -0.003161888336762786, -0.040345776826143265, 0.08576057106256485, -0.05140453949570656, 0.020687052980065346, 0.023112133145332336, -0.04934481531381607, -0.08199036866426468, 0.10058603435754776, 0.14929983019828796, 0.026758620515465736, -0.11579558253288269, 0.0778113454580307, -0.18934635818004608, -0.016071220859885216, 0.07397694885730743, -0.019250618293881416, -0.012560996226966381, 0.09078111499547958, 0.019789038226008415, 0.04325196519494057, -0.08889564126729965, -0.24859343469142914, -0.008485395461320877, -0.01316903904080391, 0.08529143780469894, 0.015116185881197453, -0.0010972145246341825, -0.2105419933795929, -0.08042071014642715, 0.12412706762552261, -0.04023183137178421, 0.04051656275987625, -0.08928649127483368, -0.022309886291623116, 0.26249682903289795, 0.0955180898308754, 0.05500561371445656, 0.016650522127747536, 0.007031741086393595, -0.04914754629135132, 0.037465110421180725, 0.04465643689036369, -0.13681164383888245, -0.07299251854419708, -0.06772947311401367, 0.09234361350536346, 0.10216247290372849, 0.02191145345568657, -0.03767137601971626, 0.1474452167749405, -0.11260522902011871, -0.009633632376790047, 0.01889931410551071, -0.06072483956813812, -0.03360285237431526, 0.07160796225070953, 0.03605387732386589, 0.10475824773311615, -0.048862241208553314, -0.061500951647758484, 0.09422484040260315, 0.23554688692092896, -0.03771080821752548, 0.06876280158758163, 0.1897028088569641, -0.019002219662070274, -0.13525304198265076, 0.191705122590065, 0.08961044996976852, 0.07430620491504669, -0.020162593573331833, -0.14702066779136658, 0.17199602723121643, 0.04296516999602318, 0.004227552097290754, 0.22108587622642517, 0.07901644706726074, -0.13058985769748688, 0.14348483085632324, -0.004469249863177538, 0.2310977280139923, -0.01216872502118349, -0.014167473651468754, -0.018301136791706085, 0.1286603957414627, 0.11279963701963425, -0.11537078022956848, 0.09106860309839249, 0.07014113664627075, -0.0375753678381443, 0.0597432516515255, -0.11299272626638412, 0.09872599691152573, 0.046176739037036896, -0.03847417235374451, -0.008420834317803383, 0.13115547597408295, 0.05301862210035324, -0.027423234656453133, 0.04892246797680855, 0.1281399428844452, 0.08195583522319794, 0.10591867566108704, -0.1381068229675293, 0.005175342317670584, 0.10645557194948196, -0.04958011209964752, -0.10977187007665634, 0.004706472624093294, 0.05425455793738365, 0.0029812308494001627, -0.01708311401307583, 0.11712510883808136, -0.06945076584815979, 0.05598149821162224, 0.14037470519542694, 0.20128558576107025, 0.19222602248191833, -0.03411080315709114, 0.058024872094392776, -0.11026100069284439, 0.07553450763225555, -0.07116461545228958, -0.05156226083636284, 0.07755260169506073, 0.11373381316661835, -0.10926572233438492, 0.038273558020591736, -0.11769787967205048, -0.07355311512947083, 0.06193762645125389, -0.09158773720264435, -0.15664330124855042, -0.0717846155166626, -0.18077850341796875, 0.12368230521678925, 0.10038097947835922, 0.0938948392868042, -0.10368628054857254, 0.05702895671129227, -0.012828623875975609, -0.0023148648906499147, -0.0397268570959568, 0.1345454752445221, 0.1529543101787567, 0.0269547738134861, -0.0898466631770134, 0.013664951547980309, -0.06783914566040039, -0.010830731131136417, 0.09350378066301346, -0.016349591314792633, -0.046742185950279236, -0.07364950329065323, 0.05130678787827492, 0.14223171770572662, -0.017420535907149315, -0.12675057351589203, -0.12660536170005798, 0.1294516623020172, 0.028934458270668983, 0.0719735324382782, 0.02500183694064617, -0.06526338309049606, -0.04762294143438339, 0.02275496907532215, -0.11612140387296677, 0.05885294824838638, 0.012937248684465885, 0.08760103583335876, -0.05749513581395149, -0.07170621305704117, -0.061339084059000015, 0.003962191753089428, -0.06954706460237503, -0.09226793050765991, 0.0073660556226968765, 0.0627317950129509, -0.06898543983697891, -0.028346531093120575, 0.03314107283949852, 0.016234194859862328, 0.021220538765192032, -0.09065773338079453, -0.07539457827806473, 0.09454125910997391, -0.08111590147018433, 0.09823060780763626, -0.04252798855304718, -0.0863637924194336, -0.05018015578389168, 0.03791600093245506, -0.011547812260687351, -0.053234945982694626, 0.0000036145959256828064, 0.0360046848654747, -0.024569328874349594, 0.037071771919727325, -0.05516587197780609, 0.0641762763261795, 0.0377277173101902, -0.008358045481145382, -0.00594691326841712, -0.010121333412826061, -0.04464638978242874, 0.03455571457743645, 0.09581062197685242, -0.04050808027386665, 0.115634486079216, -0.023880261927843094, -0.123137928545475, -0.003963941242545843, -0.12060756236314774, -0.05525849014520645, -0.04438696429133415, 0.001304642646573484, 0.23423050343990326, 0.17397119104862213, 0.01649363338947296, 0.1338963806629181, 0.0014619898283854127, -0.07445178925991058, 0.04637031629681587, 0.011198535561561584, -0.06354699283838272, 0.02441956289112568, 0.07738718390464783, -0.06741291284561157, 0.1058025062084198, -0.05187036842107773, 0.0721098855137825, -0.08073976635932922, -0.012541322037577629, 0.007985328324139118, 0.011652573011815548, 0.2823854386806488, 0.06548558175563812, 0.05258315056562424, -0.14060966670513153, 0.03982189670205116, 0.008712248876690865, -0.0011668190127238631, 0.044497568160295486, 0.11069953441619873, -0.02619631215929985, 0.03823019936680794, 0.05197237432003021, -0.04645945131778717, -0.05290082097053528, 0.04686750844120979, -0.04828691855072975, 0.04973731189966202, 0.0652984157204628, 0.03169749304652214, 0.05563609302043915, -0.025718389078974724, -0.0008987401961348951, 0.0228491872549057, -0.05863450467586517, -0.1766359657049179, -0.08641547709703445, -0.08577493578195572, -0.05892852321267128, 0.11976355314254761, -0.05838410183787346, -0.015348201617598534, -0.018019163981080055, 0.11874111741781235, 0.034453075379133224, 0.20500612258911133, 0.10804431885480881, -0.017237462103366852, -0.023363828659057617, -0.06590338796377182, -0.027652887627482414, -0.027114208787679672, -0.12261044979095459, 0.09802374988794327, -0.12005537003278732, -0.037191472947597504, -0.04824597388505936, 0.008904218673706055, 0.040346112102270126, 0.05049528554081917, -0.06990770995616913, 0.015580754727125168, 0.0403723306953907, 0.04036719724535942, 0.03236125037074089, 0.056287918239831924, -0.051328882575035095, 0.06142207607626915, 0.01002642884850502, 0.04683253541588783, -0.15586970746517181, -0.08769048750400543, 0.1915932446718216, -0.204417884349823, 0.0751795694231987, -0.041238024830818176, -0.09567756205797195, -0.01619727723300457, 0.023160217329859734, 0.13313907384872437, 0.00657283142209053, -0.06472329795360565, 0.11280492693185806, -0.023473823443055153, -0.12871691584587097, 0.04598352685570717, -0.03098123148083687, 0.09579265862703323, -0.020874805748462677, -0.0641927421092987, -0.1415039300918579, -0.07586441934108734, -0.005609135143458843, 0.02936459332704544, 0.03622617945075035, -0.11949607729911804, -0.1379450559616089, 0.2502917945384979, -0.10771580785512924, -0.06893083453178406, 0.04288943484425545, -0.050828780978918076, -0.1614222675561905, -0.10655736923217773, 0.05694383382797241, -0.09830833971500397, 0.0017199902795255184, -0.09569007903337479, 0.005520959384739399, -0.1378660500049591, -0.027344709262251854, -0.002576518803834915, 0.04759492725133896, 0.006811853963881731, 0.0006127515225671232, -0.046690210700035095, 0.0036663522478193045, 0.19717122614383698, 0.054321348667144775, 0.07485773414373398, -0.01776333712041378, 0.10638520121574402, 0.0383731983602047, -0.10454320162534714, -0.05723842605948448, 0.10190889239311218, -0.05262109637260437, 0.02690570428967476, 0.0315016470849514, -0.13123682141304016, 0.027081554755568504, 0.040012557059526443, -0.014203151687979698, -0.10656747967004776, 0.11120550334453583, -0.07360580563545227, 0.089621901512146, -0.04234406352043152, -0.012779094278812408, -0.0547390878200531, -0.09301595389842987, 0.11870621144771576, 0.021647432819008827, -0.06330536305904388, 0.030716950073838234, -0.10112572461366653, -0.05946144834160805, 0.040822140872478485, -0.01488802582025528, -0.1946020871400833, -0.020478660240769386, -0.01247361395508051, -0.08905453979969025, 0.08287765085697174, -0.044822197407484055, 0.08992630243301392, -0.006076700985431671, 0.008891244418919086, 0.14491501450538635, 0.05614021420478821, 0.021667158231139183, -0.1251002699136734, -0.08228767663240433 ]
null
null
null
See https://colab.research.google.com/drive/14MozS-9jWD3XQ0o-dZ-meqnblgHs70P2?usp=sharing
{}
null
csukuangfj/test-data-for-optimized-transducer
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
See URL
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
null
# Introduction This repo contains the benchmark results for <https://github.com/csukuangfj/transducer-loss-benchmarking> ## Usage First, install `git-lfs`. Second, use the following command to clone this repo: ```bash git lfs install git clone https://huggingface.co/csukuangfj/transducer-loss-benchmarking ``` **Caution**: You have to run `git lfs install` first. Otherwise, you will be **SAD** later. Third, ``` pip install torch-tb-profiler cd transducer-loss-benchmarking tensorboard --logdir ./log/torchaudio-30 --port 6006 tensorboard --logdir ./log/optimized_transducer-30 --port 6007 ``` Fourth, open your browser and go to - <http://localhost:6006/#pytorch_profiler> - <http://localhost:6006/#pytorch_profiler> You will see the following images: ![](./torchaudio-30.png) ![](./optimized_transducer-30.png)
{}
null
csukuangfj/transducer-loss-benchmarking
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
# Introduction This repo contains the benchmark results for <URL ## Usage First, install 'git-lfs'. Second, use the following command to clone this repo: Caution: You have to run 'git lfs install' first. Otherwise, you will be SAD later. Third, Fourth, open your browser and go to - <http://localhost:6006/#pytorch_profiler> - <http://localhost:6006/#pytorch_profiler> You will see the following images: ![](./URL) ![](./optimized_transducer-URL)
[ "# Introduction\n\nThis repo contains the benchmark results for <URL", "## Usage\n\nFirst, install 'git-lfs'.\n\nSecond, use the following command to clone this repo:\n\n\n\nCaution: You have to run 'git lfs install' first. Otherwise, you will be SAD later.\n\nThird,\n\n\nFourth, open your browser and go to\n\n- <http://localhost:6006/#pytorch_profiler>\n- <http://localhost:6006/#pytorch_profiler>\n\nYou will see the following images:\n\n![](./URL)\n\n![](./optimized_transducer-URL)" ]
[ "TAGS\n#region-us \n", "# Introduction\n\nThis repo contains the benchmark results for <URL", "## Usage\n\nFirst, install 'git-lfs'.\n\nSecond, use the following command to clone this repo:\n\n\n\nCaution: You have to run 'git lfs install' first. Otherwise, you will be SAD later.\n\nThird,\n\n\nFourth, open your browser and go to\n\n- <http://localhost:6006/#pytorch_profiler>\n- <http://localhost:6006/#pytorch_profiler>\n\nYou will see the following images:\n\n![](./URL)\n\n![](./optimized_transducer-URL)" ]
[ 6, 13, 126 ]
[ "passage: TAGS\n#region-us \n# Introduction\n\nThis repo contains the benchmark results for <URL## Usage\n\nFirst, install 'git-lfs'.\n\nSecond, use the following command to clone this repo:\n\n\n\nCaution: You have to run 'git lfs install' first. Otherwise, you will be SAD later.\n\nThird,\n\n\nFourth, open your browser and go to\n\n- <http://localhost:6006/#pytorch_profiler>\n- <http://localhost:6006/#pytorch_profiler>\n\nYou will see the following images:\n\n![](./URL)\n\n![](./optimized_transducer-URL)" ]
[ -0.02542124129831791, 0.03949194401502609, -0.00569039024412632, 0.03029646910727024, 0.15798062086105347, 0.06563792377710342, 0.005004215054214001, 0.11876966059207916, 0.1389773041009903, -0.03474846109747887, 0.10570523887872696, 0.07168398797512054, 0.03597390651702881, 0.006856535095721483, 0.09946094453334808, -0.14118137955665588, 0.03565117344260216, 0.02677096612751484, -0.027888702228665352, 0.06124399974942207, 0.027276938781142235, -0.005685562267899513, 0.10487893223762512, -0.011493954807519913, -0.09881938248872757, 0.12230752408504486, 0.0668758675456047, 0.09102863818407059, 0.07771401852369308, -0.0777004212141037, 0.08737878501415253, 0.006320204585790634, 0.02546062506735325, -0.1212213933467865, 0.031736019998788834, 0.10375207662582397, 0.0017107665771618485, 0.013533188961446285, 0.04734112694859505, -0.012345441617071629, 0.08432330936193466, 0.12501157820224762, -0.10546160489320755, 0.01811571605503559, -0.10976442694664001, -0.15065604448318481, -0.10962342470884323, -0.04082049801945686, 0.042896244674921036, 0.06779275834560394, 0.04510906711220741, 0.17093846201896667, -0.13583184778690338, 0.10052292793989182, 0.3424539864063263, -0.13635200262069702, -0.006249954923987389, 0.14926259219646454, -0.008803877048194408, 0.030699344351887703, 0.006012419238686562, 0.049080658704042435, 0.04878765717148781, 0.09442135691642761, -0.13933853805065155, -0.051625192165374756, -0.03469425067305565, -0.011297542601823807, -0.08394971489906311, -0.16070103645324707, 0.16098980605602264, -0.003186493646353483, -0.02377847209572792, 0.20606623589992523, -0.13232196867465973, -0.07832968980073929, 0.013392572291195393, 0.03551657497882843, 0.046079520136117935, 0.09721159934997559, -0.06436315178871155, -0.17887741327285767, -0.06205330044031143, -0.14596512913703918, -0.05639423429965973, 0.154863640666008, 0.04226602613925934, 0.11560418456792831, -0.029083918780088425, 0.11863987892866135, -0.04809886962175369, -0.04772686958312988, -0.03866185247898102, -0.08499163389205933, 0.0646316185593605, 0.09442473202943802, 0.03356286510825157, -0.04980652406811714, 0.18203236162662506, 0.11727700382471085, -0.06721208989620209, 0.010920548811554909, -0.07301649451255798, 0.09293748438358307, -0.06821933388710022, 0.0818755254149437, -0.047964975237846375, 0.05040851980447769, 0.09353213757276535, 0.007092399522662163, 0.023929258808493614, 0.030300544574856758, -0.0362628810107708, -0.02428545244038105, -0.009271373972296715, 0.1343303620815277, 0.11644051969051361, 0.04203709214925766, -0.05755114182829857, 0.05009700357913971, 0.06833435595035553, -0.058048151433467865, -0.012461488135159016, -0.0652666762471199, -0.06976303458213806, -0.03816479444503784, 0.1003500297665596, -0.04443192854523659, -0.12630051374435425, -0.0492045022547245, -0.020710689947009087, 0.022462937980890274, -0.11961879581212997, -0.01935717463493347, 0.02533537894487381, -0.02111201547086239, 0.0163379218429327, -0.12913769483566284, -0.1367192417383194, -0.01144877914339304, 0.04071750491857529, 0.036016952246427536, 0.057693589478731155, 0.05928829684853554, -0.022316664457321167, -0.1053871437907219, -0.003975801169872284, -0.02867055870592594, -0.10038106143474579, 0.08455590903759003, 0.016460126265883446, 0.07234293967485428, -0.13635049760341644, 0.010356729850172997, -0.08054128289222717, 0.09785942733287811, -0.2537221312522888, -0.027702229097485542, -0.14250284433364868, 0.12126395851373672, -0.09616537392139435, -0.027097193524241447, -0.10397369414567947, 0.047596171498298645, 0.11450565606355667, 0.16379325091838837, -0.07210060954093933, 0.004325130954384804, 0.19067321717739105, -0.1544007658958435, -0.0273793563246727, 0.1433413177728653, 0.05464264005422592, -0.022896286100149155, 0.011656633578240871, 0.2133939415216446, 0.06101327762007713, -0.1118059828877449, 0.06810576468706131, 0.1304563581943512, -0.22718754410743713, -0.09922246634960175, -0.04743439704179764, -0.06668106466531754, -0.1886359453201294, 0.04180271178483963, -0.09688772261142731, 0.10838035494089127, -0.05269677937030792, 0.006683373358100653, -0.04941016063094139, -0.04865690693259239, -0.10150662809610367, 0.006199758965522051, -0.025073217228055, -0.0342007651925087, 0.009314772672951221, -0.16244356334209442, 0.07771121710538864, -0.02629026398062706, 0.059290092438459396, -0.02098623290657997, 0.16409635543823242, 0.005605682730674744, 0.050041817128658295, -0.0493931844830513, -0.05657093599438667, 0.036209139972925186, 0.09424713999032974, 0.02772318571805954, -0.004245390184223652, 0.05432363972067833, -0.08407299965620041, 0.013679321855306625, -0.06848926842212677, 0.08281739801168442, -0.0822228342294693, 0.023099351674318314, -0.005058069247752428, 0.025957981124520302, 0.012298902496695518, -0.24074780941009521, 0.011696476489305496, -0.010991102084517479, 0.021719928830862045, 0.08908716589212418, -0.025280296802520752, -0.03079243376851082, 0.03371774032711983, -0.053919170051813126, 0.028808867558836937, -0.04542066901922226, 0.03527407348155975, 0.0018162665655836463, 0.011580448597669601, 0.04559032991528511, 0.12669463455677032, 0.06402084976434708, 0.12846730649471283, -0.20070764422416687, 0.02070499025285244, 0.11335578560829163, -0.05010831728577614, 0.015790803357958794, 0.024064065888524055, -0.10441389679908752, 0.08917324244976044, -0.015320184640586376, 0.01748458296060562, -0.05758540332317352, 0.04939229413866997, -0.034285131841897964, -0.0639674961566925, -0.030582791194319725, 0.06302329152822495, 0.1877654492855072, -0.07878882437944412, 0.01285009179264307, 0.03127181902527809, -0.08418890088796616, -0.04296692833304405, -0.05096244066953659, -0.03788591921329498, 0.05100804194808006, 0.13397862017154694, -0.018319357186555862, 0.11145909875631332, -0.06191515550017357, 0.020633338019251823, 0.00599531177431345, 0.007930319756269455, 0.103995680809021, -0.15217216312885284, -0.06628813594579697, -0.07537953555583954, -0.11746140569448471, -0.08299671113491058, 0.01398754958063364, -0.010264592245221138, 0.07317224144935608, -0.020194437354803085, -0.06206116825342178, 0.014323538169264793, 0.005855511408299208, -0.010992439463734627, 0.1060434877872467, -0.12403790652751923, -0.18316121399402618, -0.1390041559934616, 0.019690493121743202, -0.03793065622448921, 0.050613731145858765, 0.04511871188879013, -0.15145306289196014, -0.05745471268892288, 0.019190100952982903, 0.13513445854187012, -0.08389925956726074, -0.019729487597942352, -0.018310867249965668, 0.031031517311930656, 0.09043405205011368, -0.14924955368041992, -0.02845916710793972, -0.009680014103651047, -0.10305856913328171, 0.05945766344666481, -0.10307402163743973, 0.12349431216716766, -0.043512631207704544, 0.03703862428665161, 0.04649031162261963, 0.0707184448838234, 0.23496678471565247, -0.10600655525922775, -0.02342139557003975, 0.14187608659267426, 0.02663690596818924, 0.0909716933965683, 0.05843571946024895, 0.01143388357013464, -0.09643906354904175, -0.003895375644788146, 0.07066582888364792, -0.10154858231544495, -0.17055056989192963, 0.048838984221220016, 0.008785346522927284, 0.0028001766186207533, 0.01722424291074276, 0.09219128638505936, -0.10908050835132599, 0.06434693187475204, -0.04552764073014259, -0.0257840808480978, -0.08687549829483032, 0.01708759367465973, -0.07308731228113174, -0.06929425895214081, -0.025904634967446327, -0.052040185779333115, -0.0627511516213417, 0.10546381026506424, 0.03838774189352989, 0.12907133996486664, -0.06870044022798538, 0.03937621787190437, 0.09246775507926941, 0.16665613651275635, -0.027879171073436737, 0.16676828265190125, -0.07850228995084763, 0.025294233113527298, 0.008347397670149803, 0.020197713747620583, -0.05026829242706299, -0.02444005385041237, 0.04624190554022789, -0.0724167600274086, 0.022597631439566612, -0.024366676807403564, 0.09367228299379349, 0.1825949102640152, -0.003202797146514058, -0.16608192026615143, 0.05449492111802101, -0.028407806530594826, -0.026563020423054695, -0.05067245662212372, 0.12175020575523376, 0.04590843990445137, 0.01255951076745987, 0.023369435220956802, 0.04429003223776817, 0.07742928713560104, -0.07653791457414627, 0.040263235569000244, -0.00012902090384159237, 0.07953004539012909, -0.013195126317441463, 0.02182517573237419, -0.15790539979934692, 0.17718306183815002, 0.006093142554163933, -0.04378070682287216, -0.007371751591563225, -0.014076082035899162, 0.10092872381210327, 0.06663253903388977, 0.10329952090978622, 0.06418765336275101, -0.05497659742832184, -0.0995246097445488, -0.09043996036052704, 0.04676083102822304, 0.07275275141000748, -0.016400029882788658, 0.00774424010887742, -0.010933969169855118, -0.017893381416797638, -0.008081141859292984, -0.07498174160718918, -0.044873256236314774, 0.05502277985215187, -0.0029622141737490892, -0.014801861718297005, 0.009707305580377579, -0.06246155872941017, 0.06477110832929611, 0.06163123622536659, 0.1384488344192505, 0.05322999134659767, -0.08552147448062897, -0.02475615032017231, 0.040224313735961914, 0.13964632153511047, -0.025910774245858192, -0.043936796486377716, -0.0361718125641346, -0.09593261778354645, -0.02148343063890934, -0.1377158761024475, 0.020851461216807365, -0.05424780026078224, -0.032466646283864975, 0.031087379902601242, 0.1637202948331833, -0.02732613869011402, -0.013392877764999866, 0.014504863880574703, -0.08425328880548477, -0.02490132860839367, -0.11571141332387924, 0.11729013174772263, -0.09075281023979187, -0.008734052069485188, 0.15122510492801666, -0.05045941844582558, 0.13829204440116882, -0.051059551537036896, 0.0660296082496643, 0.04357083886861801, 0.21692471206188202, -0.011889193207025528, -0.062451910227537155, 0.06513264030218124, -0.05709492415189743, -0.25870105624198914, -0.05476486310362816, -0.11111680418252945, -0.10268969088792801, -0.03556250408291817, -0.10326461493968964, 0.12030358612537384, 0.039257727563381195, -0.02068895660340786, 0.23436596989631653, -0.14804691076278687, -0.0014080788241699338, 0.08973236382007599, 0.07447222620248795, 0.2616164982318878, -0.1581805944442749, -0.050412945449352264, 0.005102965515106916, -0.07376537472009659, -0.03169664740562439, -0.02270009182393551, 0.08170104026794434, -0.002129398984834552, 0.10771606117486954, 0.054193757474422455, -0.056345779448747635, 0.11557665467262268, -0.07692348212003708, 0.05273488536477089, -0.09282301366329193, -0.07705720514059067, -0.07215652614831924, -0.031595103442668915, 0.12359874695539474, 0.06974758952856064, 0.018404332920908928, 0.001562151126563549, 0.035649482160806656, -0.11972612887620926, 0.10880950838327408, 0.01058280747383833, 0.006376942619681358, -0.0979970246553421, -0.04816737771034241, -0.03946343809366226, 0.07461481541395187, 0.0641208365559578, 0.009905919432640076, -0.04582098871469498, 0.06339234113693237, -0.05396907776594162, -0.08225112408399582, -0.05743599683046341, 0.07830045372247696, -0.0006671942537650466, 0.0718834400177002, -0.06105370819568634, 0.07732722163200378, 0.054296448826789856, -0.039660047739744186, -0.06558429449796677, 0.037915050983428955, -0.13030517101287842, -0.037855297327041626, 0.13231632113456726, -0.13066191971302032, -0.03985043987631798, -0.06814771890640259, -0.13105224072933197, 0.05321972444653511, 0.040799401700496674, 0.1112852618098259, 0.03261219337582588, 0.018094979226589203, 0.016707615926861763, 0.04251882806420326, -0.11192989349365234, 0.019360121339559555, 0.08724123239517212, 0.0014117771061137319, -0.12154471129179001, 0.08931677043437958, -0.024227309972047806, -0.015097249299287796, -0.050071194767951965, 0.11527813971042633, -0.05157534033060074, -0.09323737770318985, -0.015011731535196304, 0.03207055851817131, 0.004348193760961294, -0.04648907855153084, -0.14422254264354706, 0.047194164246320724, -0.007313756737858057, 0.0016347793862223625, 0.07437305152416229, 0.02208271250128746, 0.12717397511005402, 0.05304620414972305, -0.0221925787627697, -0.09205697476863861, -0.002779782982543111, 0.045276954770088196, -0.13327103853225708, 0.0196037869900465, 0.010616387240588665, 0.1352834552526474, -0.04175747185945511, 0.049042508006095886, -0.19654394686222076, 0.060524843633174896, -0.11683634668588638, -0.01408008299767971, -0.016162514686584473, 0.004029357805848122, 0.059144701808691025, -0.048111442476511, -0.07747997343540192, 0.05023004859685898, -0.1441849321126938, -0.03909369558095932, 0.03764098882675171, -0.04529472067952156, -0.030489670112729073, -0.03578140586614609, 0.1138034462928772, -0.016257010400295258, 0.053262677043676376, 0.10995437949895859, -0.03349513188004494, 0.05718691647052765, -0.038012657314538956, -0.11353650689125061, 0.08379587531089783, 0.015273964032530785, -0.04656492546200752, 0.158168762922287, 0.05180281028151512, -0.0004638626123778522, 0.003399703186005354, -0.029686832800507545, 0.11520019173622131, -0.12500733137130737, 0.02171526476740837, 0.03619151934981346, -0.06631509959697723, 0.01066004578024149, -0.03168908879160881, 0.15887561440467834, 0.14102640748023987, 0.03675331547856331, -0.024793870747089386, 0.013340595178306103, -0.08672191947698593, 0.02346678264439106, 0.032988760620355606, -0.1191280335187912, 0.1178278997540474, -0.002072911011055112, 0.03122624009847641, -0.016920052468776703, 0.18331430852413177, 0.07332565635442734, 0.09227284789085388, -0.04892707243561745, 0.09799786657094955, 0.07272951304912567, -0.029249699786305428, -0.03791983053088188, -0.05414079129695892, 0.0008669908856973052, -0.06691677868366241, -0.010216495022177696, 0.10702603310346603, -0.13073912262916565, 0.11059661954641342, 0.0711246132850647, 0.11420761048793793, 0.046675220131874084, 0.007682275027036667, -0.09417540580034256, 0.003226321889087558, -0.27735987305641174, 0.014420630410313606, 0.04232138395309448, -0.006395766045898199, 0.00926254689693451, 0.06013351306319237, 0.023231761530041695, 0.06301369518041611, 0.021115003153681755, -0.028958044946193695, -0.10635848343372345, -0.0892127975821495, -0.0034427575301378965, -0.23070846498012543, 0.06751485168933868, -0.06532055884599686, -0.03626978397369385, 0.06626082956790924, -0.028629109263420105, -0.07502519339323044, 0.1422542929649353, 0.0005086823366582394, -0.09618565440177917, 0.011628407984972, -0.018845418468117714, -0.05173828452825546, 0.1255066692829132, -0.07261651754379272, 0.01841672509908676, 0.009116314351558685, 0.018805984407663345, 0.04357634112238884, -0.016129035502672195, 0.030680246651172638, -0.09700105339288712, -0.0023370555136352777, -0.04841763153672218, 0.04236746206879616, -0.04436787590384483, 0.12097995728254318, 0.020209673792123795, -0.001249745604582131, 0.02422333136200905, 0.158956378698349, -0.0386795848608017, -0.004008327145129442, -0.0431671068072319, 0.14417417347431183, -0.04673672839999199, 0.013094000518321991, -0.08734310418367386, -0.09823784977197647, -0.08022740483283997, 0.3548494577407837, 0.1809517741203308, -0.06328412145376205, -0.00647373404353857, -0.03895387798547745, 0.029175905510783195, 0.03215290233492851, 0.1221129447221756, 0.1435927450656891, 0.10263554751873016, -0.057384192943573, -0.07824314385652542, -0.011000593192875385, 0.023262236267328262, -0.156717911362648, -0.035802580416202545, 0.08758579194545746, -0.026693524792790413, -0.09704694896936417, 0.1399378478527069, -0.17539042234420776, -0.04096825420856476, -0.059287603944540024, -0.05228449031710625, 0.018210915848612785, -0.01370219886302948, 0.009020060300827026, -0.004191912245005369, 0.05249554291367531, -0.08659780025482178, 0.04662751033902168, 0.022425418719649315, -0.0356622114777565, -0.2487495094537735, -0.15451474487781525, 0.04531734809279442, -0.07342660427093506, 0.22479906678199768, 0.028345463797450066, -0.04042429104447365, 0.009024086408317089, -0.06344927102327347, -0.08631565421819687, 0.04993360862135887, -0.007006560917943716, -0.20056839287281036, -0.06861616671085358, 0.025904580950737, -0.03771137073636055, -0.1133868545293808, -0.033543702214956284, 0.013024828396737576, -0.03384968265891075, 0.1348607987165451, 0.029166748747229576, -0.09195594489574432, -0.03288853168487549, -0.22248713672161102, 0.05359429493546486, 0.11111146211624146, 0.01153834443539381, 0.06490431725978851, -0.09412461519241333, 0.05362457409501076, 0.06048983708024025, -0.0555117167532444, -0.04397399723529816, -0.07565689086914062, -0.07373347878456116, 0.08524724096059799, -0.061941955238580704, -0.020994149148464203, 0.020474478602409363, -0.017823511734604836, 0.005102504976093769, -0.05827386677265167, 0.05740026384592056, -0.0038290382362902164, 0.04940532147884369, 0.04116668924689293, -0.11312902718782425, 0.02849142625927925, 0.06447269022464752, -0.10685338079929352, -0.0942755863070488 ]
null
null
transformers
# Wav2Vec2-Large-XLSR-53-Cantonese Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on Cantonese using the [Common Voice](https://huggingface.co/datasets/common_voice). When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ```python import torch import torchaudio from datasets import load_dataset from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor test_dataset = load_dataset("common_voice", "zh-HK", split="test[:2%]") processor = Wav2Vec2Processor.from_pretrained("ctl/wav2vec2-large-xlsr-cantonese") model = Wav2Vec2ForCTC.from_pretrained("ctl/wav2vec2-large-xlsr-cantonese") resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): speech_array, sampling_rate = torchaudio.load(batch["path"]) batch["speech"] = resampler(speech_array).squeeze().numpy() return batch test_dataset = test_dataset.map(speech_file_to_array_fn) inputs = processor(test_dataset["speech"][:2], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values, attention_mask=inputs.attention_mask).logits predicted_ids = torch.argmax(logits, dim=-1) print("Prediction:", processor.batch_decode(predicted_ids)) print("Reference:", test_dataset["sentence"][:2]) ``` ## Evaluation The model can be evaluated as follows on the Chinese (Hong Kong) test data of Common Voice. ```python !pip install jiwer import torch import torchaudio from datasets import load_dataset, load_metric from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor import re import argparse lang_id = "zh-HK" model_id = "ctl/wav2vec2-large-xlsr-cantonese" chars_to_ignore_regex = '[\,\?\.\!\-\;\:"\“\%\‘\”\�\.\⋯\!\-\:\–\。\》\,\)\,\?\;\~\~\…\︰\,\(\」\‧\《\﹔\、\—\/\,\「\﹖\·\']' test_dataset = load_dataset("common_voice", f"{lang_id}", split="test") cer = load_metric("cer") processor = Wav2Vec2Processor.from_pretrained(f"{model_id}") model = Wav2Vec2ForCTC.from_pretrained(f"{model_id}") model.to("cuda") resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): batch["sentence"] = re.sub(chars_to_ignore_regex, '', batch["sentence"]).lower() speech_array, sampling_rate = torchaudio.load(batch["path"]) batch["speech"] = resampler(speech_array).squeeze().numpy() return batch test_dataset = test_dataset.map(speech_file_to_array_fn) # Preprocessing the datasets. # We need to read the aduio files as arrays def evaluate(batch): inputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits pred_ids = torch.argmax(logits, dim=-1) batch["pred_strings"] = processor.batch_decode(pred_ids) return batch result = test_dataset.map(evaluate, batched=True, batch_size=16) print("CER: {:2f}".format(100 * cer.compute(predictions=result["pred_strings"], references=result["sentence"]))) ``` **Test Result**: 15.51 % ## Training The Common Voice `train`, `validation` were used for training. The script used for training will be posted [here](https://github.com/chutaklee/CantoASR)
{"language": ["yue"], "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech", "xlsr-fine-tuning-week"], "datasets": ["common_voice"], "metrics": ["cer"], "language_bcp47": ["zh-HK"], "model-index": [{"name": "wav2vec2-large-xlsr-cantonese", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Speech Recognition"}, "dataset": {"name": "Common Voice zh-HK", "type": "common_voice", "args": "zh-HK"}, "metrics": [{"type": "cer", "value": 15.36, "name": "Test CER"}]}]}]}
automatic-speech-recognition
ctl/wav2vec2-large-xlsr-cantonese
[ "transformers", "pytorch", "jax", "wav2vec2", "automatic-speech-recognition", "audio", "speech", "xlsr-fine-tuning-week", "yue", "dataset:common_voice", "license:apache-2.0", "model-index", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "yue" ]
TAGS #transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #yue #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #has_space #region-us
# Wav2Vec2-Large-XLSR-53-Cantonese Fine-tuned facebook/wav2vec2-large-xlsr-53 on Cantonese using the Common Voice. When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ## Evaluation The model can be evaluated as follows on the Chinese (Hong Kong) test data of Common Voice. Test Result: 15.51 % ## Training The Common Voice 'train', 'validation' were used for training. The script used for training will be posted here
[ "# Wav2Vec2-Large-XLSR-53-Cantonese\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Cantonese using the Common Voice.\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:", "## Evaluation\n\nThe model can be evaluated as follows on the Chinese (Hong Kong) test data of Common Voice. \n\n\n\n\n\nTest Result: 15.51 %", "## Training\n\nThe Common Voice 'train', 'validation' were used for training.\n\nThe script used for training will be posted here" ]
[ "TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #yue #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #has_space #region-us \n", "# Wav2Vec2-Large-XLSR-53-Cantonese\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Cantonese using the Common Voice.\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:", "## Evaluation\n\nThe model can be evaluated as follows on the Chinese (Hong Kong) test data of Common Voice. \n\n\n\n\n\nTest Result: 15.51 %", "## Training\n\nThe Common Voice 'train', 'validation' were used for training.\n\nThe script used for training will be posted here" ]
[ 85, 65, 20, 32, 29 ]
[ "passage: TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #yue #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #has_space #region-us \n# Wav2Vec2-Large-XLSR-53-Cantonese\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Cantonese using the Common Voice.\nWhen using this model, make sure that your speech input is sampled at 16kHz.## Usage\n\nThe model can be used directly (without a language model) as follows:## Evaluation\n\nThe model can be evaluated as follows on the Chinese (Hong Kong) test data of Common Voice. \n\n\n\n\n\nTest Result: 15.51 %## Training\n\nThe Common Voice 'train', 'validation' were used for training.\n\nThe script used for training will be posted here" ]
[ -0.12746071815490723, 0.009867233224213123, -0.002203010953962803, -0.047473274171352386, 0.10282795876264572, -0.08455982804298401, 0.18695911765098572, 0.11480991542339325, -0.03922693431377411, 0.005476950667798519, 0.0020250030793249607, -0.03411377966403961, 0.08279486000537872, 0.08098917454481125, 0.017794588580727577, -0.2215646654367447, 0.011786747723817825, 0.057323966175317764, 0.0935540571808815, 0.12098769843578339, 0.09703639149665833, -0.07620083540678024, 0.018378769978880882, 0.1069604828953743, -0.133029967546463, 0.03383960947394371, 0.01843075454235077, -0.1140306293964386, 0.14766642451286316, 0.03350585699081421, 0.08046090602874756, 0.049755215644836426, 0.07678398489952087, -0.21656860411167145, 0.024134280160069466, 0.0342620313167572, 0.045025311410427094, 0.02592499554157257, 0.06772521138191223, 0.051542650908231735, 0.08500636368989944, 0.11455529928207397, -0.0242848452180624, 0.10411297529935837, -0.05459537357091904, -0.20345556735992432, -0.020226141437888145, 0.06053198501467705, 0.13161225616931915, 0.14668717980384827, -0.07816218584775925, 0.13324101269245148, -0.15846413373947144, 0.08707088977098465, 0.09278307110071182, -0.22948230803012848, 0.014009037055075169, 0.1314043551683426, 0.11137472093105316, 0.05054888874292374, -0.07986295223236084, -0.005406043492257595, 0.06375308334827423, 0.010662487708032131, -0.026049675419926643, -0.05175243690609932, -0.19789010286331177, -0.02729373797774315, -0.12710073590278625, -0.016212107613682747, 0.2468602955341339, -0.002693296642974019, -0.0909520611166954, -0.10784990340471268, 0.00911711435765028, -0.028875650838017464, 0.0048630028031766415, -0.07551291584968567, 0.007265950553119183, 0.051234353333711624, -0.00296372571028769, -0.057379573583602905, -0.11268537491559982, -0.15047170221805573, 0.0138632170855999, 0.012829956598579884, 0.026925547048449516, -0.01096296589821577, -0.17365294694900513, 0.09387179464101791, -0.06827438622713089, -0.06753332912921906, -0.020620150491595268, 0.04527261480689049, -0.04437718912959099, -0.008502179756760597, -0.03577717766165733, -0.12254542112350464, 0.008878174237906933, 0.0026123393326997757, 0.02584523893892765, 0.04784281551837921, -0.08518083393573761, 0.08703405410051346, 0.009676074609160423, 0.15231627225875854, -0.06883752346038818, 0.03619677200913429, 0.04351776838302612, 0.03136517480015755, -0.054156284779310226, -0.004376332275569439, -0.08205366879701614, -0.039245475083589554, 0.07161279767751694, 0.06867021322250366, -0.03798737749457359, 0.014138611033558846, -0.06206245720386505, -0.020369920879602432, 0.07333002984523773, -0.08603891730308533, -0.03334351256489754, 0.07356683909893036, -0.03370217978954315, 0.10351880639791489, 0.06503278762102127, 0.07085491716861725, -0.0784715786576271, -0.01036395039409399, 0.005909696687012911, 0.07970980554819107, -0.0038927667774260044, -0.06014368683099747, 0.020841244608163834, -0.02413717471063137, -0.024596357718110085, -0.08324981480836868, -0.10750987380743027, -0.08183608949184418, -0.05313943699002266, 0.05087575316429138, -0.035424355417490005, -0.09075375646352768, -0.04543381556868553, -0.05045556277036667, -0.0575714036822319, 0.06257046014070511, -0.03522275760769844, 0.07507907599210739, 0.07949066907167435, 0.03786410763859749, 0.04982072114944458, 0.07395675033330917, -0.09618727117776871, -0.02697264961898327, -0.025785574689507484, 0.1175551488995552, -0.003912885207682848, -0.04333268851041794, -0.08576548844575882, -0.0701189860701561, -0.06509426236152649, 0.06849490851163864, 0.034311484545469284, 0.12080971896648407, -0.2853754162788391, -0.08932636678218842, 0.21093016862869263, -0.12967392802238464, -0.04018797352910042, 0.21343524754047394, 0.002847545314580202, 0.10052747279405594, 0.10664302110671997, 0.22728334367275238, 0.05461852625012398, -0.2379201501607895, 0.030547009781003, -0.0018178577302023768, -0.02961411513388157, -0.046562351286411285, 0.087440624833107, -0.054620686918497086, -0.012970051728188992, 0.0179065503180027, -0.10403173416852951, 0.048116397112607956, -0.04537526145577431, -0.05577057972550392, -0.02621554583311081, -0.05539092421531677, 0.019584624096751213, 0.0378132201731205, 0.03240156173706055, -0.027508437633514404, -0.06322427839040756, 0.050579335540533066, 0.107469342648983, -0.1117129772901535, 0.06763087958097458, -0.12454096227884293, 0.08614616841077805, -0.019853565841913223, -0.009633881971240044, -0.12647905945777893, 0.199935644865036, 0.004600026644766331, 0.02549174800515175, 0.0753818154335022, 0.1535877287387848, -0.013215024024248123, 0.006604690570384264, -0.0558585450053215, -0.03344951942563057, 0.024145256727933884, -0.03805268555879593, -0.03212185204029083, -0.06324603408575058, -0.01641944795846939, -0.04573288559913635, 0.0983731746673584, -0.1989627480506897, 0.00013275582750793546, 0.04746904969215393, 0.020886577665805817, -0.0148461377248168, -0.0029626640025526285, 0.10408952087163925, 0.10187181085348129, -0.022141823545098305, 0.007614923641085625, 0.033729977905750275, 0.030351536348462105, -0.10072933882474899, 0.17000658810138702, -0.1651647686958313, -0.04271123558282852, 0.13307815790176392, -0.07313984632492065, -0.015889279544353485, 0.022977618500590324, -0.01689937338232994, -0.03198962286114693, -0.09532851725816727, 0.014107550494372845, 0.3200870454311371, -0.033486805856227875, 0.1535319685935974, -0.11870317161083221, 0.014568887650966644, 0.0016873261192813516, -0.09502275288105011, 0.028309771791100502, 0.05727161094546318, -0.003691375255584717, 0.032553885132074356, 0.027320193126797676, -0.08252330124378204, -0.09057573974132538, 0.2272271066904068, -0.01402521412819624, -0.06526478379964828, 0.01195299532264471, 0.040403492748737335, -0.006270790006965399, 0.044589996337890625, -0.18494360148906708, -0.06278755515813828, 0.010949689894914627, 0.05781404674053192, 0.06868959963321686, -0.19209139049053192, 0.0032034781761467457, -0.0052596330642700195, -0.14285191893577576, -0.1501619964838028, 0.056228917092084885, -0.04227873682975769, 0.04470990225672722, -0.09621012955904007, -0.025757398456335068, -0.002098778961226344, -0.030033929273486137, -0.1821015179157257, 0.10610940307378769, -0.07413607090711594, -0.24974612891674042, -0.12689542770385742, 0.04715600237250328, 0.055005814880132675, 0.024939686059951782, 0.08809228986501694, -0.1501728892326355, 0.0017184352036565542, -0.0342099629342556, 0.041425447911024094, -0.020099099725484848, -0.043646275997161865, -0.03725479170680046, 0.04521883651614189, 0.07739274948835373, -0.12787780165672302, 0.013006925582885742, -0.04673732444643974, -0.08362239599227905, 0.018116870895028114, -0.03236944228410721, -0.0069108689203858376, 0.18354792892932892, 0.03315519914031029, -0.009091177955269814, -0.04193872958421707, 0.11837702244520187, -0.10336536914110184, -0.023418037220835686, 0.22328630089759827, 0.004746129736304283, -0.017646508291363716, 0.08591236174106598, -0.0016837334260344505, -0.06344553083181381, 0.033705636858940125, -0.021080341190099716, -0.07697522640228271, -0.256820410490036, -0.07864347845315933, -0.06851265579462051, -0.03455473110079765, -0.04573523998260498, 0.018304912373423576, 0.10292859375476837, -0.010854392312467098, -0.044484943151474, -0.05706975609064102, 0.05710681527853012, 0.017024973407387733, 0.050850410014390945, 0.011524503119289875, 0.09380704164505005, -0.03528820723295212, -0.02042754366993904, 0.0001450782729079947, 0.06042521446943283, 0.08854573220014572, 0.07316013425588608, 0.04868924245238304, 0.083837129175663, 0.15402555465698242, 0.15268167853355408, 0.01655757613480091, -0.05722891911864281, -0.02320002019405365, 0.0047270930372178555, -0.06038770079612732, -0.03343654423952103, 0.058195676654577255, 0.18408919870853424, -0.05388990417122841, -0.04410609230399132, 0.002596795093268156, 0.004911500494927168, 0.16827134788036346, 0.08785233646631241, -0.15941573679447174, -0.0660964772105217, -0.01817644201219082, -0.05643367022275925, 0.014151654206216335, 0.08550605177879333, 0.18008309602737427, -0.16060906648635864, 0.024582669138908386, 0.03305364027619362, 0.10292154550552368, -0.008660621009767056, 0.03115776553750038, -0.0775982066988945, -0.0008094153017736971, 0.011672403663396835, 0.06926818937063217, -0.3304884433746338, 0.2175893932580948, -0.005982426926493645, 0.10373298078775406, -0.040512897074222565, 0.008105899207293987, 0.009279299527406693, 0.049219392240047455, 0.1270233392715454, 0.005462083965539932, 0.02793983928859234, -0.11949758976697922, -0.06705310195684433, 0.034362051635980606, 0.009118110872805119, 0.06823813170194626, 0.024716632440686226, 0.029448464512825012, -0.00814851839095354, -0.006785155739635229, -0.06368333846330643, -0.15404067933559418, -0.01151900552213192, 0.0031182188540697098, 0.16679343581199646, 0.11825362592935562, -0.02656880021095276, -0.07185079902410507, -0.10549537092447281, 0.05445234850049019, -0.12449735403060913, -0.0209230687469244, -0.03143016993999481, -0.051969800144433975, 0.08596684783697128, -0.0627909004688263, -0.017607970163226128, 0.07119470834732056, 0.09905868023633957, -0.02391345612704754, 0.005028614308685064, 0.0716281309723854, -0.10113935172557831, -0.12782467901706696, -0.022054333239793777, 0.2077244073152542, 0.06735368818044662, 0.08872399479150772, 0.0499352365732193, -0.014257288537919521, 0.002135803224518895, -0.04525597020983696, -0.014855324290692806, 0.05319685861468315, -0.05739425867795944, 0.027721164748072624, -0.02690090797841549, -0.1545579582452774, -0.16038204729557037, -0.07931819558143616, 0.11372138559818268, 0.11778703331947327, -0.034518368542194366, 0.12329907715320587, 0.21702252328395844, -0.08513771742582321, -0.17905274033546448, -0.02716694585978985, 0.0777006521821022, 0.1042976975440979, -0.03257915377616882, -0.17814777791500092, 0.057216569781303406, -0.0052196611650288105, -0.03496802970767021, -0.07379245012998581, -0.26735740900039673, -0.17376427352428436, 0.1024850532412529, -0.05227157101035118, 0.08078571408987045, -0.038833312690258026, -0.04972926527261734, 0.030033553019165993, -0.005815607029944658, -0.031093837693333626, -0.0659475177526474, 0.11305796355009079, 0.029769890010356903, 0.04989367723464966, 0.030369199812412262, -0.041287798434495926, 0.07955539971590042, 0.08887680619955063, -0.011397137306630611, -0.008275296539068222, 0.016407595947384834, 0.02121841162443161, 0.022397538647055626, 0.16842417418956757, -0.08634020388126373, 0.032464154064655304, -0.12703773379325867, -0.07772824168205261, -0.06558067351579666, 0.05627346783876419, 0.008740750141441822, -0.01631668396294117, 0.003221290884539485, -0.001129458425566554, -0.027679594233632088, 0.027916057035326958, -0.008672425523400307, -0.12800224125385284, 0.06270068138837814, 0.17255635559558868, 0.19659437239170074, -0.02678162418305874, -0.0439860038459301, -0.024133065715432167, -0.02274073287844658, 0.12328124046325684, -0.07640804350376129, 0.023198066279292107, 0.05577969178557396, 0.04460792616009712, 0.14795280992984772, 0.021314434707164764, -0.08704790472984314, 0.11448829621076584, 0.014455028809607029, -0.04245374724268913, -0.12124349921941757, -0.03367811068892479, -0.007023499347269535, 0.006981853861361742, 0.020841538906097412, 0.07725001871585846, -0.09789305925369263, -0.02729937434196472, -0.014023000374436378, 0.015256284736096859, -0.12005103379487991, 0.18404355645179749, 0.007539636921137571, 0.08857695013284683, -0.09953693300485611, 0.03544991463422775, 0.050045549869537354, -0.01615661382675171, 0.04971859231591225, -0.008397727273404598, -0.08819599449634552, -0.049784231930971146, -0.04895342141389847, 0.10730215907096863, 0.085471972823143, -0.1263781487941742, -0.061428252607584, -0.09565296024084091, -0.003688790602609515, 0.06599976122379303, 0.048039376735687256, 0.025602631270885468, -0.08654402196407318, -0.05611300840973854, -0.11708489060401917, 0.02293941006064415, 0.08725240081548691, -0.006867141928523779, -0.12026005238294601, 0.23479272425174713, 0.11925530433654785, 0.06096901744604111, -0.030576562508940697, -0.08459799736738205, -0.05571766197681427, 0.10060151666402817, -0.05657067149877548, 0.012178526259958744, -0.04793380945920944, -0.009933914057910442, -0.014166651293635368, -0.09182919561862946, -0.011205730028450489, 0.08994780480861664, -0.07490510493516922, 0.055585574358701706, -0.001447310671210289, 0.07351596653461456, -0.07738351076841354, 0.027432898059487343, 0.058151621371507645, -0.0632934644818306, 0.07063289731740952, 0.12072435021400452, -0.11354897171258926, 0.0987531915307045, -0.18992553651332855, -0.06308989226818085, 0.08635488152503967, 0.05663678050041199, -0.04662567377090454, -0.08948996663093567, 0.03359157592058182, 0.09689554572105408, 0.08925791084766388, -0.02309250831604004, 0.09379082918167114, -0.0626104474067688, -0.05736691877245903, -0.08205379545688629, -0.02995513379573822, -0.040531400591135025, 0.05724991112947464, 0.08725869655609131, 0.13094235956668854, 0.14133882522583008, -0.12254258245229721, 0.09597701579332352, -0.13797084987163544, 0.006835282780230045, -0.05512333661317825, -0.014728052541613579, -0.13265083730220795, -0.08165706694126129, 0.06903526932001114, -0.04631805419921875, 0.10135392844676971, -0.023171626031398773, 0.03865688294172287, -0.009822872467339039, -0.07537106424570084, -0.016814973205327988, -0.007295588031411171, 0.2293815165758133, 0.06276324391365051, 0.022304024547338486, 0.020021209493279457, -0.003430603537708521, 0.005009505897760391, 0.09622325748205185, 0.02951815165579319, 0.14114506542682648, -0.02908160164952278, 0.09871802479028702, 0.0838107094168663, -0.05426323413848877, -0.07832134515047073, -0.03948558494448662, -0.17205794155597687, 0.020115463063120842, -0.0670996829867363, 0.18658848106861115, 0.1605735421180725, -0.09515485912561417, 0.09224820882081985, 0.02246711403131485, -0.1054266020655632, -0.15686452388763428, -0.1205635517835617, -0.06116029620170593, -0.1492573618888855, 0.03841530531644821, -0.05499941110610962, 0.026528645306825638, 0.014884881675243378, 0.05028776079416275, -0.05184149369597435, 0.1605154126882553, 0.022837763652205467, -0.10220008343458176, 0.07632068544626236, -0.08707983791828156, 0.02067508175969124, -0.08190339058637619, 0.028214596211910248, 0.16948163509368896, -0.021335385739803314, 0.06795485317707062, 0.019066868349909782, -0.04153605177998543, 0.056004222482442856, -0.07578045129776001, -0.062077343463897705, 0.007794241886585951, -0.040271881967782974, 0.0984102115035057, 0.128941610455513, 0.11234597861766815, -0.05695551633834839, 0.02183709107339382, 0.12915821373462677, -0.032857514917850494, -0.12034349143505096, -0.14496932923793793, 0.18736408650875092, 0.027602363377809525, -0.009473422542214394, -0.02008584700524807, -0.030391588807106018, 0.0014382685767486691, 0.26397886872291565, 0.18849323689937592, 0.03509325534105301, 0.020540202036499977, -0.035794299095869064, -0.009614736773073673, -0.042181696742773056, 0.026994889602065086, 0.11537710577249527, 0.17195577919483185, 0.0033777104690670967, 0.02011292800307274, -0.0733630433678627, -0.0878719612956047, -0.016280626878142357, -0.0031844843178987503, -0.07025963813066483, -0.09079641103744507, -0.013028711080551147, 0.1116478443145752, -0.08200601488351822, -0.08149518072605133, -0.09943146258592606, -0.08783300966024399, -0.05778159946203232, 0.017467377707362175, 0.030130615457892418, 0.11828924715518951, -0.0013920519268140197, -0.05944781005382538, 0.0413462333381176, 0.11170587688684464, 0.002438825787976384, -0.02644984796643257, -0.06965066492557526, 0.03735581040382385, -0.08819371461868286, 0.02834842912852764, 0.002463347977027297, 0.14860846102237701, 0.008365350775420666, 0.10223882645368576, 0.0016303478041663766, 0.16919763386249542, -0.02807922288775444, -0.045084066689014435, 0.013764540664851665, 0.14234791696071625, -0.014960071071982384, 0.12251440435647964, 0.009218120947480202, -0.09781023859977722, 0.06277918815612793, -0.13623754680156708, 0.014386257156729698, -0.13658595085144043, 0.07753930240869522, -0.04842418432235718, 0.1017984002828598, 0.08620229363441467, -0.06963185220956802, -0.030822616070508957, -0.04120481386780739, 0.048898790031671524, 0.000311294017592445, -0.05832398310303688, -0.04970679432153702, -0.23466132581233978, 0.027778606861829758, -0.11267463117837906, -0.008928108960390091, -0.19655638933181763, -0.017889106646180153, -0.009715316817164421, -0.09296458214521408, -0.014206902123987675, 0.02159084565937519, 0.09425710141658783, 0.03407401964068413, -0.009493691846728325, -0.08804582804441452, 0.05373304709792137, 0.13355903327465057, -0.17060115933418274, -0.11093630641698837 ]
null
null
transformers
# My Awesome Model
{"tags": ["conversational"]}
text-generation
cumtowndiscord/DialoGPT-small-joshua
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# My Awesome Model
[ "# My Awesome Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# My Awesome Model" ]
[ 51, 4 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# My Awesome Model" ]
[ -0.05259015038609505, 0.05521034821867943, -0.005910294596105814, 0.017722278833389282, 0.15250112116336823, 0.02286236733198166, 0.07657632976770401, 0.09513414651155472, -0.025391526520252228, -0.047348517924547195, 0.15119488537311554, 0.19781284034252167, -0.020334534347057343, 0.101333387196064, -0.04688440263271332, -0.3143521845340729, 0.06439975649118423, 0.05463787540793419, -0.015605635941028595, 0.12023304402828217, 0.09468326717615128, -0.0530015267431736, 0.08742043375968933, -0.012155864387750626, -0.1293085366487503, -0.0027921805158257484, -0.002384399762377143, -0.10180269181728363, 0.11194873601198196, 0.033712033182382584, 0.05166437849402428, 0.0182647667825222, -0.05843055993318558, -0.139859139919281, 0.03845210000872612, -0.015005595050752163, -0.05602653697133064, 0.05648263916373253, 0.059830192476511, -0.07164353132247925, 0.1669619083404541, 0.13275989890098572, -0.04237370565533638, 0.056127581745386124, -0.17620700597763062, 0.017941240221261978, 0.01800798624753952, 0.019184142351150513, 0.05306641012430191, 0.10830496996641159, -0.03932326287031174, 0.09217294305562973, -0.11410652846097946, 0.08313368260860443, 0.07800983637571335, -0.29151955246925354, -0.025312699377536774, 0.10440942645072937, 0.06437138468027115, 0.048375632613897324, -0.013386772945523262, 0.0621674507856369, 0.02149512618780136, 0.008602659218013287, 0.02225899137556553, -0.06727100163698196, -0.05789240449666977, 0.032748885452747345, -0.0967593789100647, -0.03634428232908249, 0.19753605127334595, -0.024647634476423264, 0.053590498864650726, -0.06265407055616379, -0.11300963163375854, -0.039751436561346054, -0.050429005175828934, -0.029761891812086105, -0.05090925097465515, 0.09489558637142181, 0.004352911841124296, -0.09534718841314316, -0.13405443727970123, -0.01370926946401596, -0.1618979275226593, 0.15892250835895538, 0.012579603120684624, 0.046201955527067184, -0.19210097193717957, 0.11465331166982651, -0.03857925534248352, -0.08259090781211853, 0.030513519421219826, -0.12010065466165543, 0.03160654753446579, -0.008132083341479301, -0.019599268212914467, -0.049325279891490936, 0.061037879437208176, 0.08101806789636612, 0.018783701583743095, 0.005755073390901089, 0.018167443573474884, 0.05343452841043472, 0.05891622602939606, 0.10033947974443436, -0.02891627699136734, -0.0625043511390686, 0.0025436533614993095, -0.12051084637641907, -0.01122665498405695, -0.05357983708381653, -0.18095199763774872, 0.002246231772005558, 0.02455340512096882, 0.05192234739661217, 0.011778532527387142, 0.09955989569425583, -0.028496338054537773, -0.026898741722106934, 0.06898727267980576, 0.002862759632989764, -0.015707949176430702, -0.005368964280933142, -0.010934269987046719, 0.11485416442155838, -0.023099146783351898, 0.04774846136569977, -0.12022071331739426, 0.020393015816807747, -0.07851235568523407, -0.0019349842332303524, -0.06214260309934616, -0.04864754155278206, -0.0019346009939908981, -0.06985589861869812, 0.021118074655532837, -0.14833110570907593, -0.17990200221538544, -0.005064866971224546, 0.021302316337823868, -0.052403319627046585, -0.09162671118974686, -0.0982397273182869, -0.02586611732840538, 0.03574685752391815, -0.05873546749353409, 0.013170980848371983, -0.06884536147117615, 0.06542801111936569, 0.0029820678755640984, 0.05682007595896721, -0.14085575938224792, 0.08719147741794586, -0.12582023441791534, -0.023288866505026817, -0.061977192759513855, 0.1109607070684433, 0.024780582636594772, 0.1267160177230835, 0.004311583004891872, -0.0033308975398540497, -0.08729329705238342, 0.08271238207817078, -0.04243258014321327, 0.22770646214485168, -0.10479787737131119, -0.08809807151556015, 0.2632525563240051, -0.05423165112733841, -0.16432519257068634, 0.10179096460342407, -0.014350244775414467, 0.12198644131422043, 0.13850919902324677, 0.16080057621002197, 0.007628654129803181, 0.03313867375254631, 0.10115300863981247, 0.08631709218025208, -0.08573295921087265, -0.0611947737634182, 0.023627014830708504, -0.011463395319879055, -0.10670105367898941, 0.046802595257759094, 0.04794782027602196, 0.08188598603010178, -0.04982871189713478, -0.028600862249732018, -0.01972118206322193, -0.044152840971946716, 0.05264130234718323, 0.007675500120967627, 0.13217447698116302, -0.03674980252981186, -0.03692879155278206, -0.023745311424136162, 0.01699630729854107, -0.03115241602063179, 0.007061392068862915, -0.05687357112765312, 0.11091547459363937, -0.03406180441379547, 0.051789235323667526, -0.16953988373279572, -0.04873261600732803, -0.02087729424238205, 0.1402055323123932, 0.04973345249891281, 0.1329866498708725, 0.06287940591573715, -0.010758201591670513, 0.00859389640390873, 0.007998145185410976, 0.13181665539741516, 0.007865442894399166, -0.07660657912492752, -0.047718439251184464, 0.09176599979400635, -0.05973208695650101, 0.06147782504558563, -0.098741315305233, -0.004747362341731787, -0.01433002483099699, 0.08674649894237518, 0.006352655589580536, 0.029382232576608658, -0.006192679051309824, 0.003654100699350238, -0.06161240115761757, 0.017873648554086685, 0.12492607533931732, -0.01421504095196724, -0.07439801841974258, 0.22084392607212067, -0.15798072516918182, 0.18006981909275055, 0.18165533244609833, -0.3081994652748108, 0.024602634832262993, -0.08860466629266739, -0.036338552832603455, 0.03426366671919823, 0.0491504967212677, -0.034147560596466064, 0.16587987542152405, -0.016766328364610672, 0.201018825173378, -0.03547777235507965, -0.01287798210978508, -0.010399105958640575, -0.03656993433833122, -0.010632630437612534, 0.09065473079681396, 0.15122920274734497, -0.1677125245332718, 0.18270380795001984, 0.1660280078649521, 0.06873020529747009, 0.17776396870613098, 0.034313347190618515, -0.006856906693428755, 0.07112615555524826, -0.022670727223157883, -0.07675548642873764, -0.049287427216768265, -0.26302891969680786, -0.027947327122092247, 0.06471601128578186, 0.04510856419801712, 0.11924877762794495, -0.10971947014331818, -0.037208184599876404, 0.010892451740801334, -0.013165894895792007, 0.02132410928606987, 0.09682225435972214, 0.01171150617301464, 0.11804302036762238, -0.021027036011219025, -0.05209195241332054, 0.0898953229188919, 0.02727191150188446, -0.0787680521607399, 0.19168277084827423, -0.10074768215417862, -0.3233809769153595, -0.11354339867830276, -0.18166927993297577, -0.017843691632151604, 0.05878754332661629, 0.08049646019935608, -0.09228580445051193, -0.02625267766416073, -0.01639235019683838, 0.0758359357714653, -0.09145816415548325, -0.015880629420280457, -0.09367848187685013, 0.034986745566129684, -0.10827737301588058, -0.07011983543634415, -0.05141967162489891, -0.03368452936410904, -0.04457031562924385, 0.13157756626605988, -0.12242637574672699, 0.06396433711051941, 0.2076517641544342, 0.06227295100688934, 0.05622440204024315, -0.0229496993124485, 0.23288212716579437, -0.10842552781105042, 0.02383521944284439, 0.1717897206544876, -0.03566030040383339, 0.0727933868765831, 0.13435456156730652, 0.006721907295286655, -0.08144525438547134, 0.03465581312775612, -0.04592517390847206, -0.08630958944559097, -0.20441576838493347, -0.14156180620193481, -0.12814727425575256, 0.07913564145565033, 0.03285396471619606, 0.05478321388363838, 0.15024253726005554, 0.11386489123106003, 0.007987297140061855, 0.00976672861725092, -0.006888182368129492, 0.05438044294714928, 0.17482298612594604, -0.05838097631931305, 0.10041683167219162, -0.037591226398944855, -0.1924494504928589, 0.08022978901863098, 0.04309763014316559, 0.08280511945486069, 0.07474655658006668, 0.0856199786067009, 0.013537914492189884, 0.03723837807774544, 0.10897084325551987, 0.1165735274553299, 0.031679023057222366, -0.038079675287008286, -0.04882059991359711, -0.026300756260752678, -0.03285675123333931, 0.05745977535843849, 0.07790146768093109, -0.1608346849679947, -0.06348084658384323, -0.06350091099739075, 0.07662643492221832, 0.09017108380794525, 0.11811108142137527, -0.21219493448734283, 0.01579318381845951, 0.092556893825531, -0.0494147390127182, -0.1304239183664322, 0.07402537018060684, -0.00466050673276186, -0.1397053301334381, 0.037663187831640244, -0.014095795340836048, 0.1359514445066452, -0.0778401643037796, 0.10336452722549438, -0.08307972550392151, -0.06147889420390129, 0.03632286190986633, 0.1355396956205368, -0.30774354934692383, 0.2137020230293274, -0.022472934797406197, -0.05296783149242401, -0.10508129745721817, -0.011727629229426384, 0.020913105458021164, 0.09079049527645111, 0.10090240091085434, -0.0025442070327699184, 0.0061299679800868034, -0.0345483273267746, -0.053218815475702286, 0.024456629529595375, 0.07957815378904343, -0.08542889356613159, 0.0017540202243253589, -0.02361489273607731, -0.004407065454870462, -0.032844748347997665, -0.01189463958144188, -0.011617658659815788, -0.16786961257457733, 0.06556065380573273, -0.002625665394589305, 0.11129079759120941, 0.03491498529911041, 0.0024013579823076725, -0.1009332686662674, 0.19977013766765594, 0.01796281896531582, -0.08052749931812286, -0.08830537647008896, -0.03254766762256622, 0.03660419583320618, -0.06121435388922691, 0.027481911703944206, -0.06916457414627075, 0.033381566405296326, -0.06441576033830643, -0.18325145542621613, 0.1268530637025833, -0.10945470631122589, -0.03609596937894821, -0.04321056231856346, 0.18323224782943726, -0.00929707009345293, -0.0011623724130913615, 0.05866571143269539, 0.0032208464108407497, -0.1347510665655136, -0.10740556567907333, 0.020214511081576347, -0.015275230631232262, 0.009142245166003704, 0.05559912323951721, -0.009665844030678272, 0.00045268211397342384, -0.039558928459882736, -0.023234419524669647, 0.32348164916038513, 0.10732097923755646, -0.04944206401705742, 0.17007054388523102, 0.13087597489356995, -0.0827672928571701, -0.30699312686920166, -0.10971353948116302, -0.10529600828886032, -0.026918673887848854, -0.037983208894729614, -0.19617970287799835, 0.09504909813404083, -0.03528566658496857, -0.022136637941002846, 0.11253651231527328, -0.2759084105491638, -0.0770430713891983, 0.1826775223016739, 0.003314757253974676, 0.3998824954032898, -0.10265109688043594, -0.08777514100074768, -0.06741699576377869, -0.1120782196521759, 0.2033512443304062, -0.05560711398720741, 0.08663415163755417, -0.00517998356372118, 0.15513743460178375, 0.055607251822948456, -0.02176513522863388, 0.08932057023048401, -0.005811662413179874, -0.0546204075217247, -0.1219351515173912, -0.03444604203104973, -0.009159418754279613, 0.007239421829581261, 0.03589896112680435, -0.04242607578635216, 0.01279151439666748, -0.1399589478969574, -0.045490626245737076, -0.0764620453119278, 0.024699507281184196, 0.021008269861340523, -0.0652410089969635, -0.01643640361726284, -0.03945036977529526, -0.012804778292775154, 0.03164318576455116, 0.15236099064350128, -0.06478006392717361, 0.1476556956768036, 0.04904455319046974, 0.15412139892578125, -0.14745712280273438, -0.02258288487792015, -0.06896031647920609, -0.05498642474412918, 0.04900865629315376, -0.10053684562444687, 0.050061121582984924, 0.1202658861875534, -0.0742902010679245, 0.0987328365445137, 0.0922594666481018, -0.01938629150390625, 0.0012483424507081509, 0.1226617842912674, -0.2489612102508545, -0.07742628455162048, -0.10509459674358368, 0.013337249867618084, 0.10138551890850067, 0.06995654851198196, 0.17304721474647522, -0.0037713919300585985, -0.036284226924180984, -0.0064643872901797295, 0.025414984673261642, -0.03540204465389252, 0.05724727362394333, -0.002706433180719614, 0.016663886606693268, -0.15213344991207123, 0.060368724167346954, -0.00024176653823815286, -0.1438901126384735, -0.013603870756924152, 0.16073721647262573, -0.11208858340978622, -0.15145981311798096, -0.007263668347150087, 0.13685113191604614, -0.13171035051345825, -0.03302847594022751, -0.03708777576684952, -0.170182466506958, 0.07439173012971878, 0.1024777740240097, 0.08549231290817261, 0.08025266975164413, -0.06620611250400543, -0.00807863101363182, -0.011656313203275204, -0.026087598875164986, 0.031810320913791656, -0.023377234116196632, -0.09044221043586731, 0.03872343525290489, -0.026654237881302834, 0.13591371476650238, -0.09607382118701935, -0.09331836551427841, -0.135749951004982, 0.039314381778240204, -0.12405620515346527, -0.08138058334589005, -0.12200927734375, -0.0591500885784626, 0.00224387738853693, -0.0001289021165575832, -0.035674065351486206, -0.06687422841787338, -0.13582271337509155, 0.04366770386695862, -0.04484611004590988, 0.0013091047294437885, -0.040241483598947525, 0.04561002552509308, 0.06766383349895477, -0.03493715822696686, 0.13722217082977295, 0.11722734570503235, -0.07864081114530563, 0.08946478366851807, -0.16657429933547974, -0.0683990865945816, 0.08854512125253677, 0.008173754438757896, 0.06165994703769684, 0.06743349134922028, 0.033807408064603806, 0.06109451875090599, 0.04151686280965805, 0.03488299250602722, 0.01739438995718956, -0.09271225333213806, 0.015541021712124348, 0.022296719253063202, -0.1294609159231186, -0.04801803454756737, -0.029226921498775482, 0.00939185917377472, 0.008117396384477615, 0.11003357172012329, -0.0426274873316288, 0.09439733624458313, -0.05888751894235611, 0.036728594452142715, 0.016222506761550903, -0.16461637616157532, -0.020102784037590027, -0.11915475130081177, 0.028684545308351517, -0.0033096212428063154, 0.25625869631767273, 0.06346847862005234, 0.020517030730843544, 0.01250078622251749, 0.08567021042108536, 0.07241600006818771, 0.02562166005373001, 0.1956365555524826, 0.10854171961545944, -0.05020022392272949, -0.12334850430488586, 0.09686340391635895, 0.034720368683338165, 0.06432123482227325, 0.13385434448719025, -0.026959087699651718, 0.002498799469321966, 0.11019360274076462, 0.011678861454129219, 0.04961980879306793, -0.09859088063240051, -0.16400282084941864, -0.00994415208697319, 0.061864156275987625, -0.04559077322483063, 0.12240655720233917, 0.11382720619440079, -0.020697353407740593, 0.03180128335952759, -0.010503606870770454, -0.05694027617573738, -0.16998925805091858, -0.1630837321281433, -0.08357038348913193, -0.11794789135456085, -0.0027763545513153076, -0.11386270076036453, 0.013879159465432167, 0.06452289968729019, 0.0604364387691021, -0.09019444137811661, 0.08891061693429947, 0.0687386617064476, -0.11843101680278778, 0.08828350901603699, -0.033263903111219406, 0.07249268144369125, 0.0015160300536081195, 0.003872724948450923, -0.13800905644893646, 0.032393742352724075, -0.008493867702782154, 0.04159298539161682, -0.09244006127119064, 0.022458361461758614, -0.11297028511762619, -0.07659684121608734, -0.07971972227096558, 0.05093973129987717, -0.03541257977485657, 0.1390930563211441, 0.001295371213927865, -0.035233911126852036, 0.024190181866288185, 0.22729112207889557, -0.06350252777338028, -0.030667411163449287, -0.0618741400539875, 0.21414142847061157, 0.024466563016176224, 0.10703565180301666, -0.016775688156485558, 0.019240234047174454, -0.0764411985874176, 0.3689337372779846, 0.344390869140625, -0.1225387305021286, -0.0015968306688591838, 0.031062176451086998, 0.036916591227054596, 0.11621878296136856, 0.12602226436138153, 0.057955991476774216, 0.2995031177997589, -0.08396036922931671, -0.002026971662417054, -0.02688612788915634, -0.03624163940548897, -0.04409930482506752, 0.10547586530447006, 0.06835740804672241, -0.03330419585108757, -0.027012333273887634, 0.1376710683107376, -0.2966996431350708, 0.12323499470949173, -0.15714547038078308, -0.1487535685300827, -0.06873904913663864, -0.005042468197643757, 0.08589684963226318, 0.04748665541410446, 0.1069009080529213, -0.019124338403344154, -0.08203735202550888, 0.05766449123620987, 0.0320524163544178, -0.22844897210597992, 0.011852608993649483, 0.08361081779003143, -0.06153005734086037, 0.011767351068556309, -0.017906347289681435, 0.038472190499305725, 0.07790610194206238, 0.025976579636335373, -0.032770540565252304, 0.06325861811637878, -0.005814229138195515, -0.05033424496650696, 0.04302205145359039, 0.05059972032904625, 0.017107632011175156, -0.1511564701795578, 0.07320158183574677, -0.1762860119342804, 0.0566408596932888, -0.005331212189048529, -0.04948166385293007, 0.000018263708625454456, 0.01998119056224823, -0.06808236241340637, 0.05880929157137871, 0.0952666699886322, -0.012173139490187168, -0.002317852806299925, -0.056667573750019073, 0.007662574760615826, -0.0679154172539711, -0.0747012197971344, -0.10497893393039703, -0.1338900774717331, -0.11392296850681305, 0.10846775025129318, -0.011928223073482513, -0.19833622872829437, 0.02906924858689308, -0.11258108913898468, 0.04933213070034981, -0.13360801339149475, 0.08599711954593658, 0.1282832771539688, 0.021543797105550766, -0.01265349704772234, 0.04020093381404877, 0.01591683179140091, 0.08550478518009186, -0.09200563281774521, -0.10515180230140686 ]
null
null
transformers
Fine tuning LayoutLMv2 model on Vietnamese bill dataset ```python from transformers import LayoutLMv2ForTokenClassification model = LayoutLMv2ForTokenClassification.from_pretrained('cuongngm/layoutlm-bill', num_labels=len(labels)) ``` labels = ['price', 'storename', 'total_cost', 'phone', 'address', 'unitprice', 'item', 'subitem', 'other', 'time', 'unit', 'total refunds', 'total_qty', 'seller', 'total_received']
{}
token-classification
cuongngm/layoutlm-bill
[ "transformers", "pytorch", "layoutlmv2", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #layoutlmv2 #token-classification #autotrain_compatible #endpoints_compatible #region-us
Fine tuning LayoutLMv2 model on Vietnamese bill dataset labels = ['price', 'storename', 'total_cost', 'phone', 'address', 'unitprice', 'item', 'subitem', 'other', 'time', 'unit', 'total refunds', 'total_qty', 'seller', 'total_received']
[]
[ "TAGS\n#transformers #pytorch #layoutlmv2 #token-classification #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 41 ]
[ "passage: TAGS\n#transformers #pytorch #layoutlmv2 #token-classification #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.05778699368238449, 0.03974124416708946, -0.008774658665060997, 0.03134288638830185, 0.16732493042945862, 0.034163497388362885, 0.07008584588766098, 0.09666585177183151, 0.01093202829360962, -0.03180955722928047, 0.11920414119958878, 0.2536323070526123, -0.02541540376842022, 0.12405043840408325, -0.08940679579973221, -0.2751603424549103, 0.05289534479379654, 0.07812897115945816, -0.0509544312953949, 0.10437314957380295, 0.09426096081733704, -0.096477210521698, 0.08267372101545334, -0.017689958214759827, -0.15998992323875427, 0.031330011785030365, 0.031133079901337624, -0.11463280022144318, 0.09397059679031372, 0.0392887182533741, 0.17064033448696136, 0.03290946036577225, -0.023194191977381706, -0.1294083297252655, 0.026020493358373642, 0.026482179760932922, -0.06576782464981079, 0.06989692151546478, 0.09297885000705719, -0.0724194347858429, 0.0004459601186681539, 0.023980289697647095, 0.02180878445506096, 0.04127785190939903, -0.1154012456536293, -0.11993054300546646, -0.028065886348485947, 0.06803034991025925, 0.06359178572893143, 0.028339453041553497, 0.04528860002756119, 0.21416284143924713, -0.12959837913513184, 0.1112191379070282, 0.11915595829486847, -0.2852734923362732, -0.005756948608905077, 0.14852507412433624, -0.0030514465179294348, -0.03204598277807236, -0.02021097019314766, 0.04491802304983139, 0.021199705079197884, 0.018341125920414925, 0.027534527704119682, -0.06974580883979797, -0.10134192556142807, 0.02801455929875374, -0.09212496131658554, -0.03259341046214104, 0.19970397651195526, -0.03201518580317497, 0.06338144093751907, -0.0024487206246703863, -0.10643959790468216, -0.02835744433104992, -0.020721422508358955, 0.0026460534427314997, -0.027708260342478752, 0.0392601303756237, 0.006277269683778286, 0.015453265979886055, -0.10328508913516998, 0.008988311514258385, -0.2241227626800537, 0.22883152961730957, 0.021964799612760544, 0.07942602038383484, -0.16387812793254852, 0.07133140414953232, 0.00432487390935421, -0.09671732783317566, 0.0443706251680851, -0.10392092913389206, 0.00154713389929384, -0.050665080547332764, -0.02832726575434208, 0.016673464328050613, 0.07684864103794098, 0.11847390979528427, 0.09207979589700699, 0.04180806502699852, 0.0005521529237739742, 0.09081310778856277, 0.031043238937854767, 0.10453872382640839, -0.008934730663895607, -0.026695383712649345, 0.05356279015541077, -0.11922793090343475, 0.013831875286996365, -0.04324895516037941, -0.1353645622730255, -0.052963029593229294, 0.07332616299390793, 0.10282445698976517, 0.025591719895601273, 0.06446130573749542, -0.06997820734977722, -0.05468494072556496, 0.10225611925125122, -0.07712759077548981, 0.02380007691681385, 0.0021684167440980673, 0.010550598613917828, 0.11851963400840759, -0.007207420188933611, 0.0061905160546302795, -0.04666688293218613, 0.11484462022781372, -0.05645641312003136, 0.006140651181340218, -0.045840680599212646, -0.06910344213247299, 0.046140436083078384, -0.14090818166732788, 0.044356945902109146, -0.16669055819511414, -0.11685968935489655, 0.035181231796741486, 0.041286733001470566, 0.0077555058524012566, -0.04102035239338875, 0.020996762439608574, 0.0021928229834884405, 0.0021646725945174694, -0.06196162849664688, -0.014183941297233105, -0.04863809049129486, 0.06290029734373093, -0.011364633217453957, 0.06171398237347603, -0.08672089129686356, 0.05835375934839249, -0.09694076329469681, 0.02440287359058857, -0.10379631072282791, -0.0006411506328731775, -0.052533239126205444, 0.16130277514457703, -0.03289404883980751, -0.06531120091676712, -0.040451355278491974, 0.00815553031861782, -0.042518362402915955, 0.11008085310459137, -0.08943292498588562, -0.1067400798201561, 0.1347787082195282, -0.11807852983474731, -0.1288880556821823, 0.06096221134066582, 0.006109313108026981, -0.010931028053164482, 0.06116866692900658, 0.09868606925010681, 0.12345214188098907, -0.034919194877147675, 0.05594572052359581, 0.12384092807769775, -0.15172681212425232, -0.15482929348945618, 0.01971510797739029, 0.012406782247126102, -0.12012148648500443, 0.05494942516088486, 0.06585130095481873, 0.07786384969949722, -0.07147765904664993, -0.04589385911822319, -0.04318201541900635, -0.015889273956418037, 0.10509875416755676, 0.054638952016830444, 0.09694869816303253, -0.0420546755194664, 0.02332139015197754, 0.04700779914855957, 0.03256576508283615, 0.034252576529979706, 0.015309659764170647, -0.08613849431276321, 0.11357367783784866, -0.07669250667095184, 0.009631022810935974, -0.19131550192832947, -0.1182670146226883, 0.005608411505818367, 0.059924058616161346, -0.033211253583431244, 0.13797996938228607, 0.07247842103242874, -0.03458826616406441, -0.0023120781406760216, -0.027080101892352104, 0.15149423480033875, 0.03204483911395073, -0.06698699295520782, -0.10485642403364182, -0.0035018303897231817, -0.06614898145198822, -0.03379574418067932, -0.03812050446867943, 0.013367078267037868, 0.07384876906871796, 0.16941611468791962, 0.01876668632030487, 0.07587146013975143, -0.015248815529048443, 0.05340377986431122, -0.06478475034236908, -0.005989662371575832, 0.11225032806396484, -0.008588879369199276, -0.05581550672650337, 0.12216658890247345, -0.11563436686992645, 0.3249482214450836, 0.185505673289299, -0.2773588299751282, 0.017563272267580032, -0.04416828602552414, -0.023483986034989357, -0.0004700309655163437, 0.023501502349972725, 0.05075979232788086, 0.02893635630607605, 0.013826107606291771, 0.16484588384628296, -0.017082910984754562, -0.05047663673758507, 0.015549123287200928, -0.06421256810426712, -0.04786786064505577, 0.06996762007474899, 0.101493239402771, -0.19506941735744476, 0.1725781261920929, 0.21628913283348083, 0.0018555746646597981, 0.09177366644144058, -0.013344105333089828, 0.023524761199951172, 0.031723566353321075, -0.03686145320534706, -0.027301767840981483, -0.008122175000607967, -0.1615874469280243, -0.02990984171628952, 0.08935689181089401, 0.029235607013106346, 0.05835922434926033, -0.1354506015777588, -0.01868385449051857, 0.016246246173977852, 0.039081692695617676, -0.015885232016444206, 0.09966475516557693, 0.051478609442710876, 0.07956003397703171, -0.00868115946650505, -0.07854697853326797, 0.10971519351005554, 0.004792613442987204, -0.06752726435661316, 0.15722130239009857, -0.12969624996185303, -0.2947699725627899, -0.14390388131141663, -0.19127389788627625, -0.05046501010656357, 0.04131399095058441, 0.048620935529470444, -0.09950754046440125, -0.059498172253370285, 0.06847665458917618, 0.001623176853172481, -0.06055719032883644, 0.07062265276908875, -0.03540179878473282, 0.08509432524442673, -0.018519936129450798, -0.06954783946275711, -0.055927615612745285, -0.0414201058447361, -0.02613709680736065, 0.13159692287445068, -0.07070932537317276, 0.061695270240306854, 0.16283351182937622, -0.013026673346757889, 0.06530477106571198, -0.008078161627054214, 0.14188720285892487, -0.04140574112534523, -0.013613970950245857, 0.22200967371463776, -0.05633757263422012, 0.08353669196367264, 0.1540241837501526, 0.041764017194509506, -0.0403645895421505, -0.00221289717592299, -0.044953569769859314, -0.11235775053501129, -0.1929023265838623, -0.14358146488666534, -0.11976822465658188, 0.022163430228829384, 0.0670013353228569, 0.06063150241971016, 0.10678660124540329, 0.10436863452196121, 0.02886897139251232, 0.03517468273639679, -0.05179718881845474, 0.07515324652194977, 0.2329932153224945, 0.0010029739933088422, 0.13521945476531982, -0.06918307393789291, -0.12653949856758118, 0.0786515474319458, 0.06753218173980713, 0.16002078354358673, 0.08846481889486313, -0.03066476806998253, 0.0255486648529768, 0.13737152516841888, 0.16751514375209808, 0.1297721564769745, 0.019602876156568527, -0.03937277942895889, 0.005858137737959623, 0.00009884304745355621, -0.050046294927597046, 0.01805928908288479, 0.11923974007368088, -0.11102869361639023, -0.049453530460596085, -0.08008626103401184, 0.0741138905286789, 0.10086892545223236, 0.05682622641324997, -0.22243084013462067, 0.023685850203037262, 0.07992563396692276, 0.00009739531378727406, -0.07107511907815933, 0.041660066694021225, -0.05711185559630394, -0.13931119441986084, 0.09855468571186066, -0.04457836225628853, 0.1122964546084404, -0.08670590817928314, 0.04690837860107422, -0.012499743141233921, -0.044510915875434875, 0.04440256580710411, 0.11083681136369705, -0.23993900418281555, 0.23940879106521606, 0.013277639634907246, -0.06741482764482498, -0.08327994495630264, 0.0008145290776155889, 0.038868650794029236, 0.2019420564174652, 0.054901283234357834, 0.018034491688013077, -0.1020551547408104, -0.18951259553432465, -0.04604224115610123, 0.011967078782618046, 0.08163987845182419, -0.0021220948547124863, -0.02166498824954033, -0.03929366171360016, -0.026156701147556305, -0.018793383613228798, -0.033498529344797134, -0.0002000758977374062, -0.11958488076925278, 0.06002609431743622, 0.04885181412100792, 0.01973840221762657, 0.0027429915498942137, -0.06634359061717987, -0.11346053332090378, 0.23760053515434265, -0.08375486731529236, -0.07633186131715775, -0.12529438734054565, -0.08247952908277512, 0.07333134859800339, -0.08693837374448776, 0.07178293913602829, -0.09247510880231857, 0.046413157135248184, -0.03039172664284706, -0.20090904831886292, 0.1285516619682312, -0.12863905727863312, -0.041191618889570236, -0.06193387880921364, 0.14642776548862457, -0.09815216064453125, 0.018241681158542633, 0.01747041381895542, 0.017957033589482307, -0.07901791483163834, -0.08617274463176727, 0.016419190913438797, 0.032105859369039536, 0.03878277912735939, 0.050481166690588, -0.057005301117897034, -0.03258875384926796, 0.018971407786011696, 0.02980881556868553, 0.24965596199035645, 0.18610970675945282, -0.07782752066850662, 0.12343915551900864, 0.11683381348848343, -0.044916391372680664, -0.3175329566001892, -0.07109321653842926, -0.09933280944824219, -0.040871161967515945, -0.03415766730904579, -0.13326579332351685, 0.1367558240890503, 0.022940225899219513, -0.04264963045716286, 0.07325160503387451, -0.16267172992229462, -0.08590833842754364, 0.1943003535270691, -0.0010218152310699224, 0.3578220307826996, -0.06566820293664932, -0.0798448771238327, -0.04600640386343002, -0.15766489505767822, 0.09459023177623749, 0.002373378723859787, 0.06756804883480072, -0.04350960999727249, 0.041615575551986694, 0.03665178641676903, -0.06291329115629196, 0.10352667421102524, 0.03191421926021576, 0.04651443660259247, -0.09885291755199432, -0.0890415608882904, 0.046507179737091064, -0.036033984273672104, 0.009482534602284431, 0.049101680517196655, 0.03696424141526222, -0.1246163547039032, -0.014286420308053493, -0.06409062445163727, 0.08560869097709656, 0.04267823323607445, -0.057284217327833176, 0.007375302724540234, -0.029944051057100296, -0.005789119750261307, -0.00421120272949338, 0.2532790005207062, 0.02436981536448002, 0.11846763640642166, 0.14298470318317413, 0.09473973512649536, -0.1688377559185028, -0.027834711596369743, -0.07952225208282471, -0.060129642486572266, 0.07923906296491623, -0.04621187970042229, 0.08262167870998383, 0.1278066486120224, -0.029770877212285995, 0.03190390393137932, 0.11209224164485931, 0.05402187630534172, -0.043240662664175034, 0.153057262301445, -0.1937270313501358, 0.0442107692360878, -0.027363670989871025, 0.0004532164894044399, 0.048240721225738525, 0.12736783921718597, 0.1153838038444519, 0.041781287640333176, -0.023382162675261497, 0.006082751788198948, -0.011888482607901096, -0.053514204919338226, 0.0843736082315445, 0.07893380522727966, 0.048407651484012604, -0.1457613855600357, 0.05378333106637001, 0.027309643104672432, -0.09572086483240128, -0.05806012824177742, 0.06238840892910957, -0.1553935408592224, -0.11065811663866043, 0.011939547024667263, 0.09843193739652634, -0.13321523368358612, -0.06422793120145798, -0.053000614047050476, -0.1297502964735031, 0.07072881609201431, 0.1531151980161667, 0.13062144815921783, 0.09846726804971695, -0.04803524166345596, -0.04502645507454872, -0.02932196669280529, -0.018670959398150444, 0.010611818172037601, 0.060962654650211334, -0.18999697268009186, 0.038105450570583344, -0.00807112455368042, 0.16742603480815887, -0.09692341834306717, -0.06846752017736435, -0.1336631029844284, 0.050985608249902725, -0.08091622591018677, -0.06551729142665863, -0.0959804505109787, -0.016431493684649467, 0.024287445470690727, -0.06683551520109177, -0.04708725959062576, -0.014623597264289856, -0.11040210723876953, 0.04908880591392517, 0.003515340853482485, 0.03214099630713463, -0.053075410425662994, -0.046283118426799774, 0.0833362266421318, -0.03921474888920784, 0.08558426797389984, 0.08461696654558182, -0.05467032268643379, 0.06570801883935928, -0.07886151969432831, -0.13805243372917175, 0.14046891033649445, 0.04486068710684776, 0.09652195870876312, 0.017176570370793343, 0.038593705743551254, 0.073255255818367, 0.011806592345237732, 0.050993017852306366, 0.07536526024341583, -0.12348739057779312, 0.032782990485429764, -0.06000329554080963, -0.1613915115594864, -0.02635570615530014, -0.05736926570534706, 0.09324377775192261, 0.017503922805190086, 0.15200082957744598, -0.004327565431594849, 0.07368715852499008, -0.06430529057979584, 0.002900370629504323, -0.031497515738010406, -0.18479296565055847, -0.03858397901058197, -0.04583295062184334, 0.012515569105744362, -0.006835192907601595, 0.2555631995201111, 0.04336102306842804, 0.01708845980465412, 0.04016496613621712, 0.09534046053886414, -0.011532729491591454, 0.017750654369592667, 0.15677224099636078, 0.09642043709754944, -0.01634051837027073, -0.03656552731990814, 0.08647118508815765, 0.016823289915919304, -0.0778827890753746, 0.11005629599094391, 0.03772580996155739, -0.04265972599387169, 0.05467662960290909, 0.032995134592056274, 0.01356444600969553, -0.16403166949748993, -0.1561090648174286, -0.059449054300785065, 0.07617752999067307, 0.023158937692642212, 0.02678116224706173, 0.11130081862211227, -0.028325339779257774, 0.02580459602177143, -0.015071915462613106, -0.02922098897397518, -0.18761996924877167, -0.12038294225931168, -0.09932973980903625, -0.12261207401752472, 0.011479394510388374, -0.03784283250570297, -0.028312478214502335, 0.11179319769144058, 0.049811072647571564, -0.024361172690987587, 0.05118126422166824, 0.0034409374929964542, -0.011275487020611763, 0.008979369886219501, -0.01128071453422308, 0.010722801089286804, -0.01220316719263792, -0.014854183420538902, -0.13901731371879578, -0.02910418063402176, -0.04148200526833534, 0.005580550990998745, -0.04760715737938881, 0.022339073941111565, -0.09673412889242172, -0.11421101540327072, -0.04191211983561516, 0.023931344971060753, -0.05582542344927788, 0.08719328790903091, -0.005910171195864677, 0.02947339415550232, 0.018003767356276512, 0.14212849736213684, -0.07178422808647156, -0.08833976089954376, -0.047104042023420334, 0.24008573591709137, 0.05422874167561531, 0.09437024593353271, 0.007490250747650862, 0.011500077322125435, -0.08013186603784561, 0.277520090341568, 0.2691807746887207, -0.03454088047146797, 0.05437979847192764, 0.020099295303225517, 0.014366094022989273, 0.08129112422466278, 0.12801529467105865, 0.0743185356259346, 0.20950961112976074, -0.08197618275880814, -0.052265483886003494, -0.051328178495168686, -0.016310935840010643, -0.11299603432416916, 0.0442340113222599, 0.028652755543589592, -0.05005773529410362, -0.06117720156908035, 0.07672857493162155, -0.1584158092737198, 0.17652533948421478, 0.05530504137277603, -0.17637746036052704, -0.0781833678483963, -0.02176763489842415, 0.1454465687274933, -0.003720478154718876, 0.06764408946037292, -0.04851401224732399, -0.0823097974061966, 0.05232630670070648, 0.010331254452466965, -0.22684168815612793, -0.05646951496601105, 0.08493104577064514, 0.016064852476119995, 0.019817477092146873, -0.032334569841623306, 0.07010794430971146, 0.07780664414167404, 0.05638008192181587, -0.058445267379283905, 0.020888328552246094, -0.00633151875808835, -0.08150964975357056, -0.00883753877133131, -0.0008433987968601286, 0.0026467349380254745, -0.06789134442806244, 0.03673473745584488, -0.13965782523155212, 0.017946675419807434, -0.07979290932416916, -0.01338474452495575, -0.015799034386873245, 0.02234521508216858, -0.017149455845355988, 0.05813609063625336, 0.06571108847856522, 0.0014315495500341058, -0.04241492599248886, -0.06686784327030182, -0.009505338035523891, 0.03682727366685867, -0.11386525630950928, -0.1286592334508896, -0.0847516804933548, -0.04587195813655853, 0.06421384960412979, -0.0035427496768534184, -0.05032660812139511, -0.049098946154117584, -0.09952948987483978, 0.005699475761502981, -0.14046946167945862, 0.06746573001146317, 0.05839042738080025, 0.03217732906341553, -0.007235072087496519, -0.03762112557888031, 0.02165811136364937, 0.03594008460640907, -0.13543765246868134, -0.10064350068569183 ]
null
null
transformers
# Harry Potter DialoGPT Model
{"tags": ["conversational"]}
text-generation
cutiebunny639/DialoGPT-small-harry
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Harry Potter DialoGPT Model
[ "# Harry Potter DialoGPT Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Harry Potter DialoGPT Model" ]
[ 51, 8 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Harry Potter DialoGPT Model" ]
[ -0.0009023238671943545, 0.07815738022327423, -0.006546166725456715, 0.07792752981185913, 0.10655936598777771, 0.048972971737384796, 0.17639793455600739, 0.12185695022344589, 0.016568755730986595, -0.04774167761206627, 0.11647630482912064, 0.2130284160375595, -0.002118367003276944, 0.024608047679066658, -0.05022026598453522, -0.3065771162509918, 0.0474756620824337, 0.014356585219502449, -0.07174845039844513, 0.11724270135164261, 0.09064973145723343, -0.046179238706827164, 0.08330509811639786, -0.009135239757597446, -0.13198648393154144, -0.039482954889535904, 0.019292812794446945, -0.11745545268058777, 0.1662212759256363, 0.05298272892832756, 0.02469746209681034, -0.008447164669632912, -0.06598151475191116, -0.15036040544509888, 0.037190426141023636, -0.027472136542201042, -0.01080626156181097, 0.05462246760725975, 0.023526115342974663, -0.07521048933267593, 0.170567125082016, 0.17678891122341156, 0.0833497866988182, 0.0349111407995224, -0.14917024970054626, -0.045548245310783386, 0.008950977586209774, 0.05421316996216774, -0.017893504351377487, 0.09349167346954346, -0.019903047010302544, 0.11801653355360031, -0.04491448402404785, 0.09210366010665894, 0.15255063772201538, -0.4016275703907013, -0.027563704177737236, 0.08920855820178986, 0.05989706888794899, 0.12076901644468307, -0.10560955852270126, 0.03972794860601425, -0.0039703017100691795, 0.01236654631793499, -0.014540530741214752, -0.08304883539676666, -0.07308239489793777, 0.032504837960004807, -0.1272556483745575, 0.008525865152478218, 0.23756256699562073, -0.10643257945775986, 0.037069112062454224, -0.09791990369558334, -0.07414398342370987, 0.048336777836084366, -0.053761593997478485, -0.081727035343647, -0.054839808493852615, 0.06347949057817459, 0.004366500303149223, -0.06301609426736832, -0.08326146006584167, -0.0006536149303428829, -0.12781435251235962, 0.17595994472503662, 0.061243366450071335, 0.041611745953559875, -0.21322020888328552, 0.08940251916646957, 0.04477722570300102, -0.04711297154426575, 0.007116159424185753, -0.11796226352453232, 0.04023287072777748, 0.005483259446918964, -0.03256071358919144, -0.021854614838957787, 0.0393419973552227, 0.13909944891929626, -0.01777748204767704, 0.03252175822854042, 0.006831915583461523, 0.05811219662427902, 0.08162496984004974, 0.02222144603729248, 0.019291909411549568, -0.0818009302020073, 0.019385190680623055, -0.08128736168146133, -0.0030400939285755157, -0.048940129578113556, -0.17071883380413055, -0.07477642595767975, 0.052610911428928375, 0.020047198981046677, 0.03746970370411873, 0.08054786175489426, -0.0017944995779544115, -0.05560554191470146, 0.03284840285778046, 0.01671096310019493, -0.020622212439775467, -0.010361049324274063, -0.02412462793290615, 0.19123271107673645, 0.019619356840848923, 0.014111656695604324, -0.12379156798124313, 0.10023640841245651, -0.08179095387458801, 0.0037731381598860025, 0.02743307314813137, -0.04204464703798294, -0.004716555587947369, 0.02917117439210415, 0.023101668804883957, -0.1252521574497223, -0.1099385917186737, -0.0030569476075470448, -0.012054097838699818, -0.036421261727809906, -0.10490952432155609, -0.08483029156923294, -0.012153145857155323, 0.0449371263384819, -0.013397793285548687, 0.007936403155326843, -0.05143149942159653, 0.0985720232129097, -0.0514979362487793, 0.09873400628566742, -0.08342572301626205, 0.06359215080738068, -0.09124887734651566, -0.061886150389909744, -0.11452563107013702, 0.05216052383184433, 0.012905281968414783, 0.066250741481781, 0.016998225823044777, -0.044836658984422684, -0.014836243353784084, 0.05253177136182785, -0.07656687498092651, 0.1940697431564331, -0.041674621403217316, -0.12459053844213486, 0.24146439135074615, -0.09138800948858261, -0.1802034229040146, 0.12973085045814514, -0.022254703566432, 0.08523941785097122, 0.12802475690841675, 0.20380465686321259, -0.00019822151807602495, -0.01302915159612894, 0.07281201332807541, 0.07031642645597458, -0.09803894907236099, 0.06239739805459976, 0.029653839766979218, -0.008071083575487137, -0.08906278014183044, 0.05762826278805733, 0.046033453196287155, -0.010650773532688618, -0.035073768347501755, -0.001896020956337452, -0.012895751744508743, -0.022185025736689568, 0.14126582443714142, -0.02006692811846733, 0.1300428807735443, -0.06926563382148743, -0.03515486419200897, -0.009500149637460709, 0.03533667325973511, -0.04091939330101013, 0.08151165395975113, -0.0436173714697361, 0.10586477071046829, 0.09034156054258347, 0.053724925965070724, -0.13120363652706146, 0.00466286763548851, -0.015246815048158169, 0.17014820873737335, 0.08964069187641144, 0.05222717300057411, 0.06265474855899811, -0.0020888058934360743, -0.06708643585443497, 0.045407816767692566, 0.13778303563594818, -0.037020038813352585, -0.12218865007162094, -0.1755627691745758, 0.051157694309949875, -0.045444171875715256, 0.10855234414339066, -0.10010123997926712, 0.022670533508062363, -0.055906031280756, 0.07772238552570343, -0.024998966604471207, 0.020512236282229424, -0.0013405600329861045, -0.021700702607631683, -0.08356887847185135, -0.002377772703766823, 0.08597290515899658, -0.02048647589981556, -0.06707409024238586, 0.16556480526924133, -0.16400809586048126, 0.1631954461336136, 0.2116095870733261, -0.28542569279670715, -0.005696662236005068, -0.15163889527320862, -0.0208092350512743, 0.019645055755972862, 0.07834604382514954, 0.026225795969367027, 0.2044338881969452, -0.012928472831845284, 0.16565458476543427, -0.05699567869305611, -0.07730039209127426, -0.06881127506494522, -0.048101142048835754, 0.013522743247449398, 0.09095205366611481, 0.04542696103453636, -0.11962861567735672, 0.13119758665561676, 0.1054433062672615, 0.06484298408031464, 0.12711186707019806, 0.1030748188495636, -0.008113685995340347, 0.07252490520477295, -0.03624548763036728, -0.03462279960513115, -0.09254947304725647, -0.30446043610572815, -0.04840317741036415, 0.0939924493432045, 0.007963384501636028, 0.09285714477300644, -0.0919896736741066, -0.03311870992183685, 0.006042704917490482, 0.009473444893956184, 0.028337622061371803, 0.09653715789318085, 0.013490920886397362, 0.15320514142513275, -0.008011690340936184, -0.03430786728858948, 0.05891305208206177, 0.017982570454478264, -0.09147711098194122, 0.17280617356300354, -0.17050009965896606, -0.27190929651260376, -0.06990014761686325, -0.21745692193508148, -0.013139115646481514, 0.05258983001112938, 0.0786920040845871, -0.11818131804466248, -0.018352627754211426, -0.006239492911845446, 0.05685517191886902, -0.2425733357667923, 0.0004911290016025305, -0.1354890614748001, 0.0501418262720108, -0.1974833607673645, -0.09718500077724457, -0.02271542325615883, -0.013450481928884983, -0.0464281290769577, 0.13365240395069122, -0.1448695808649063, -0.011572926305234432, 0.2329535037279129, 0.032479673624038696, 0.027794739231467247, -0.05020907148718834, 0.19788463413715363, -0.0958966314792633, -0.023973820731043816, 0.11024576425552368, -0.05038975924253464, 0.04834126681089401, 0.06649978458881378, -0.012981836684048176, -0.08557141572237015, 0.023789849132299423, -0.068336620926857, -0.03150583803653717, -0.27926525473594666, -0.0930178239941597, -0.09319330751895905, 0.11305391043424606, 0.04079577326774597, 0.06421639025211334, 0.16545771062374115, 0.05191578343510628, -0.024325082078576088, -0.03006586618721485, 0.11609793454408646, 0.12905290722846985, 0.2277202159166336, -0.06067761778831482, 0.10221996158361435, 0.009445492178201675, -0.08203992247581482, 0.06062209978699684, 0.056782789528369904, 0.06324724853038788, 0.02584579586982727, 0.03694582358002663, -0.030939655378460884, 0.1121687963604927, 0.12571842968463898, 0.05258069559931755, 0.0481170229613781, 0.0002127334737451747, -0.0561506561934948, -0.008168719708919525, -0.05726633965969086, 0.06774696707725525, 0.061340972781181335, -0.12918008863925934, -0.08061543852090836, 0.0011613310780376196, 0.06660808622837067, -0.016230419278144836, 0.06823775917291641, -0.13560809195041656, -0.03582429885864258, 0.0790911465883255, -0.07693151384592056, -0.14156894385814667, 0.11972879618406296, -0.026570770889520645, -0.19904157519340515, 0.05265914276242256, 0.007704653777182102, 0.0908159390091896, -0.06360849738121033, 0.05343840271234512, -0.13023801147937775, -0.12935101985931396, -0.018437571823596954, 0.07945099472999573, -0.3450873792171478, 0.13536721467971802, -0.013286802917718887, -0.02876877970993519, -0.06474969536066055, -0.02640824392437935, 0.013905409723520279, 0.12719078361988068, 0.08667250722646713, 0.0008821099763736129, 0.0991629809141159, 0.03823768347501755, 0.04188435152173042, -0.002011700300499797, 0.10950417071580887, 0.0050011589191854, 0.004797275178134441, -0.04982118681073189, 0.007274609990417957, -0.05164213851094246, -0.07472953200340271, 0.08393982797861099, -0.20678792893886566, 0.09087453782558441, -0.03378438204526901, 0.08427679538726807, 0.04304937273263931, -0.018965769559144974, -0.1001204177737236, 0.19745583832263947, -0.012206900864839554, -0.11405988782644272, -0.07517550885677338, -0.02810264565050602, 0.09103139489889145, -0.013817726634442806, 0.012886416167020798, -0.045470476150512695, 0.032183047384023666, -0.1263762265443802, -0.1597503274679184, 0.08734500408172607, -0.04441224783658981, -0.10894393920898438, -0.025462759658694267, 0.20382575690746307, -0.007266622502356768, 0.08242089301347733, 0.01605331338942051, 0.010653935372829437, -0.18066231906414032, -0.04018142446875572, 0.02645772136747837, -0.0016437612939625978, 0.005979063920676708, 0.047698814421892166, 0.019091911613941193, 0.06207629665732384, -0.1069745197892189, -0.013920160941779613, 0.3158324360847473, 0.15978319942951202, -0.00912671908736229, 0.14943915605545044, 0.1093616932630539, -0.08669080585241318, -0.17238758504390717, -0.1171615794301033, -0.1210922971367836, -0.08425768464803696, -0.10681738704442978, -0.1525043100118637, 0.09535340964794159, -0.03392014652490616, 0.03498011827468872, 0.14615866541862488, -0.280263751745224, -0.10949636250734329, 0.13820378482341766, 0.010744688101112843, 0.3510635495185852, -0.12303631007671356, -0.044944874942302704, -0.06214528530836105, -0.16933435201644897, 0.08021392673254013, -0.031203703954815865, 0.11581093072891235, -0.0744495838880539, 0.19395925104618073, 0.01719796098768711, 0.014287159778177738, 0.0916559100151062, 0.05038322135806084, -0.05808406323194504, -0.07368700206279755, -0.10248131304979324, 0.010812131687998772, 0.03546109423041344, 0.010252019390463829, -0.008802837692201138, 0.0211968794465065, -0.11341743916273117, -0.050869911909103394, -0.06302189081907272, 0.0072614275850355625, -0.01001308299601078, -0.042155615985393524, -0.05533592775464058, -0.022557416930794716, -0.020093943923711777, 0.02266426384449005, 0.14185629785060883, -0.07527699321508408, 0.18586260080337524, 0.02357078716158867, 0.1586609035730362, -0.11956068128347397, -0.06724818795919418, -0.029193658381700516, -0.05280323326587677, 0.06468886137008667, -0.08884575963020325, -0.027708567678928375, 0.1332162618637085, -0.01903904788196087, 0.04655366763472557, 0.12936700880527496, 0.02046884410083294, 0.015383756719529629, 0.034968774765729904, -0.2578005790710449, -0.07463036477565765, -0.03505445644259453, -0.012416874058544636, 0.05272092670202255, 0.05525677278637886, 0.19735674560070038, -0.03551921248435974, -0.08521962910890579, 0.020131373777985573, 0.02735883742570877, -0.02776256389915943, 0.10749414563179016, 0.019579345360398293, -0.004837906453758478, -0.16151933372020721, 0.08257976174354553, -0.005964108742773533, -0.08297000825405121, 0.028665626421570778, 0.2024049311876297, -0.12141239643096924, -0.10309756547212601, -0.06804922968149185, 0.07315051555633545, -0.09220825880765915, 0.016043387353420258, -0.005091092549264431, -0.1521538347005844, 0.06916408240795135, 0.07598215341567993, 0.04075418785214424, 0.06513199955224991, -0.11743064224720001, -0.015730571001768112, -0.04170290008187294, -0.002195435343310237, 0.03521120920777321, 0.01863143965601921, -0.057492829859256744, 0.15846455097198486, -0.0676199421286583, 0.08538917452096939, -0.0744810476899147, -0.1058846190571785, -0.1395980566740036, 0.04660497233271599, -0.08038312196731567, -0.07247276604175568, -0.12832807004451752, -0.052204377949237823, -0.0067099276930093765, -0.03388519585132599, 0.006552806124091148, -0.06627799570560455, -0.10922821611166, 0.01822470687329769, -0.00743203004822135, -0.009385870769619942, -0.06096754968166351, 0.026706209406256676, 0.06246216222643852, -0.039788868278265, 0.15730851888656616, 0.22509248554706573, -0.13591648638248444, 0.11564400047063828, -0.09797432273626328, -0.105463907122612, 0.046008042991161346, 0.009427277371287346, 0.03594303876161575, 0.0503489226102829, -0.03594081476330757, 0.0044484552927315235, 0.03905477747321129, 0.08074651658535004, 0.08456914126873016, -0.06776505708694458, 0.020801106467843056, -0.05122765153646469, -0.14904099702835083, -0.016655439510941505, -0.0464773029088974, 0.06876829266548157, -0.006725262850522995, 0.11020535975694656, -0.0515950471162796, 0.07739507406949997, -0.07558431476354599, 0.050614211708307266, 0.021146971732378006, -0.14688286185264587, -0.006612539757043123, -0.07093682140111923, 0.042144812643527985, -0.008834975771605968, 0.20241086184978485, -0.03228091076016426, 0.010342049412429333, 0.033811055123806, 0.06203942745923996, -0.01957780309021473, 0.009357001632452011, 0.2014283686876297, 0.12640917301177979, -0.08496357500553131, -0.02679651789367199, 0.06793134659528732, 0.07248228788375854, 0.07093550264835358, 0.10807815194129944, -0.015352966263890266, 0.028434239327907562, 0.07829629629850388, -0.060215238481760025, 0.07576877623796463, -0.08603982627391815, -0.11668483167886734, 0.05793621391057968, 0.012955795042216778, -0.055695828050374985, 0.20305177569389343, 0.19142870604991913, -0.026278704404830933, 0.018410727381706238, -0.0029499190859496593, -0.10117456316947937, -0.15619947016239166, -0.05423750728368759, -0.07170962542295456, -0.1319410353899002, -0.004549739416688681, -0.16646917164325714, 0.022016216069459915, -0.01132756657898426, 0.09506805986166, -0.06855440139770508, -0.01345991250127554, 0.1364889293909073, -0.1055467277765274, 0.0847758799791336, -0.024517204612493515, 0.07877567410469055, -0.03746940940618515, -0.018209461122751236, -0.10342709720134735, 0.007514837197959423, 0.01131442841142416, 0.06840907037258148, -0.10897937417030334, 0.02432350255548954, -0.12208317965269089, -0.08617185056209564, -0.026142612099647522, 0.09279687702655792, -0.0403008833527565, 0.15116846561431885, 0.02645145356655121, -0.06710928678512573, -0.004313822835683823, 0.2646709978580475, -0.08046227693557739, -0.08319197595119476, -0.030799202620983124, 0.2152107208967209, 0.04053696244955063, 0.06396269053220749, 0.019140036776661873, 0.038027774542570114, -0.07184682041406631, 0.2957373559474945, 0.34401440620422363, -0.1318037211894989, -0.007773484103381634, 0.04225075617432594, 0.04406323283910751, 0.14687567949295044, 0.07998795062303543, 0.11360671371221542, 0.2849363386631012, -0.09197647124528885, 0.016657205298542976, -0.04230864346027374, -0.01424806285649538, -0.06908884644508362, 0.045314885675907135, 0.08216670155525208, -0.09241747111082077, -0.022950593382120132, 0.08125471323728561, -0.29741767048835754, 0.10791494697332382, -0.15600289404392242, -0.14948409795761108, -0.05027429759502411, -0.008771711029112339, 0.014683255925774574, 0.019041186198592186, 0.09663030505180359, 0.025651484727859497, -0.07275258749723434, 0.07816889137029648, 0.024486342445015907, -0.23020237684249878, -0.01345184724777937, 0.1456068754196167, -0.06789913028478622, -0.025938833132386208, -0.021313713863492012, 0.051610056310892105, 0.05763651058077812, 0.09027529507875443, -0.03809558227658272, -0.0746568813920021, -0.007141788024455309, -0.022818787023425102, 0.01914946548640728, 0.0597183033823967, 0.06841408461332321, -0.0920223817229271, 0.1167774423956871, -0.07350476831197739, 0.0650370642542839, 0.037623800337314606, -0.022277191281318665, 0.0018526542698964477, 0.013183658011257648, -0.06512464582920074, 0.05533479526638985, 0.1295643299818039, -0.025459708645939827, -0.002524374984204769, -0.028180841356515884, -0.0767761766910553, -0.024015206843614578, -0.04643676429986954, -0.09101243317127228, -0.18130090832710266, -0.12738600373268127, 0.041754670441150665, -0.03240608796477318, -0.2046082615852356, 0.0060346988029778, -0.1128578633069992, 0.03700976446270943, -0.14154092967510223, 0.10004086047410965, 0.07216610759496689, 0.004716616589576006, 0.006774604320526123, 0.0675399899482727, 0.045677728950977325, 0.14796748757362366, -0.16543124616146088, -0.04919974133372307 ]
null
null
transformers
**Disclaimer**: *This model is still under testing and may change in the future, we will try to keep backwards compatibility. For any questions reach us at [email protected]* # MediaWatch News Topics (Greek) Fine-tuned model for multi-label text-classification (SequenceClassification), based on [roberta-el-news](https://huggingface.co/cvcio/roberta-el-news), using [Hugging Face's](https://huggingface.co/) [Transformers](https://github.com/huggingface/transformers) library. This model is to classify news in real-time on upto 33 topics including: *AFFAIRS*, *AGRICULTURE*, *ARTS_AND_CULTURE*, *BREAKING_NEWS*, *BUSINESS*, *COVID*, *ECONOMY*, *EDUCATION*, *ELECTIONS*, *ENTERTAINMENT*, *ENVIRONMENT*, *FOOD*, *HEALTH*, *INTERNATIONAL*, *LAW_AND_ORDER*, *MILITARY*, *NON_PAPER*, *OPINION*, *POLITICS*, *REFUGEE*, *REGIONAL*, *RELIGION*, *SCIENCE*, *SOCIAL_MEDIA*, *SOCIETY*, *SPORTS*, *TECH*, *TOURISM*, *TRANSPORT*, *TRAVEL*, *WEATHER*, *CRIME*, *JUSTICE*. ## How to use You can use this model directly with a pipeline for text-classification: ```python from transformers import pipeline pipe = pipeline( task="text-classification", model="cvcio/mediawatch-el-topics", tokenizer="cvcio/roberta-el-news" # or cvcio/mediawatch-el-topics ) topics = pipe( "Η βιασύνη αρκετών χωρών να άρουν τους περιορισμούς κατά του κορονοϊού, "+ "αν όχι να κηρύξουν το τέλος της πανδημίας, με το σκεπτικό ότι έφτασε "+ "πλέον η ώρα να συμβιώσουμε με την Covid-19, έχει κάνει μερικούς πιο "+ "επιφυλακτικούς επιστήμονες να προειδοποιούν ότι πρόκειται μάλλον "+ "για «ενδημική αυταπάτη» και ότι είναι πρόωρη τέτοια υπερβολική "+ "χαλάρωση. Καθώς τα κρούσματα της Covid-19, μετά το αιφνιδιαστικό "+ "μαζικό κύμα της παραλλαγής Όμικρον, εμφανίζουν τάση υποχώρησης σε "+ "Ευρώπη και Βόρεια Αμερική, όπου περισσεύει η κόπωση μεταξύ των "+ "πολιτών μετά από δύο χρόνια πανδημίας, ειδικοί και μη αδημονούν να "+ "«ξεμπερδέψουν» με τον κορονοϊό.", padding=True, truncation=True, max_length=512, return_all_scores=True ) print(topics) # outputs [ [ {'label': 'AFFAIRS', 'score': 0.0018806682201102376}, {'label': 'AGRICULTURE', 'score': 0.00014653144171461463}, {'label': 'ARTS_AND_CULTURE', 'score': 0.0012948638759553432}, {'label': 'BREAKING_NEWS', 'score': 0.0001729220530251041}, {'label': 'BUSINESS', 'score': 0.0028276608791202307}, {'label': 'COVID', 'score': 0.4407998025417328}, {'label': 'ECONOMY', 'score': 0.039826102554798126}, {'label': 'EDUCATION', 'score': 0.0019098613411188126}, {'label': 'ELECTIONS', 'score': 0.0003333651984576136}, {'label': 'ENTERTAINMENT', 'score': 0.004249618388712406}, {'label': 'ENVIRONMENT', 'score': 0.0015828514005988836}, {'label': 'FOOD', 'score': 0.0018390495097264647}, {'label': 'HEALTH', 'score': 0.1204477995634079}, {'label': 'INTERNATIONAL', 'score': 0.25892165303230286}, {'label': 'LAW_AND_ORDER', 'score': 0.07646272331476212}, {'label': 'MILITARY', 'score': 0.00033025629818439484}, {'label': 'NON_PAPER', 'score': 0.011991199105978012}, {'label': 'OPINION', 'score': 0.16166265308856964}, {'label': 'POLITICS', 'score': 0.0008890336030162871}, {'label': 'REFUGEE', 'score': 0.0011504743015393615}, {'label': 'REGIONAL', 'score': 0.0008734092116355896}, {'label': 'RELIGION', 'score': 0.0009001944563351572}, {'label': 'SCIENCE', 'score': 0.05075162276625633}, {'label': 'SOCIAL_MEDIA', 'score': 0.00039615994319319725}, {'label': 'SOCIETY', 'score': 0.0043518817983567715}, {'label': 'SPORTS', 'score': 0.002416545059531927}, {'label': 'TECH', 'score': 0.0007818648009561002}, {'label': 'TOURISM', 'score': 0.011870541609823704}, {'label': 'TRANSPORT', 'score': 0.0009422845905646682}, {'label': 'TRAVEL', 'score': 0.03004464879631996}, {'label': 'WEATHER', 'score': 0.00040286066359840333}, {'label': 'CRIME', 'score': 0.0005416403291746974}, {'label': 'JUSTICE', 'score': 0.000990519649349153} ] ] ``` ## Labels All labels, except *NON_PAPER*, retrieved by source articles during the data collection step, without any preprocessing, assuming that journalists and newsrooms assign correct tags to the articles. We disregarded all articles with more than 6 tags to reduce bias and tag manipulation. | label | roc_auc | samples | |-------:|--------:|--------:| | AFFAIRS | 0.9872 | 6,314 | | AGRICULTURE | 0.9799 | 1,254 | | ARTS_AND_CULTURE | 0.9838 | 15,968 | | BREAKING_NEWS | 0.9675 | 827 | | BUSINESS | 0.9811 | 6,507 | | COVID | 0.9620 | 50,000 | | CRIME | 0.9885 | 34,421 | | ECONOMY | 0.9765 | 45,474 | | EDUCATION | 0.9865 | 10,111 | | ELECTIONS | 0.9940 | 7,571 | | ENTERTAINMENT | 0.9925 | 23,323 | | ENVIRONMENT | 0.9847 | 23,060 | | FOOD | 0.9934 | 3,712 | | HEALTH | 0.9723 | 16,852 | | INTERNATIONAL | 0.9624 | 50,000 | | JUSTICE | 0.9862 | 4,860 | | LAW_AND_ORDER | 0.9177 | 50,000 | | MILITARY | 0.9838 | 6,536 | | NON_PAPER | 0.9595 | 4,589 | | OPINION | 0.9624 | 6,296 | | POLITICS | 0.9773 | 50,000 | | REFUGEE | 0.9949 | 4,536 | | REGIONAL | 0.9520 | 50,000 | | RELIGION | 0.9922 | 11,533 | | SCIENCE | 0.9837 | 1,998 | | SOCIAL_MEDIA | 0.991 | 6,212 | | SOCIETY | 0.9439 | 50,000 | | SPORTS | 0.9939 | 31,396 | | TECH | 0.9923 | 8,225 | | TOURISM | 0.9900 | 8,081 | | TRANSPORT | 0.9879 | 3,211 | | TRAVEL | 0.9832 | 4,638 | | WEATHER | 0.9950 | 19,931 | | loss | 0.0533 | - | | roc_auc | 0.9855 | - | ## Pretraining The model was pretrained using an NVIDIA A10 GPU for 15 epochs (~ approx 59K steps, 8 hours training) with a batch size of 128. The optimizer used is Adam with a learning rate of 1e-5, and weight decay 0.01. We used roc_auc_micro to evaluate the results. ### Framework versions - Transformers 4.13.0 - Pytorch 1.9.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3 ## Authors Dimitris Papaevagelou - [@andefined](https://github.com/andefined) ## About Us [Civic Information Office](https://cvcio.org/) is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest.
{"language": "el", "license": "gpl-3.0", "tags": ["roberta", "Greek", "news", "transformers", "text-classification"], "pipeline_tag": "text-classification", "widget": [{"text": "\u03a0\u03b1\u03c1\u2019 \u03bf\u03bb\u03af\u03b3\u03bf\u03bd \u00ab\u03b8\u03b5\u03c1\u03bc\u03cc\u00bb \u03b5\u03c0\u03b5\u03b9\u03c3\u03cc\u03b4\u03b9\u03bf \u03c4\u03bf\u03c5\u03c1\u03ba\u03b9\u03ba\u03bf\u03cd \u03c0\u03bf\u03bb\u03b5\u03bc\u03b9\u03ba\u03bf\u03cd \u03c0\u03bb\u03bf\u03af\u03bf\u03c5 \u03bc\u03b5 \u03b5\u03bb\u03bb\u03b7\u03bd\u03b9\u03ba\u03cc \u03c9\u03ba\u03b5\u03b1\u03bd\u03bf\u03b3\u03c1\u03b1\u03c6\u03b9\u03ba\u03cc \u03c3\u03c4\u03b7\u03bd \u03c0\u03b5\u03c1\u03b9\u03bf\u03c7\u03ae \u03bc\u03b5\u03c4\u03b1\u03be\u03cd \u03a1\u03cc\u03b4\u03bf\u03c5 \u03ba\u03b1\u03b9 \u039a\u03b1\u03c3\u03c4\u03b5\u03bb\u03cc\u03c1\u03b9\u03b6\u03bf\u03c5, \u03c3\u03c4\u03bf \u03b4\u03b9\u03ac\u03c3\u03c4\u03b7\u03bc\u03b1 20-23 \u03a3\u03b5\u03c0\u03c4\u03b5\u03bc\u03b2\u03c1\u03af\u03bf\u03c5, \u03b1\u03c0\u03bf\u03ba\u03ac\u03bb\u03c5\u03c8\u03b5 \u03c4\u03bf \u039f\u03a1\u0395\u039d. \u03a3\u03cd\u03bc\u03c6\u03c9\u03bd\u03b1 \u03bc\u03b5 \u03c0\u03bb\u03b7\u03c1\u03bf\u03c6\u03bf\u03c1\u03af\u03b5\u03c2 \u03c0\u03bf\u03c5 \u03bc\u03b5\u03c4\u03ad\u03b4\u03c9\u03c3\u03b5 \u03c4\u03bf \u03ba\u03b5\u03bd\u03c4\u03c1\u03b9\u03ba\u03cc \u03b4\u03b5\u03bb\u03c4\u03af\u03bf \u03b5\u03b9\u03b4\u03ae\u03c3\u03b5\u03c9\u03bd, \u03cc\u03c4\u03b1\u03bd \u03c4\u03bf \u03b5\u03bb\u03bb\u03b7\u03bd\u03b9\u03ba\u03cc \u03b5\u03c1\u03b5\u03c5\u03bd\u03b7\u03c4\u03b9\u03ba\u03cc \u00ab \u0391\u0399\u0393\u0391\u0399\u039f \u00bb \u03c0\u03bf\u03c5 \u03b1\u03bd\u03ae\u03ba\u03b5\u03b9 \u03c3\u03c4\u03bf \u0395\u03bb\u03bb\u03b7\u03bd\u03b9\u03ba\u03cc \u039a\u03ad\u03bd\u03c4\u03c1\u03bf \u0398\u03b1\u03bb\u03b1\u03c3\u03c3\u03af\u03c9\u03bd \u0395\u03c1\u03b5\u03c5\u03bd\u03ce\u03bd \u03b2\u03b3\u03ae\u03ba\u03b5 \u03ad\u03be\u03c9 \u03b1\u03c0\u03cc \u03c4\u03b1 6 \u03bd.\u03bc, \u03c3\u03b5 \u03b4\u03b9\u03b5\u03b8\u03bd\u03ae \u03cd\u03b4\u03b1\u03c4\u03b1, \u03c4\u03bf \u03c0\u03c1\u03bf\u03c3\u03ad\u03b3\u03b3\u03b9\u03c3\u03b5 \u03c4\u03bf\u03c5\u03c1\u03ba\u03b9\u03ba\u03cc \u03c0\u03bf\u03bb\u03b5\u03bc\u03b9\u03ba\u03cc \u03c0\u03bb\u03bf\u03af\u03bf, \u03bf \u03ba\u03c5\u03b2\u03b5\u03c1\u03bd\u03ae\u03c4\u03b7\u03c2 \u03c4\u03bf\u03c5 \u03bf\u03c0\u03bf\u03af\u03bf\u03c5 \u03b6\u03ae\u03c4\u03b7\u03c3\u03b5 \u03b4\u03cd\u03bf \u03c6\u03bf\u03c1\u03ad\u03c2 \u03bc\u03ad\u03c3\u03c9 \u03b1\u03c3\u03c5\u03c1\u03bc\u03ac\u03c4\u03bf\u03c5 \u03bd\u03b1 \u03b5\u03bd\u03b7\u03bc\u03b5\u03c1\u03c9\u03b8\u03b5\u03af \u03b3\u03b9\u03b1 \u03c4\u03b1 \u03c3\u03c4\u03bf\u03b9\u03c7\u03b5\u03af\u03b1 \u03c4\u03bf\u03c5 \u03c0\u03bb\u03bf\u03af\u03bf\u03c5, \u03b1\u03bb\u03bb\u03ac \u03ba\u03b1\u03b9 \u03b3\u03b9\u03b1 \u03c4\u03b7\u03bd \u03b1\u03c0\u03bf\u03c3\u03c4\u03bf\u03bb\u03ae \u03c4\u03bf\u03c5. \u039f \u03c0\u03bb\u03bf\u03af\u03b1\u03c1\u03c7\u03bf\u03c2 \u03c4\u03bf\u03c5 \u03b5\u03bb\u03bb\u03b7\u03bd\u03b9\u03ba\u03bf\u03cd \u03b5\u03c1\u03b5\u03c5\u03bd\u03b7\u03c4\u03b9\u03ba\u03bf\u03cd \u03b4\u03b5\u03bd \u03b1\u03c0\u03ac\u03bd\u03c4\u03b7\u03c3\u03b5 \u03ba\u03b1\u03b9 \u03c4\u03b5\u03bb\u03b9\u03ba\u03ac \u03c4\u03bf \u03c4\u03bf\u03c5\u03c1\u03ba\u03b9\u03ba\u03cc \u03c0\u03bf\u03bb\u03b5\u03bc\u03b9\u03ba\u03cc \u03b1\u03c0\u03bf\u03bc\u03b1\u03ba\u03c1\u03cd\u03bd\u03b8\u03b7\u03ba\u03b5.", "example_title": "Topic AFFAIRS"}, {"text": "\u0397 \u03ba\u03c5\u03b2\u03b5\u03c1\u03bd\u03b7\u03c4\u03b9\u03ba\u03ae \u03b1\u03bd\u03b9\u03ba\u03b1\u03bd\u03cc\u03c4\u03b7\u03c4\u03b1 \u03bf\u03b4\u03b7\u03b3\u03b5\u03af \u03c4\u03b7\u03bd \u03c7\u03ce\u03c1\u03b1 \u03c3\u03c4\u03bf \u03c7\u03ac\u03bf\u03c2. \u0397 \u03ba\u03c5\u03b2\u03b5\u03c1\u03bd\u03b7\u03c3\u03b7 \u039c\u03b7\u03c4\u03c3\u03bf\u03c4\u03b1\u03ba\u03b7 \u03b1\u03b4\u03c5\u03bd\u03b1\u03c4\u03b5\u03af \u03bd\u03b1 \u03b4\u03b9\u03b1\u03c7\u03b5\u03b9\u03c1\u03b9\u03c3\u03c4\u03b5\u03af \u03c4\u03b7\u03bd \u03c0\u03b1\u03bd\u03b4\u03b7\u03bc\u03af\u03b1. \u0394\u03b5\u03bd \u03bc\u03c0\u03bf\u03c1\u03b5\u03b9 \u03bf\u03cd\u03c4\u03b5 \u03bd\u03b1 \u03c0\u03b5\u03af\u03c3\u03b5\u03b9 \u03c4\u03bf\u03bd \u03ba\u03cc\u03c3\u03bc\u03bf \u03bd\u03b1 \u03b5\u03bc\u03b2\u03bf\u03bb\u03b9\u03b1\u03c3\u03c4\u03b5\u03af, \u03c0\u03bf\u03c5 \u03ae\u03c4\u03b1\u03bd \u03c4\u03bf \u03c0\u03b9\u03bf \u03b1\u03c0\u03bb\u03bf \u03c0\u03c1\u03ac\u03b3\u03bc\u03b1. \u03a3\u03b7\u03bc\u03b5\u03c1\u03b1 \u03bb\u03bf\u03b9\u03c0\u03cc\u03bd \u03c6\u03c4\u03ac\u03c3\u03b1\u03bc\u03b5 \u03c3\u03c4\u03bf \u03c3\u03b7\u03bc\u03b5\u03af\u03bf \u03bd\u03b1 \u03bc\u03b9\u03bb\u03ac\u03bc\u03b5 \u03b3\u03b9\u03b1 \u03b5\u03c0\u03b1\u03bd\u03b1\u03c6\u03bf\u03c1\u03ac \u03c4\u03b7\u03c2 \u03c7\u03c1\u03ae\u03c3\u03b7\u03c2 \u03bc\u03ac\u03c3\u03ba\u03b1\u03c2 \u03c3\u03b5 \u03b5\u03be\u03c9\u03c4\u03b5\u03c1\u03b9\u03ba\u03bf\u03cd\u03c2 \u03c7\u03ce\u03c1\u03bf\u03c5\u03c2 \u03b1\u03ba\u03cc\u03bc\u03b7 \u03ba\u03b1\u03b9 \u03cc\u03c0\u03bf\u03c5 \u03b4\u03b5\u03bd \u03c5\u03c0\u03ac\u03c1\u03c7\u03b5\u03b9 \u03c3\u03c5\u03b3\u03c7\u03c1\u03c9\u03c4\u03b9\u03c3\u03bc\u03cc\u03c2. \u03a3\u03c4\u03b9\u03c2 \u03c3\u03c5\u03b6\u03b7\u03c4\u03ae\u03c3\u03b5\u03b9\u03c2 \u03c4\u03c9\u03bd \u03b5\u03b9\u03b4\u03b9\u03ba\u03ce\u03bd \u03b8\u03b1 \u03b2\u03c1\u03b5\u03b8\u03b5\u03af \u03b5\u03c0\u03af\u03c3\u03b7\u03c2 \u03c4\u03bf \u03b5\u03bd\u03b4\u03b5\u03c7\u03cc\u03bc\u03b5\u03bd\u03bf \u03b3\u03b9\u03b1 \u03c4\u03bf\u03c0\u03b9\u03ba\u03ac lockdown \u03c3\u03b5 \u03c0\u03b5\u03c1\u03b9\u03bf\u03c7\u03ad\u03c2 \u03bc\u03b5 \u03b2\u03b1\u03c1\u03cd \u03b9\u03b9\u03ba\u03cc \u03c6\u03bf\u03c1\u03c4\u03af\u03bf \u03b3\u03b9\u03b1 \u03bd\u03b1 \u03bc\u03b7\u03bd \u03be\u03b5\u03c6\u03cd\u03b3\u03b5\u03b9 \u03b7 \u03ba\u03b1\u03c4\u03ac\u03c3\u03c4\u03b1\u03c3\u03b7, \u03b5\u03bd\u03ce \u03b8\u03b1 \u03c7\u03c1\u03b5\u03b9\u03ac\u03b6\u03b5\u03c4\u03b1\u03b9 \u03ba\u03ac\u03c0\u03bf\u03b9\u03bf\u03c2 \u03b3\u03b9\u03b1 \u03c4\u03b9\u03c2 \u03bc\u03b5\u03c4\u03b1\u03ba\u03b9\u03bd\u03ae\u03c3\u03b5\u03b9\u03c2 \u03c4\u03bf\u03c5 \u03b5\u03af\u03c4\u03b5 \u03c0\u03b9\u03c3\u03c4\u03bf\u03c0\u03bf\u03b9\u03b7\u03c4\u03b9\u03ba\u03cc \u03b5\u03bc\u03b2\u03bf\u03bb\u03b9\u03b1\u03c3\u03bc\u03bf\u03cd \u03ae \u03bd\u03cc\u03c3\u03b7\u03c3\u03b7\u03c2 \u03ba\u03b1\u03b9 \u03bf\u03b9 \u03b1\u03bd\u03b5\u03bc\u03b2\u03bf\u03bb\u03af\u03b1\u03c3\u03c4\u03bf\u03b9 rapid \u03ae \u03bc\u03bf\u03c1\u03b9\u03b1\u03ba\u03cc \u03c4\u03b5\u03c3\u03c4.", "example_title": "Topic COVID"}, {"text": "\u0397 \u00ab\u03c9\u03c1\u03b1\u03af\u03b1 \u0395\u03bb\u03ad\u03bd\u03b7\u00bb \u03b5\u03c0\u03ad\u03c3\u03c4\u03c1\u03b5\u03c8\u03b5 \u03c3\u03c4\u03b7\u03bd \u03c4\u03b7\u03bb\u03b5\u03cc\u03c1\u03b1\u03c3\u03b7, \u03bc\u03ad\u03c3\u03b1 \u03b1\u03c0\u03cc \u03c4\u03b7 \u03c3\u03c5\u03c7\u03bd\u03cc\u03c4\u03b7\u03c4\u03b1 \u03c4\u03bf\u03c5 MEGA \u03ba\u03b1\u03b9 \u03ac\u03c6\u03b7\u03c3\u03b5 \u03c4\u03b9\u03c2 \u03ba\u03b1\u03bb\u03cd\u03c4\u03b5\u03c1\u03b5\u03c2 \u03b5\u03bd\u03c4\u03c5\u03c0\u03ce\u03c3\u03b5\u03b9\u03c2. \u03a4\u03bf \u03c0\u03bb\u03b1\u03c4\u03cc \u03b1\u03c0\u03cc \u03c4\u03bf \u03bf\u03c0\u03bf\u03af\u03bf \u03b5\u03bc\u03c6\u03b1\u03bd\u03af\u03b6\u03b5\u03c4\u03b1\u03b9 \u03b7 \u0395\u03bb\u03ad\u03bd\u03b7 \u039c\u03b5\u03bd\u03b5\u03b3\u03ac\u03ba\u03b7 \u03ad\u03c7\u03b5\u03b9 \u03c6\u03c4\u03b9\u03b1\u03c7\u03c4\u03b5\u03af \u03b1\u03c0\u03cc \u03c4\u03b7\u03bd \u03b1\u03c1\u03c7\u03ae \u03b3\u03b9\u03b1 \u03c4\u03b7\u03bd \u03b5\u03ba\u03c0\u03bf\u03bc\u03c0\u03ae \u03c4\u03b7\u03c2. \u03a3\u03ae\u03bc\u03b5\u03c1\u03b1, \u03c3\u03c4\u03bf \u03ba\u03bb\u03b5\u03af\u03c3\u03b9\u03bc\u03bf \u03c4\u03b7\u03c2 \u03b5\u03ba\u03c0\u03bf\u03bc\u03c0\u03ae\u03c2 \u03b7 \u0395\u03bb\u03ad\u03bd\u03b7 \u03c0\u03ad\u03c1\u03b1\u03c3\u03b5 \u03b1\u03bd\u03ac\u03bc\u03b5\u03c3\u03b1 \u03b1\u03c0\u03cc \u03c4\u03b9\u03c2 \u03ba\u03ac\u03bc\u03b5\u03c1\u03b5\u03c2 \u03b3\u03b9\u03b1 \u03bd\u03b1 \u03bc\u03c0\u03b5\u03b9 \u03c3\u03c4\u03bf \u03ba\u03b1\u03bc\u03b1\u03c1\u03af\u03bd\u03b9 \u03c4\u03b7\u03c2 \u00ab\u039c\u03b7\u03bd \u03c4\u03c1\u03bf\u03bc\u03bf\u03ba\u03c1\u03b1\u03c4\u03b5\u03af\u03c3\u03c4\u03b5, \u03b5\u03af\u03bc\u03b1\u03b9 \u03b7 \u0395\u03bb\u03ad\u03bd\u03b7 \u039c\u03b5\u03bd\u03b5\u03b3\u03ac\u03ba\u03b7, \u03c4\u03b1 \u03ba\u03ac\u03bd\u03c9 \u03b1\u03c5\u03c4\u03ac. \u039c\u03b5 \u03c3\u03c5\u03b3\u03c7\u03c9\u03c1\u03b5\u03af\u03c4\u03b1\u03b9, \u03ad\u03c7\u03c9 \u03c8\u03c5\u03c7\u03bf\u03bb\u03bf\u03b3\u03b9\u03ba\u03ac \u03b1\u03bd \u03b4\u03b5\u03bd \u03b5\u03af\u03bc\u03b1\u03b9 \u03b5\u03bb\u03b5\u03cd\u03b8\u03b5\u03c1\u03b7\u00bb \u03b5\u03af\u03c0\u03b5 \u03b1\u03c1\u03c7\u03b9\u03ba\u03ac \u03b7 \u03c0\u03b1\u03c1\u03bf\u03c5\u03c3\u03b9\u03ac\u03c3\u03c4\u03c1\u03b9\u03b1 \u03c3\u03c4\u03bf\u03c5\u03c2 \u03c3\u03c5\u03bd\u03b5\u03c1\u03b3\u03ac\u03c4\u03b5\u03c2 \u03c4\u03b7\u03c2 \u03ba\u03b1\u03b9 \u03c0\u03c1\u03cc\u03c3\u03b8\u03b5\u03c3\u03b5 \u03c3\u03c4\u03b7 \u03c3\u03c5\u03bd\u03ad\u03c7\u03b5\u03b9\u03b1: \u00ab\u0397 \u0395\u03bb\u03ad\u03bd\u03b7 \u03bf\u03bb\u03bf\u03ba\u03bb\u03ae\u03c1\u03c9\u03c3\u03b5. \u039c\u03c0\u03bf\u03c1\u03b5\u03af\u03c4\u03b5 \u03bd\u03b1 \u03c3\u03c5\u03bd\u03b5\u03c7\u03af\u03c3\u03b5\u03c4\u03b5 \u03bc\u03b5 \u03c4\u03bf \u03c5\u03c0\u03cc\u03bb\u03bf\u03b9\u03c0\u03bf \u03c0\u03c1\u03cc\u03b3\u03c1\u03b1\u03bc\u03bc\u03b1 \u03c4\u03bf\u03c5 Mega. \u0395\u03b3\u03ce \u03b1\u03bd\u03bf\u03af\u03b3\u03c9 \u03c4\u03bf \u03ba\u03b1\u03bc\u03b1\u03c1\u03af\u03bd\u03b9, \u03b1\u03bd \u03bc\u03b5 \u03b1\u03c6\u03ae\u03c3\u03bf\u03c5\u03bd. \u039c\u03c0\u03b1\u03af\u03bd\u03c9 \u03ba\u03b1\u03bc\u03b1\u03c1\u03af\u03bd\u03b9\u00bb. \u0394\u03b5\u03af\u03c4\u03b5 \u03c4\u03bf \u03b1\u03c0\u03cc\u03c3\u03c0\u03b1\u03c3\u03bc\u03b1!", "example_title": "Topic ENTERTAINMENT"}, {"text": "\u0388\u03bd\u03b1 \u03b5\u03be\u03b1\u03b9\u03c1\u03b5\u03c4\u03b9\u03ba\u03ac \u03b5\u03bd\u03b4\u03b9\u03b1\u03c6\u03ad\u03c1\u03bf\u03bd \u00ab\u03ba\u03bf\u03c5\u03c4\u03c3\u03bf\u03bc\u03c0\u03bf\u03bb\u03b9\u03cc\u00bb \u03b5\u03bd\u03c4\u03cc\u03c0\u03b9\u03c3\u03b1\u03bd \u03bf\u03b9 \u03ba\u03b5\u03c1\u03b1\u03af\u03b5\u03c2 \u03c4\u03b7\u03c2 \u03c3\u03c4\u03ae\u03bb\u03b7\u03c2 \u03c0\u03ad\u03c1\u03b9\u03be \u03c4\u03bf\u03c5 \u039c\u03b5\u03b3\u03ac\u03c1\u03bf\u03c5 \u039c\u03b1\u03be\u03af\u03bc\u03bf\u03c5 : \u03c4\u03bf \u03ba\u03b1\u03c4\u03ac \u03c0\u03cc\u03c3\u03bf\u03bd, \u03b4\u03b7\u03bb\u03b1\u03b4\u03ae, \u03bf \u00ab\u03b5\u03be \u03b1\u03c0\u03bf\u03c1\u03c1\u03ae\u03c4\u03c9\u03bd\u00bb \u03c4\u03bf\u03c5 \u039a\u03c5\u03c1\u03b9\u03ac\u03ba\u03bf\u03c5 \u039c\u03b7\u03c4\u03c3\u03bf\u03c4\u03ac\u03ba\u03b7 , \u0393\u03b9\u03ce\u03c1\u03b3\u03bf\u03c2 \u0393\u03b5\u03c1\u03b1\u03c0\u03b5\u03c4\u03c1\u03af\u03c4\u03b7\u03c2 \u03bc\u03b5\u03c4\u03ad\u03c7\u03b5\u03b9 \u03c3\u03c4\u03b7 \u03b4\u03b9\u03b1\u03c7\u03b5\u03af\u03c1\u03b9\u03c3\u03b7 \u03c4\u03b7\u03c2 \u03c0\u03b1\u03bd\u03b4\u03b7\u03bc\u03af\u03b1\u03c2 \u03ba\u03b1\u03b9 \u03c3\u03c4\u03b7\u03bd \u03b4\u03b9\u03b1\u03b4\u03b9\u03ba\u03b1\u03c3\u03af\u03b1 \u03bb\u03ae\u03c8\u03b7\u03c2 \u03b1\u03c0\u03bf\u03c6\u03ac\u03c3\u03b5\u03c9\u03bd. \u03a4\u03bf \u03b5\u03bd \u03bb\u03cc\u03b3\u03c9 \u00ab\u03ba\u03bf\u03c5\u03c4\u03c3\u03bf\u03bc\u03c0\u03bf\u03bb\u03b9\u03cc\u00bb \u03c0\u03c5\u03c1\u03bf\u03b4\u03cc\u03c4\u03b7\u03c3\u03b5 \u03c4\u03bf \u03b3\u03b5\u03b3\u03bf\u03bd\u03cc\u03c2 \u03cc\u03c4\u03b9 \u03c3\u03b5 \u03c3\u03b1\u03b2\u03b2\u03b1\u03c4\u03b9\u03ac\u03c4\u03b9\u03ba\u03b7 \u03b5\u03c6\u03b7\u03bc\u03b5\u03c1\u03af\u03b4\u03b1 \u03b4\u03b7\u03bc\u03bf\u03c3\u03b9\u03b5\u03cd\u03b8\u03b7\u03ba\u03b1\u03bd \u03c0\u03c1\u03bf\u03c7\u03b8\u03ad\u03c2 \u03b4\u03b7\u03bb\u03ce\u03c3\u03b5\u03b9\u03c2 \u03c4\u03bf\u03c5 \u03c5\u03c0\u03bf\u03c5\u03c1\u03b3\u03bf\u03cd \u0395\u03c0\u03b9\u03ba\u03c1\u03b1\u03c4\u03b5\u03af\u03b1\u03c2 \u03bc\u03b5 \u03c4\u03b9\u03c2 \u03bf\u03c0\u03bf\u03af\u03b5\u03c2 \u03b1\u03c0\u03ad\u03ba\u03bb\u03b5\u03b9\u03b5 \u03ba\u03ac\u03b8\u03b5 \u03c3\u03b5\u03bd\u03ac\u03c1\u03b9\u03bf \u03bd\u03ad\u03c9\u03bd \u03bf\u03c1\u03b9\u03b6\u03cc\u03bd\u03c4\u03b9\u03c9\u03bd \u03bc\u03ad\u03c4\u03c1\u03c9\u03bd \u03ba\u03b1\u03b9 \u03c4\u03b7\u03bd \u03af\u03b4\u03b9\u03b1 \u03ce\u03c1\u03b1, \u03c4\u03bf \u039c\u03b1\u03be\u03af\u03bc\u03bf\u03c5 \u03b1\u03bd\u03ae\u03b3\u03b3\u03b5\u03bb\u03bb\u03b5\u2026 \u03ba\u03b1\u03c1\u03b1\u03bd\u03c4\u03af\u03bd\u03b1 \u03c3\u03c4\u03b7 \u039c\u03cd\u03ba\u03bf\u03bd\u03bf. \u00ab\u0395\u03af\u03bd\u03b1\u03b9 \u03b1\u03c5\u03c4\u03bf\u03bd\u03cc\u03b7\u03c4\u03bf \u03cc\u03c4\u03b9 \u03b7 \u03ba\u03bf\u03b9\u03bd\u03c9\u03bd\u03af\u03b1 \u03ba\u03b1\u03b9 \u03b7 \u03bf\u03b9\u03ba\u03bf\u03bd\u03bf\u03bc\u03af\u03b1 \u03b4\u03b5\u03bd \u03b1\u03bd\u03c4\u03ad\u03c7\u03bf\u03c5\u03bd \u03bf\u03c1\u03b9\u03b6\u03cc\u03bd\u03c4\u03b9\u03bf\u03c5\u03c2 \u03c0\u03b5\u03c1\u03b9\u03bf\u03c1\u03b9\u03c3\u03bc\u03bf\u03cd\u03c2\u00bb, \u03ad\u03bb\u03b5\u03b3\u03b5 \u03c7\u03b1\u03c1\u03b1\u03ba\u03c4\u03b7\u03c1\u03b9\u03c3\u03c4\u03b9\u03ba\u03ac \u03bf \u0393\u03b5\u03c1\u03b1\u03c0\u03b5\u03c4\u03c1\u03af\u03c4\u03b7\u03c2, \u03c4\u03b7\u03bd \u03ce\u03c1\u03b1 \u03c0\u03bf\u03c5 \u03b7 \u03ba\u03c5\u03b2\u03ad\u03c1\u03bd\u03b7\u03c3\u03b7 \u03b1\u03bd\u03b1\u03ba\u03bf\u03af\u03bd\u03c9\u03bd\u03b5\u2026 \u03b1\u03c5\u03c4\u03bf\u03cd\u03c2 \u03c4\u03bf\u03c5\u03c2 \u03bf\u03c1\u03b9\u03b6\u03cc\u03bd\u03c4\u03b9\u03bf\u03c5\u03c2 \u03c0\u03b5\u03c1\u03b9\u03bf\u03c1\u03b9\u03c3\u03bc\u03bf\u03cd\u03c2. \u03a9\u03c2 \u03b5\u03ba \u03c4\u03bf\u03cd\u03c4\u03c9\u03bd, \u03b4\u03cd\u03bf \u03c4\u03b9\u03bd\u03ac \u03bc\u03c0\u03bf\u03c1\u03b5\u03af \u03bd\u03b1 \u03c3\u03c5\u03bc\u03b2\u03b1\u03af\u03bd\u03bf\u03c5\u03bd: \u03b5\u03af\u03c4\u03b5 \u03bf \u03c5\u03c0\u03bf\u03c5\u03c1\u03b3\u03cc\u03c2 \u0395\u03c0\u03b9\u03ba\u03c1\u03b1\u03c4\u03b5\u03af\u03b1\u03c2 \u03b4\u03b5\u03bd \u03bc\u03b5\u03c4\u03ad\u03c7\u03b5\u03b9 \u03c0\u03bb\u03ad\u03bf\u03bd \u03c3\u03c4\u03b7 \u03bb\u03ae\u03c8\u03b7 \u03c4\u03c9\u03bd \u03b1\u03c0\u03bf\u03c6\u03ac\u03c3\u03b5\u03c9\u03bd, \u03b5\u03af\u03c4\u03b5 \u03b7 \u03b1\u03c0\u03cc\u03c6\u03b1\u03c3\u03b7 \u03b3\u03b9\u03b1 \u03bf\u03c1\u03b9\u03b6\u03cc\u03bd\u03c4\u03b9\u03b1 \u03bc\u03ad\u03c4\u03c1\u03b1 \u03b5\u03bb\u03ae\u03c6\u03b8\u03b7 \u03c5\u03c0\u03cc \u03c4\u03bf \u03ba\u03c1\u03ac\u03c4\u03bf\u03c2 \u03c0\u03b1\u03bd\u03b9\u03ba\u03bf\u03cd \u03c4\u03bf \u03c0\u03c1\u03c9\u03af \u03c4\u03bf\u03c5 \u03a3\u03b1\u03b2\u03b2\u03ac\u03c4\u03bf\u03c5, \u03cc\u03c4\u03b1\u03bd \u03ad\u03c6\u03c4\u03b1\u03c3\u03b5 \u03c3\u03c4\u03bf \u039c\u03b1\u03be\u03af\u03bc\u03bf\u03c5 \u03b7 \u03c4\u03b5\u03bb\u03b5\u03c5\u03c4\u03b1\u03af\u03b1 \u00ab\u03c6\u03bf\u03c5\u03c1\u03bd\u03b9\u03ac\u00bb \u03c4\u03c9\u03bd \u03b5\u03c0\u03b9\u03b4\u03b7\u03bc\u03b9\u03bf\u03bb\u03bf\u03b3\u03b9\u03ba\u03ce\u03bd \u03b4\u03b5\u03b4\u03bf\u03bc\u03ad\u03bd\u03c9\u03bd \u03b3\u03b9\u03b1 \u03c4\u03bf \u03bd\u03b7\u03c3\u03af \u03c4\u03c9\u03bd \u03b1\u03bd\u03ad\u03bc\u03c9\u03bd\u2026", "example_title": "Topic NON_PAPER"}, {"text": "\u0395\u03af\u03bd\u03b1\u03b9 \u03be\u03b5\u03ba\u03ac\u03b8\u03b1\u03c1\u03bf \u03cc\u03c4\u03b9 \u03bc\u03b5\u03c4\u03ac \u03c4\u03bf \u03c0\u03bb\u03ae\u03b3\u03bc\u03b1 \u03c0\u03bf\u03c5 \u03b4\u03ad\u03c7\u03b8\u03b7\u03ba\u03b5 \u03b7 \u03ba\u03c5\u03b2\u03ad\u03c1\u03bd\u03b7\u03c3\u03ae \u03c4\u03bf\u03c5 \u03b1\u03c0\u03cc \u03c4\u03b9\u03c2 \u03b1\u03b4\u03c5\u03bd\u03b1\u03bc\u03af\u03b5\u03c2 \u03c3\u03c4\u03b7\u03bd \u03b1\u03bd\u03c4\u03b9\u03bc\u03b5\u03c4\u03ce\u03c0\u03b9\u03c3\u03b7 \u03c4\u03c9\u03bd \u03ba\u03b1\u03c4\u03b1\u03c3\u03c4\u03c1\u03bf\u03c6\u03b9\u03ba\u03ce\u03bd \u03c0\u03c5\u03c1\u03ba\u03b1\u03b3\u03b9\u03ce\u03bd \u03c4\u03bf \u03bc\u03b5\u03b3\u03ac\u03bb\u03bf \u03c3\u03c4\u03bf\u03af\u03c7\u03b7\u03bc\u03b1 \u03b3\u03b9\u03b1 \u03c4\u03bf\u03bd \u039a\u03c5\u03c1\u03b9\u03ac\u03ba\u03bf \u039c\u03b7\u03c4\u03c3\u03bf\u03c4\u03ac\u03ba\u03b7 \u03b5\u03af\u03bd\u03b1\u03b9 \u03bd\u03b1 \u03c0\u03c1\u03bf\u03c7\u03c9\u03c1\u03ae\u03c3\u03b5\u03b9 \u03c3\u03c5\u03bd\u03c4\u03b5\u03c4\u03b1\u03b3\u03bc\u03ad\u03bd\u03b1 \u03ba\u03b1\u03b9 \u03c7\u03c9\u03c1\u03af\u03c2 \u03c0\u03b1\u03c1\u03b1\u03c4\u03c1\u03ac\u03b3\u03bf\u03c5\u03b4\u03b1 \u03bf \u03c3\u03c7\u03b5\u03b4\u03b9\u03b1\u03c3\u03bc\u03cc\u03c2 \u03b3\u03b9\u03b1 \u03c4\u03b7\u03bd \u03b1\u03c0\u03bf\u03ba\u03b1\u03c4\u03ac\u03c3\u03c4\u03b1\u03c3\u03b7 \u03c4\u03c9\u03bd \u03b6\u03b7\u03bc\u03b9\u03ce\u03bd. \u039f \u03a0\u03c1\u03c9\u03b8\u03c5\u03c0\u03bf\u03c5\u03c1\u03b3\u03cc\u03c2 \u03ad\u03c7\u03b5\u03b9 \u03ae\u03b4\u03b7 \u03c6\u03c4\u03b9\u03ac\u03be\u03b5\u03b9 \u03bc\u03b9\u03b1 \u03bf\u03bc\u03ac\u03b4\u03b1 \u03ba\u03c1\u03bf\u03cd\u03c3\u03b7\u03c2 \u03c4\u03b7\u03bd \u03bf\u03c0\u03bf\u03af\u03b1 \u03b1\u03c0\u03bf\u03c4\u03b5\u03bb\u03bf\u03cd\u03bd 9 \u03c5\u03c0\u03bf\u03c5\u03c1\u03b3\u03bf\u03af. \u03a4\u03b1 \u03bc\u03ad\u03bb\u03b7 \u03c0\u03bf\u03c5 \u03b1\u03c0\u03b1\u03c1\u03c4\u03af\u03b6\u03bf\u03c5\u03bd \u03c4\u03b7\u03bd \u03bf\u03bc\u03ac\u03b4\u03b1 \u03ba\u03c1\u03bf\u03cd\u03c3\u03b7\u03c2 \u03ba\u03b1\u03b9 \u03c4\u03b1 \u03bf\u03c0\u03bf\u03af\u03b1 \u03b2\u03c1\u03af\u03c3\u03ba\u03bf\u03bd\u03c4\u03b1\u03b9 \u03c3\u03b5 \u03c3\u03c5\u03bd\u03b5\u03c7\u03ae, \u03ba\u03b1\u03b8\u03b7\u03bc\u03b5\u03c1\u03b9\u03bd\u03ae \u03b5\u03c0\u03b1\u03c6\u03ae \u03bc\u03b5 \u03c4\u03bf\u03bd \u039a\u03c5\u03c1\u03b9\u03ac\u03ba\u03bf \u039c\u03b7\u03c4\u03c3\u03bf\u03c4\u03ac\u03ba\u03b7 \u03b5\u03af\u03bd\u03b1\u03b9, \u03cc\u03c0\u03c9\u03c2 \u03bc\u03b1\u03c2 \u03c0\u03bb\u03b7\u03c1\u03bf\u03c6\u03bf\u03c1\u03b5\u03af \u03b7 \u03c3\u03c4\u03ae\u03bb\u03b7 \u00ab\u0398\u03b5\u03c9\u03c1\u03b5\u03af\u03bf\u00bb \u03c4\u03b7\u03c2 \u00ab\u039a\u03b1\u03b8\u03b7\u03bc\u03b5\u03c1\u03b9\u03bd\u03ae\u03c2\u00bb \u03b5\u03af\u03bd\u03b1\u03b9 \u03bf\u03b9: \u0393. \u0393\u03b5\u03c1\u03b1\u03c0\u03b5\u03c4\u03c1\u03af\u03c4\u03b7\u03c2, \u0391. \u03a3\u03ba\u03ad\u03c1\u03c4\u03c3\u03bf\u03c2, \u03a7\u03c1. \u03a4\u03c1\u03b9\u03b1\u03bd\u03c4\u03cc\u03c0\u03bf\u03c5\u03bb\u03bf\u03c2, \u039a. \u039a\u03b1\u03c1\u03b1\u03bc\u03b1\u03bd\u03bb\u03ae\u03c2, \u039a. \u03a3\u03ba\u03c1\u03ad\u03ba\u03b1\u03c2, \u03a3\u03c4. \u03a0\u03ad\u03c4\u03c3\u03b1\u03c2, \u03a3\u03c0. \u039b\u03b9\u03b2\u03b1\u03bd\u03cc\u03c2 \u03ba\u03b1\u03b9 \u03c6\u03c5\u03c3\u03b9\u03ba\u03ac \u03bf\u03b9 \u03a7\u03c1. \u03a3\u03c4\u03b1\u03b9\u03ba\u03bf\u03cd\u03c1\u03b1\u03c2 \u03ba\u03b1\u03b9 \u0398. \u03a3\u03ba\u03c5\u03bb\u03b1\u03ba\u03ac\u03ba\u03b7\u03c2.", "example_title": "Topic OPINION"}]}
text-classification
cvcio/mediawatch-el-topics
[ "transformers", "pytorch", "safetensors", "roberta", "text-classification", "Greek", "news", "el", "doi:10.57967/hf/0711", "license:gpl-3.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "el" ]
TAGS #transformers #pytorch #safetensors #roberta #text-classification #Greek #news #el #doi-10.57967/hf/0711 #license-gpl-3.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
Disclaimer: *This model is still under testing and may change in the future, we will try to keep backwards compatibility. For any questions reach us at info@URL* MediaWatch News Topics (Greek) ============================== Fine-tuned model for multi-label text-classification (SequenceClassification), based on roberta-el-news, using Hugging Face's Transformers library. This model is to classify news in real-time on upto 33 topics including: *AFFAIRS*, *AGRICULTURE*, *ARTS\_AND\_CULTURE*, *BREAKING\_NEWS*, *BUSINESS*, *COVID*, *ECONOMY*, *EDUCATION*, *ELECTIONS*, *ENTERTAINMENT*, *ENVIRONMENT*, *FOOD*, *HEALTH*, *INTERNATIONAL*, *LAW\_AND\_ORDER*, *MILITARY*, *NON\_PAPER*, *OPINION*, *POLITICS*, *REFUGEE*, *REGIONAL*, *RELIGION*, *SCIENCE*, *SOCIAL\_MEDIA*, *SOCIETY*, *SPORTS*, *TECH*, *TOURISM*, *TRANSPORT*, *TRAVEL*, *WEATHER*, *CRIME*, *JUSTICE*. How to use ---------- You can use this model directly with a pipeline for text-classification: Labels ------ All labels, except *NON\_PAPER*, retrieved by source articles during the data collection step, without any preprocessing, assuming that journalists and newsrooms assign correct tags to the articles. We disregarded all articles with more than 6 tags to reduce bias and tag manipulation. Pretraining ----------- The model was pretrained using an NVIDIA A10 GPU for 15 epochs (~ approx 59K steps, 8 hours training) with a batch size of 128. The optimizer used is Adam with a learning rate of 1e-5, and weight decay 0.01. We used roc\_auc\_micro to evaluate the results. ### Framework versions * Transformers 4.13.0 * Pytorch 1.9.0+cu111 * Datasets 1.16.1 * Tokenizers 0.10.3 Authors ------- Dimitris Papaevagelou - @andefined About Us -------- Civic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest.
[ "### Framework versions\n\n\n* Transformers 4.13.0\n* Pytorch 1.9.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3\n\n\nAuthors\n-------\n\n\nDimitris Papaevagelou - @andefined\n\n\nAbout Us\n--------\n\n\nCivic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest." ]
[ "TAGS\n#transformers #pytorch #safetensors #roberta #text-classification #Greek #news #el #doi-10.57967/hf/0711 #license-gpl-3.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Framework versions\n\n\n* Transformers 4.13.0\n* Pytorch 1.9.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3\n\n\nAuthors\n-------\n\n\nDimitris Papaevagelou - @andefined\n\n\nAbout Us\n--------\n\n\nCivic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest." ]
[ 73, 80 ]
[ "passage: TAGS\n#transformers #pytorch #safetensors #roberta #text-classification #Greek #news #el #doi-10.57967/hf/0711 #license-gpl-3.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Framework versions\n\n\n* Transformers 4.13.0\n* Pytorch 1.9.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3\n\n\nAuthors\n-------\n\n\nDimitris Papaevagelou - @andefined\n\n\nAbout Us\n--------\n\n\nCivic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest." ]
[ -0.07905156910419464, 0.016066186130046844, -0.0049019064754247665, 0.03880559280514717, 0.11747734248638153, -0.025776969268918037, 0.20655949413776398, 0.023515531793236732, 0.09398451447486877, -0.035853855311870575, 0.08108482509851456, 0.13080668449401855, -0.01706824265420437, 0.13460232317447662, -0.10768435150384903, -0.2179214358329773, 0.04876385256648064, 0.059240736067295074, 0.038918256759643555, 0.10573556274175644, 0.11775710433721542, -0.07309464365243912, 0.05233168601989746, 0.011716045439243317, -0.029098110273480415, -0.01694837585091591, 0.0437735989689827, -0.14913766086101532, 0.13871905207633972, 0.01901128515601158, 0.06971719115972519, 0.030787194147706032, -0.03959740325808525, -0.020768949761986732, 0.02976277843117714, -0.05001736432313919, -0.07685453444719315, 0.03856084123253822, 0.0774054080247879, -0.09264584630727768, 0.22410902380943298, -0.04826197028160095, -0.02507856860756874, -0.026140892878174782, -0.18929359316825867, -0.17477203905582428, -0.08207891881465912, 0.08641457557678223, -0.0029855973552912474, 0.17250950634479523, -0.0005893466295674443, 0.23581741750240326, -0.1179843321442604, 0.05142107978463173, 0.09087765961885452, -0.22526168823242188, -0.04224886745214462, -0.01954675279557705, -0.011130494996905327, 0.07322602719068527, 0.051999080926179886, 0.10758133977651596, -0.012330609373748302, 0.05367635563015938, -0.010189964435994625, -0.015050264075398445, -0.020826661959290504, -0.048403043299913406, -0.10805384814739227, -0.07228998839855194, 0.40614813566207886, 0.007922004908323288, 0.025292828679084778, 0.011750380508601665, 0.0372244156897068, 0.08725155144929886, 0.019502777606248856, 0.008639977313578129, -0.07492609322071075, -0.0020763948559761047, -0.0047498526982963085, -0.0020464530680328608, -0.1405879557132721, 0.05894821882247925, -0.2148965746164322, 0.30826687812805176, 0.004124253056943417, 0.08928117901086807, -0.08919161558151245, 0.0682256817817688, 0.004526443313807249, -0.07995238900184631, 0.03974155709147453, -0.014473062008619308, 0.18287824094295502, -0.01804495044052601, 0.03627626225352287, -0.06715558469295502, 0.06998106092214584, 0.05266755819320679, -0.12373145669698715, -0.11770708858966827, 0.06013720855116844, 0.1580992341041565, 0.003073663916438818, 0.03428022563457489, -0.02064763754606247, 0.11379121989011765, 0.007440127432346344, -0.050111956894397736, 0.10481560975313187, 0.040511976927518845, -0.07614447921514511, 0.04119361191987991, 0.008364083245396614, 0.06572917103767395, 0.08163964003324509, 0.10945655405521393, -0.044597379863262177, 0.03301619738340378, 0.030676262453198433, -0.004262801259756088, -0.03644109517335892, -0.09167786687612534, -0.04038318246603012, 0.051797185093164444, 0.019621284678578377, -0.0063141887076199055, -0.10253237187862396, 0.18034130334854126, -0.023524435237050056, -0.0037325453013181686, 0.04880881309509277, -0.008259469643235207, 0.08567515760660172, -0.06980317085981369, 0.07870353013277054, -0.21037639677524567, -0.01590961590409279, -0.018545446917414665, 0.033460211008787155, -0.039072178304195404, -0.06771659851074219, 0.04810429364442825, -0.011538718827068806, -0.07987897098064423, -0.07828878611326218, -0.10040206462144852, -0.0833248645067215, 0.14627806842327118, -0.025289874523878098, -0.002541313413530588, -0.15797701478004456, 0.01438634842634201, -0.12299038469791412, -0.025580275803804398, 0.01404849998652935, -0.060020484030246735, -0.06140865013003349, 0.04915881156921387, -0.014268248341977596, 0.01477106474339962, -0.005782968830317259, 0.02736896462738514, 0.018744798377156258, 0.2065264880657196, 0.1033283993601799, -0.049460697919130325, 0.07388360053300858, -0.10293258726596832, -0.17130506038665771, 0.10728342831134796, 0.021540803834795952, 0.2002110630273819, 0.15043146908283234, 0.12083660066127777, -0.04143926873803139, -0.2440742701292038, -0.124014712870121, -0.03180209919810295, -0.13156302273273468, -0.16875553131103516, 0.0792856365442276, -0.008437369018793106, -0.048763908445835114, 0.06337876617908478, -0.17701639235019684, 0.07775087654590607, -0.1047544777393341, -0.017046216875314713, -0.006067186594009399, -0.04492441192269325, 0.07228617370128632, 0.029269343242049217, 0.12957000732421875, -0.10294214636087418, -0.11609885096549988, -0.007408870384097099, -0.018585985526442528, 0.02860712632536888, 0.04171581193804741, -0.1279069036245346, 0.2010825276374817, 0.0026780462358146906, 0.011963294818997383, -0.14728105068206787, -0.025612127035856247, -0.038988590240478516, 0.017139218747615814, 0.13663256168365479, 0.03289603441953659, 0.06497713178396225, -0.055197395384311676, -0.07474981993436813, 0.026658669114112854, -0.021680859848856926, 0.020950034260749817, 0.012049364857375622, -0.10211198776960373, 0.1379554718732834, -0.0030765950214117765, 0.06184045970439911, -0.027830421924591064, 0.039787400513887405, 0.1227339580655098, 0.06321415305137634, -0.01956787146627903, 0.07797710597515106, -0.09574095904827118, 0.012861215509474277, 0.043793316930532455, 0.05039913207292557, 0.10464818775653839, 0.04198932275176048, -0.04288307577371597, 0.05320959538221359, -0.0715266615152359, 0.30254417657852173, 0.1649775356054306, -0.15363357961177826, -0.00864849891513586, 0.028866786509752274, -0.06867839395999908, 0.008006897754967213, -0.05050034075975418, 0.06152502819895744, 0.07091879099607468, 0.01131339929997921, 0.048810362815856934, -0.04114878922700882, -0.009644967503845692, 0.06370881199836731, -0.0809469074010849, -0.07761562615633011, 0.113695427775383, 0.08861785382032394, -0.11550793796777725, 0.12867817282676697, 0.24427343904972076, -0.025449300184845924, 0.16465333104133606, -0.028325561434030533, -0.006184936501085758, -0.017112651839852333, -0.038291990756988525, 0.030565768480300903, 0.1008598580956459, -0.08202656358480453, 0.06379181891679764, 0.02702038176357746, 0.02969461865723133, 0.04719728231430054, -0.14034394919872284, -0.09229160845279694, -0.031031835824251175, 0.0031919104512780905, -0.041403379291296005, 0.10363079607486725, -0.04811647906899452, 0.13456983864307404, 0.024305203929543495, -0.12250987440347672, 0.08113261312246323, 0.025028610602021217, -0.053999755531549454, 0.11433570832014084, 0.05179646983742714, -0.2513314485549927, -0.1288934051990509, -0.04801210016012192, -0.06301472336053848, -0.005332339555025101, 0.10348119586706161, 0.0018156748265028, -0.014680111780762672, 0.006388868670910597, 0.08230960369110107, 0.0011456909123808146, -0.044404998421669006, -0.02346412092447281, 0.07631855458021164, -0.04890120401978493, -0.04689262807369232, -0.09067222476005554, -0.14195440709590912, -0.0469246543943882, 0.07898595184087753, -0.11124217510223389, 0.13365332782268524, 0.01845884509384632, -0.0207054540514946, 0.05476762726902962, -0.02992958016693592, 0.1642698347568512, -0.12289495766162872, 0.00785911176353693, 0.16223673522472382, 0.03571459278464317, 0.0487816259264946, 0.16499149799346924, 0.07327986508607864, -0.015464264899492264, -0.018215207383036613, -0.06948459148406982, -0.06237003207206726, -0.23353180289268494, -0.14150427281856537, -0.06326974928379059, 0.08560456335544586, -0.03722201660275459, 0.049858711659908295, 0.09736707806587219, 0.09369911253452301, -0.0016173750627785921, -0.04920295625925064, 0.043477632105350494, 0.058272477239370346, 0.32008683681488037, 0.045560624450445175, 0.05827372893691063, -0.08194613456726074, -0.090177521109581, 0.09432409703731537, -0.03725990280508995, 0.1414913535118103, 0.09309114515781403, 0.010401099920272827, 0.0883086770772934, 0.04105585068464279, 0.054453618824481964, 0.07624442875385284, 0.1265261024236679, -0.012536932714283466, 0.021291881799697876, -0.04359199479222298, 0.012473474256694317, -0.06126183271408081, -0.061331845819950104, -0.06539339572191238, -0.0579284131526947, -0.16420705616474152, 0.07231360673904419, 0.17198772728443146, -0.04180096462368965, -0.15339700877666473, -0.11631231755018234, -0.01774965599179268, 0.021945443004369736, -0.05020976811647415, -0.05482683330774307, -0.08455388993024826, -0.12936528027057648, 0.043442949652671814, -0.06576204299926758, 0.08092467486858368, -0.10213358700275421, 0.05048695206642151, -0.03785102069377899, -0.17559689283370972, 0.030457906424999237, 0.07661973685026169, -0.11421297490596771, 0.40565258264541626, 0.003894887864589691, 0.09705089032649994, -0.09226953238248825, -0.05685539171099663, -0.03080804832279682, 0.24695883691310883, 0.1527511030435562, 0.005106863100081682, 0.09046876430511475, -0.06607294082641602, -0.10175330936908722, 0.0376276895403862, 0.011601217091083527, -0.10967247188091278, -0.011045828461647034, -0.021886296570301056, 0.004188053775578737, -0.0045255753211677074, -0.11314139515161514, -0.06476585566997528, -0.1146467924118042, 0.07927031069993973, 0.026423675939440727, 0.0890086367726326, -0.005756240803748369, -0.13909828662872314, -0.1050180196762085, 0.07441014051437378, -0.07633800059556961, -0.11048093438148499, -0.07214498519897461, -0.06491955369710922, -0.059838779270648956, -0.10936790704727173, 0.01974494941532612, -0.034727003425359726, -0.13410361111164093, -0.03592384606599808, -0.09146445244550705, 0.08265800774097443, -0.04895348846912384, -0.07285014539957047, 0.011054795235395432, 0.12401725351810455, 0.04754355177283287, -0.006722358986735344, -0.047325197607278824, 0.061488308012485504, -0.060731858015060425, -0.08301005512475967, 0.007680921815335751, -0.04606366157531738, 0.1697828471660614, -0.017076054587960243, -0.0022847021464258432, -0.1729886680841446, -0.06473414599895477, -0.11108708381652832, 0.11489366739988327, 0.2472728043794632, -0.03954218700528145, -0.03150080889463425, 0.18619106709957123, -0.035497523844242096, -0.19222146272659302, -0.11347322165966034, -0.07648374885320663, -0.018484655767679214, 0.04742422699928284, -0.058283593505620956, 0.03614451363682747, 0.18936705589294434, -0.12764230370521545, -0.01026194915175438, -0.182393878698349, -0.06700186431407928, 0.13412779569625854, -0.021112365648150444, 0.3150877058506012, -0.09788988530635834, -0.013313760980963707, 0.049299802631139755, -0.1421067863702774, 0.194784015417099, -0.0711418017745018, 0.048901788890361786, -0.04514353349804878, 0.025388553738594055, -0.014314047060906887, -0.020132800564169884, 0.10155848413705826, -0.08169084787368774, 0.009927231818437576, -0.11294453591108322, -0.218669131398201, 0.1525668501853943, 0.02312736213207245, 0.07664075493812561, 0.07905647158622742, -0.014181428588926792, -0.03973982855677605, -0.05419604852795601, -0.09464540332555771, 0.12931156158447266, 0.019267085939645767, -0.051164042204618454, -0.05409229174256325, 0.048249468207359314, -0.09136934578418732, -0.025251328945159912, 0.25070270895957947, -0.06632523983716965, 0.14726491272449493, 0.09508731216192245, 0.16815604269504547, -0.07712742686271667, -0.029186977073550224, -0.04886763542890549, -0.07865157723426819, 0.029957298189401627, 0.007722865324467421, -0.05791911482810974, 0.107343889772892, -0.04054728150367737, -0.016876760870218277, 0.045852094888687134, -0.1120837852358818, 0.017131030559539795, 0.12331654131412506, -0.19800019264221191, -0.11464234441518784, -0.06697020679712296, 0.050328608602285385, 0.03459818288683891, 0.21057164669036865, 0.10516494512557983, -0.03607027605175972, -0.018009625375270844, -0.004584568552672863, -0.03264394402503967, -0.008898952975869179, 0.0018283737590536475, 0.036131780594587326, -0.05163019523024559, -0.06712224334478378, 0.05517550930380821, 0.14291729032993317, -0.13375425338745117, -0.10168039053678513, 0.08095203340053558, -0.09714141488075256, -0.07709605991840363, -0.016606910154223442, -0.017011696472764015, -0.11627133935689926, -0.09274189919233322, -0.13169480860233307, -0.12963362038135529, 0.04311990737915039, 0.18794186413288116, 0.09437926113605499, 0.018005549907684326, -0.030760375782847404, -0.017421698197722435, -0.0790015235543251, 0.0357574000954628, 0.057184722274541855, -0.06418772041797638, -0.07586827874183655, -0.06342267245054245, 0.06200478971004486, 0.09574098140001297, -0.07904942333698273, -0.05426356568932533, -0.11647078394889832, 0.06089983135461807, -0.00015918503049761057, -0.030481796711683273, -0.07619906216859818, -0.02037152089178562, -0.06675498187541962, -0.09689316153526306, -0.07947942614555359, -0.009867089800536633, -0.08119391649961472, 0.06663035601377487, 0.02238517440855503, 0.041794151067733765, -0.08405487984418869, -0.0015534409321844578, 0.06697762757539749, 0.007446419447660446, 0.11802634596824646, 0.07375215739011765, -0.006978836841881275, 0.08092793822288513, -0.13336312770843506, 0.037245094776153564, 0.05402139201760292, -0.0427747517824173, 0.050406116992235184, -0.025980481877923012, -0.026337526738643646, 0.10011487454175949, 0.07630239427089691, 0.0948743000626564, 0.014126522466540337, -0.1116320863366127, 0.12348209321498871, 0.09533653408288956, -0.09367065876722336, -0.06078478321433067, 0.0041611408814787865, 0.019288521260023117, 0.035025209188461304, 0.18027453124523163, -0.07326167076826096, 0.015191925689578056, -0.02430771477520466, 0.039028316736221313, 0.0009159499895758927, -0.18432989716529846, -0.08108390867710114, -0.07052313536405563, 0.010983340442180634, 0.024525538086891174, 0.2038526087999344, 0.11190135031938553, -0.0318143367767334, 0.059344783425331116, 0.11365635693073273, -0.03965449333190918, -0.04950934648513794, 0.025841284543275833, -0.0341835618019104, 0.03185869753360748, -0.041344307363033295, 0.0011842164676636457, -0.0552426278591156, -0.10125155001878738, 0.16131652891635895, 0.03482813760638237, 0.030197633430361748, 0.007759596221148968, 0.03625420853495598, 0.09041175991296768, -0.058205123990774155, -0.25125470757484436, -0.02007182314991951, -0.0028701634146273136, -0.038413722068071365, 0.034852284938097, 0.14274650812149048, -0.00892299972474575, -0.008887450210750103, -0.04873444139957428, -0.012241778895258904, -0.06304444372653961, -0.20082087814807892, -0.06638459861278534, -0.08093824237585068, 0.010114840231835842, -0.035679832100868225, -0.020641272887587547, 0.1406691074371338, -0.0034296601079404354, -0.12716999650001526, 0.03229699283838272, -0.01665939949452877, -0.06792793422937393, 0.08303274214267731, -0.038935381919145584, 0.040127936750650406, -0.22674869000911713, 0.016170579940080643, -0.1815677136182785, 0.050135087221860886, 0.0017431817250326276, 0.08342792093753815, -0.10371152311563492, -0.045436255633831024, -0.1379595249891281, -0.07426227629184723, -0.01799686625599861, 0.0884542167186737, 0.031040219590067863, 0.09513135254383087, -0.00196011527441442, 0.03625982999801636, 0.0985291451215744, 0.23088963329792023, 0.007066212594509125, -0.12133582681417465, -0.02953982539474964, 0.1274951547384262, -0.022778771817684174, 0.04515240341424942, 0.02626165747642517, -0.04615398123860359, 0.020020464435219765, 0.22301128506660461, 0.28949734568595886, 0.05030130222439766, 0.05545840412378311, -0.026307767257094383, 0.03788668289780617, 0.06186242401599884, 0.0498608760535717, 0.008102988824248314, 0.26661738753318787, -0.10350821167230606, 0.04527818411588669, -0.020013712346553802, 0.0858309343457222, -0.0981597974896431, 0.005000196397304535, 0.006694204639643431, -0.08067399263381958, -0.09468644112348557, 0.16954025626182556, -0.11147204786539078, 0.0933748111128807, -0.05141610652208328, -0.1513429433107376, -0.11686163395643234, -0.005179329309612513, 0.024015169590711594, -0.018160145729780197, 0.10851892828941345, -0.03280145674943924, -0.05001852661371231, -0.04298075661063194, 0.07752435654401779, -0.16285473108291626, -0.043853290379047394, 0.13766008615493774, 0.06729619204998016, 0.16848713159561157, 0.011142611503601074, 0.11874738335609436, 0.05429711937904358, -0.04777509719133377, -0.06885569542646408, 0.13780082762241364, 0.03173977509140968, -0.025108702480793, 0.03558271750807762, -0.1447448432445526, -0.0023040957748889923, -0.034646037966012955, 0.057562634348869324, -0.07454302161931992, 0.0972728356719017, -0.05725298076868057, -0.002785838209092617, -0.042283229529857635, 0.06518951058387756, -0.06917330622673035, 0.02568676881492138, 0.06168215349316597, -0.0242376197129488, -0.07041893899440765, -0.027002492919564247, 0.04758426919579506, 0.1241825595498085, -0.10297702252864838, -0.018490131944417953, -0.0838925763964653, -0.04463762417435646, 0.03344569727778435, 0.06391146034002304, -0.07080045342445374, -0.0033302591182291508, -0.0945274606347084, 0.00631204666569829, -0.1274549812078476, -0.0466446615755558, 0.05659324675798416, -0.06101299077272415, -0.0368962436914444, -0.01062831748276949, 0.01462324894964695, 0.038160860538482666, -0.1153833419084549, -0.06712779402732849 ]
null
null
transformers
# RoBERTa Greek base model Pretrained model on Greek language with the Masked Language Modeling (MLM) objective using [Hugging Face's](https://huggingface.co/) [Transformers](https://github.com/huggingface/transformers) library. This model is *NOT* case-sensitive and all Greek diacritics retained. ### How to use You can use this model directly with a pipeline for masked language modeling: ```python # example url # https://www.news247.gr/politiki/misologa-maximoy-gia-tin-ekthesi-tsiodra-lytra-gia-ti-thnitotita-ektos-meth.9462425.html # not present in train/eval set from transformers import pipeline pipe = pipeline('fill-mask', model='cvcio/roberta-el-news') pipe( 'Η κυβέρνηση μουδιασμένη από τη <mask> της έκθεσης Τσιόδρα-Λύτρα, ' 'επιχειρεί χωρίς να απαντά ουσιαστικά να ρίξει ευθύνες στον ΣΥΡΙΖΑ, που κυβερνούσε πριν... 2 χρόνια.' ) # outputs [ { 'sequence': 'Η κυβέρνηση μουδιασμένη από τη δημοσιοποίηση της έκθεσης Τσιόδρα-Λύτρα, επιχειρεί χωρίς να απαντά ουσιαστικά να ρίξει ευθύνες στον ΣΥΡΙΖΑ, που κυβερνούσε πριν... 2 χρόνια.', 'score': 0.5881184339523315, 'token': 20235, 'token_str': ' δημοσιοποίηση' }, { 'sequence': 'Η κυβέρνηση μουδιασμένη από τη δημοσίευση της έκθεσης Τσιόδρα-Λύτρα, επιχειρεί χωρίς να απαντά ουσιαστικά να ρίξει ευθύνες στον ΣΥΡΙΖΑ, που κυβερνούσε πριν... 2 χρόνια.', 'score': 0.05952141433954239, 'token': 9696, 'token_str': ' δημοσίευση' }, { 'sequence': 'Η κυβέρνηση μουδιασμένη από τη διαχείριση της έκθεσης Τσιόδρα-Λύτρα, επιχειρεί χωρίς να απαντά ουσιαστικά να ρίξει ευθύνες στον ΣΥΡΙΖΑ, που κυβερνούσε πριν... 2 χρόνια.', 'score': 0.029887061566114426, 'token': 4315, 'token_str': ' διαχείριση' }, { 'sequence': 'Η κυβέρνηση μουδιασμένη από τη διαρροή της έκθεσης Τσιόδρα-Λύτρα, επιχειρεί χωρίς να απαντά ουσιαστικά να ρίξει ευθύνες στον ΣΥΡΙΖΑ, που κυβερνούσε πριν... 2 χρόνια.', 'score': 0.022848669439554214, 'token': 24940, 'token_str': ' διαρροή' }, { 'sequence': 'Η κυβέρνηση μουδιασμένη από τη ματαίωση της έκθεσης Τσιόδρα-Λύτρα, επιχειρεί χωρίς να απαντά ουσιαστικά να ρίξει ευθύνες στον ΣΥΡΙΖΑ, που κυβερνούσε πριν... 2 χρόνια.', 'score': 0.01729060709476471, 'token': 46913, 'token_str': ' ματαίωση' } ] ``` ## Training data The model was pretrained on 8 millon unique news articles (~ approx 160M sentences, 33GB of text), collected with [MediaWatch](https://mediawatch.io/), from October 2016 upto December 2021. ## Preprocessing The texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50,265. During the preprocessing we only unescaped html text to the correspoing Unicode characters (ex. `&amp;` => `&`). ## Pretraining The model was pretrained using an NVIDIA A10 GPU for 3 epochs (~ approx 760K steps, 182 hours) with a batch size of 14 (x2 gradient accumulation steps = 28) and a sequence length of 512 tokens. The optimizer used is Adam with a learning rate of 5e-5, and linear decay of the learning rate. ### Training results | epochs | steps | train/train_loss | train/loss | eval/loss | |-------:|--------:|-----------------:|------------:|----------:| | 3 | 765,414 | 0.3960 | 1.2356 | 0.9028 | ### Evaluation results The model fine-tuned on ner task using the [elNER](https://github.com/nmpartzio/elner) dataset and achieved the following results: | task | epochs | lr | batch | dataset | precision | recall | f1 | accuracy | |-----:|-------:|-----:|------:|--------:|----------:|-------:|-------:|---------:| | ner | 5 | 1e-5 | 16/16 | elNER4 | 0.8954 | 0.9280 | 0.9114 | 0.9872 | | ner | 5 | 1e-4 | 16/16 | elNER18 | 0.9069 | 0.9268 | 0.9168 | 0.9823 | ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-5 - train_batch_size: 14 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 28 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Framework versions - Transformers 4.13.0 - Pytorch 1.9.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3 ## Authors Dimitris Papaevagelou - [@andefined](https://github.com/andefined) ## About Us [Civic Information Office](https://cvcio.org/) is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest.
{"language": "el", "license": "gpl-3.0", "tags": ["generated_from_trainer", "roberta", "Greek", "news", "transformers"], "widget": [{"text": "\u0397 \u03ba\u03c5\u03b2\u03ad\u03c1\u03bd\u03b7\u03c3\u03b7 \u03bc\u03bf\u03c5\u03b4\u03b9\u03b1\u03c3\u03bc\u03ad\u03bd\u03b7 \u03b1\u03c0\u03cc \u03c4\u03b7 <mask> \u03c4\u03b7\u03c2 \u03ad\u03ba\u03b8\u03b5\u03c3\u03b7\u03c2 \u03a4\u03c3\u03b9\u03cc\u03b4\u03c1\u03b1-\u039b\u03cd\u03c4\u03c1\u03b1, \u03b5\u03c0\u03b9\u03c7\u03b5\u03b9\u03c1\u03b5\u03af \u03c7\u03c9\u03c1\u03af\u03c2 \u03bd\u03b1 \u03b1\u03c0\u03b1\u03bd\u03c4\u03ac \u03bf\u03c5\u03c3\u03b9\u03b1\u03c3\u03c4\u03b9\u03ba\u03ac \u03bd\u03b1 \u03c1\u03af\u03be\u03b5\u03b9 \u03b5\u03c5\u03b8\u03cd\u03bd\u03b5\u03c2 \u03c3\u03c4\u03bf\u03bd \u03a3\u03a5\u03a1\u0399\u0396\u0391, \u03c0\u03bf\u03c5 \u03ba\u03c5\u03b2\u03b5\u03c1\u03bd\u03bf\u03cd\u03c3\u03b5 \u03c0\u03c1\u03b9\u03bd... 2 \u03c7\u03c1\u03cc\u03bd\u03b9\u03b1."}], "model-index": [{"name": "roberta-el-news", "results": []}]}
fill-mask
cvcio/roberta-el-news
[ "transformers", "pytorch", "safetensors", "roberta", "fill-mask", "generated_from_trainer", "Greek", "news", "el", "doi:10.57967/hf/0712", "license:gpl-3.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "el" ]
TAGS #transformers #pytorch #safetensors #roberta #fill-mask #generated_from_trainer #Greek #news #el #doi-10.57967/hf/0712 #license-gpl-3.0 #autotrain_compatible #endpoints_compatible #region-us
RoBERTa Greek base model ======================== Pretrained model on Greek language with the Masked Language Modeling (MLM) objective using Hugging Face's Transformers library. This model is *NOT* case-sensitive and all Greek diacritics retained. ### How to use You can use this model directly with a pipeline for masked language modeling: Training data ------------- The model was pretrained on 8 millon unique news articles (~ approx 160M sentences, 33GB of text), collected with MediaWatch, from October 2016 upto December 2021. Preprocessing ------------- The texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50,265. During the preprocessing we only unescaped html text to the correspoing Unicode characters (ex. '&' => '&'). Pretraining ----------- The model was pretrained using an NVIDIA A10 GPU for 3 epochs (~ approx 760K steps, 182 hours) with a batch size of 14 (x2 gradient accumulation steps = 28) and a sequence length of 512 tokens. The optimizer used is Adam with a learning rate of 5e-5, and linear decay of the learning rate. ### Training results ### Evaluation results The model fine-tuned on ner task using the elNER dataset and achieved the following results: ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 5e-5 * train\_batch\_size: 14 * eval\_batch\_size: 8 * seed: 42 * gradient\_accumulation\_steps: 2 * total\_train\_batch\_size: 28 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Framework versions * Transformers 4.13.0 * Pytorch 1.9.0+cu111 * Datasets 1.16.1 * Tokenizers 0.10.3 Authors ------- Dimitris Papaevagelou - @andefined About Us -------- Civic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest.
[ "### How to use\n\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\nTraining data\n-------------\n\n\nThe model was pretrained on 8 millon unique news articles (~ approx 160M sentences, 33GB of text), collected with MediaWatch, from October 2016 upto December 2021.\n\n\nPreprocessing\n-------------\n\n\nThe texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50,265. During the preprocessing we only unescaped html text to the correspoing Unicode characters (ex. '&' => '&').\n\n\nPretraining\n-----------\n\n\nThe model was pretrained using an NVIDIA A10 GPU for 3 epochs (~ approx 760K steps, 182 hours) with a batch size of 14 (x2 gradient accumulation steps = 28) and a sequence length of 512 tokens. The optimizer used is Adam with a learning rate of 5e-5, and linear decay of the learning rate.", "### Training results", "### Evaluation results\n\n\nThe model fine-tuned on ner task using the elNER dataset and achieved the following results:", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-5\n* train\\_batch\\_size: 14\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 28\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Framework versions\n\n\n* Transformers 4.13.0\n* Pytorch 1.9.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3\n\n\nAuthors\n-------\n\n\nDimitris Papaevagelou - @andefined\n\n\nAbout Us\n--------\n\n\nCivic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest." ]
[ "TAGS\n#transformers #pytorch #safetensors #roberta #fill-mask #generated_from_trainer #Greek #news #el #doi-10.57967/hf/0712 #license-gpl-3.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### How to use\n\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\nTraining data\n-------------\n\n\nThe model was pretrained on 8 millon unique news articles (~ approx 160M sentences, 33GB of text), collected with MediaWatch, from October 2016 upto December 2021.\n\n\nPreprocessing\n-------------\n\n\nThe texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50,265. During the preprocessing we only unescaped html text to the correspoing Unicode characters (ex. '&' => '&').\n\n\nPretraining\n-----------\n\n\nThe model was pretrained using an NVIDIA A10 GPU for 3 epochs (~ approx 760K steps, 182 hours) with a batch size of 14 (x2 gradient accumulation steps = 28) and a sequence length of 512 tokens. The optimizer used is Adam with a learning rate of 5e-5, and linear decay of the learning rate.", "### Training results", "### Evaluation results\n\n\nThe model fine-tuned on ner task using the elNER dataset and achieved the following results:", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-5\n* train\\_batch\\_size: 14\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 28\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Framework versions\n\n\n* Transformers 4.13.0\n* Pytorch 1.9.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3\n\n\nAuthors\n-------\n\n\nDimitris Papaevagelou - @andefined\n\n\nAbout Us\n--------\n\n\nCivic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest." ]
[ 76, 226, 4, 27, 126, 80 ]
[ "passage: TAGS\n#transformers #pytorch #safetensors #roberta #fill-mask #generated_from_trainer #Greek #news #el #doi-10.57967/hf/0712 #license-gpl-3.0 #autotrain_compatible #endpoints_compatible #region-us \n### How to use\n\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\nTraining data\n-------------\n\n\nThe model was pretrained on 8 millon unique news articles (~ approx 160M sentences, 33GB of text), collected with MediaWatch, from October 2016 upto December 2021.\n\n\nPreprocessing\n-------------\n\n\nThe texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50,265. During the preprocessing we only unescaped html text to the correspoing Unicode characters (ex. '&' => '&').\n\n\nPretraining\n-----------\n\n\nThe model was pretrained using an NVIDIA A10 GPU for 3 epochs (~ approx 760K steps, 182 hours) with a batch size of 14 (x2 gradient accumulation steps = 28) and a sequence length of 512 tokens. The optimizer used is Adam with a learning rate of 5e-5, and linear decay of the learning rate.### Training results### Evaluation results\n\n\nThe model fine-tuned on ner task using the elNER dataset and achieved the following results:### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-5\n* train\\_batch\\_size: 14\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 28\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0" ]
[ -0.07185406982898712, 0.11020053178071976, -0.0020101454574614763, 0.07151404768228531, 0.09931200742721558, 0.03591727465391159, 0.08194401860237122, 0.094209223985672, -0.05608896166086197, 0.1316375434398651, 0.12507982552051544, -0.02822117507457733, 0.10092224925756454, 0.12001820653676987, 0.03529959172010422, -0.2515507936477661, 0.052757762372493744, -0.04426714777946472, 0.009512585587799549, 0.12931206822395325, 0.08196597546339035, -0.12235694378614426, 0.07491415739059448, 0.01327538676559925, -0.07076792418956757, -0.05833911895751953, -0.03820396959781647, -0.03132692351937294, 0.09277646988630295, 0.04503678157925606, 0.05884741619229317, 0.0020982606802135706, 0.09628027677536011, -0.22324785590171814, 0.011861863546073437, 0.09503134340047836, -0.010271716862916946, 0.053635887801647186, 0.08407444506883621, 0.03167155757546425, 0.15461379289627075, -0.10139526426792145, 0.06784902513027191, 0.015596142038702965, -0.1412007063627243, -0.1642000675201416, -0.09539168328046799, -0.025481585413217545, 0.14440006017684937, 0.1097228154540062, -0.050172947347164154, 0.07752158492803574, -0.06950666755437851, 0.03488064929842949, 0.09671923518180847, -0.30072346329689026, -0.06944169849157333, -0.039398279041051865, 0.01907891407608986, 0.048086389899253845, -0.10234393924474716, -0.0004109617730136961, 0.023874029517173767, 0.012304898351430893, 0.11129923164844513, 0.022033169865608215, 0.16414964199066162, -0.02672925963997841, -0.15574049949645996, -0.11093519628047943, 0.04375205561518669, 0.05959800258278847, -0.02653542347252369, -0.15382297337055206, -0.0668570026755333, -0.19411307573318481, -0.022222958505153656, 0.007135681342333555, 0.01255261804908514, -0.029796579852700233, -0.014489947818219662, -0.046348124742507935, -0.09493106603622437, -0.07088738679885864, 0.005201473366469145, 0.2106115221977234, 0.07870911806821823, 0.018948256969451904, 0.02405151166021824, 0.12197966128587723, -0.02305910363793373, -0.12074458599090576, -0.04494404047727585, 0.012937341816723347, -0.13571526110172272, 0.013535771518945694, -0.036050695925951004, -0.0577268972992897, -0.04473615810275078, 0.18846024572849274, -0.03200173005461693, 0.06144675612449646, 0.12016166001558304, -0.014138953760266304, -0.02456015720963478, 0.149918332695961, -0.08196529000997543, -0.08572033047676086, -0.01733807474374771, 0.132459819316864, 0.048143960535526276, -0.05116267874836922, -0.03264826536178589, 0.043841518461704254, 0.06975394487380981, 0.011679037474095821, 0.01255602017045021, 0.01977311074733734, -0.05533013120293617, -0.038596030324697495, 0.12867365777492523, -0.10663089901208878, 0.02737193927168846, 0.02880108915269375, -0.058407012373209, -0.056902751326560974, -0.01650478132069111, -0.01833200454711914, -0.0589773990213871, 0.1166563555598259, -0.07182776927947998, -0.04328848421573639, -0.1309424489736557, -0.05936560034751892, -0.013762394897639751, -0.051752299070358276, -0.041275180876255035, -0.07643824815750122, -0.12946456670761108, -0.0602986104786396, 0.0659528598189354, -0.03755760192871094, -0.06623371690511703, -0.05247502401471138, -0.03188144043087959, 0.04886458069086075, -0.02712554857134819, 0.11720864474773407, -0.04686165973544121, 0.07971690595149994, 0.03266407176852226, 0.031956952065229416, 0.05212508514523506, 0.030347708612680435, -0.08331653475761414, 0.057328712195158005, -0.02266489528119564, 0.09355362504720688, 0.0019262864952906966, -0.0032297559082508087, -0.07618514448404312, -0.09286241978406906, -0.09859472513198853, -0.013445399701595306, 0.09084899723529816, 0.07298643887042999, -0.13307073712348938, -0.035141605883836746, 0.13234049081802368, -0.09735114127397537, -0.07283042371273041, 0.16803759336471558, -0.04533328488469124, -0.08233422040939331, 0.05385136976838112, 0.08622001856565475, 0.08196540176868439, -0.09472561627626419, -0.05510856956243515, -0.07693305611610413, -0.005619229283183813, 0.06793057918548584, 0.1037115752696991, -0.006399104371666908, 0.029749583452939987, 0.0029883799143135548, -0.09823591262102127, 0.021184438839554787, -0.02934424579143524, -0.08680412918329239, -0.003142555942758918, -0.04496948421001434, 0.014539177529513836, 0.06526894867420197, -0.013183975592255592, -0.07654117047786713, -0.16912102699279785, -0.10717600584030151, 0.0962219387292862, -0.09612838178873062, 0.04031955823302269, -0.08471278101205826, 0.06717902421951294, -0.041435904800891876, 0.010772467590868473, -0.13629208505153656, -0.06657683849334717, 0.04721017926931381, -0.11646462231874466, 0.019068481400609016, -0.05355122685432434, 0.040732432156801224, 0.0753236934542656, -0.04633935168385506, -0.06475894898176193, -0.09743712097406387, -0.04126061871647835, -0.06609248369932175, -0.08483579754829407, -0.049631938338279724, -0.0271261278539896, 0.156561017036438, -0.1096983551979065, -0.011253024451434612, 0.02315649390220642, 0.08140744268894196, 0.000609612266998738, -0.07652103900909424, -0.023821450769901276, 0.0180530846118927, 0.0006471815868280828, -0.05998864024877548, 0.029118487611413002, 0.008508373983204365, -0.02961329184472561, -0.007702960167080164, -0.202151358127594, -0.04679238796234131, 0.11255981028079987, 0.06739436089992523, -0.05387398228049278, -0.0016396676655858755, -0.029719369485974312, -0.04506360739469528, -0.04032915458083153, 0.017191054299473763, 0.27443933486938477, 0.031346987932920456, 0.10944178700447083, -0.08952943235635757, -0.012640354223549366, 0.010811815038323402, 0.029385752975940704, -0.04524534195661545, 0.0773715004324913, 0.06761052459478378, -0.12247780710458755, 0.057287558913230896, 0.04664841666817665, 0.005071609281003475, 0.11644437909126282, 0.013497387990355492, -0.11169003695249557, -0.04063928499817848, 0.008566707372665405, 0.005566836334764957, 0.12797926366329193, -0.03999416157603264, 0.0066584753803908825, 0.024178562685847282, 0.00605736905708909, 0.02599770948290825, -0.17498908936977386, 0.03988979011774063, 0.04808010905981064, -0.03842262923717499, 0.0519871823489666, -0.04207285866141319, -0.057573989033699036, 0.05236804857850075, 0.0950121283531189, -0.018040457740426064, 0.0023588489275425673, -0.05434110760688782, -0.08546552062034607, 0.18474993109703064, -0.061822883784770966, -0.21794798970222473, -0.11180149763822556, 0.0395452044904232, -0.053364239633083344, 0.008728304877877235, 0.0022266958840191364, -0.03169891983270645, -0.087907575070858, -0.0970178171992302, 0.03632175177335739, 0.004741096403449774, -0.0012732025934383273, -0.01164042018353939, -0.03791671246290207, 0.009114168584346771, -0.06810378283262253, 0.035368483513593674, -0.044340234249830246, -0.046273373067379, 0.01155314315110445, 0.07294822484254837, 0.13136333227157593, 0.12861214578151703, 0.01863664574921131, 0.016033664345741272, -0.0015709275612607598, 0.20627471804618835, -0.1566697359085083, 0.004104440566152334, 0.10678649693727493, 0.0004288779746275395, 0.04893965646624565, 0.11355797946453094, 0.016615407541394234, -0.08407532423734665, 0.035623084753751755, 0.04617719352245331, -0.0561102032661438, -0.13496607542037964, -0.059429462999105453, -0.050410594791173935, -0.07429693639278412, 0.16089887917041779, 0.04086637124419212, -0.038491081446409225, 0.027763020247220993, -0.06599388271570206, -0.04462018236517906, 0.04957875236868858, 0.08230289071798325, 0.0543704517185688, 0.057318028062582016, 0.07495813816785812, -0.030311236158013344, -0.004535183776170015, 0.07087840139865875, -0.011272366158664227, 0.2127874344587326, -0.04961210489273071, 0.23132984340190887, 0.0422450490295887, 0.10803171247243881, 0.021112889051437378, 0.014942602254450321, -0.01637766696512699, 0.00719436164945364, 0.0030950908549129963, -0.019939148798584938, -0.07945790141820908, 0.07111254334449768, 0.062002357095479965, 0.0206292811781168, -0.06740671396255493, 0.022013336420059204, 0.05455024540424347, 0.28054219484329224, 0.07995034754276276, -0.27806568145751953, -0.05922145023941994, 0.010180174373090267, -0.06256093084812164, -0.07615336775779724, 0.04506111890077591, 0.09154027700424194, -0.08829803764820099, 0.05521168187260628, -0.09438443928956985, 0.06251194328069687, -0.1134408712387085, -0.017006507143378258, 0.13589709997177124, 0.13072898983955383, 0.00011486838775454089, 0.06885562092065811, -0.23723819851875305, 0.17503713071346283, -0.0058779241517186165, 0.116221122443676, -0.03710933029651642, 0.007874177768826485, -0.002497006906196475, -0.012902414426207542, 0.10582059621810913, -0.027827871963381767, -0.00841281097382307, -0.03151471167802811, -0.14578858017921448, 0.026819318532943726, 0.11042775958776474, -0.06589142978191376, 0.12510435283184052, -0.023144938051700592, -0.005120978225022554, -0.010875081643462181, -0.02734730765223503, -0.03302321583032608, -0.16355153918266296, 0.05511185899376869, -0.07265496999025345, 0.01947750337421894, -0.07540475577116013, -0.06735359877347946, -0.12881304323673248, 0.12852007150650024, -0.07091439515352249, -0.061805419623851776, -0.11325322091579437, 0.03202259913086891, 0.08847687393426895, -0.07462440431118011, 0.0601089671254158, 0.003155659418553114, 0.11886188387870789, -0.02533089555799961, -0.03274589776992798, 0.05769838020205498, -0.013179282657802105, -0.2088620364665985, -0.0060546682216227055, 0.13127769529819489, 0.07089327275753021, 0.04799012839794159, -0.026848623529076576, 0.0830584466457367, -0.05059683322906494, -0.09636177867650986, 0.026689067482948303, -0.005584605038166046, 0.036915089935064316, -0.006521482486277819, 0.08801400661468506, 0.00362120708450675, -0.09257543832063675, 0.014782004058361053, 0.07829855382442474, 0.2352587729692459, -0.07278896868228912, 0.05247014760971069, 0.11733114719390869, -0.02284904569387436, -0.20959973335266113, -0.04864414036273956, 0.06195041537284851, 0.0628746747970581, -0.05600442364811897, -0.16103416681289673, -0.037113506346940994, 0.06303604692220688, -0.033449672162532806, 0.0680554136633873, -0.32333579659461975, -0.1378476619720459, 0.04512961953878403, 0.10613708198070526, -0.02218526042997837, -0.09761545062065125, -0.032237302511930466, 0.002894036704674363, -0.019831843674182892, 0.12437979876995087, 0.043804533779621124, 0.10352488607168198, -0.004831692669540644, -0.030692532658576965, 0.021999893710017204, -0.07414553314447403, 0.17011956870555878, -0.03736650198698044, 0.07712319493293762, -0.0398935042321682, 0.06979871541261673, -0.0017706938087940216, -0.07099968940019608, 0.05675681680440903, -0.047075286507606506, 0.01777949556708336, -0.08542147278785706, -0.06554347276687622, -0.08292677253484726, -0.010893544182181358, -0.03974643722176552, -0.04012518376111984, -0.06877780705690384, 0.07618860900402069, 0.07335074245929718, 0.016311118379235268, 0.09559816122055054, -0.0034471589606255293, -0.05484217777848244, 0.08221183717250824, 0.10545933246612549, 0.009208356030285358, -0.11820897459983826, -0.03278358653187752, 0.02292104810476303, 0.06932924687862396, -0.13164030015468597, 0.0020080709364265203, 0.10572721809148788, 0.016455451026558876, 0.12429090589284897, 0.004073291085660458, -0.13831011950969696, 0.011572230607271194, 0.06334107369184494, -0.14335380494594574, -0.19313569366931915, -0.010593285784125328, -0.07028192281723022, -0.11160703003406525, -0.06711044907569885, 0.06471017003059387, -0.05115637555718422, -0.00023821981449145824, 0.0015217306790873408, 0.06182873249053955, -0.0041116951033473015, 0.12039858847856522, 0.02416480891406536, 0.04302436485886574, -0.10719870030879974, 0.12950082123279572, 0.09132536500692368, -0.14904208481311798, 0.005179012659937143, 0.13954216241836548, -0.08054418116807938, -0.05602901056408882, -0.06866953521966934, 0.03592974692583084, 0.0471491701900959, -0.05981709808111191, -0.0730459913611412, -0.07262987643480301, 0.06225256994366646, -0.022792832925915718, 0.05104200169444084, 0.01945771835744381, -0.00900298822671175, 0.019178356975317, -0.13554875552654266, 0.1141137182712555, 0.0098203020170331, 0.026542698964476585, -0.06437524408102036, 0.08795344829559326, 0.04418971762061119, 0.010493339039385319, -0.015240214765071869, -0.006913907825946808, -0.02565569244325161, 0.01572325825691223, -0.00911996141076088, -0.020428434014320374, -0.05911092832684517, 0.020671894773840904, -0.018609002232551575, -0.01785651408135891, -0.06191953271627426, 0.016813868656754494, -0.06329427659511566, -0.08189120888710022, -0.012657606042921543, -0.016648756340146065, -0.08220615237951279, -0.021921096369624138, 0.037334274500608444, -0.09351279586553574, 0.10599227249622345, -0.0092931455001235, 0.036491334438323975, 0.027427135035395622, -0.0020149715710431337, 0.0450349897146225, -0.010839872062206268, 0.02439805306494236, -0.01656108908355236, -0.11191315948963165, 0.02171798050403595, -0.014594229869544506, -0.056297991424798965, -0.009111369960010052, 0.04781639575958252, -0.12610794603824615, 0.06495877355337143, -0.012408577837049961, -0.014421703293919563, -0.04386652261018753, 0.04201646149158478, -0.012321169488132, 0.03638618066906929, 0.1691243201494217, -0.05192379280924797, 0.06096911057829857, -0.16536101698875427, -0.014872882515192032, 0.02120707742869854, -0.02551104500889778, -0.076688751578331, -0.03280771151185036, 0.08394236117601395, -0.03126969933509827, 0.11082910746335983, 0.019033627584576607, -0.0032831435091793537, 0.04644467309117317, 0.00909892376512289, -0.025222910568118095, 0.05253493785858154, 0.08927629142999649, 0.022039640694856644, -0.021908387541770935, 0.0785214975476265, 0.019688589498400688, 0.023374607786536217, 0.0719025582075119, 0.16097977757453918, 0.16848769783973694, 0.16991586983203888, 0.09371142834424973, -0.01783278025686741, -0.11835605651140213, -0.13233010470867157, 0.1455545872449875, -0.10853997617959976, 0.07041566073894501, 0.0058260271325707436, 0.08961547911167145, 0.10356831550598145, -0.18717069923877716, 0.054933223873376846, 0.006534218788146973, -0.08190929889678955, -0.05040914937853813, -0.12769974768161774, -0.0627150684595108, -0.06976951658725739, 0.027344688773155212, -0.08574602007865906, 0.024551142007112503, 0.10028550773859024, 0.0111938351765275, -0.04643639177083969, 0.15799428522586823, -0.010947237722575665, -0.016681576147675514, 0.08049069344997406, 0.04294682294130325, 0.03497375175356865, 0.04579649493098259, -0.05680013447999954, 0.009166440926492214, -0.04929622262716293, 0.10461067408323288, -0.020668303593993187, -0.014975650236010551, 0.08785098791122437, 0.05988873913884163, -0.07958970963954926, 0.026606380939483643, 0.008396962657570839, 0.08661576360464096, 0.14947055280208588, 0.04090038686990738, 0.002188713289797306, -0.03311668336391449, 0.14396515488624573, -0.02390092797577381, -0.007086784113198519, -0.13713234663009644, 0.13126240670681, 0.01802097260951996, -0.00045952745131216943, 0.03339806944131851, -0.10666674375534058, 0.008592077530920506, 0.17924313247203827, 0.09499464929103851, -0.09825854003429413, -0.02391960285604, -0.0006441191653721035, -0.003384453011676669, -0.0023597220424562693, 0.09916000068187714, 0.029252037405967712, 0.15118245780467987, -0.06597130745649338, -0.02311490848660469, -0.052561551332473755, -0.040618058294057846, -0.029713500291109085, 0.0919921025633812, -0.0001755131670506671, 0.032362766563892365, -0.10686829686164856, 0.03251758962869644, 0.022777942940592766, -0.11082077771425247, 0.09259284287691116, -0.1309712529182434, -0.14821919798851013, 0.03192056342959404, 0.047064848244190216, -0.013387485407292843, 0.07577630877494812, 0.03366602957248688, 0.009281572885811329, 0.077886663377285, 0.018017303198575974, -0.06747723370790482, -0.07112441211938858, 0.07172827422618866, -0.03465615212917328, 0.27842777967453003, -0.014862722717225552, 0.04433949664235115, 0.09323861449956894, 0.012378190644085407, -0.13580498099327087, 0.09562138468027115, 0.04148133471608162, -0.06887788325548172, 0.048702891916036606, 0.151450514793396, -0.002663266146555543, 0.0560603030025959, 0.07632419466972351, -0.016926869750022888, 0.033881381154060364, -0.09983348846435547, -0.0305529423058033, -0.07771182060241699, 0.049290671944618225, -0.05329463630914688, 0.16638629138469696, 0.18977369368076324, -0.020364413037896156, -0.011340725235641003, -0.01777900941669941, 0.03907085210084915, -0.025303157046437263, 0.20080798864364624, -0.00263361819088459, -0.1861673891544342, 0.01832820102572441, -0.05068984627723694, -0.007943380624055862, -0.220947727560997, -0.0748271718621254, 0.020516974851489067, -0.06341355293989182, -0.019357847049832344, 0.11116600036621094, 0.017225580289959908, 0.013630336150527, -0.06398793309926987, -0.015186019241809845, -0.028850525617599487, 0.09815862774848938, -0.14240211248397827, -0.09659236669540405 ]
null
null
transformers
# Greek RoBERTa Uncased (v1) Pretrained model on Greek language using a masked language modeling (MLM) objective using [Hugging Face's](https://huggingface.co/) [Transformers](https://github.com/huggingface/transformers) library. This model is case-sensitive and has no Greek diacritics (uncased, no-accents). ### Training data This model was pretrained on almost 18M unique tweets, all Greek, collected between 2008-2021, from almost 450K distinct users. ### Preprocessing The texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50256. For the tokenizer we splited strings containing any numbers (ex. EU2019 ==> EU 2019). The tweet normalization logic described in the example listed bellow. ```python import unicodedata from transformers import pipeline def normalize_tweet(tweet, do_lower = True, do_strip_accents = True, do_split_word_numbers = False, user_fill = '', url_fill = ''): # your tweet pre-processing logic goes here # example... # remove extra spaces, escape HTML, replace non-standard punctuation # replace any @user with blank # replace any link with blank # explode hashtags to strings (ex. #EU2019 ==> EU 2019) # remove all emojis # if do_split_word_numbers: # splited strings containing any numbers # standardize punctuation # remove unicode symbols if do_lower: tweet = tweet.lower() if do_strip_accents: tweet = strip_accents(tweet) return tweet.strip() def strip_accents(s): return ''.join(c for c in unicodedata.normalize('NFD', s) if unicodedata.category(c) != 'Mn') nlp = pipeline('fill-mask', model = 'cvcio/roberta-el-uncased-twitter-v1') print( nlp( normalize_tweet( '<mask>: Μεγάλη υποχώρηση του ιικού φορτίου σε Αττική και Θεσσαλονίκη' ) ) ) ``` ### Pretraining The model was pretrained on a T4 GPU for 1.2M steps with a batch size of 96 and a sequence length of 96. The optimizer used is Adam with a learning rate of 1e-5, gradient accumulation steps of 8, learning rate warmup for 50000 steps and linear decay of the learning rate after. ### Authors Dimitris Papaevagelou - [@andefined](https://github.com/andefined) ### About Us [Civic Information Office](https://cvcio.org/) is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest.
{"language": "el", "tags": ["roberta", "twitter", "Greek"], "widget": [{"text": "<mask>: \u03bc\u03b5\u03b3\u03b1\u03bb\u03b7 \u03c5\u03c0\u03bf\u03c7\u03c9\u03c1\u03b7\u03c3\u03b7 \u03c4\u03bf\u03c5 \u03b9\u03b9\u03ba\u03bf\u03c5 \u03c6\u03bf\u03c1\u03c4\u03b9\u03bf\u03c5 \u03c3\u03b5 \u03b1\u03c4\u03c4\u03b9\u03ba\u03b7 \u03ba\u03b1\u03b9 \u03b8\u03b5\u03c3\u03c3\u03b1\u03bb\u03bf\u03bd\u03b9\u03ba\u03b7"}]}
fill-mask
cvcio/roberta-el-uncased-twitter-v1
[ "transformers", "pytorch", "safetensors", "roberta", "fill-mask", "twitter", "Greek", "el", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "el" ]
TAGS #transformers #pytorch #safetensors #roberta #fill-mask #twitter #Greek #el #autotrain_compatible #endpoints_compatible #region-us
# Greek RoBERTa Uncased (v1) Pretrained model on Greek language using a masked language modeling (MLM) objective using Hugging Face's Transformers library. This model is case-sensitive and has no Greek diacritics (uncased, no-accents). ### Training data This model was pretrained on almost 18M unique tweets, all Greek, collected between 2008-2021, from almost 450K distinct users. ### Preprocessing The texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50256. For the tokenizer we splited strings containing any numbers (ex. EU2019 ==> EU 2019). The tweet normalization logic described in the example listed bellow. ### Pretraining The model was pretrained on a T4 GPU for 1.2M steps with a batch size of 96 and a sequence length of 96. The optimizer used is Adam with a learning rate of 1e-5, gradient accumulation steps of 8, learning rate warmup for 50000 steps and linear decay of the learning rate after. ### Authors Dimitris Papaevagelou - @andefined ### About Us Civic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest.
[ "# Greek RoBERTa Uncased (v1)\n\nPretrained model on Greek language using a masked language modeling (MLM) objective using Hugging Face's Transformers library. This model is case-sensitive and has no Greek diacritics (uncased, no-accents).", "### Training data\n\nThis model was pretrained on almost 18M unique tweets, all Greek, collected between 2008-2021, from almost 450K distinct users.", "### Preprocessing\n\nThe texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50256. For the tokenizer we splited strings containing any numbers (ex. EU2019 ==> EU 2019). The tweet normalization logic described in the example listed bellow.", "### Pretraining\n\nThe model was pretrained on a T4 GPU for 1.2M steps with a batch size of 96 and a sequence length of 96. The optimizer used is Adam with a learning rate of 1e-5, gradient accumulation steps of 8, learning rate warmup for 50000 steps and linear decay of the learning rate after.", "### Authors\n\nDimitris Papaevagelou - @andefined", "### About Us\n\nCivic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest." ]
[ "TAGS\n#transformers #pytorch #safetensors #roberta #fill-mask #twitter #Greek #el #autotrain_compatible #endpoints_compatible #region-us \n", "# Greek RoBERTa Uncased (v1)\n\nPretrained model on Greek language using a masked language modeling (MLM) objective using Hugging Face's Transformers library. This model is case-sensitive and has no Greek diacritics (uncased, no-accents).", "### Training data\n\nThis model was pretrained on almost 18M unique tweets, all Greek, collected between 2008-2021, from almost 450K distinct users.", "### Preprocessing\n\nThe texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50256. For the tokenizer we splited strings containing any numbers (ex. EU2019 ==> EU 2019). The tweet normalization logic described in the example listed bellow.", "### Pretraining\n\nThe model was pretrained on a T4 GPU for 1.2M steps with a batch size of 96 and a sequence length of 96. The optimizer used is Adam with a learning rate of 1e-5, gradient accumulation steps of 8, learning rate warmup for 50000 steps and linear decay of the learning rate after.", "### Authors\n\nDimitris Papaevagelou - @andefined", "### About Us\n\nCivic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest." ]
[ 49, 66, 35, 77, 77, 15, 31 ]
[ "passage: TAGS\n#transformers #pytorch #safetensors #roberta #fill-mask #twitter #Greek #el #autotrain_compatible #endpoints_compatible #region-us \n# Greek RoBERTa Uncased (v1)\n\nPretrained model on Greek language using a masked language modeling (MLM) objective using Hugging Face's Transformers library. This model is case-sensitive and has no Greek diacritics (uncased, no-accents).### Training data\n\nThis model was pretrained on almost 18M unique tweets, all Greek, collected between 2008-2021, from almost 450K distinct users.### Preprocessing\n\nThe texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50256. For the tokenizer we splited strings containing any numbers (ex. EU2019 ==> EU 2019). The tweet normalization logic described in the example listed bellow.### Pretraining\n\nThe model was pretrained on a T4 GPU for 1.2M steps with a batch size of 96 and a sequence length of 96. The optimizer used is Adam with a learning rate of 1e-5, gradient accumulation steps of 8, learning rate warmup for 50000 steps and linear decay of the learning rate after.### Authors\n\nDimitris Papaevagelou - @andefined### About Us\n\nCivic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest." ]
[ -0.13889215886592865, 0.030834985896945, -0.0015109189553186297, 0.08547939360141754, 0.08499906212091446, -0.016043299809098244, 0.06009538099169731, 0.04357997700572014, 0.011621802113950253, 0.11339075118303299, 0.11407652497291565, -0.0004191815387457609, 0.053026653826236725, 0.18937434256076813, 0.009774249978363514, -0.28796571493148804, 0.058847442269325256, -0.07105852663516998, 0.009257730096578598, 0.0975346639752388, 0.08654413372278214, -0.0694483295083046, 0.06383232772350311, -0.027119213715195656, -0.013890231028199196, -0.021831821650266647, -0.01621975377202034, -0.0564420185983181, 0.10838375985622406, 0.0049655805341899395, 0.0001803670747904107, 0.016937028616666794, 0.0835934728384018, -0.12039882689714432, 0.012864972464740276, 0.07720614224672318, 0.020155994221568108, 0.03367253392934799, 0.08931837230920792, -0.03399285674095154, 0.18137940764427185, -0.09555332362651825, 0.09859590977430344, 0.010343542322516441, -0.10839520394802094, -0.06632605195045471, -0.0924772322177887, 0.050536300987005234, 0.131745845079422, 0.08190163224935532, -0.04173910617828369, 0.050641998648643494, -0.06242506206035614, 0.0668710470199585, 0.05042340233922005, -0.22819258272647858, -0.04150019586086273, 0.050753459334373474, 0.005399129819124937, 0.060707468539476395, -0.010835074819624424, 0.026827286928892136, 0.02524695172905922, 0.010281451046466827, 0.03504930064082146, 0.021263262256979942, 0.16130752861499786, -0.03491296246647835, -0.15042772889137268, -0.042819540947675705, 0.07277262955904007, 0.018377628177404404, -0.04303625226020813, -0.12056078761816025, -0.03297596052289009, -0.05710771307349205, -0.013027737848460674, -0.010352504439651966, -0.02997669391334057, 0.020856505259871483, 0.024063678458333015, -0.01080864854156971, -0.07287675142288208, 0.015986425802111626, -0.03026767633855343, 0.2834857404232025, 0.05434795841574669, 0.03548193722963333, 0.02100761979818344, 0.07527301460504532, -0.02980560064315796, -0.06328009814023972, 0.033617500215768814, 0.03359132260084152, -0.11746644228696823, -0.013973170891404152, -0.03572867438197136, -0.08954455703496933, -0.03466826304793358, 0.12650208175182343, -0.01594570279121399, 0.052815984934568405, 0.14070016145706177, 0.00918540544807911, 0.0020594701636582613, 0.09310995787382126, -0.07356396317481995, -0.04964419826865196, 0.03709018602967262, 0.10069368779659271, 0.007404905278235674, -0.0627676323056221, -0.05409747362136841, 0.025055183097720146, 0.10519139468669891, 0.05127861350774765, 0.004602259956300259, 0.05049071088433266, -0.04753684625029564, -0.03072742000222206, 0.09258073568344116, -0.12902753055095673, -0.025553859770298004, -0.035591766238212585, -0.04197143018245697, 0.004831272643059492, -0.006055430043488741, -0.03155745938420296, -0.11586140841245651, 0.12312450259923935, -0.05851628631353378, -0.06189403682947159, -0.04144364222884178, -0.05945318937301636, 0.0027333530597388744, -0.11997803300619125, -0.04176472872495651, -0.10706999152898788, -0.15946917235851288, -0.04144585505127907, 0.0837322250008583, -0.02009168453514576, -0.062369443476200104, -0.06051623076200485, -0.03822134807705879, -0.009675348177552223, 0.00838912557810545, 0.06466296315193176, -0.051081299781799316, 0.07884889841079712, -0.09043184667825699, 0.05664859339594841, -0.013861074112355709, -0.00041311135282739997, -0.13878974318504333, -0.01608668826520443, -0.09728749096393585, 0.1162363737821579, -0.028988424688577652, 0.0036310849245637655, -0.1029307171702385, -0.04355468973517418, -0.03494392707943916, 0.01365678384900093, 0.06540026515722275, 0.17536702752113342, -0.1680482029914856, -0.06050771847367287, 0.03588971868157387, -0.11005803942680359, 0.0008418802754022181, 0.16519129276275635, -0.02476557530462742, 0.021853096783161163, 0.09719745814800262, 0.1862206608057022, -0.023417485877871513, -0.11329659819602966, -0.04620855301618576, -0.05440265312790871, -0.023827925324440002, 0.061634719371795654, 0.016633963212370872, -0.048907600343227386, -0.053946685045957565, 0.032640594989061356, -0.0857607051730156, 0.040263157337903976, -0.045312270522117615, -0.030846424400806427, 0.016885191202163696, -0.0744595155119896, 0.04293488711118698, 0.010941319167613983, 0.022692883387207985, -0.07335828989744186, -0.13566993176937103, -0.060882359743118286, 0.05777646601200104, -0.08503541350364685, 0.03406120091676712, -0.08917705714702606, 0.11662416160106659, -0.024414854124188423, 0.019507277756929398, -0.1829083114862442, -0.037066686898469925, 0.06082339212298393, -0.09653367847204208, 0.08871281892061234, -0.09694564342498779, 0.07442518323659897, 0.030173325911164284, -0.06027163565158844, -0.033112578094005585, -0.021510349586606026, -0.011259582825005054, -0.11905843019485474, -0.13567613065242767, 0.012154909782111645, -0.04174782708287239, 0.007858308963477612, -0.08387603610754013, 0.018306149169802666, 0.07018029689788818, 0.11774113029241562, 0.03007475845515728, -0.10898873955011368, -0.015244349837303162, 0.01696367934346199, 0.036803122609853745, -0.08647317439317703, 0.006020023487508297, 0.00848044641315937, -0.06306672096252441, 0.07039991766214371, -0.23210154473781586, -0.07518859207630157, 0.09138783812522888, 0.09664598852396011, -0.11621090769767761, -0.011662313714623451, -0.035641446709632874, -0.01941792480647564, -0.0536186508834362, 0.05447180196642876, 0.27448293566703796, -0.0005915474030189216, 0.0866001695394516, -0.09551713615655899, -0.03383766859769821, 0.05487916246056557, 0.010953330434858799, -0.03260405734181404, 0.09229989349842072, 0.04865572974085808, -0.12871767580509186, -0.00874178484082222, 0.026395713910460472, 0.013699891045689583, 0.13973166048526764, -0.020058520138263702, -0.08617880940437317, 0.004346927162259817, 0.027522427961230278, 0.023857159540057182, 0.1628355234861374, -0.0682796984910965, 0.0027571478858590126, -0.0024120875168591738, 0.07058870047330856, 0.04642999917268753, -0.114796943962574, 0.03486732393503189, 0.0037705181166529655, -0.031301308423280716, 0.009647940285503864, 0.024880075827240944, -0.051382940262556076, 0.0994889959692955, 0.022970642894506454, -0.05435511842370033, -0.023533614352345467, 0.004022557288408279, -0.11988946795463562, 0.21260084211826324, -0.053811147809028625, -0.17164839804172516, -0.06023256108164787, 0.06613385677337646, 0.0010106588015332818, -0.03896765038371086, -0.005513446405529976, -0.08464598655700684, -0.08009810745716095, -0.11454927176237106, 0.08793626725673676, -0.016452474519610405, 0.07449153065681458, -0.02331405319273472, -0.03443829342722893, -0.038323499262332916, -0.057199664413928986, -0.03537757322192192, -0.1045909970998764, -0.03914112225174904, 0.004698712378740311, -0.057384103536605835, 0.05192112177610397, 0.1942073553800583, -0.0680900514125824, 0.03437689691781998, -0.04542793333530426, 0.15743064880371094, -0.10589908808469772, 0.10582911223173141, 0.09739401191473007, 0.007095348555594683, 0.03229719027876854, 0.10170156508684158, -0.0017564500449225307, -0.11614775657653809, 0.049617379903793335, 0.05505840480327606, -0.09056475758552551, -0.18526312708854675, -0.12036076933145523, -0.05746270716190338, -0.011569581925868988, 0.1128380224108696, 0.04673232510685921, -0.058146655559539795, 0.004124309867620468, -0.014849056489765644, -0.015985406935214996, 0.09379352629184723, 0.0848848968744278, 0.06191689148545265, 0.05820617079734802, 0.07360504567623138, -0.022065257653594017, 0.01496030017733574, 0.11895903944969177, -0.048157695680856705, 0.25741150975227356, -0.0858839899301529, 0.029251694679260254, 0.014543390832841396, 0.035674989223480225, -0.0003799172700382769, 0.05970175564289093, 0.01793246902525425, 0.019914662465453148, -0.04041799530386925, -0.03910832107067108, -0.008001098409295082, 0.04749684035778046, -0.006671961396932602, -0.10548973828554153, -0.10477648675441742, 0.06226939707994461, 0.07230488955974579, 0.25524598360061646, 0.07386273890733719, -0.223506897687912, -0.11473111063241959, -0.02628651075065136, -0.05994514748454094, -0.04070160537958145, 0.05113445967435837, 0.0212734192609787, -0.05210626870393753, 0.0669521912932396, -0.04132915660738945, 0.06253974884748459, -0.11208408325910568, -0.028040161356329918, 0.000924514839425683, 0.10250742733478546, -0.042886193841695786, 0.01669193245470524, -0.17128998041152954, 0.19777309894561768, 0.014604667201638222, 0.15845274925231934, -0.04020419716835022, 0.015461409464478493, -0.003248240565881133, 0.0401575081050396, 0.12174463272094727, 0.04079097881913185, -0.04133627936244011, -0.006465288810431957, -0.14903151988983154, 0.03773225098848343, 0.10414217412471771, -0.05092526972293854, 0.029764046892523766, -0.01029679924249649, 0.02596236579120159, -0.027392111718654633, -0.19671744108200073, -0.08213359862565994, -0.147467240691185, 0.05088776722550392, -0.11649809032678604, -0.07326240092515945, -0.054815877228975296, -0.10591621696949005, 0.009133022278547287, 0.1953299194574356, 0.03417939320206642, -0.06766388565301895, -0.11981935799121857, -0.014713726006448269, 0.13230928778648376, -0.07034007459878922, 0.054055724292993546, 0.0027560596354305744, 0.008802722208201885, -0.08662381768226624, -0.06806470453739166, 0.13252179324626923, -0.06977096945047379, -0.19745363295078278, -0.03751804679632187, 0.12691713869571686, 0.09710227698087692, 0.003912351094186306, -0.0029478538781404495, 0.056875716894865036, 0.006085332948714495, -0.10549645125865936, 0.02242925576865673, 0.017879903316497803, 0.003912637941539288, 0.058802492916584015, 0.08836042135953903, -0.14412078261375427, -0.08851473033428192, -0.011216102167963982, -0.017044605687260628, 0.30608510971069336, -0.04454493522644043, 0.05891108140349388, 0.2315647304058075, -0.05275919288396835, -0.18885394930839539, 0.011645983904600143, 0.015861550346016884, 0.03526277467608452, -0.10113989561796188, -0.09448014944791794, 0.01462764386087656, 0.07183791697025299, -0.029614388942718506, -0.02073974721133709, -0.2644074559211731, -0.16140323877334595, 0.07155248522758484, 0.060483936220407486, 0.21439853310585022, -0.0458269864320755, 0.02689659409224987, 0.03901373967528343, 0.05086878687143326, 0.08379879593849182, 0.11475405097007751, 0.05636497214436531, 0.024090619757771492, -0.020657572895288467, 0.035110149532556534, -0.05366680771112442, 0.09693602472543716, 0.03696287423372269, 0.009229999966919422, -0.033254001289606094, -0.01400453969836235, -0.025979695841670036, -0.037248414009809494, 0.11161135137081146, 0.009551459923386574, 0.05877988785505295, -0.09477690607309341, -0.07864458858966827, -0.0666072741150856, 0.05877253785729408, -0.009792137891054153, -0.08317770063877106, -0.07168682664632797, 0.08146300166845322, 0.0488695353269577, -0.008364320732653141, -0.051629941910505295, -0.05967927724123001, -0.054841142147779465, 0.0699208676815033, 0.15894049406051636, -0.08498968929052353, -0.03450244292616844, 0.02885415218770504, -0.05186077207326889, 0.05749012902379036, -0.16626684367656708, -0.045946668833494186, 0.10089036822319031, 0.011310816742479801, 0.07516943663358688, 0.014034229330718517, -0.16668252646923065, 0.005119022913277149, 0.05670760199427605, -0.13262589275836945, -0.14759545028209686, 0.03166773170232773, -0.020151961594820023, -0.05694343522191048, 0.06014106422662735, 0.09671883285045624, -0.10075151920318604, -0.008959256112575531, -0.010561985895037651, 0.07667050510644913, 0.0017134346999228, 0.1091192215681076, 0.027595899999141693, 0.010182269848883152, -0.07375553250312805, 0.15510039031505585, 0.06968935579061508, -0.1798023134469986, 0.03273286670446396, 0.14710085093975067, -0.12567606568336487, -0.026467997580766678, 0.028500445187091827, -0.005493116565048695, 0.009767541661858559, -0.006053413730114698, -0.068646639585495, -0.08446177840232849, 0.07818963378667831, 0.09299354255199432, 0.032967325299978256, -0.015752481296658516, -0.0793476402759552, 0.008755392394959927, -0.10526282340288162, 0.04278825595974922, 0.1259368658065796, 0.012197923846542835, -0.015460792928934097, 0.0864538699388504, 0.01906404085457325, -0.022213350981473923, -0.03301222249865532, 0.00034356117248535156, -0.08030682802200317, -0.0154968136921525, 0.06133369728922844, -0.03495471924543381, -0.07730121910572052, 0.016471313312649727, -0.0360553152859211, -0.05333331599831581, -0.005540343467146158, 0.01870899647474289, -0.04236839339137077, -0.007842550054192543, -0.0053185042925179005, -0.027561809867620468, -0.05814288929104805, -0.004766130354255438, 0.03889811038970947, -0.08628130704164505, 0.178212970495224, -0.012286379933357239, -0.02678641676902771, 0.09452126920223236, -0.08431747555732727, 0.090990349650383, 0.03721470385789871, 0.0038274338003247976, 0.0028489951509982347, -0.05611877143383026, 0.01853238046169281, -0.014109376817941666, 0.02459586225450039, 0.04231778159737587, 0.10296720266342163, -0.06613322347402573, 0.11179281026124954, -0.009177609346807003, -0.018171051517128944, -0.09601890295743942, 0.07786044478416443, -0.0006152652786113322, 0.0850493535399437, 0.05934595689177513, -0.06858760863542557, 0.014521765522658825, -0.12303241342306137, -0.014619279652833939, 0.02826182171702385, -0.05174340680241585, -0.04559171944856644, -0.0036679457407444715, 0.11322351545095444, -0.008440245874226093, 0.1104946881532669, 0.02270655892789364, -0.013155879452824593, 0.03725620359182358, 0.024561448022723198, 0.0031236037611961365, 0.04467571899294853, 0.03185216709971428, -0.04174024239182472, -0.0046828798949718475, -0.049553826451301575, -0.001386473304592073, -0.00031340442365035415, 0.047034502029418945, 0.18116480112075806, 0.06473365426063538, 0.0742889866232872, 0.08664427697658539, -0.0454479455947876, -0.04651385918259621, -0.14235077798366547, 0.04538944363594055, -0.08137072622776031, 0.09477076679468155, 0.012328378856182098, 0.11615276336669922, 0.1516081988811493, -0.1500978171825409, 0.044408414512872696, 0.02679440565407276, -0.10936439782381058, -0.08180243521928787, -0.17902618646621704, -0.019815364852547646, -0.02534199319779873, 0.0027534663677215576, -0.08128737658262253, -0.0008985170279629529, 0.0721656009554863, 0.03387413173913956, 0.008194572292268276, 0.0727866068482399, -0.025558069348335266, -0.006133462768048048, 0.1373906433582306, 0.008048200979828835, 0.02414938434958458, 0.10265649110078812, -0.08629734069108963, -0.04807807877659798, -0.017796291038393974, 0.060495927929878235, 0.02473585493862629, -0.059596385806798935, 0.06130366772413254, 0.01274432148784399, -0.0410507395863533, 0.01775008998811245, -0.012676114216446877, 0.036547254770994186, 0.1664261668920517, 0.06863336265087128, -0.054542578756809235, -0.0016500625060871243, 0.22080670297145844, -0.00030185465584509075, -0.07510529458522797, -0.13354940712451935, 0.1500498354434967, -0.013743002898991108, 0.03673967346549034, 0.050718553364276886, -0.08562523871660233, 0.0014097228413447738, 0.14164437353610992, 0.12958692014217377, -0.016614500433206558, -0.006028094794601202, -0.02674633450806141, -0.019631410017609596, 0.017613288015127182, 0.12533612549304962, -0.009127453900873661, 0.22829194366931915, -0.06395868957042694, 0.05709519609808922, -0.060017745941877365, -0.060041166841983795, -0.09386958181858063, 0.10072840750217438, 0.02160189114511013, -0.02783544920384884, -0.08564281463623047, 0.12098096311092377, -0.023614676669239998, -0.15995189547538757, -0.03217460960149765, -0.12081661820411682, -0.12870751321315765, -0.0165437962859869, -0.02873748354613781, -0.02570880576968193, 0.11883123219013214, 0.007501387502998114, 0.02042268216609955, -0.053202223032712936, 0.04284951463341713, -0.06990204006433487, -0.08126174658536911, 0.04896772652864456, 0.0948154404759407, 0.19075095653533936, 0.03095564804971218, 0.09717079997062683, 0.06333848834037781, 0.02926539070904255, -0.1010890081524849, 0.05723383650183678, 0.001958828419446945, 0.07219676673412323, 0.041773322969675064, 0.0718727633357048, -0.034371595829725266, -0.004037057980895042, 0.07638771831989288, -0.011987896636128426, 0.056716252118349075, -0.15766558051109314, -0.0388057604432106, -0.06856923550367355, 0.06822583079338074, -0.05377673730254173, 0.10045503824949265, 0.1340218186378479, -0.013896612450480461, 0.01700476184487343, -0.04576883092522621, 0.004195115063339472, -0.0006409371271729469, 0.03142477571964264, -0.010731512680649757, -0.13830289244651794, -0.020136887207627296, -0.029735183343291283, 0.03414566069841385, -0.09466246515512466, -0.01722305826842785, -0.0966418907046318, -0.01781177893280983, -0.05387331172823906, 0.1676945984363556, 0.02598477154970169, -0.04100999981164932, -0.032887816429138184, 0.025507282465696335, 0.039965614676475525, 0.06651737540960312, -0.11982947587966919, -0.061949241906404495 ]
null
null
transformers
## Hello World
{}
token-classification
cwtpc/wangchanberta-ner-8989
[ "transformers", "pytorch", "camembert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #camembert #token-classification #autotrain_compatible #endpoints_compatible #region-us
## Hello World
[ "## Hello World" ]
[ "TAGS\n#transformers #pytorch #camembert #token-classification #autotrain_compatible #endpoints_compatible #region-us \n", "## Hello World" ]
[ 39, 3 ]
[ "passage: TAGS\n#transformers #pytorch #camembert #token-classification #autotrain_compatible #endpoints_compatible #region-us \n## Hello World" ]
[ -0.020713642239570618, 0.10025706142187119, -0.008806237950921059, 0.03443806990981102, 0.1759691834449768, 0.013023915700614452, 0.051259372383356094, 0.08922743797302246, 0.06123132258653641, -0.004442824982106686, 0.1286446899175644, 0.22500000894069672, -0.02998882532119751, 0.11595501750707626, -0.05177784338593483, -0.3157079517841339, 0.0666830763220787, 0.054703060537576675, -0.07739651948213577, 0.09963671863079071, 0.08838541805744171, -0.09702940285205841, 0.09185191988945007, -0.03266112133860588, -0.11767882108688354, 0.05003638193011284, 0.010805368423461914, -0.09966360032558441, 0.10540498048067093, 0.02655402198433876, 0.15496188402175903, 0.039344191551208496, -0.04194609075784683, -0.13709315657615662, 0.03231203183531761, -0.010647383518517017, -0.043654270470142365, 0.04409871995449066, 0.04905543103814125, -0.09302467107772827, 0.004680538084357977, 0.046962328255176544, 0.022645652294158936, 0.06551717966794968, -0.14305689930915833, -0.1424890011548996, -0.03294418007135391, 0.020055016502738, 0.028606481850147247, 0.049244627356529236, 0.02853645198047161, 0.22696399688720703, -0.11606154590845108, 0.10424420982599258, 0.10246043652296066, -0.26055219769477844, -0.005851516034454107, 0.10068424046039581, 0.017376504838466644, 0.003740768413990736, -0.031285203993320465, 0.03467675298452377, 0.032232724130153656, 0.007534687872976065, 0.005132865626364946, -0.06762788444757462, -0.0619046725332737, 0.00778287835419178, -0.09605979919433594, -0.01625429093837738, 0.1957307904958725, -0.03221413493156433, 0.03792569786310196, 0.02621643990278244, -0.09021186083555222, -0.09153009206056595, -0.02880677953362465, -0.02577345445752144, -0.015620185993611813, 0.0492507703602314, 0.028436673805117607, 0.009643921628594398, -0.10939407348632812, 0.04698970913887024, -0.21852096915245056, 0.2455911636352539, 0.01844179816544056, 0.04956270754337311, -0.18072767555713654, 0.03295471891760826, -0.004912005737423897, -0.07361463457345963, 0.017371149733662605, -0.10063320398330688, 0.023588255047798157, -0.043418917804956436, 0.003337407950311899, 0.03444160148501396, 0.09316720068454742, 0.14404872059822083, 0.04844031110405922, 0.07339771836996078, 0.00003969831232097931, 0.07872821390628815, 0.06872697174549103, 0.06915637105703354, 0.017412113025784492, -0.09729529172182083, 0.04192199558019638, -0.1475607454776764, -0.017041297629475594, -0.05506838485598564, -0.11554975062608719, -0.0159319955855608, 0.049245428293943405, 0.08820835500955582, 0.04297875985503197, 0.08941751718521118, -0.03105776198208332, -0.029473816975951195, 0.12078247964382172, -0.06752034276723862, 0.022869793698191643, 0.014581876806914806, -0.020684242248535156, 0.10731153935194016, -0.05739589408040047, -0.004364282824099064, -0.04104932025074959, 0.09061844646930695, -0.07272090762853622, -0.012453820556402206, -0.037840619683265686, -0.032529961317777634, 0.04809851199388504, -0.13318097591400146, 0.04288773611187935, -0.1690087914466858, -0.04472148045897484, 0.024060124531388283, 0.024774761870503426, 0.016080940142273903, -0.03183286264538765, -0.012304326519370079, 0.008822249248623848, 0.009222088381648064, -0.04431065171957016, -0.06483852863311768, -0.07180867344141006, 0.08584655076265335, -0.0018455618992447853, 0.07158781588077545, -0.1629096418619156, 0.06694493442773819, -0.11375857144594193, 0.021222036331892014, -0.11069060117006302, -0.01607503369450569, -0.08891256153583527, 0.15100371837615967, -0.0035545174032449722, -0.04584035649895668, -0.033399004489183426, 0.034961897879838943, -0.04265034571290016, 0.1513187438249588, -0.073786661028862, -0.10726507753133774, 0.17530547082424164, -0.11254021525382996, -0.14282234013080597, 0.07108346372842789, 0.005023092031478882, -0.002250480931252241, 0.04734962806105614, 0.10155566036701202, 0.05017751827836037, -0.0009348875028081238, 0.041054435074329376, 0.07190616428852081, -0.14960771799087524, -0.11221160739660263, -0.006965389009565115, 0.001447196933440864, -0.12095314264297485, 0.065846748650074, 0.08743921667337418, 0.07194831222295761, -0.08080489933490753, -0.02914266474545002, -0.014715813100337982, -0.010685648769140244, 0.10104677081108093, 0.06878065317869186, 0.0948602706193924, -0.0829242467880249, 0.029559001326560974, -0.004609512630850077, 0.03443969041109085, 0.045053571462631226, 0.005128407385200262, -0.0933147519826889, 0.14765727519989014, 0.008999976329505444, 0.004132147878408432, -0.1566007137298584, -0.06273510307073593, 0.010939753614366055, 0.10357926040887833, 0.012308657169342041, 0.13809436559677124, 0.08407579362392426, -0.041781485080718994, 0.015817072242498398, -0.04216063395142555, 0.1812426745891571, 0.01259149145334959, -0.03554088994860649, -0.08627676218748093, 0.07397095859050751, -0.060726314783096313, 0.006247902754694223, -0.05278143659234047, 0.008820130489766598, 0.11348483711481094, 0.15358762443065643, 0.021858956664800644, 0.06042046099901199, -0.03262457251548767, 0.051080409437417984, -0.09330898523330688, 0.008726900443434715, 0.1054287925362587, -0.015805194154381752, -0.07656118273735046, 0.13032017648220062, -0.14830201864242554, 0.2823004126548767, 0.2150687277317047, -0.3542865216732025, -0.02235523797571659, -0.004951581358909607, -0.005400826223194599, 0.0117062171921134, 0.056819625198841095, 0.06605638563632965, 0.0244549922645092, -0.005566845647990704, 0.1680983304977417, 0.005398008041083813, -0.013982637785375118, 0.015258053317666054, -0.07670781761407852, -0.05635349079966545, 0.05855857953429222, 0.049835868179798126, -0.1815086305141449, 0.18188084661960602, 0.22186391055583954, 0.008949787355959415, 0.1249474361538887, 0.0002269129327032715, 0.03897581994533539, 0.06337998062372208, -0.0118575319647789, -0.018605271354317665, -0.014490635134279728, -0.219120591878891, -0.040474649518728256, 0.05525413900613785, 0.019941408187150955, 0.032248493283987045, -0.08600815385580063, -0.03597691282629967, 0.0068766712211072445, 0.05025280639529228, 0.03763992711901665, 0.09085102379322052, 0.04612063243985176, 0.09765227138996124, 0.020610833540558815, -0.09944778680801392, 0.08449171483516693, 0.009522094391286373, -0.04823427274823189, 0.17175476253032684, -0.11070796847343445, -0.3022056519985199, -0.08747333288192749, -0.2122466266155243, -0.03772692009806633, 0.04600506275892258, 0.06692832708358765, -0.08840576559305191, -0.04835090786218643, 0.06905083358287811, -0.005860600620508194, -0.07049162685871124, 0.05978662893176079, -0.0652894377708435, 0.07188611477613449, -0.02904854714870453, -0.03839289769530296, -0.059211667627096176, -0.031544603407382965, -0.038829948753118515, 0.1504257619380951, -0.07631048560142517, 0.09016036987304688, 0.15911734104156494, -0.018633155152201653, 0.05312039703130722, -0.013438583351671696, 0.1879112422466278, -0.08537967503070831, -0.026246367022395134, 0.13793042302131653, -0.06505429744720459, 0.08559147268533707, 0.1441076099872589, 0.05202966555953026, -0.0733344629406929, 0.023693496361374855, -0.017835363745689392, -0.11205238848924637, -0.13722074031829834, -0.1297808289527893, -0.0921013131737709, 0.07737217098474503, 0.03852121904492378, 0.07958061993122101, 0.1442013680934906, 0.0929555669426918, 0.04273306950926781, -0.0337039940059185, -0.050543297082185745, 0.08779928088188171, 0.15276868641376495, -0.023936448618769646, 0.13166190683841705, -0.05969808250665665, -0.15984736382961273, 0.07792395353317261, 0.07268506288528442, 0.06611090153455734, 0.12471627444028854, -0.01089635118842125, 0.022145921364426613, 0.1313825249671936, 0.15682563185691833, 0.10584491491317749, 0.030644699931144714, -0.023952431976795197, -0.0037815854884684086, -0.0023720005992799997, -0.08931370079517365, 0.012191473506391048, 0.15358787775039673, -0.10905009508132935, -0.06942235678434372, -0.12985356152057648, 0.05432172492146492, 0.11613300442695618, 0.0591675266623497, -0.14988401532173157, 0.020110178738832474, 0.07833822071552277, -0.009368429891765118, -0.06330853700637817, 0.07762768119573593, -0.07587604224681854, -0.12451773881912231, 0.08275100588798523, -0.01360382977873087, 0.10531105846166611, -0.08659174293279648, 0.05745347589254379, -0.04997837170958519, -0.10137386620044708, 0.037645161151885986, 0.07064469158649445, -0.24529077112674713, 0.2502822279930115, -0.0028862268663942814, -0.08947984874248505, -0.08015529811382294, -0.016405945643782616, 0.03678901121020317, 0.21500301361083984, 0.10434021800756454, 0.010352516546845436, -0.11373793333768845, -0.19400548934936523, 0.009370436891913414, -0.02567845582962036, 0.11108183860778809, -0.009282871149480343, -0.008996783755719662, -0.03829415887594223, -0.0260236207395792, -0.05468940734863281, -0.026748383417725563, 0.03069904074072838, -0.1169932633638382, 0.06183454394340515, 0.03738247975707054, 0.06252629309892654, 0.015991536900401115, -0.02790035866200924, -0.13001935184001923, 0.1628253310918808, -0.07771766185760498, -0.022129608318209648, -0.10633914172649384, -0.11818438023328781, 0.03016403503715992, -0.09198429435491562, 0.06251797825098038, -0.10365168750286102, -0.01877828873693943, -0.029879380017518997, -0.14972320199012756, 0.1461184024810791, -0.07163476198911667, -0.05796762555837631, -0.06619498878717422, 0.12738247215747833, -0.050333499908447266, 0.0000485123491671402, -0.0012934646802023053, 0.01606133207678795, -0.06243335083127022, -0.07223925739526749, 0.027741001918911934, -0.0366252101957798, 0.027264641597867012, 0.05685688182711601, -0.04516279324889183, -0.047843363136053085, -0.02252037078142166, 0.02736760303378105, 0.24212627112865448, 0.21672853827476501, -0.06996624171733856, 0.10443873703479767, 0.16671624779701233, -0.03726040944457054, -0.3421591818332672, -0.06910421699285507, -0.13674597442150116, -0.061615727841854095, -0.06236092001199722, -0.13815806806087494, 0.11003949493169785, 0.019402239471673965, -0.058313872665166855, 0.0995226502418518, -0.12131083756685257, -0.059225600212812424, 0.17481385171413422, 0.018278468400239944, 0.3587537407875061, -0.09558431059122086, -0.05680680647492409, -0.018896881490945816, -0.14855614304542542, 0.1158064752817154, 0.06817933917045593, 0.09693427383899689, -0.030711622908711433, 0.023491734638810158, 0.045115403831005096, -0.05320701748132706, 0.08714737743139267, -0.010712971910834312, 0.037921879440546036, -0.12850987911224365, -0.09939440339803696, -0.0284144077450037, -0.010073098354041576, 0.0007203428540378809, 0.034742049872875214, 0.014208880253136158, -0.10686993598937988, -0.014508910477161407, -0.0921170637011528, 0.12706205248832703, 0.029085543006658554, -0.07623688131570816, -0.008392123505473137, -0.023684682324528694, -0.022706402465701103, 0.023025358095765114, 0.2847757339477539, -0.04533744230866432, 0.14722940325737, 0.113134004175663, 0.10718613117933273, -0.1832704097032547, 0.01908540539443493, -0.06889280676841736, -0.053723402321338654, 0.06515437364578247, -0.06213353946805, 0.07258941978216171, 0.12684616446495056, -0.04594774916768074, 0.025150850415229797, 0.10704398900270462, 0.018880033865571022, -0.02282615937292576, 0.15382783114910126, -0.21087245643138885, -0.05144799128174782, -0.01887974701821804, 0.03558744862675667, 0.09622368961572647, 0.11664950102567673, 0.1020505502820015, 0.04160699248313904, -0.03918534889817238, 0.0061564575880765915, -0.021960211917757988, -0.035340264439582825, 0.04256027191877365, 0.06650233268737793, 0.029446953907608986, -0.13197557628154755, 0.049299269914627075, 0.0318022258579731, -0.20102903246879578, -0.04306856915354729, 0.16430051624774933, -0.11977240443229675, -0.13310669362545013, 0.005645920522511005, 0.11983425915241241, -0.14886312186717987, -0.037850309163331985, -0.012266790494322777, -0.11563581973314285, 0.07378631085157394, 0.1960964798927307, 0.12024813890457153, 0.0926031842827797, 0.0034893148113042116, -0.016391199082136154, -0.0018585233483463526, -0.038484424352645874, -0.0383073128759861, 0.046073075383901596, -0.1370113044977188, -0.008825665339827538, -0.004770656116306782, 0.16718018054962158, -0.10605547577142715, -0.0929407998919487, -0.14693240821361542, 0.018168341368436813, -0.0625934898853302, -0.10274574160575867, -0.11409033089876175, -0.03313785418868065, 0.020585304126143456, -0.10201369971036911, -0.010639951564371586, -0.0626838281750679, -0.11526976525783539, 0.0367356613278389, 0.02645130269229412, 0.020174484699964523, -0.04644716903567314, -0.026963436976075172, 0.11505013704299927, -0.025109151378273964, 0.12781652808189392, 0.10290892422199249, -0.06445075571537018, 0.07864794880151749, -0.08940636366605759, -0.0846819058060646, 0.10623971372842789, 0.013580354861915112, 0.11034788936376572, 0.07662410289049149, 0.01721748523414135, 0.03700981289148331, 0.03688126057386398, 0.05881321057677269, 0.08363072574138641, -0.11449918895959854, 0.03720887750387192, -0.0013651682529598475, -0.17211057245731354, -0.010408339090645313, -0.06377691775560379, 0.12145452201366425, -0.008172819390892982, 0.12612970173358917, -0.025000743567943573, 0.10061054676771164, -0.019242456182837486, 0.01686469465494156, -0.01795363985002041, -0.21729062497615814, -0.031327612698078156, -0.05904213339090347, 0.009234683588147163, -0.02017296478152275, 0.19593703746795654, -0.010924240574240685, 0.03002255968749523, 0.057841986417770386, 0.04793616011738777, -0.0015675697941333055, 0.04097912460565567, 0.15516862273216248, 0.08325263857841492, -0.061982154846191406, -0.05456039682030678, 0.08805672079324722, 0.0035377680324018, -0.043313585221767426, 0.09956426918506622, 0.05261862277984619, -0.022016452625393867, 0.04049248620867729, 0.01638689637184143, 0.03358166292309761, -0.13766269385814667, -0.2156456708908081, -0.0430288091301918, 0.043902039527893066, 0.004637292120605707, 0.06893377751111984, 0.0908796489238739, 0.009886275045573711, 0.03651933744549751, -0.05278649553656578, -0.016945116221904755, -0.16621345281600952, -0.06385431438684464, -0.09997095912694931, -0.09246046841144562, 0.017705345526337624, -0.039673224091529846, -0.024821901693940163, 0.03401292487978935, 0.06557337939739227, -0.062298733741045, 0.06484554708003998, -0.006162852514535189, 0.0013467546086758375, 0.01756233163177967, 0.0050670490600168705, -0.010809245519340038, -0.0038332566618919373, -0.03155558556318283, -0.19792859256267548, -0.0020624431781470776, -0.06680745631456375, 0.017848435789346695, -0.09242182970046997, 0.007360225077718496, -0.11882893741130829, -0.11388389766216278, -0.020694613456726074, 0.032417166978120804, -0.05591730400919914, 0.08648976683616638, -0.03550132364034653, 0.038113903254270554, 0.011762780137360096, 0.17190726101398468, -0.06932131946086884, -0.06117447465658188, -0.031157679855823517, 0.23836812376976013, 0.03568531572818756, 0.13365277647972107, -0.03453025221824646, 0.01994147337973118, -0.08931102603673935, 0.2965608835220337, 0.2826343774795532, -0.031191963702440262, 0.05637151747941971, 0.023857155814766884, 0.021968532353639603, 0.0711701288819313, 0.09498167037963867, 0.0929773822426796, 0.20383574068546295, -0.08021513372659683, -0.02829308807849884, -0.05191989988088608, 0.018051808699965477, -0.11114205420017242, 0.08119910955429077, 0.05078539252281189, -0.027734387665987015, -0.0779368132352829, 0.04326765984296799, -0.2169463187456131, 0.1460457295179367, 0.021897103637456894, -0.21945783495903015, -0.07023286074399948, -0.021584227681159973, 0.11698152869939804, 0.020614709705114365, 0.050038762390613556, 0.006201825104653835, -0.07722686976194382, 0.03224457800388336, 0.028164273127913475, -0.22368861734867096, -0.053519509732723236, 0.05870307609438896, -0.043208759278059006, 0.07573892921209335, -0.017659390345215797, -0.01913662627339363, 0.09053095430135727, 0.06246627867221832, -0.039296478033065796, 0.01776166632771492, 0.020526211708784103, -0.037481632083654404, -0.04506540298461914, 0.03539389744400978, 0.028919387608766556, -0.07367279380559921, 0.04664115235209465, -0.16296660900115967, 0.042235199362039566, -0.07359538972377777, -0.017969438806176186, -0.01596948318183422, 0.06203071400523186, -0.03910883143544197, 0.04858998581767082, 0.06391781568527222, 0.019815683364868164, -0.04402141645550728, -0.046524062752723694, -0.0350087471306324, 0.0188580509275198, -0.11477411538362503, -0.14824460446834564, -0.12408174574375153, -0.059207335114479065, 0.06891012191772461, -0.005253707990050316, -0.06321454793214798, -0.034614671021699905, -0.08397065848112106, 0.04314541816711426, -0.11227713525295258, 0.07630158960819244, 0.056686945259571075, 0.039795875549316406, -0.013803495094180107, -0.03960544243454933, 0.03436680883169174, 0.07656688988208771, -0.12503498792648315, -0.08998812735080719 ]
null
null
transformers
## Cyclone Chinese NER This model provides simplified Chinese NER model based on pretrained model BERT (specifically BERT + CRF) Currently, we only support 8 general type of entities ("address", "company", "government", "name", "organization", "position", "scene", "time") ### Usage from transformers import BertConfig config = BertConfig.from_pretrained("bert-base-chinese", num_labels=num_labels) model_path = "cyclone/cyclone-ner" tokenizer = CNerTokenizer.from_pretrained(model_path, do_lower_case=True) model = BertCrfForNer.from_pretrained(model_path, config=config)
{}
null
cyclone/cyclone-ner
[ "transformers", "pytorch", "bert", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #bert #endpoints_compatible #region-us
## Cyclone Chinese NER This model provides simplified Chinese NER model based on pretrained model BERT (specifically BERT + CRF) Currently, we only support 8 general type of entities ("address", "company", "government", "name", "organization", "position", "scene", "time") ### Usage from transformers import BertConfig config = BertConfig.from_pretrained("bert-base-chinese", num_labels=num_labels) model_path = "cyclone/cyclone-ner" tokenizer = CNerTokenizer.from_pretrained(model_path, do_lower_case=True) model = BertCrfForNer.from_pretrained(model_path, config=config)
[ "## Cyclone Chinese NER\r\n\r\nThis model provides simplified Chinese NER model based on pretrained model BERT (specifically BERT + CRF)\r\nCurrently, we only support 8 general type of entities (\"address\", \"company\", \"government\", \"name\", \"organization\", \"position\", \"scene\", \"time\")", "### Usage\r\n from transformers import BertConfig\r\n\r\n config = BertConfig.from_pretrained(\"bert-base-chinese\", num_labels=num_labels)\r\n\r\n model_path = \"cyclone/cyclone-ner\"\r\n\r\n tokenizer = CNerTokenizer.from_pretrained(model_path, do_lower_case=True)\r\n model = BertCrfForNer.from_pretrained(model_path, config=config)" ]
[ "TAGS\n#transformers #pytorch #bert #endpoints_compatible #region-us \n", "## Cyclone Chinese NER\r\n\r\nThis model provides simplified Chinese NER model based on pretrained model BERT (specifically BERT + CRF)\r\nCurrently, we only support 8 general type of entities (\"address\", \"company\", \"government\", \"name\", \"organization\", \"position\", \"scene\", \"time\")", "### Usage\r\n from transformers import BertConfig\r\n\r\n config = BertConfig.from_pretrained(\"bert-base-chinese\", num_labels=num_labels)\r\n\r\n model_path = \"cyclone/cyclone-ner\"\r\n\r\n tokenizer = CNerTokenizer.from_pretrained(model_path, do_lower_case=True)\r\n model = BertCrfForNer.from_pretrained(model_path, config=config)" ]
[ 23, 74, 110 ]
[ "passage: TAGS\n#transformers #pytorch #bert #endpoints_compatible #region-us \n## Cyclone Chinese NER\r\n\r\nThis model provides simplified Chinese NER model based on pretrained model BERT (specifically BERT + CRF)\r\nCurrently, we only support 8 general type of entities (\"address\", \"company\", \"government\", \"name\", \"organization\", \"position\", \"scene\", \"time\")### Usage\r\n from transformers import BertConfig\r\n\r\n config = BertConfig.from_pretrained(\"bert-base-chinese\", num_labels=num_labels)\r\n\r\n model_path = \"cyclone/cyclone-ner\"\r\n\r\n tokenizer = CNerTokenizer.from_pretrained(model_path, do_lower_case=True)\r\n model = BertCrfForNer.from_pretrained(model_path, config=config)" ]
[ -0.00883391685783863, 0.16348592936992645, -0.0021209942642599344, 0.07804320752620697, 0.1171920970082283, -0.04894353821873665, 0.06639860570430756, -0.09048741310834885, -0.04604075849056244, 0.11351130902767181, 0.05356454849243164, -0.12329444289207458, -0.012968155555427074, -0.06474047154188156, 0.0033601305913180113, -0.3353304862976074, 0.1149047389626503, 0.07340380549430847, 0.09062208235263824, 0.1066189631819725, 0.027814073488116264, -0.07614495605230331, 0.10974299162626266, 0.1388358622789383, -0.10443173348903656, 0.03679715469479561, -0.11491724103689194, -0.06516018509864807, 0.03823067247867584, -0.05092986300587654, 0.20471946895122528, -0.003203411353752017, 0.033864833414554596, -0.0622108094394207, 0.009578125551342964, -0.015694817528128624, -0.035839494317770004, 0.056278567761182785, 0.11570657044649124, 0.06222368776798248, 0.08738869428634644, 0.027467031031847, 0.03628715127706528, 0.044785402715206146, -0.08763279020786285, -0.027777371928095818, -0.0532035231590271, 0.22580306231975555, 0.10338688641786575, 0.11352836340665817, 0.029132230207324028, 0.07875959575176239, -0.16616208851337433, 0.016675742343068123, 0.0816519483923912, -0.22252541780471802, -0.05822927504777908, 0.15204337239265442, 0.015568469651043415, -0.07558964192867279, -0.0205996073782444, -0.035726457834243774, 0.03595423325896263, 0.013455281965434551, -0.01310708373785019, -0.05046100541949272, -0.13950709998607635, 0.016185307875275612, -0.15553800761699677, 0.07763293385505676, 0.1135963499546051, 0.08694218844175339, -0.03716200217604637, 0.052429672330617905, -0.03178425878286362, -0.06951792538166046, -0.05738508701324463, -0.09031131118535995, -0.030628742650151253, -0.036129746586084366, 0.13929763436317444, -0.021880285814404488, -0.014240792021155357, -0.07720693200826645, -0.14127053320407867, 0.18778881430625916, 0.02091752178966999, -0.004237720742821693, -0.1613209992647171, 0.021601341664791107, -0.13609236478805542, -0.10934461653232574, -0.016704287379980087, -0.0537119135260582, -0.11225147545337677, 0.08239202201366425, -0.02741137146949768, 0.0503174290060997, 0.0029329785611480474, 0.14720657467842102, -0.04524151608347893, 0.035695288330316544, 0.0461336225271225, -0.010777432471513748, -0.02755799889564514, 0.07401483505964279, -0.12579509615898132, -0.019331924617290497, 0.011030266061425209, -0.11613353341817856, 0.01577788218855858, -0.03908366709947586, -0.13948652148246765, -0.07092422991991043, 0.03738290071487427, -0.0308510884642601, -0.05904029309749603, 0.09481066465377808, -0.020644094794988632, -0.09894420951604843, 0.3524816930294037, -0.00818937923759222, 0.07965563237667084, 0.09280315041542053, -0.05663742870092392, 0.14428569376468658, -0.04840005189180374, 0.01698233373463154, 0.033631402999162674, 0.037167906761169434, -0.10928191989660263, -0.07108522206544876, -0.05777784809470177, -0.09962747991085052, 0.06790393590927124, -0.1795709729194641, 0.002895439276471734, -0.15513037145137787, -0.16947340965270996, 0.08206579834222794, 0.013418465852737427, -0.07754019647836685, -0.10086633265018463, -0.025435298681259155, -0.0512932650744915, 0.0005449238233268261, -0.040090810507535934, -0.18312327563762665, -0.01444131601601839, -0.1275091916322708, -0.03905076906085014, 0.02420087531208992, -0.13081307709217072, 0.010841581970453262, -0.08009142428636551, 0.09666513651609421, -0.12587134540081024, -0.021499034017324448, -0.10067395120859146, 0.06693798303604126, -0.14473176002502441, -0.0784997045993805, -0.026866411790251732, -0.07552085816860199, 0.05872194096446037, 0.14475587010383606, -0.08986339718103409, -0.028122445568442345, -0.025684639811515808, -0.1304430216550827, -0.004223653580993414, 0.04902242496609688, -0.023589691147208214, -0.06226129084825516, -0.016673773527145386, 0.09648003429174423, 0.14027345180511475, 0.006683229468762875, -0.013802730478346348, -0.001559535856358707, -0.021564947441220284, -0.12935499846935272, 0.024583298712968826, 0.07398906350135803, -0.21222490072250366, 0.06620828807353973, -0.03153783082962036, 0.032122135162353516, -0.03512285277247429, -0.012800252996385098, -0.008842184208333492, -0.015660902485251427, 0.04122309386730194, -0.05087422579526901, 0.08729606866836548, 0.018864529207348824, 0.009317140094935894, 0.11817973107099533, 0.045763663947582245, -0.013406718149781227, 0.042073484510183334, -0.045026469975709915, 0.11554250866174698, -0.0872509554028511, -0.010915880091488361, -0.104170061647892, -0.19831274449825287, 0.10849201679229736, -0.14496302604675293, 0.05372917652130127, 0.005071135237812996, 0.03516253083944321, -0.0005857457872480154, 0.024219553917646408, 0.013668647967278957, 0.035281695425510406, 0.06959434598684311, -0.013203050009906292, -0.0725175067782402, -0.07877416163682938, -0.0671524852514267, 0.025077274069190025, -0.20447877049446106, 0.0570642463862896, 0.13961753249168396, 0.052317604422569275, -0.05681491643190384, 0.02646043337881565, 0.08817833662033081, 0.032802119851112366, -0.03985697403550148, 0.05021775886416435, 0.047316938638687134, 0.04955228045582771, -0.09757793694734573, 0.145896315574646, -0.015902426093816757, -0.12079472094774246, 0.017947230488061905, 0.08768980950117111, -0.11219435185194016, 0.03714539483189583, -0.03063557855784893, -0.060677967965602875, 0.0760369673371315, 0.146461620926857, 0.23030562698841095, 0.035912103950977325, 0.10729829967021942, -0.01496745366603136, 0.018314702436327934, 0.09884124249219894, -0.05074106901884079, -0.04462447389960289, 0.10905405879020691, 0.15365666151046753, -0.02792956493794918, 0.05522140488028526, -0.008813739754259586, -0.14037902653217316, 0.03003675304353237, 0.09379846602678299, -0.02108561061322689, -0.01835591159760952, -0.08007372170686722, 0.057631053030490875, 0.043816372752189636, -0.08592838048934937, 0.008006539195775986, 0.05212601646780968, 0.03591401129961014, 0.02033975161612034, -0.10628042370080948, -0.00946846790611744, 0.06338655203580856, 0.041892196983098984, 0.02940540947020054, 0.055978477001190186, -0.05315738543868065, 0.02483529783785343, -0.014353875070810318, -0.2762748599052429, 0.0008832390885800123, 0.04209081083536148, -0.04311278834939003, 0.1814439743757248, -0.01940566673874855, -0.22433029115200043, -0.18090112507343292, -0.18848609924316406, -0.22870860993862152, -0.06268364191055298, 0.006729637272655964, -0.0836026594042778, -0.0919453576207161, -0.0862848311662674, -0.143731027841568, -0.10098537057638168, -0.008473072201013565, -0.02595525048673153, -0.11237248033285141, -0.02116020768880844, -0.06280360370874405, 0.07123667001724243, -0.028781631961464882, -0.07482564449310303, 0.04038241133093834, -0.09099069237709045, 0.14474204182624817, 0.19324196875095367, -0.048969924449920654, 0.006879524327814579, 0.09563831239938736, 0.16417580842971802, 0.014383723959326744, 0.08638133853673935, 0.02734452486038208, -0.04255013167858124, 0.05716660991311073, 0.1322832852602005, 0.09902167320251465, -0.006003084592521191, -0.029485588893294334, 0.03281179815530777, -0.062398750334978104, -0.14226235449314117, -0.11882056295871735, -0.07551755756139755, -0.03670107573270798, 0.03583512827754021, 0.0932830274105072, 0.1331232488155365, 0.09244266152381897, 0.1047682985663414, -0.019711695611476898, -0.1172478049993515, 0.14421723783016205, 0.05024704337120056, 0.05729905888438225, 0.0552230067551136, -0.017807234078645706, -0.0319703072309494, 0.0324861966073513, 0.06362643837928772, 0.1394285261631012, 0.11014983057975769, 0.12192095071077347, 0.03544853627681732, 0.15413548052310944, 0.2226042002439499, 0.09875956177711487, -0.1464317888021469, 0.024421794340014458, -0.0063840229995548725, -0.01476768497377634, 0.1654629111289978, -0.020791955292224884, -0.10788964480161667, 0.017385175451636314, 0.006087517365813255, -0.10587777942419052, -0.04945632070302963, 0.0817384347319603, 0.07700443267822266, -0.256815105676651, -0.04650413990020752, -0.07388637214899063, 0.031279392540454865, -0.023973243311047554, 0.046334199607372284, 0.011212090030312538, -0.01227999571710825, 0.10524161905050278, 0.00750342383980751, 0.10297038406133652, 0.08345409482717514, -0.00008558992703910917, -0.0029944165144115686, 0.07999174296855927, 0.0010234039509668946, 0.055272333323955536, 0.06614978611469269, 0.18908336758613586, -0.008780206553637981, -0.09975089132785797, 0.019223535433411598, -0.06681974232196808, -0.05813560634851456, 0.2585012912750244, 0.11277859658002853, 0.016729775816202164, 0.031627267599105835, -0.02018752135336399, -0.13557876646518707, 0.04000290110707283, 0.02095722407102585, -0.025769775733351707, 0.05030403658747673, 0.10163144022226334, -0.060210082679986954, 0.02828136645257473, 0.08102765679359436, -0.06625525653362274, -0.10510093718767166, 0.002030544448643923, 0.09787876158952713, -0.002875476609915495, -0.004174551460891962, -0.033325739204883575, -0.1129191592335701, 0.14514873921871185, -0.010941362008452415, 0.06696612387895584, -0.09252320975065231, -0.18566562235355377, 0.15285395085811615, -0.0992901474237442, 0.08628367632627487, -0.10946537554264069, 0.06146196648478508, 0.034815698862075806, -0.08175361901521683, 0.052268270403146744, -0.033818915486335754, -0.009148882701992989, -0.015993136912584305, -0.044673021882772446, 0.08585918694734573, 0.01712396740913391, 0.030427319929003716, 0.028792139142751694, -0.04424772784113884, -0.14527687430381775, -0.17082878947257996, 0.01737898588180542, 0.044866226613521576, 0.04400871694087982, -0.04125792533159256, -0.04550414904952049, -0.0025473693385720253, 0.0590326189994812, 0.13606108725070953, 0.033490944653749466, -0.12638302147388458, 0.06999962776899338, 0.32979679107666016, 0.020566513761878014, -0.19249576330184937, -0.013265850022435188, 0.06295210868120193, -0.00029567867750301957, -0.14664238691329956, -0.1150754764676094, 0.17842867970466614, 0.1448170393705368, -0.033421147614717484, -0.036965321749448776, 0.01656465046107769, -0.044453151524066925, 0.13732682168483734, -0.010792823508381844, 0.15490415692329407, -0.05035579949617386, -0.04308201000094414, 0.06829240173101425, -0.1313418596982956, 0.19983471930027008, 0.03826028108596802, 0.0472685731947422, -0.042134106159210205, -0.004968202672898769, 0.022379552945494652, -0.060708560049533844, 0.13440974056720734, 0.1490289270877838, 0.01558211725205183, 0.021877694875001907, -0.10088210552930832, 0.08941220492124557, -0.07426739484071732, 0.11557139456272125, 0.07391120493412018, 0.0741826742887497, -0.1134292483329773, -0.07504597306251526, -0.033457159996032715, 0.05484107881784439, 0.02868897095322609, 0.003151918528601527, -0.020508727058768272, 0.05456109717488289, 0.04623107984662056, -0.0020905910059809685, 0.244338721036911, -0.02203908935189247, -0.021979084238409996, 0.22520150244235992, 0.007795632816851139, -0.02610641159117222, 0.015996215865015984, -0.05972186475992203, -0.0419321246445179, 0.15492579340934753, -0.05560680106282234, -0.013139207847416401, 0.0742802619934082, 0.042630236595869064, 0.029709452763199806, 0.01794574409723282, -0.019390858709812164, -0.026086661964654922, 0.04519408196210861, -0.15290957689285278, -0.0881403386592865, -0.06757918745279312, -0.05875294655561447, -0.03351186588406563, 0.15315119922161102, 0.020846083760261536, -0.1672956794500351, 0.0202204417437315, 0.041985489428043365, -0.027449283748865128, -0.02772814966738224, -0.026643233373761177, 0.005643839947879314, 0.04087342694401741, -0.09079495072364807, 0.029906947165727615, 0.1376098245382309, -0.010766222141683102, -0.026863059028983116, 0.0701095461845398, -0.1516682356595993, -0.03261586278676987, -0.07163142412900925, 0.17141488194465637, -0.15247975289821625, -0.06590243428945541, -0.002415867755189538, -0.06785690039396286, 0.03430604189634323, 0.21858952939510345, 0.055371325463056564, -0.04585885629057884, -0.10874046385288239, -0.0668044164776802, -0.1140807643532753, -0.03056999109685421, 0.048200253397226334, 0.11315734684467316, -0.1264745444059372, -0.05933491885662079, 0.09495503455400467, 0.0679112896323204, -0.04208357632160187, -0.08787265419960022, -0.1508154571056366, -0.027709055691957474, -0.17947298288345337, 0.054583802819252014, -0.05480792000889778, 0.04542410373687744, -0.042057156562805176, -0.031077438965439796, -0.04601123929023743, -0.011377681978046894, -0.052509699016809464, 0.043884243816137314, -0.01352686807513237, 0.09320466965436935, 0.005718536209315062, 0.040034063160419464, 0.03064812161028385, -0.028029127046465874, 0.022371001541614532, 0.061833057552576065, -0.03823760524392128, 0.04668314382433891, -0.1580805778503418, -0.01769396662712097, 0.011000622063875198, 0.08091079443693161, 0.10871273279190063, 0.02363027073442936, -0.00008447138679912314, 0.0038385901134461164, -0.03161042928695679, -0.09283122420310974, -0.05585474148392677, -0.08645689487457275, -0.08528915792703629, -0.060141921043395996, -0.11002420634031296, -0.07767754793167114, -0.04325411841273308, 0.049629777669906616, 0.134361132979393, 0.07890800386667252, 0.024976028129458427, 0.07208441942930222, 0.038760531693696976, -0.045194290578365326, -0.013643646612763405, -0.08465485274791718, -0.06250950694084167, -0.062591552734375, 0.014725103043019772, -0.014986323192715645, 0.150737002491951, -0.07082123309373856, -0.04721999540925026, -0.06720330566167831, 0.01608899049460888, -0.0044603548012673855, -0.02260289341211319, 0.19038543105125427, 0.12319029122591019, 0.01927514374256134, 0.038534827530384064, 0.06951585412025452, -0.029075900092720985, 0.09037573635578156, 0.05678071454167366, 0.043425388634204865, -0.2070557177066803, 0.06778737157583237, 0.12371062487363815, -0.0273590087890625, 0.05523725971579552, -0.005736070219427347, -0.17287957668304443, 0.07884339988231659, 0.028389638289809227, 0.1412317007780075, 0.15556001663208008, -0.0047406964004039764, 0.05472429841756821, 0.12730394303798676, -0.08700279146432877, -0.11734253168106079, -0.29605230689048767, -0.08300289511680603, -0.09476065635681152, 0.011762529611587524, -0.05492977797985077, -0.025360632687807083, 0.07518624514341354, 0.06170244142413139, 0.056990575045347214, 0.09026072919368744, 0.01870608516037464, 0.09164175391197205, 0.020346952602267265, 0.0038824304938316345, -0.035737767815589905, 0.0569143071770668, 0.04389292374253273, -0.020261110737919807, 0.005099548492580652, -0.010210114531219006, 0.022040409967303276, 0.14700709283351898, 0.03381533920764923, 0.04749048873782158, -0.08962485939264297, -0.033549949526786804, -0.046185944229364395, -0.05184585601091385, 0.04692189767956734, 0.04133811220526695, -0.01255352795124054, -0.01081941556185484, -0.021265635266900063, -0.034188076853752136, 0.0015658069169148803, -0.020458156242966652, 0.29926443099975586, 0.026995215564966202, 0.009420214220881462, 0.05748005956411362, -0.017368579283356667, -0.09893328696489334, 0.2229585498571396, 0.13357260823249817, -0.028109634295105934, 0.014176450669765472, 0.035654548555612564, 0.0047594825737178326, 0.016413556411862373, 0.06816580891609192, 0.0035045649856328964, 0.23892654478549957, 0.007355669979006052, -0.06955362111330032, -0.10695943981409073, -0.01650063320994377, -0.09652462601661682, -0.008220906369388103, 0.10057258605957031, -0.13457763195037842, -0.14856290817260742, 0.051710546016693115, -0.23665127158164978, -0.01587158627808094, 0.042537931352853775, 0.058480944484472275, -0.09229299426078796, -0.03863649442791939, -0.017020218074321747, 0.0020889306906610727, 0.02523707039654255, -0.05601124465465546, 0.026427602395415306, 0.034208767116069794, 0.02163274586200714, -0.08796228468418121, -0.045942433178424835, 0.115231953561306, 0.1375461220741272, 0.1398613005876541, 0.03999631479382515, -0.012331875041127205, 0.03788674622774124, 0.05617082864046097, -0.14915959537029266, 0.07152508199214935, -0.0017424849793314934, -0.11887053400278091, -0.05440588295459747, -0.009779517538845539, -0.029525049030780792, -0.02327602542936802, -0.07597402483224869, -0.13780422508716583, 0.044316940009593964, 0.07142841815948486, 0.015319714322686195, -0.02246369607746601, 0.043814852833747864, 0.028538979589939117, 0.09589656442403793, 0.15342313051223755, -0.015491630882024765, -0.09868676215410233, -0.05329442769289017, -0.011209186166524887, 0.005977707449346781, -0.03644721955060959, 0.03505965694785118, -0.07836025953292847, -0.010563082993030548, -0.11734850704669952, 0.09071695059537888, -0.15666712820529938, -0.028867915272712708, 0.019595175981521606, 0.048655200749635696, 0.007279754616320133, -0.020533209666609764, 0.06519631296396255, 0.056114450097084045, -0.01895919255912304, -0.03386515751481056, 0.0011861514067277312, 0.04531729593873024, -0.14098328351974487, -0.1133762001991272 ]
null
null
transformers
## Cyclone SIMCSE RoBERTa WWM Ext Chinese This model provides simplified Chinese sentence embeddings encoding based on [Simple Contrastive Learning](https://arxiv.org/abs/2104.08821). The pretrained model(Chinese RoBERTa WWM Ext) is used for token encoding. ### Usage Please use [SentenceTransformer](https://github.com/UKPLab/sentence-transformers) to load the model. from sentence_transformers import SentenceTransformer encoder = SentenceTransformer('cyclone/simcse-chinese-roberta-wwm-ext')
{}
feature-extraction
cyclone/simcse-chinese-roberta-wwm-ext
[ "transformers", "pytorch", "bert", "feature-extraction", "arxiv:2104.08821", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2104.08821" ]
[]
TAGS #transformers #pytorch #bert #feature-extraction #arxiv-2104.08821 #endpoints_compatible #has_space #region-us
## Cyclone SIMCSE RoBERTa WWM Ext Chinese This model provides simplified Chinese sentence embeddings encoding based on Simple Contrastive Learning. The pretrained model(Chinese RoBERTa WWM Ext) is used for token encoding. ### Usage Please use SentenceTransformer to load the model. from sentence_transformers import SentenceTransformer encoder = SentenceTransformer('cyclone/simcse-chinese-roberta-wwm-ext')
[ "## Cyclone SIMCSE RoBERTa WWM Ext Chinese\r\n\r\nThis model provides simplified Chinese sentence embeddings encoding based on Simple Contrastive Learning.\r\nThe pretrained model(Chinese RoBERTa WWM Ext) is used for token encoding.", "### Usage\r\nPlease use SentenceTransformer to load the model.\r\n\r\n from sentence_transformers import SentenceTransformer\r\n \r\n encoder = SentenceTransformer('cyclone/simcse-chinese-roberta-wwm-ext')" ]
[ "TAGS\n#transformers #pytorch #bert #feature-extraction #arxiv-2104.08821 #endpoints_compatible #has_space #region-us \n", "## Cyclone SIMCSE RoBERTa WWM Ext Chinese\r\n\r\nThis model provides simplified Chinese sentence embeddings encoding based on Simple Contrastive Learning.\r\nThe pretrained model(Chinese RoBERTa WWM Ext) is used for token encoding.", "### Usage\r\nPlease use SentenceTransformer to load the model.\r\n\r\n from sentence_transformers import SentenceTransformer\r\n \r\n encoder = SentenceTransformer('cyclone/simcse-chinese-roberta-wwm-ext')" ]
[ 42, 63, 54 ]
[ "passage: TAGS\n#transformers #pytorch #bert #feature-extraction #arxiv-2104.08821 #endpoints_compatible #has_space #region-us \n## Cyclone SIMCSE RoBERTa WWM Ext Chinese\r\n\r\nThis model provides simplified Chinese sentence embeddings encoding based on Simple Contrastive Learning.\r\nThe pretrained model(Chinese RoBERTa WWM Ext) is used for token encoding.### Usage\r\nPlease use SentenceTransformer to load the model.\r\n\r\n from sentence_transformers import SentenceTransformer\r\n \r\n encoder = SentenceTransformer('cyclone/simcse-chinese-roberta-wwm-ext')" ]
[ -0.028554080054163933, -0.05265853926539421, -0.0018137121805921197, 0.03653391823172569, 0.12859664857387543, -0.018765108659863472, 0.004289775621145964, -0.011611961759626865, -0.0632324144244194, 0.012004598043859005, 0.10620573908090591, -0.0015284570399671793, 0.046450722962617874, -0.07722622901201248, -0.025195837020874023, -0.336261510848999, 0.09935060888528824, 0.05692271888256073, -0.029211755841970444, 0.12336000055074692, 0.11886604875326157, -0.07289236783981323, 0.11483168601989746, 0.0844116061925888, -0.07990805804729462, 0.10707252472639084, -0.0997481644153595, -0.08387737721204758, 0.03641323372721672, 0.011655351147055626, 0.13959354162216187, 0.042524319142103195, 0.02819414995610714, -0.08177418261766434, -0.013230511918663979, -0.05726635456085205, -0.07030005753040314, -0.007794570177793503, 0.007728426717221737, -0.04961279407143593, 0.12478499859571457, -0.06071997806429863, 0.03727329149842262, 0.03151329234242439, -0.09592186659574509, -0.05395086482167244, 0.01061064563691616, 0.018184617161750793, 0.18632420897483826, 0.1643422394990921, -0.027103623375296593, 0.07203170657157898, -0.1276538223028183, 0.07829044759273529, 0.0941571369767189, -0.2709347903728485, 0.01792575605213642, 0.06402808427810669, 0.14312276244163513, -0.056575559079647064, -0.06324461102485657, 0.0003469485091045499, 0.06056957319378853, 0.047981638461351395, -0.03507796302437782, -0.10393670201301575, -0.04862479120492935, 0.06670326739549637, -0.14761702716350555, 0.08613467961549759, 0.2490658462047577, -0.0012308473233133554, -0.07120515406131744, -0.01768689602613449, -0.014781669713556767, -0.09888111054897308, -0.01838277094066143, -0.09485369920730591, -0.007439131382852793, 0.05624143406748772, 0.04316550865769386, 0.0028199716471135616, -0.09182246029376984, -0.03975829482078552, -0.17681579291820526, 0.2278364896774292, -0.016466546803712845, 0.012392746284604073, -0.20823612809181213, 0.04418637976050377, 0.03664528951048851, -0.1414201259613037, -0.038434166461229324, -0.047473516315221786, -0.056076791137456894, 0.1010257825255394, -0.08074494451284409, -0.1453487128019333, -0.06695928424596786, 0.08865798264741898, -0.0631721094250679, 0.05931340903043747, 0.14814415574073792, 0.03158188983798027, 0.028389565646648407, 0.2059755176305771, -0.057035382837057114, -0.10625506937503815, -0.05387764051556587, -0.09393737465143204, -0.022329088300466537, -0.059798695147037506, -0.18364140391349792, -0.022669104859232903, -0.0778726190328598, 0.02347579225897789, -0.059100788086652756, 0.051952484995126724, -0.07081833481788635, -0.08205163478851318, 0.24861067533493042, -0.07975571602582932, 0.03486604616045952, -0.006547190714627504, -0.015851693227887154, 0.20657138526439667, 0.014810599386692047, 0.025620901957154274, -0.0060340105555951595, 0.0047826748341321945, -0.07942261546850204, 0.0014673139667138457, -0.03708310052752495, -0.1331237107515335, -0.05597119778394699, -0.047285642474889755, 0.05817592889070511, -0.1441667079925537, -0.1476108282804489, 0.02506878413259983, -0.05086884647607803, -0.027116307988762856, -0.029271485283970833, -0.10783398151397705, -0.07893465459346771, 0.043636124581098557, -0.016029750928282738, -0.1429409235715866, 0.02519022673368454, -0.06644728034734726, -0.02345479279756546, 0.06210560351610184, -0.1985948532819748, 0.014726425521075726, -0.15611398220062256, 0.04860301315784454, -0.02490077167749405, 0.08104516565799713, -0.024987520650029182, 0.09768649935722351, -0.08799136430025101, -0.06805640459060669, -0.022040506824851036, 0.03811465948820114, 0.007087404374033213, 0.16585247218608856, -0.23564167320728302, -0.03096463903784752, 0.07353729009628296, -0.1546444445848465, -0.02583317644894123, 0.07294319570064545, -0.018440159037709236, 0.06416332721710205, 0.05775143951177597, 0.06766245514154434, -0.011863128282129765, 0.015588336624205112, 0.003053799271583557, 0.10569481551647186, -0.09212550520896912, 0.07409712672233582, 0.0882255882024765, 0.052207693457603455, -0.08877768367528915, 0.01578286848962307, 0.01281706802546978, 0.04446591064333916, 0.011585998348891735, -0.0418853722512722, 0.008697301149368286, -0.02007039450109005, 0.07072444260120392, -0.034251514822244644, 0.05878858268260956, -0.03143363073468208, -0.07551507651805878, 0.07463924586772919, 0.08876726776361465, -0.05019879713654518, 0.07194610685110092, -0.14354026317596436, -0.027039041742682457, 0.012370524927973747, -0.020934708416461945, -0.11880430579185486, 0.0626557469367981, 0.002972130896523595, 0.09432724118232727, 0.013447299599647522, 0.04685625061392784, 0.0577484667301178, -0.016605990007519722, 0.03702881932258606, -0.048437654972076416, 0.060705721378326416, 0.07131627202033997, -0.07351797819137573, -0.03576355427503586, -0.04745744913816452, -0.0696883350610733, 0.03193676471710205, -0.08218434453010559, 0.004894951358437538, 0.1435016393661499, -0.029737655073404312, -0.029690783470869064, 0.03174235671758652, -0.012122434563934803, 0.04217135161161423, -0.06172099709510803, 0.027768082916736603, 0.006299990229308605, 0.02106938511133194, -0.2046656459569931, 0.2743980288505554, -0.15841926634311676, -0.1504749059677124, 0.08902309834957123, -0.09196034073829651, -0.10030601918697357, 0.05243365466594696, 0.028545653447508812, 0.0023706667125225067, 0.05711603909730911, -0.01004449836909771, 0.30580443143844604, -0.0616888701915741, 0.14076317846775055, -0.04595628380775452, 0.03153489530086517, -0.0014744950458407402, -0.03238334879279137, -0.009770937263965607, 0.15129221975803375, -0.01875651627779007, -0.18552260100841522, 0.03336269035935402, 0.15125030279159546, -0.037926916033029556, 0.03542780876159668, 0.045411963015794754, -0.011871443130075932, -0.07054996490478516, -0.04961884394288063, -0.000028735888918163255, 0.03191017359495163, -0.11057586222887039, -0.019851673394441605, 0.028595618903636932, 0.029796605929732323, 0.04230821877717972, -0.1080506443977356, -0.05131703242659569, 0.13623537123203278, 0.056160785257816315, 0.006220954470336437, 0.026697786524891853, -0.057960379868745804, 0.0043800813145935535, -0.06712402403354645, -0.11862947791814804, -0.010847237892448902, 0.016649411991238594, -0.09160031378269196, 0.23083670437335968, -0.106538325548172, -0.2970646321773529, -0.12933214008808136, -0.127070352435112, 0.012471036985516548, -0.01653752289712429, 0.08066755533218384, -0.1580963432788849, -0.10652310401201248, -0.12501975893974304, -0.009687747806310654, -0.1872607171535492, 0.001706346869468689, 0.03904769569635391, -0.060417115688323975, -0.059557538479566574, -0.0870155543088913, 0.00479641230776906, 0.005613026209175587, -0.08638027310371399, 0.0358906053006649, -0.06581546366214752, 0.09599330276250839, 0.18519079685211182, -0.0006188054103404284, 0.01355703640729189, 0.017238087952136993, 0.2193848192691803, -0.018834659829735756, 0.03041720949113369, 0.08937172591686249, -0.11138813942670822, 0.028958017006516457, 0.003573603928089142, 0.023446472361683846, -0.06837836652994156, -0.004410624969750643, -0.05958646163344383, -0.058587461709976196, -0.1452970802783966, -0.11607512831687927, -0.07276449352502823, 0.018615519627928734, -0.02500985562801361, 0.03873768448829651, 0.09215667843818665, 0.10144171118736267, 0.0688202828168869, 0.004301843233406544, -0.09094186127185822, 0.06628037244081497, -0.02529226616024971, -0.01595231704413891, 0.10607825219631195, -0.06981340795755386, -0.09109243005514145, 0.02810552343726158, -0.030084017664194107, 0.08024221658706665, 0.10857710242271423, 0.14462095499038696, 0.030204184353351593, 0.06946121901273727, 0.13509753346443176, 0.10931218415498734, -0.1279791295528412, -0.05241833254694939, -0.037718575447797775, -0.020788423717021942, 0.11157722026109695, 0.09828516840934753, 0.03357618674635887, 0.023432152345776558, -0.08205505460500717, 0.027725564315915108, 0.01924886181950569, -0.050149571150541306, 0.08973336964845657, -0.14888058602809906, -0.03232160955667496, 0.05593320354819298, -0.005431205965578556, -0.010573669336736202, 0.10648512840270996, 0.06792301684617996, -0.04916700720787048, 0.08757100999355316, 0.014724217355251312, 0.11016914248466492, 0.12237569689750671, 0.004065375775098801, -0.07963571697473526, 0.022781288251280785, 0.03808610886335373, 0.07327651977539062, -0.11530778557062149, 0.11667045205831528, 0.012852412648499012, -0.05921176075935364, -0.04763185232877731, -0.00004634155266103335, -0.03576110303401947, 0.17809416353702545, 0.18168413639068604, 0.010568063706159592, -0.12784844636917114, -0.02469017170369625, -0.06744131445884705, 0.013821878470480442, 0.12990069389343262, -0.026604261249303818, 0.05219230428338051, 0.029933864250779152, 0.007591355126351118, 0.03351060673594475, 0.09711827337741852, 0.01522031705826521, -0.1851930320262909, 0.009077007882297039, 0.056083496659994125, 0.012388921342790127, 0.007132034283131361, 0.020573195070028305, -0.08857641369104385, 0.10833170264959335, -0.09139471501111984, 0.008584904484450817, -0.1025598794221878, -0.11149031668901443, 0.133989617228508, -0.09990469366312027, 0.03184884414076805, 0.021654248237609863, -0.018279055133461952, -0.02584991417825222, -0.06818464398384094, 0.11018829792737961, -0.05169561132788658, 0.014567111618816853, -0.005228639114648104, -0.017514174804091454, 0.07076042890548706, 0.06108803302049637, 0.051532745361328125, 0.044428735971450806, -0.10162658244371414, -0.1666816771030426, -0.12140217423439026, -0.087103933095932, -0.046067334711551666, 0.12329791486263275, 0.020518776029348373, -0.013722213916480541, -0.05480894073843956, 0.0029430733993649483, 0.2777245342731476, 0.08762418478727341, -0.11529836803674698, 0.12056899070739746, 0.09672082960605621, -0.024853000417351723, -0.27188441157341003, -0.08189088851213455, -0.027901871129870415, 0.09741096198558807, 0.05826409161090851, -0.023992864415049553, 0.1111774817109108, -0.01057279109954834, 0.010985243134200573, -0.07081631571054459, -0.198762446641922, -0.0897345170378685, 0.13268648087978363, -0.05525638535618782, 0.1980586051940918, -0.11382845789194107, -0.03600718826055527, 0.005016077309846878, -0.127798393368721, 0.1573350876569748, 0.02642645128071308, 0.0704907700419426, 0.019485624507069588, -0.04403137415647507, 0.021316412836313248, 0.006437281612306833, 0.19769945740699768, 0.07388767600059509, 0.02109670266509056, -0.004503563046455383, -0.10134641826152802, 0.13139723241329193, 0.03188442438840866, 0.10773658752441406, -0.1030254140496254, 0.04505109786987305, -0.2481520175933838, -0.056376587599515915, 0.0013444715877994895, 0.008372259326279163, 0.015007957816123962, -0.008538810536265373, -0.05351082608103752, 0.018764007836580276, 0.08724860101938248, -0.01476339902728796, 0.19827066361904144, -0.03862372785806656, -0.09327656775712967, 0.15161004662513733, 0.13713997602462769, -0.007364931516349316, -0.05112944170832634, -0.022585105150938034, -0.03206939995288849, 0.11950196325778961, -0.13194715976715088, 0.005323527380824089, 0.0015040163416415453, 0.020258383825421333, 0.09056442975997925, 0.06383555382490158, 0.0011089193867519498, 0.06282540410757065, 0.059457313269376755, -0.1051386147737503, -0.039845313876867294, -0.12983925640583038, 0.028257964178919792, 0.06808996945619583, 0.0597837008535862, 0.05755084753036499, -0.12000826001167297, 0.009038764983415604, -0.03681228309869766, 0.039911966770887375, -0.10567508637905121, 0.029260458424687386, 0.014384759590029716, 0.03007824905216694, -0.1201416552066803, 0.035273630172014236, 0.0661335214972496, -0.14891234040260315, 0.07775920629501343, 0.10447096079587936, -0.06324685364961624, -0.05790479853749275, -0.1499793529510498, 0.09509885311126709, -0.1722574234008789, -0.049012184143066406, -0.03530192747712135, -0.1585608869791031, -0.0020063663832843304, 0.06785351037979126, 0.10528763383626938, 0.0016525924438610673, -0.11814191937446594, -0.03947191312909126, -0.07613006979227066, -0.03451929986476898, 0.11910589039325714, 0.029985418543219566, -0.06537564098834991, 0.09379979968070984, 0.09637060761451721, 0.12572512030601501, -0.05392558500170708, -0.06278687715530396, -0.05255541205406189, 0.014549506828188896, -0.10639381408691406, 0.017303315922617912, -0.0836903303861618, -0.027040237560868263, 0.0019835636485368013, -0.042488500475883484, -0.042058251798152924, 0.006421484984457493, -0.05836177244782448, 0.02304888516664505, -0.03458524867892265, 0.02109677530825138, 0.07761142402887344, -0.0036607992369681597, 0.048368457704782486, -0.035062480717897415, 0.00538516603410244, 0.08575626462697983, -0.09509007632732391, 0.047820672392845154, -0.1352050006389618, -0.020805705338716507, 0.05103360861539841, 0.03457112982869148, 0.08598483353853226, 0.044723328202962875, -0.02341841347515583, 0.024430986493825912, -0.0029801891651004553, -0.058883920311927795, -0.0014586274046450853, -0.04521581530570984, -0.05211729556322098, -0.10447581112384796, -0.08270181715488434, -0.07622569799423218, -0.04873146116733551, 0.04442548006772995, 0.03336840122938156, 0.15338212251663208, 0.003795547178015113, 0.127537339925766, -0.036449532955884933, -0.02483913116157055, -0.03406333178281784, -0.04557514563202858, -0.05321895331144333, -0.08960004150867462, 0.0360647477209568, -0.03977464511990547, 0.1665407121181488, -0.09224707633256912, -0.024264980107545853, -0.02254582569003105, 0.04323096200823784, 0.10038851201534271, 0.03284026309847832, 0.26751598715782166, 0.09816057980060577, 0.01994429901242256, -0.12759289145469666, 0.04874265566468239, 0.037268657237291336, 0.10797896236181259, -0.009253021329641342, 0.03444815054535866, -0.09033303707838058, 0.22483383119106293, -0.010294448584318161, 0.034903720021247864, 0.0424368716776371, -0.05996602401137352, -0.10859231650829315, 0.021566815674304962, 0.0268521960824728, 0.06506744027137756, 0.16005998849868774, 0.00408919295296073, 0.09532929956912994, 0.10713735222816467, -0.09572161734104156, -0.1395399570465088, -0.10632681846618652, -0.06639260053634644, -0.10496421903371811, -0.0034818986896425486, -0.12106571346521378, 0.04064825549721718, 0.03073231875896454, 0.02456234022974968, 0.03268531337380409, 0.1320710927248001, -0.021972794085741043, -0.01482109073549509, 0.07792782783508301, -0.03162577003240585, 0.06200997158885002, 0.15181033313274384, 0.053919825702905655, 0.031089046970009804, -0.017142940312623978, 0.02317793481051922, 0.017793014645576477, 0.02024920843541622, 0.031181026250123978, -0.03685006871819496, -0.11704324930906296, -0.041150324046611786, 0.03336191922426224, -0.030050378292798996, 0.11827144026756287, 0.0618823766708374, 0.007910322397947311, -0.030415041372179985, 0.002732502529397607, -0.07113370299339294, -0.07490155100822449, -0.08051985502243042, 0.16381649672985077, 0.11241519451141357, 0.11187484860420227, 0.026086006313562393, -0.046532705426216125, -0.04802589863538742, 0.24728617072105408, 0.13325995206832886, -0.10686445981264114, 0.017686622217297554, 0.04459148645401001, 0.023245971649885178, 0.06552821397781372, 0.06002194806933403, 0.017792677506804466, 0.3122994601726532, -0.054981350898742676, -0.061166033148765564, -0.1696951538324356, -0.05731309577822685, -0.05504193902015686, 0.03212856501340866, 0.09968579560518265, -0.10518928617238998, -0.023353353142738342, 0.08642531931400299, -0.16564400494098663, 0.07223304361104965, 0.033861156553030014, -0.12324117869138718, -0.04159051179885864, 0.009412085637450218, 0.15891246497631073, 0.1234574019908905, 0.07621511816978455, -0.0031577274203300476, 0.010425662621855736, 0.06940478086471558, 0.03653779998421669, -0.14718684554100037, 0.1104818657040596, 0.07711056619882584, -0.1248268112540245, -0.024740777909755707, 0.007425549440085888, 0.04336364194750786, 0.042017970234155655, 0.10589101910591125, -0.021945960819721222, 0.1228252574801445, -0.03262008726596832, -0.021187057718634605, 0.08536650985479355, 0.06643002480268478, -0.004966977518051863, -0.019930534064769745, -0.018473999574780464, -0.17470063269138336, 0.11451516300439835, 0.11301539838314056, -0.04108269140124321, -0.012447907589375973, -0.01188506931066513, -0.0040296283550560474, 0.10312604159116745, 0.1407899707555771, -0.026900891214609146, -0.03061670809984207, 0.0012754366034641862, -0.004361702594906092, -0.04622119665145874, -0.05718992277979851, -0.049828317016363144, -0.09782666712999344, -0.03861868381500244, 0.013082960620522499, 0.06213219463825226, -0.1631295531988144, -0.02962542697787285, -0.058445531874895096, -0.013389287516474724, -0.009974230080842972, 0.02627825178205967, 0.12528014183044434, 0.010740464553236961, -0.02097354643046856, 0.04808202385902405, 0.07183939218521118, 0.03521266579627991, -0.12103113532066345, -0.13070207834243774 ]
null
null
transformers
# About This is a sample repo.
{}
fill-mask
cylee/tutorial
[ "transformers", "tf", "bert", "fill-mask", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #tf #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
# About This is a sample repo.
[ "# About\n\nThis is a sample repo." ]
[ "TAGS\n#transformers #tf #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n", "# About\n\nThis is a sample repo." ]
[ 35, 8 ]
[ "passage: TAGS\n#transformers #tf #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n# About\n\nThis is a sample repo." ]
[ -0.020638464018702507, -0.04526073858141899, -0.004770228173583746, 0.0014838868519291282, 0.10005791485309601, 0.0338241346180439, 0.09835878014564514, 0.08801118284463882, 0.029648128896951675, -0.03379072993993759, 0.1531848907470703, 0.13872165977954865, -0.028896372765302658, 0.21480605006217957, -0.06427262723445892, -0.2564944922924042, 0.05779002979397774, 0.04405888542532921, -0.10111920535564423, 0.0830889493227005, 0.07054884731769562, -0.064909428358078, 0.08429653942584991, -0.036377567797899246, -0.12653031945228577, 0.08280526101589203, 0.07343019545078278, -0.10839973390102386, 0.134135901927948, 0.10209453850984573, 0.20319393277168274, 0.04592067003250122, -0.03285146504640579, -0.06954608112573624, 0.05843556672334671, 0.01912878453731537, -0.06747812777757645, 0.04090375453233719, 0.020071323961019516, -0.010681383311748505, 0.056457579135894775, 0.1526065319776535, 0.017002910375595093, 0.08558686077594757, -0.15771658718585968, -0.11301999539136887, -0.02559477649629116, -0.02004312165081501, 0.018359532579779625, 0.04852477088570595, 0.0020172931253910065, 0.26829463243484497, -0.15308307111263275, 0.10696929693222046, 0.20043031871318817, -0.23793333768844604, -0.0028316196985542774, 0.0904601514339447, 0.07612587511539459, -0.08559839427471161, -0.02544979378581047, 0.04741762951016426, 0.09985996037721634, 0.03825804591178894, -0.007233008276671171, -0.07028944790363312, -0.12097861617803574, 0.004248871933668852, -0.08399491757154465, -0.03823775425553322, 0.21030612289905548, 0.028088543564081192, -0.004382567014545202, 0.0464450865983963, -0.09857963770627975, 0.002242234768345952, 0.0037664391566067934, -0.04362202808260918, -0.07796148210763931, 0.04794059321284294, -0.09586690366268158, -0.0039565712213516235, -0.08636413514614105, -0.03661603853106499, -0.2586638629436493, 0.22413961589336395, -0.012860853224992752, 0.06729239225387573, -0.15108740329742432, 0.053095199167728424, -0.053203754127025604, -0.1053212583065033, 0.018302472308278084, -0.0981588363647461, -0.025783687829971313, -0.06239490211009979, -0.09151691198348999, -0.12490551173686981, 0.11915712803602219, 0.24423404037952423, 0.0712323933839798, 0.00542096234858036, -0.08940641582012177, 0.03892061114311218, 0.03064960613846779, 0.03359363600611687, -0.014226265251636505, -0.004457438830286264, 0.0351538360118866, -0.13886019587516785, -0.04682942107319832, -0.0552871972322464, -0.12201213091611862, -0.009864922612905502, 0.0020369342528283596, 0.0680900439620018, 0.06029688939452171, 0.0766008049249649, -0.05104264244437218, -0.019329437986016273, 0.05552689731121063, -0.078093983232975, -0.047489702701568604, -0.031451862305402756, 0.045884061604738235, 0.05736669525504112, 0.04202933982014656, 0.014117380604147911, 0.027191750705242157, 0.02204611524939537, -0.07637125253677368, -0.09899304062128067, -0.05424252897500992, -0.09563472121953964, 0.03187945485115051, -0.062764473259449, 0.05283556133508682, -0.2242327779531479, -0.10372186452150345, 0.04945734143257141, 0.04565800353884697, -0.01933915726840496, -0.011303355917334557, 0.018657684326171875, -0.011733991093933582, 0.029940424486994743, -0.029993772506713867, 0.05706217885017395, -0.07331940531730652, 0.09153001010417938, -0.012133529409766197, 0.1242019385099411, -0.10899204015731812, 0.019415685907006264, -0.06786307692527771, 0.0035835690796375275, -0.16919364035129547, -0.02000764198601246, -0.04618607088923454, 0.1735881268978119, -0.06334850937128067, -0.016138944774866104, -0.1644633263349533, 0.031540412455797195, -0.030858565121889114, 0.171010360121727, -0.1372336745262146, -0.10122884809970856, 0.21048006415367126, -0.12789253890514374, -0.11565631628036499, 0.09264744818210602, -0.01495369989424944, 0.05486151948571205, 0.03852052986621857, 0.08585153520107269, 0.0543067567050457, -0.13131888210773468, 0.08173246681690216, 0.1321915090084076, -0.12771795690059662, -0.03297298029065132, 0.005200146231800318, -0.0017359136836603284, -0.1657947599887848, 0.029110342264175415, 0.10621645301580429, 0.14434485137462616, -0.09255307912826538, -0.022239647805690765, -0.020990025252103806, -0.06182858720421791, 0.08548519760370255, 0.03181752189993858, 0.09763478487730026, -0.07307291030883789, -0.04520689696073532, -0.015608086250722408, 0.015269991010427475, 0.05225202441215515, 0.00806522835046053, -0.13513454794883728, 0.11748164892196655, -0.052085794508457184, 0.0019670245237648487, -0.17492525279521942, -0.12753723561763763, 0.0019191503524780273, 0.10624560713768005, -0.01465243473649025, 0.23093606531620026, 0.11888088285923004, -0.06675355136394501, -0.015736235305666924, 0.023919323459267616, 0.08064392954111099, 0.04809902235865593, -0.0014421007363125682, -0.11875361949205399, 0.02627851814031601, -0.10671203583478928, -0.06991101056337357, -0.004614669363945723, -0.04211418330669403, 0.03621794655919075, 0.21150550246238708, 0.05531386286020279, 0.035803135484457016, -0.02918100357055664, -0.0023681097663939, -0.05380008742213249, -0.03105071745812893, 0.022643905133008957, 0.010222511366009712, -0.053125474601984024, 0.14990423619747162, -0.06784448772668839, 0.28559738397598267, 0.18189503252506256, -0.22580625116825104, -0.08306197077035904, 0.040265511721372604, -0.036275144666433334, 0.006764399353414774, 0.026825202628970146, -0.04138622805476189, 0.06762538105249405, -0.020113732665777206, 0.11926358938217163, -0.0045202868059277534, -0.010695114731788635, 0.028493741527199745, -0.038538459688425064, -0.04043921083211899, 0.022380029782652855, 0.11302640289068222, -0.1906518191099167, 0.13263672590255737, 0.20236220955848694, 0.05764956772327423, 0.1346011757850647, -0.040712181478738785, -0.06637027114629745, 0.018114572390913963, -0.0322340652346611, 0.003096441039815545, 0.06543438136577606, -0.12667813897132874, -0.0026154578663408756, 0.07583598047494888, -0.03307846933603287, 0.04512356221675873, -0.06946029514074326, -0.024867791682481766, 0.03357677161693573, 0.0174387339502573, -0.07151643931865692, 0.11887408047914505, 0.011668744497001171, 0.07698667794466019, 0.020764408633112907, -0.0873173251748085, 0.08963430672883987, 0.01864396035671234, -0.08302129805088043, 0.1729007214307785, -0.11461217701435089, -0.2905164957046509, -0.09630388021469116, -0.1613900363445282, 0.044441163539886475, 0.06808650493621826, 0.05368661880493164, -0.11668174713850021, -0.1049017533659935, 0.053734190762043, 0.05478675663471222, 0.015502508729696274, 0.10014822334051132, -0.038901135325431824, 0.02781647816300392, 0.0025857798755168915, -0.06986098736524582, -0.034762393683195114, 0.01434298511594534, -0.029876770451664925, 0.14037871360778809, -0.17076954245567322, 0.0825507864356041, 0.1335289478302002, 0.03020785003900528, 0.07823547720909119, -0.02060268074274063, 0.2283136546611786, -0.0937584713101387, 0.007528198417276144, 0.14289237558841705, -0.05993090197443962, 0.012894549407064915, 0.16230887174606323, -0.0005658446461893618, -0.0620291493833065, 0.07352832704782486, -0.008656895719468594, -0.12492774426937103, -0.13076554238796234, -0.06121625751256943, -0.09274028241634369, -0.01192986499518156, 0.0684160515666008, 0.03786822035908699, 0.1650833785533905, 0.06022380292415619, 0.04315483197569847, 0.026021292433142662, -0.10334702581167221, 0.012695228680968285, 0.08549732714891434, -0.057250864803791046, 0.12792626023292542, -0.06719677150249481, -0.15088196098804474, 0.10027607530355453, -0.0635102167725563, 0.13224521279335022, 0.11084934324026108, -0.008262407034635544, 0.02277718484401703, 0.10222909599542618, 0.13635724782943726, 0.17997793853282928, 0.05685535818338394, -0.09208337217569351, -0.026208745315670967, -0.0179587509483099, -0.03988543152809143, 0.06694933772087097, 0.11399666965007782, -0.11942823976278305, -0.050956618040800095, -0.10861597955226898, 0.030256908386945724, 0.15666210651397705, 0.11109250783920288, -0.26186686754226685, -0.019157709553837776, 0.008513352833688259, -0.04765915870666504, -0.0645303726196289, 0.048812881112098694, 0.02741658128798008, -0.09513694792985916, 0.033470772206783295, -0.025694791227579117, 0.0964333638548851, 0.10321322828531265, 0.054411448538303375, -0.008284061215817928, -0.0842454805970192, -0.03354126587510109, 0.0566696934401989, -0.2782317101955414, 0.27515000104904175, -0.01016558799892664, -0.04593979939818382, -0.09881193190813065, -0.016508307307958603, 0.06206893548369408, 0.1079106256365776, 0.09999005496501923, -0.006494046654552221, -0.14250190556049347, -0.15325872600078583, -0.013277361169457436, 0.06332948058843613, 0.10027562826871872, -0.00243855407461524, 0.018110990524291992, -0.04434330016374588, -0.018266664817929268, 0.04309516400098801, 0.14265723526477814, -0.05457423999905586, -0.045975372195243835, 0.0641486644744873, 0.004787772428244352, 0.03683453053236008, -0.04889858886599541, -0.06182840093970299, -0.04244852066040039, 0.15385058522224426, 0.002186993369832635, -0.007641430012881756, -0.1446448415517807, 0.010328131727874279, 0.13259302079677582, -0.06688529253005981, 0.12302999198436737, -0.030522217974066734, 0.028021007776260376, -0.03566089645028114, -0.21988730132579803, 0.1290862113237381, -0.1446457952260971, -0.015765009447932243, -0.060969993472099304, 0.06644421070814133, -0.050436556339263916, 0.05359421297907829, 0.037106774747371674, 0.0032880280632525682, -0.02391975000500679, -0.02058585360646248, -0.002986230654641986, -0.12060830742120743, 0.04130121320486069, 0.03277308866381645, -0.030752582475543022, -0.03232359513640404, 0.012583254836499691, 0.009571905247867107, 0.18893033266067505, 0.18657660484313965, -0.07027825713157654, 0.1039186343550682, 0.07866182923316956, -0.009840846061706543, -0.3097914159297943, -0.06118353456258774, -0.07759015262126923, 0.01277625560760498, 0.010034519247710705, -0.12072498351335526, 0.13402318954467773, -0.048304833471775055, -0.01080978661775589, 0.11147766560316086, -0.11387280374765396, -0.11328095942735672, 0.29053398966789246, 0.010719874873757362, 0.38116803765296936, -0.10856204479932785, -0.06997020542621613, -0.03777197375893593, -0.13801439106464386, 0.0843663364648819, -0.020650748163461685, 0.05087010934948921, 0.023384878411889076, 0.06996983289718628, 0.043651893734931946, -0.06657912582159042, 0.08374866843223572, -0.07241640985012054, 0.01520271971821785, -0.15506552159786224, -0.08848509192466736, -0.023774081841111183, 0.011511479504406452, -0.036034148186445236, 0.0721561461687088, 0.028976531699299812, -0.025008656084537506, -0.020763549953699112, -0.06965016573667526, 0.11772362887859344, 0.04865168407559395, -0.08418596535921097, -0.028577685356140137, -0.0018436623504385352, -0.02190971188247204, -0.0247767586261034, 0.13460059463977814, -0.07623261958360672, 0.24662993848323822, 0.09712981432676315, 0.08616308122873306, -0.18647433817386627, -0.00853667501360178, -0.003607884282246232, -0.08599067479372025, 0.07983455061912537, -0.05240541696548462, 0.07452008128166199, 0.08955749869346619, 0.01981748640537262, 0.08886528015136719, 0.11080878973007202, -0.008541845716536045, -0.024832597002387047, 0.18267756700515747, -0.18377229571342468, 0.05420858412981033, -0.07339087873697281, -0.036655131727457047, 0.011988529935479164, -0.014469988644123077, 0.10823020339012146, 0.012440677732229233, 0.013977005146443844, -0.007391748484224081, -0.04270554706454277, -0.07903293520212173, 0.06440825015306473, 0.11268201470375061, 0.04197923094034195, -0.12098194658756256, -0.013280150480568409, -0.0034102106001228094, -0.1626819521188736, 0.012065340764820576, -0.015798499807715416, -0.09504900872707367, -0.10021037608385086, 0.06259552389383316, 0.10644295811653137, -0.1291263997554779, -0.040158115327358246, -0.11430942267179489, -0.11904085427522659, 0.09068558365106583, 0.31644853949546814, 0.08677760511636734, 0.10012417286634445, 0.031366318464279175, -0.005782293621450663, -0.001768449554219842, -0.005622320808470249, 0.006916365120559931, 0.02220650389790535, -0.10641536116600037, 0.11033403873443604, -0.055655498057603836, 0.1333911418914795, -0.13713371753692627, -0.030170608311891556, -0.18058747053146362, 0.02135332114994526, -0.11590111255645752, -0.0422915481030941, -0.07284621894359589, -0.05581505596637726, 0.05785138159990311, -0.09088589251041412, -0.07888615131378174, -0.034667741507291794, -0.13626335561275482, 0.04524976760149002, 0.05795403569936752, -0.005972633138298988, -0.02274378575384617, -0.012868447229266167, 0.13148395717144012, -0.03458724543452263, 0.05368667095899582, 0.1135440394282341, -0.06553711742162704, 0.11817150563001633, -0.09726561605930328, -0.10096146166324615, 0.07600395381450653, 0.043523259460926056, 0.0619838684797287, 0.034587033092975616, 0.01229378767311573, 0.039380524307489395, 0.04963087663054466, 0.043830014765262604, 0.024640554562211037, -0.08749712258577347, 0.021984189748764038, -0.012316819280385971, -0.13999585807323456, -0.03081514500081539, -0.060124751180410385, 0.09107375890016556, 0.019467653706669807, 0.07518960535526276, -0.07167627662420273, 0.08419368416070938, -0.042876288294792175, 0.01660837046802044, -0.030686190351843834, -0.10572521388530731, 0.022983277216553688, -0.026656627655029297, 0.01251986064016819, -0.014156930148601532, 0.2132682204246521, -0.00628779549151659, 0.05454711616039276, 0.04149198904633522, 0.01970505528151989, 0.10290651023387909, 0.0024518154095858335, 0.27581292390823364, 0.09299074113368988, -0.016475440934300423, -0.13366058468818665, 0.0931694284081459, -0.0031840309966355562, -0.10320062935352325, 0.14420278370380402, 0.07799676805734634, -0.07678233087062836, 0.06233298406004906, -0.0012125575449317694, -0.05739833042025566, 0.06437022984027863, -0.15090450644493103, -0.04037998989224434, 0.09464351087808609, 0.03541694954037666, -0.029542040079832077, 0.14420783519744873, -0.04129832237958908, 0.03785943612456322, -0.012674197554588318, 0.006610057316720486, -0.18569014966487885, -0.10675752907991409, -0.10199112445116043, -0.13667325675487518, 0.032409824430942535, -0.06317730247974396, -0.017300376668572426, 0.0803971216082573, 0.04962587356567383, -0.0005138966371305287, 0.18775111436843872, -0.0956338495016098, -0.07236050069332123, 0.04231424629688263, -0.009723044000566006, -0.04206345975399017, -0.08007223159074783, -0.035000160336494446, -0.06545595079660416, 0.028011104092001915, -0.1088850274682045, -0.04262785241007805, -0.05196830630302429, 0.028862489387392998, -0.09196444600820541, -0.08700177818536758, -0.06893384456634521, 0.022616561502218246, -0.07622350752353668, 0.07738178223371506, 0.0010606672149151564, 0.06625279039144516, 0.014977718703448772, 0.11302351206541061, -0.0710948035120964, -0.13553334772586823, -0.17081569135189056, 0.17477484047412872, -0.015586421824991703, 0.09106175601482391, -0.030886072665452957, -0.0023205445613712072, -0.07019472867250443, 0.3485925495624542, 0.24955683946609497, -0.026121770963072777, 0.05311055853962898, 0.04206584766507149, 0.0401461161673069, 0.045228634029626846, 0.1494004875421524, 0.058998625725507736, 0.2065235674381256, -0.05751710757613182, -0.15357810258865356, -0.00707464013248682, -0.021719615906476974, -0.06663738191127777, -0.008245185017585754, 0.08680449426174164, -0.046184051781892776, -0.032751306891441345, 0.11718534678220749, -0.14824621379375458, 0.06742644309997559, 0.015025177039206028, -0.13683275878429413, -0.05389387533068657, -0.06299667805433273, 0.04276658967137337, 0.04083248972892761, 0.0849292203783989, -0.0768495500087738, -0.03941527381539345, 0.05695216357707977, 0.014200659468770027, -0.2243623286485672, -0.10158994048833847, 0.12319184094667435, -0.046445563435554504, 0.08290817588567734, -0.019328314810991287, 0.05993558093905449, 0.08491167426109314, 0.03692527487874031, -0.02249038778245449, 0.023084858432412148, 0.03654235973954201, -0.12671387195587158, -0.040198683738708496, 0.04083153232932091, 0.014049685560166836, -0.03471129387617111, 0.009849331341683865, -0.1937323808670044, 0.06009393185377121, -0.030061572790145874, -0.03268570452928543, -0.029644256457686424, 0.0792255625128746, -0.10289949923753738, 0.11105287075042725, 0.1167161837220192, 0.002029495080932975, -0.002681157784536481, -0.06176215037703514, 0.05465411767363548, 0.07483833283185959, -0.11109264940023422, -0.14462910592556, -0.1179095134139061, -0.05808859318494797, -0.07839802652597427, -0.08100049197673798, -0.17128737270832062, -0.043028078973293304, -0.08405254036188126, 0.016455641016364098, -0.058819618076086044, 0.07526671141386032, 0.08395033329725266, 0.03977546468377113, 0.007307989057153463, -0.00710875540971756, 0.038840632885694504, 0.0497129000723362, -0.12808674573898315, -0.07670791447162628 ]
null
null
transformers
# Description: This is a smaller per-trained model on Sinhalese Language using Masked Language Modeling(MLM). And the model is trained on Oscar Sinhala dataset. # How to Use: The model can be used directly with a pipeline for masked language modeling: ```python >>> from transformers import AutoTokenizer, AutoModelForMaskedLM, pipeline >>> tokenizer = AutoTokenizer.from_pretrained("d42kw01f/Sinhala-RoBERTa") >>> model = AutoModelForMaskedLM.from_pretrained("d42kw01f/Sinhala-RoBERTa") >>> fill_mask = pipeline('fill-mask', model=model, tokenizer=tokenizer) >>> fill_mask("මම ගෙදර <mask>.") [{'score': 0.1822454035282135, 'sequence': 'මම ගෙදර ආව.', 'token': 701, 'token_str': ' ආව'}, {'score': 0.10513380169868469, 'sequence': 'මම ගෙදර ය.', 'token': 310, 'token_str': ' ය'}, {'score': 0.06417194753885269, 'sequence': 'මම ගෙදර එක.', 'token': 328, 'token_str': ' එක'}, {'score': 0.05026362091302872, 'sequence': 'මම ගෙදර ඇත.', 'token': 330, 'token_str': ' ඇත'}, {'score': 0.029960114508867264, 'sequence': 'මම ගෙදර යනව.', 'token': 834, 'token_str': ' යනව'}] ```
{}
fill-mask
d42kw01f/Sinhala-RoBERTa
[ "transformers", "pytorch", "roberta", "fill-mask", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us
# Description: This is a smaller per-trained model on Sinhalese Language using Masked Language Modeling(MLM). And the model is trained on Oscar Sinhala dataset. # How to Use: The model can be used directly with a pipeline for masked language modeling:
[ "# Description:\n\nThis is a smaller per-trained model on Sinhalese Language using Masked Language Modeling(MLM). And the model is trained on Oscar Sinhala dataset.", "# How to Use:\nThe model can be used directly with a pipeline for masked language modeling:" ]
[ "TAGS\n#transformers #pytorch #roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n", "# Description:\n\nThis is a smaller per-trained model on Sinhalese Language using Masked Language Modeling(MLM). And the model is trained on Oscar Sinhala dataset.", "# How to Use:\nThe model can be used directly with a pipeline for masked language modeling:" ]
[ 37, 39, 22 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n# Description:\n\nThis is a smaller per-trained model on Sinhalese Language using Masked Language Modeling(MLM). And the model is trained on Oscar Sinhala dataset.# How to Use:\nThe model can be used directly with a pipeline for masked language modeling:" ]
[ -0.06981303542852402, -0.013871806673705578, -0.002937681507319212, 0.06777280569076538, 0.16878779232501984, 0.00494441669434309, 0.12198388576507568, 0.04897332563996315, -0.018968595191836357, -0.03944399207830429, 0.13632826507091522, 0.04446830973029137, 0.013817831873893738, 0.16273614764213562, 0.053546447306871414, -0.4239024221897125, 0.00979489367455244, -0.0225876085460186, -0.04638388007879257, 0.10715381056070328, 0.12413541227579117, -0.02528974786400795, 0.0771169438958168, 0.04932312294840813, -0.07936476916074753, 0.029843905940651894, -0.022205574437975883, -0.09870719164609909, 0.06832592934370041, 0.020515939220786095, 0.0986592248082161, -0.057686697691679, 0.07245297729969025, -0.08628211915493011, 0.04990250617265701, -0.05078044906258583, 0.02937581017613411, -0.04104109853506088, -0.03275354206562042, -0.05463511496782303, 0.25868353247642517, 0.019529519602656364, 0.023048100993037224, 0.03822454810142517, -0.1305742710828781, -0.10555802285671234, 0.07398613542318344, -0.120207279920578, 0.04613327607512474, 0.1433180272579193, -0.03589393198490143, 0.16687226295471191, -0.15382301807403564, 0.08084648102521896, -0.04510420933365822, -0.17193610966205597, -0.011939623393118382, 0.16038918495178223, 0.1320858895778656, -0.060887206345796585, -0.01902216486632824, 0.09474441409111023, 0.02161748707294464, 0.07923702895641327, 0.05464504659175873, -0.0870261937379837, 0.12107778340578079, -0.09012101590633392, -0.10114505141973495, 0.021397756412625313, 0.15463288128376007, -0.05901305004954338, -0.030826086178421974, -0.09283945709466934, -0.027226822450757027, 0.08314958959817886, -0.030422393232584, -0.0005212990217842162, -0.04937286302447319, 0.04753800109028816, 0.050784628838300705, -0.044220682233572006, -0.05687043070793152, -0.014370834454894066, -0.09971508383750916, 0.16880248486995697, 0.011891392059624195, 0.04939929023385048, -0.24299417436122894, -0.03704894334077835, -0.05800619721412659, -0.11846154928207397, 0.01020907424390316, -0.06858015805482864, 0.023423422127962112, 0.08615496009588242, -0.009800590574741364, -0.10411541908979416, 0.06254685670137405, 0.03461705893278122, 0.07326367497444153, 0.028334802016615868, 0.06767719238996506, 0.030377497896552086, 0.05321294069290161, 0.15920300781726837, -0.11225254088640213, -0.03405619040131569, 0.08450593799352646, -0.08550221472978592, 0.06575038284063339, -0.05711361765861511, -0.1318221241235733, -0.041712261736392975, -0.0856914222240448, 0.10619644820690155, -0.003600445343181491, 0.11152949184179306, -0.05085936188697815, -0.07302562147378922, -0.11953794211149216, -0.14181064069271088, -0.052827611565589905, -0.02966049313545227, -0.020601026713848114, 0.05161895602941513, 0.06776269525289536, 0.0018765988061204553, -0.07259766757488251, -0.04061496630311012, -0.018799640238285065, 0.024129245430231094, -0.13681986927986145, -0.15489868819713593, -0.011783597059547901, -0.12194237112998962, 0.0465032160282135, -0.15152134001255035, -0.1304561197757721, 0.013891668058931828, 0.050511132925748825, -0.03283613175153732, -0.022140907123684883, -0.09171362221240997, 0.029308870434761047, -0.0015938583528622985, 0.012761018238961697, -0.03052178956568241, -0.03874775767326355, -0.031876496970653534, 0.03341059014201164, 0.149908185005188, -0.0628058984875679, 0.03604450821876526, -0.024017399176955223, 0.08518582582473755, -0.09855884313583374, 0.07131461799144745, -0.04432553052902222, 0.07287783175706863, -0.026319727301597595, -0.07450217008590698, -0.07329607754945755, 0.06993042677640915, 0.006653843447566032, 0.19128815829753876, -0.19464631378650665, -0.04908447340130806, 0.2176087200641632, -0.13732534646987915, -0.06714663654565811, 0.09961587190628052, -0.0014596089022234082, 0.13176384568214417, 0.0161573588848114, 0.08668382465839386, 0.02228495106101036, -0.12477706372737885, 0.17551299929618835, 0.08307643979787827, -0.11099719256162643, -0.04715033620595932, 0.10792861878871918, 0.01342079322785139, -0.04786711931228638, 0.04617442190647125, 0.03429707884788513, 0.10083793103694916, -0.061471253633499146, -0.05933103337883949, 0.028813069686293602, -0.07108630239963531, 0.00556509755551815, 0.024656744673848152, 0.06972961872816086, -0.03682412952184677, -0.002290786011144519, 0.03967765346169472, 0.10976827889680862, -0.04583033546805382, 0.045886922627687454, -0.19151130318641663, 0.04099065065383911, -0.025300942361354828, -0.013313822448253632, -0.1543540209531784, -0.005938657559454441, -0.03954683616757393, 0.04529602825641632, 0.101474329829216, -0.006020621862262487, 0.04775169864296913, 0.015370815061032772, -0.02497561275959015, 0.08216515928506851, 0.08079473674297333, 0.027773968875408173, -0.021387208253145218, -0.08774126321077347, 0.03415834903717041, -0.06915759295225143, 0.15077495574951172, -0.016075564548373222, -0.005308669991791248, -0.14509235322475433, 0.05549193546175957, -0.017700249329209328, 0.020356709137558937, 0.10053727775812149, 0.07830825448036194, 0.0012671208241954446, -0.048678360879421234, 0.07464785873889923, -0.010385793633759022, -0.14266328513622284, 0.1906343698501587, -0.18866412341594696, 0.1193932294845581, 0.13800960779190063, -0.15194469690322876, 0.0442998930811882, 0.11703355610370636, 0.03890621289610863, 0.01901061274111271, 0.019539937376976013, 0.13116420805454254, 0.16899169981479645, 0.016003932803869247, 0.14410416781902313, -0.031709685921669006, 0.07953306287527084, 0.0012117731384932995, -0.14844220876693726, -0.04220759496092796, 0.07287084311246872, 0.11312824487686157, -0.189293771982193, 0.07846080511808395, 0.0672534853219986, -0.027889426797628403, 0.21223007142543793, 0.026852065697312355, -0.0030595725402235985, -0.061612989753484726, -0.08369507640600204, 0.020523060113191605, 0.06546889990568161, -0.21103490889072418, -0.053424060344696045, 0.027621697634458542, -0.034523651003837585, 0.08203525096178055, -0.019154522567987442, -0.05070457234978676, 0.05294349044561386, 0.010886368341743946, -0.08545072376728058, 0.13888613879680634, -0.08097920566797256, 0.013315419666469097, -0.019067032262682915, -0.11093518882989883, 0.01257253997027874, -0.008372005075216293, -0.07319387048482895, 0.18970900774002075, -0.16260787844657898, -0.37465378642082214, -0.13550755381584167, -0.12635695934295654, 0.005812535528093576, 0.031675126403570175, 0.006254440639168024, -0.11712143570184708, -0.08527054637670517, 0.006639304105192423, -0.016973422840237617, -0.10826610773801804, -0.0563894547522068, 0.026127852499485016, 0.002840016968548298, -0.018789339810609818, -0.10269014537334442, -0.02262425608932972, -0.012410328723490238, -0.030902186408638954, 0.09841323643922806, -0.20094063878059387, 0.054279014468193054, 0.18174642324447632, 0.013270239345729351, 0.08451477438211441, -0.006945273373275995, 0.20843784511089325, -0.06812519580125809, -0.04624941945075989, 0.22769108414649963, -0.015416008420288563, 0.00309136719442904, 0.10818295180797577, -0.012097490020096302, -0.0747574046254158, 0.017006976529955864, -0.023321153596043587, -0.13402345776557922, -0.16680952906608582, -0.1280718445777893, -0.09152710437774658, -0.006352666765451431, -0.044666580855846405, 0.0029424382373690605, -0.007380107883363962, 0.07333163917064667, 0.06910490244626999, -0.04687570035457611, 0.003559195902198553, 0.033822350203990936, 0.09370304644107819, -0.05874676629900932, 0.0963868722319603, -0.05147299915552139, -0.03388892486691475, 0.0285227932035923, -0.07867797464132309, 0.08715210855007172, 0.12635083496570587, 0.04671206697821617, 0.10385799407958984, 0.12830978631973267, 0.13540245592594147, 0.08221876621246338, -0.023542679846286774, -0.009952718392014503, -0.020009977743029594, -0.025224508717656136, -0.05320194363594055, 0.0037842877209186554, 0.08443736284971237, -0.05407571792602539, -0.038943514227867126, 0.22783365845680237, -0.018308676779270172, 0.047492288053035736, 0.0672779232263565, -0.16845257580280304, -0.02930283360183239, 0.04018695652484894, 0.056013524532318115, -0.08576422184705734, 0.05692620202898979, 0.045474570244550705, -0.11907589435577393, 0.0859183818101883, 0.0018999370513483882, 0.0879063606262207, 0.05377911031246185, 0.044471390545368195, -0.14334458112716675, 0.04405646771192551, 0.014069139957427979, 0.10192739218473434, -0.2842293083667755, 0.2967989444732666, 0.02129185199737549, 0.03779105097055435, -0.06817659735679626, -0.02865181863307953, 0.06220215559005737, 0.19240248203277588, 0.14305540919303894, 0.044083304703235626, -0.005308435298502445, -0.08963818848133087, -0.10364580154418945, 0.03129034489393234, 0.08174624294042587, 0.060059256851673126, 0.015850959345698357, 0.01797373592853546, -0.003488609567284584, -0.0015732628526166081, 0.031241364777088165, -0.10044240951538086, -0.10045665502548218, 0.03455459699034691, 0.07342387735843658, -0.09436299651861191, -0.03396255895495415, -0.10102007538080215, -0.027765469625592232, 0.1761481910943985, 0.04563947767019272, -0.055925291031599045, -0.10733362287282944, -0.01583012193441391, 0.10303926467895508, -0.11534187942743301, 0.0859491378068924, -0.07057876139879227, 0.07162322849035263, -0.057468730956315994, -0.07035384327173233, 0.1068655475974083, -0.07607413828372955, 0.07442024350166321, 0.020959671586751938, 0.082126185297966, 0.03496561199426651, 0.02508598193526268, 0.09536189585924149, -0.011157013475894928, -0.020121775567531586, -0.0842025950551033, -0.08510537445545197, 0.14193648099899292, -0.025056567043066025, 0.12740880250930786, -0.08947454392910004, -0.11093201488256454, -0.017328087240457535, -0.07610005140304565, 0.25269782543182373, 0.08380085229873657, -0.05947375297546387, 0.08079514652490616, 0.17392116785049438, -0.10820655524730682, -0.30985620617866516, -0.11265663802623749, -0.05654701590538025, 0.15244024991989136, 0.034610189497470856, -0.12355618178844452, 0.033762846142053604, -0.07905113697052002, 0.005360300187021494, -0.1272306740283966, -0.17537814378738403, -0.16753803193569183, 0.22839690744876862, 0.08033040910959244, 0.2753322422504425, -0.04144240543246269, -0.0078846774995327, -0.14361068606376648, -0.05595748871564865, -0.02754274383187294, -0.05664630979299545, 0.12739421427249908, -0.010751232504844666, 0.06433285772800446, 0.03157836198806763, -0.020144013687968254, 0.13382719457149506, -0.0300921443849802, -0.007155481725931168, -0.11458905786275864, -0.09206024557352066, 0.07440536469221115, 0.006017157342284918, 0.10585229843854904, 0.0384911447763443, -0.017833037301898003, -0.12470695376396179, -0.08807612210512161, -0.014424207620322704, 0.040718916803598404, 0.035002194344997406, -0.049592744559049606, -0.0019469661638140678, 0.07561645656824112, 0.009402700699865818, 0.013668196275830269, -0.03202653303742409, -0.1368483006954193, -0.006628453731536865, 0.0076835667714476585, 0.24811510741710663, -0.09351256489753723, 0.02898881398141384, -0.02161608450114727, -0.008783620782196522, 0.008379669860005379, -0.0022153714671730995, -0.029264526441693306, 0.055540360510349274, -0.001413254882209003, 0.08502465486526489, 0.03232283145189285, -0.031914424151182175, 0.06298459321260452, 0.09697291254997253, 0.05013052001595497, -0.12497101724147797, -0.02513756975531578, -0.07694441080093384, 0.058687470853328705, -0.039932981133461, 0.06700453907251358, -0.0396491140127182, -0.02542102336883545, -0.05117995664477348, -0.0029914104379713535, -0.06612729281187057, 0.06443246454000473, 0.06918802112340927, 0.02804923988878727, -0.07959523797035217, 0.06521640717983246, -0.019621916115283966, -0.006398398894816637, 0.04086456075310707, 0.10619234293699265, -0.10648805648088455, -0.07938257604837418, 0.03589673712849617, 0.196670800447464, -0.04694245010614395, -0.073465995490551, -0.020500024780631065, -0.10703641176223755, -0.021488474681973457, 0.15994028747081757, 0.0773802101612091, -0.0227949321269989, -0.13549935817718506, -0.04319974035024643, -0.06494088470935822, 0.018689285963773727, 0.13675346970558167, -0.10095765441656113, -0.09054374694824219, 0.1403590440750122, 0.07777746766805649, 0.15021046996116638, -0.09057047218084335, -0.11753034591674805, -0.06771652400493622, 0.08718051016330719, -0.12744997441768646, 0.06747966259717941, -0.10013841837644577, -0.014785357750952244, -0.020686564967036247, -0.06577055156230927, -0.06959710270166397, 0.03175614774227142, -0.08804670721292496, 0.033095620572566986, -0.04168369621038437, -0.0202584657818079, 0.04335080087184906, -0.058555010706186295, 0.05663291737437248, 0.010091989301145077, 0.03339652344584465, 0.115078866481781, -0.09049095213413239, 0.07406463474035263, -0.12174998223781586, -0.05892553552985191, 0.06757483631372452, 0.02956615388393402, 0.056378643959760666, -0.09290260076522827, -0.009314191527664661, 0.047689665108919144, 0.06394411623477936, 0.03629130870103836, 0.07573705911636353, -0.03789807856082916, -0.02614927478134632, 0.026711298152804375, -0.04772467166185379, 0.0017705532955005765, 0.02259417437016964, 0.0338849201798439, 0.06664168834686279, 0.10988550633192062, -0.0647491067647934, 0.0982324630022049, -0.05597507208585739, 0.023574290797114372, -0.09655457735061646, -0.010422144085168839, -0.013406053185462952, -0.0621713325381279, 0.05822232365608215, -0.03247874230146408, 0.18895785510540009, 0.09050537645816803, 0.07911795377731323, -0.018011586740612984, 0.05070146918296814, 0.031050967052578926, -0.0325736440718174, 0.1869802325963974, 0.09277777373790741, 0.07548048347234726, 0.0745464339852333, 0.025229545310139656, -0.007974823005497456, 0.1707540601491928, 0.0059904782101511955, 0.019961170852184296, 0.09936022013425827, 0.1498953253030777, 0.033338773995637894, 0.10847322642803192, -0.061034440994262695, -0.23265819251537323, -0.04999305680394173, 0.058327566832304, -0.023955088108778, -0.05058005824685097, 0.16566181182861328, -0.04174375534057617, 0.08521850407123566, -0.01863657869398594, -0.12663014233112335, -0.23415566980838776, -0.2629638612270355, -0.07853604108095169, 0.02835944853723049, 0.02417578734457493, -0.11699226498603821, -0.030252784490585327, 0.05915513634681702, 0.1038769856095314, -0.04004363343119621, 0.1618667095899582, -0.1528978794813156, -0.05505026876926422, 0.04162309318780899, -0.0120856249704957, 0.04972366616129875, 0.021495800465345383, 0.05522678419947624, -0.11506564915180206, -0.007906110025942326, 0.013088681735098362, -0.03737097606062889, -0.048528190702199936, 0.03407822549343109, -0.09849001467227936, -0.04345886781811714, -0.09345710277557373, 0.024389799684286118, 0.05918976664543152, 0.1555771380662918, 0.009825484827160835, -0.08015049248933792, -0.011753348633646965, 0.08923321217298508, 0.025620151311159134, -0.2840748429298401, -0.1341552734375, 0.2841452360153198, -0.0006231832085177302, 0.03215399757027626, -0.10542232543230057, 0.05108438804745674, -0.01697799190878868, 0.31983521580696106, 0.3449276089668274, -0.10300271958112717, 0.009785987436771393, -0.039946265518665314, 0.029502110555768013, -0.040291495621204376, 0.14634636044502258, 0.033985935151576996, 0.16231882572174072, -0.05498005077242851, 0.016917120665311813, -0.1287693828344345, -0.0634557232260704, -0.19240877032279968, -0.0717082992196083, 0.07472455501556396, -0.04718128964304924, 0.016637951135635376, 0.13772791624069214, -0.1737225502729416, 0.13639529049396515, -0.07779930531978607, -0.07929279655218124, -0.09532133489847183, -0.07971932739019394, -0.0780882015824318, 0.12131232768297195, 0.07760221511125565, -0.0401923805475235, -0.0002456881629768759, 0.01023957785218954, 0.04437786340713501, -0.16234251856803894, -0.05833618715405464, 0.13829457759857178, 0.02636900544166565, 0.0517725795507431, 0.029007500037550926, 0.047565240412950516, 0.05981863662600517, 0.05394461750984192, -0.005608620122075081, 0.15336838364601135, -0.03615490719676018, 0.09731926023960114, 0.05202067270874977, 0.06769012659788132, -0.09578811377286911, -0.09437751024961472, -0.030333684757351875, 0.0022147540003061295, 0.07850459218025208, -0.09615244716405869, -0.03816070035099983, 0.014258679002523422, 0.10375243425369263, -0.10241423547267914, 0.09992170333862305, 0.12646184861660004, -0.009689854457974434, -0.076097272336483, -0.03708403557538986, 0.058505069464445114, -0.033339302986860275, -0.09849344193935394, -0.1286320835351944, -0.07759249210357666, -0.08330436795949936, -0.10009738802909851, -0.018204843625426292, -0.17317692935466766, -0.03585680574178696, -0.10608850419521332, -0.026195736601948738, -0.014424741268157959, 0.06844572722911835, 0.062000572681427, 0.04333117976784706, -0.014479577541351318, -0.050344452261924744, 0.06213580071926117, 0.07075780630111694, -0.19198167324066162, -0.12232889235019684 ]
null
null
transformers
# Description: This is a smaller per-trained model on Tamil Language using Masked Language Modeling(MLM). And the model is trained on Oscar Tamil dataset. # How to Use: The model can be used directly with a pipeline for masked language modeling: ```python >>> from transformers import AutoTokenizer, AutoModelForMaskedLM, pipeline >>> tokenizer = AutoTokenizer.from_pretrained("d42kw01f/Tamil-RoBERTa") >>> model = AutoModelForMaskedLM.from_pretrained("d42kw01f/Tamil-RoBERTa") >>> fill_mask = pipeline('fill-mask', model=model, tokenizer=tokenizer) >>> fill_mask("நான் வீட்டு <mask>.") ```
{}
fill-mask
d42kw01f/Tamil-RoBERTa
[ "transformers", "pytorch", "roberta", "fill-mask", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us
# Description: This is a smaller per-trained model on Tamil Language using Masked Language Modeling(MLM). And the model is trained on Oscar Tamil dataset. # How to Use: The model can be used directly with a pipeline for masked language modeling:
[ "# Description:\n\nThis is a smaller per-trained model on Tamil Language using Masked Language Modeling(MLM). And the model is trained on Oscar Tamil dataset.", "# How to Use:\nThe model can be used directly with a pipeline for masked language modeling:" ]
[ "TAGS\n#transformers #pytorch #roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n", "# Description:\n\nThis is a smaller per-trained model on Tamil Language using Masked Language Modeling(MLM). And the model is trained on Oscar Tamil dataset.", "# How to Use:\nThe model can be used directly with a pipeline for masked language modeling:" ]
[ 37, 37, 22 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n# Description:\n\nThis is a smaller per-trained model on Tamil Language using Masked Language Modeling(MLM). And the model is trained on Oscar Tamil dataset.# How to Use:\nThe model can be used directly with a pipeline for masked language modeling:" ]
[ -0.05470907688140869, -0.05960835888981819, -0.0018824213184416294, 0.06922741234302521, 0.1661115437746048, -0.017173541709780693, 0.16198600828647614, 0.052911125123500824, 0.03793082758784294, -0.03748409450054169, 0.12512698769569397, 0.11311330646276474, 0.0008931029005907476, 0.15146887302398682, 0.03670521453022957, -0.4382869601249695, -0.0016318145208060741, -0.0216978769749403, -0.05161536857485771, 0.0969729796051979, 0.1340651959180832, -0.03427586704492569, 0.08342784643173218, 0.03197740763425827, -0.05772296339273453, 0.002233500126749277, -0.018273701891303062, -0.08558827638626099, 0.0671960860490799, -0.011822703294456005, 0.08210847526788712, -0.06231231614947319, 0.07855113595724106, -0.07380597293376923, 0.053546734154224396, -0.049269355833530426, 0.05751131847500801, -0.00848042406141758, 0.027612505480647087, -0.02117978036403656, 0.32032328844070435, 0.022043563425540924, 0.059950657188892365, 0.02203219383955002, -0.08594223111867905, -0.00498528266325593, 0.052581027150154114, -0.11161057651042938, 0.05600805580615997, 0.16920772194862366, -0.032987676560878754, 0.23485156893730164, -0.10586287826299667, 0.11652710288763046, -0.05148641765117645, -0.2313985675573349, -0.02533225528895855, 0.18577007949352264, 0.141518235206604, -0.03833044692873955, -0.0004793820553459227, 0.11216562241315842, 0.005106575321406126, 0.08826375752687454, 0.04255671799182892, -0.053260210901498795, 0.005597040057182312, -0.0713481679558754, -0.09578956663608551, 0.04906518757343292, 0.11872253566980362, -0.09924400597810745, 0.009651332162320614, -0.09273356199264526, -0.049998246133327484, 0.02407675050199032, -0.030407272279262543, 0.0006808124599047005, -0.03702101483941078, 0.0627845823764801, 0.07366258651018143, -0.01833275519311428, -0.06142260506749153, -0.012135868892073631, -0.08868918567895889, 0.14109130203723907, 0.02668546698987484, 0.06436646729707718, -0.2722378969192505, -0.024872275069355965, -0.0666184350848198, -0.08976247906684875, 0.032113008201122284, -0.07063554227352142, 0.026589589193463326, 0.051795121282339096, -0.035061728209257126, -0.13686594367027283, 0.01243554800748825, 0.011246302165091038, 0.04592601954936981, -0.009500158950686455, 0.055931925773620605, 0.03926053270697594, 0.05225742235779762, 0.1649980992078781, -0.17684471607208252, 0.042661987245082855, 0.10008089989423752, -0.055537380278110504, 0.09357725083827972, -0.058145295828580856, -0.11653009057044983, 0.00010735169053077698, -0.07217703759670258, 0.10839906334877014, -0.008310756646096706, 0.14391666650772095, -0.04451220855116844, -0.08656120300292969, -0.07378077507019043, -0.13339774310588837, -0.03369242697954178, -0.05964156985282898, -0.022208398208022118, 0.023993775248527527, 0.06734819710254669, 0.015429181046783924, -0.05234583467245102, -0.05064805597066879, -0.060958340764045715, 0.06755348294973373, -0.13854171335697174, -0.17337308824062347, -0.012320440262556076, -0.1624804437160492, 0.04091868922114372, -0.1350836455821991, -0.12184115499258041, 0.06131918728351593, 0.04719647020101547, -0.07199189811944962, -0.020836228504776955, -0.07197000831365585, 0.009832327254116535, -0.009213629178702831, 0.02849096804857254, -0.03244148567318916, -0.018144771456718445, -0.04380837455391884, 0.010624543763697147, 0.15644162893295288, -0.011263446882367134, 0.03310434892773628, -0.045585617423057556, 0.05232040956616402, -0.05976366624236107, 0.07339473068714142, -0.03805239871144295, 0.11347883939743042, -0.001873405883088708, -0.047380901873111725, -0.030809126794338226, 0.05074906349182129, 0.01743883080780506, 0.1730729341506958, -0.20529970526695251, -0.026651300489902496, 0.1307196021080017, -0.12780368328094482, -0.12306808680295944, 0.06412270665168762, -0.0072016241028904915, 0.11726126819849014, 0.01873469166457653, 0.04846130311489105, -0.03647361323237419, -0.09393813461065292, 0.10772547870874405, 0.0820944607257843, -0.11752520501613617, -0.023431673645973206, 0.10385570675134659, 0.03871864452958107, -0.06846389174461365, 0.035075198858976364, 0.01615588180720806, 0.06839988380670547, -0.09486784785985947, -0.05676453560590744, 0.003678669687360525, -0.08837009221315384, 0.010239314287900925, 0.014936699531972408, 0.11968083679676056, -0.021168813109397888, 0.002192472107708454, -0.02296590805053711, 0.10726247727870941, -0.029091045260429382, 0.030402444303035736, -0.22280986607074738, 0.04509250819683075, -0.06738890707492828, 0.010760589502751827, -0.21285323798656464, -0.010582932271063328, 0.009867189452052116, 0.09920988231897354, 0.0805811956524849, -0.038387883454561234, 0.05254087597131729, 0.04172882065176964, 0.004649993032217026, 0.09357380121946335, 0.08531971275806427, 0.04244087636470795, -0.06568510830402374, -0.11918900161981583, 0.013146919198334217, -0.08921844512224197, 0.11132655292749405, -0.030752940103411674, 0.0009372485219500959, -0.10042306035757065, 0.06139775738120079, 0.01882277987897396, 0.011669035069644451, 0.09600738435983658, 0.06283106654882431, -0.005345492158085108, -0.06425618380308151, 0.06683772802352905, -0.016582563519477844, -0.10110731422901154, 0.1629505604505539, -0.1562114655971527, 0.0994575098156929, 0.11934725940227509, -0.13983675837516785, 0.05175141245126724, 0.2010537087917328, 0.014175018295645714, 0.02152564562857151, 0.02199883759021759, 0.15309593081474304, 0.18038342893123627, 0.03463246300816536, 0.1663077175617218, -0.027172105386853218, 0.0782204121351242, -0.006506034638732672, -0.14421814680099487, -0.04551391676068306, 0.06875989586114883, 0.18055371940135956, -0.18671490252017975, 0.06675016134977341, -0.0016811555251479149, -0.010444141924381256, 0.16641713678836823, 0.069273941218853, -0.006078590638935566, -0.04627438634634018, -0.12808829545974731, 0.01999475061893463, 0.018668048083782196, -0.16872020065784454, -0.0534202866256237, 0.028000440448522568, -0.05176851153373718, 0.08970082551240921, -0.06678209453821182, -0.05043603852391243, 0.05276208370923996, 0.03353898599743843, -0.11055230349302292, 0.17110243439674377, -0.07610686123371124, 0.04360020533204079, -0.03093993104994297, -0.08706190437078476, 0.00542545085772872, 0.007926680147647858, -0.12476666271686554, 0.18095719814300537, -0.1503944993019104, -0.39813628792762756, -0.1443590372800827, -0.12273567914962769, -0.009066280908882618, 0.03629279509186745, -0.01998179219663143, -0.128207728266716, -0.10604387521743774, -0.02379244565963745, 0.0013893023133277893, -0.06909549981355667, -0.033664774149656296, 0.03187307342886925, 0.015014891512691975, -0.02118060737848282, -0.09267962723970413, -0.028702717274427414, -0.02046694979071617, -0.07742297649383545, 0.0869424119591713, -0.15517011284828186, 0.041564665734767914, 0.16822278499603271, -0.028378179296851158, 0.08042066544294357, 0.018262360244989395, 0.21476441621780396, -0.05835657939314842, -0.009393049404025078, 0.23334026336669922, 0.027489008381962776, -0.009977507404983044, 0.16735440492630005, 0.015971677377820015, -0.060625847429037094, 0.020696556195616722, -0.028628723695874214, -0.10505901277065277, -0.20048055052757263, -0.14726419746875763, -0.0724082961678505, -0.027887897565960884, -0.021650927141308784, 0.018794596195220947, 0.028393881395459175, 0.07762712240219116, 0.06966081261634827, 0.002358559286221862, -0.0007221364066936076, 0.024753529578447342, 0.027688756585121155, -0.04378032684326172, 0.118248850107193, -0.06938058137893677, -0.03068488836288452, 0.019398372620344162, -0.09559008479118347, 0.16733859479427338, 0.12140040099620819, -0.0124610411003232, 0.09341533482074738, 0.060780927538871765, 0.16136018931865692, 0.10476436465978622, -0.037302806973457336, -0.04072825610637665, -0.0014047938166186213, -0.029794931411743164, -0.026749135926365852, 0.04081879183650017, 0.04002274572849274, -0.0811694785952568, -0.036639824509620667, 0.242766335606575, -0.0018147295340895653, 0.0910331979393959, 0.04826464131474495, -0.23325997591018677, -0.026663530617952347, 0.03683413565158844, 0.050756510347127914, -0.06731332838535309, 0.02720138244330883, 0.008171254768967628, -0.11390470713376999, 0.07529740780591965, -0.0362260602414608, 0.10274410247802734, 0.021822582930326462, 0.04639635235071182, -0.10271140187978745, 0.028450652956962585, 0.04233722761273384, 0.1010417565703392, -0.22346346080303192, 0.3088914155960083, 0.02438518963754177, 0.03930164501070976, -0.048229657113552094, -0.014858769252896309, 0.07704320549964905, 0.25367555022239685, 0.1221834123134613, 0.017090747132897377, 0.020732345059514046, -0.14607270061969757, -0.09807131439447403, 0.01668313890695572, 0.05396176874637604, 0.030108265578746796, 0.051039084792137146, -0.01881863735616207, -0.02964039333164692, -0.006298616528511047, 0.00410460727289319, -0.13209199905395508, -0.0931176245212555, 0.07767115533351898, 0.059787701815366745, -0.10409180819988251, -0.04113946855068207, -0.13636983931064606, -0.028895972296595573, 0.17957332730293274, 0.05121477693319321, -0.07007133960723877, -0.05636760592460632, 0.013145795091986656, 0.09719406068325043, -0.10545169562101364, 0.06779941916465759, -0.0785512775182724, 0.0922354981303215, -0.06162041053175926, -0.08822758495807648, 0.10471931099891663, -0.0792984813451767, 0.08371701091527939, 0.03768949955701828, 0.035246215760707855, -0.006200993433594704, 0.004893151111900806, 0.051283542066812515, 0.02106725238263607, -0.01412571594119072, -0.06602934002876282, -0.1037839949131012, 0.21046972274780273, -0.05704593285918236, 0.10890568792819977, -0.1578439474105835, -0.08706425875425339, -0.05289813503623009, -0.06547291576862335, 0.2459067702293396, 0.06599624454975128, -0.08718974888324738, 0.08300311118364334, 0.11829529702663422, -0.10427607595920563, -0.2977031469345093, -0.09408951550722122, -0.057912860065698624, 0.13995979726314545, 0.07665543258190155, -0.1416826844215393, 0.024131832644343376, -0.09704822301864624, -0.0003389979829080403, -0.09147077798843384, -0.17133642733097076, -0.15646867454051971, 0.18613915145397186, 0.08867257088422775, 0.2246176153421402, -0.04760761559009552, -0.0050115566700696945, -0.10256312042474747, 0.022693870589137077, -0.03182420879602432, 0.012759603559970856, 0.1252053678035736, -0.009845070540904999, 0.0416535884141922, 0.03459608182311058, -0.028185658156871796, 0.11960572004318237, -0.06448767334222794, -0.007938604801893234, -0.10906194895505905, -0.11775942146778107, 0.0814235582947731, 0.02609417587518692, 0.1726374626159668, 0.0677919015288353, 0.00804061908274889, -0.08898533135652542, -0.08613542467355728, 0.008143571205437183, -0.01792503520846367, 0.06198234483599663, -0.026161406189203262, -0.003882448421791196, 0.07328367978334427, -0.0044461991637945175, 0.010525495745241642, -0.03434258699417114, -0.10778944194316864, 0.0006252369494177401, -0.011345941573381424, 0.23107992112636566, -0.1386377364397049, 0.039691027253866196, -0.017895476892590523, -0.0011833836324512959, 0.009180154651403427, -0.015772070735692978, -0.06031882390379906, 0.061062898486852646, -0.02788521721959114, 0.08874355256557465, 0.04307638108730316, 0.004839784000068903, 0.03142441436648369, 0.11896596848964691, 0.048598531633615494, -0.1415558159351349, -0.043278809636831284, 0.02576340176165104, 0.05438574403524399, -0.029399719089269638, 0.06887590140104294, -0.08019950985908508, -0.02382521890103817, -0.057635121047496796, 0.020634155720472336, -0.0436147004365921, 0.03840181604027748, 0.03300134092569351, 0.03224819526076317, -0.09371161460876465, 0.06165500730276108, 0.01094481348991394, -0.008821483701467514, 0.033532194793224335, 0.11835259199142456, -0.1432919055223465, -0.09383539110422134, -0.004076181445270777, 0.1691485345363617, -0.024595364928245544, -0.026416154578328133, -0.050277624279260635, -0.10046710073947906, -0.004105022177100182, 0.12006903439760208, 0.0747578889131546, 0.006321767345070839, -0.19471293687820435, -0.0748923048377037, -0.06706461310386658, 0.030234016478061676, 0.11164754629135132, -0.09732991456985474, -0.12688224017620087, 0.098687544465065, 0.09348946809768677, 0.13393887877464294, -0.09394941478967667, -0.09773881733417511, 0.008119504898786545, 0.05483822897076607, -0.07741593569517136, 0.04114016890525818, -0.09103123843669891, -0.03341449424624443, -0.012335197068750858, -0.051570385694503784, -0.06162216141819954, 0.02016761712729931, -0.07093937695026398, 0.022241219878196716, -0.08940389007329941, -0.01255261804908514, 0.03381747379899025, -0.07289872318506241, 0.06080833449959755, -0.005969781428575516, 0.04109444469213486, 0.08186545968055725, -0.10731445997953415, 0.04772002249956131, -0.09986725449562073, -0.08211629092693329, 0.03451176732778549, 0.04378829523921013, 0.06429600715637207, -0.10776594281196594, 0.00364294508472085, 0.047343913465738297, 0.06461150199174881, 0.0542486272752285, 0.08940652012825012, -0.03988073021173477, -0.05547548830509186, 0.032934125512838364, -0.009042522870004177, -0.010830452665686607, 0.03100249543786049, 0.009135144762694836, 0.08027292042970657, 0.1098882406949997, -0.07149089872837067, 0.09734955430030823, -0.054544903337955475, 0.0460338369011879, -0.10274019092321396, -0.02142753079533577, -0.042923711240291595, -0.05816500261425972, 0.0672895684838295, -0.05962096527218819, 0.19202573597431183, 0.13290423154830933, 0.06170928478240967, 0.001944023766554892, 0.13265395164489746, 0.060607973486185074, -0.01588711328804493, 0.14458538591861725, 0.10672629624605179, 0.07872603088617325, 0.02178521454334259, 0.04610838741064072, -0.006738552823662758, 0.12065097689628601, 0.01929732970893383, 0.0021590429823845625, 0.09680673480033875, 0.1155959814786911, 0.05965888872742653, 0.0991581380367279, -0.07916926592588425, -0.24029624462127686, -0.08411019295454025, 0.040987495332956314, -0.04715872183442116, -0.04448936879634857, 0.21106581389904022, -0.038809794932603836, 0.08042740821838379, -0.025360815227031708, -0.11709784716367722, -0.20076583325862885, -0.23171910643577576, -0.1031196191906929, 0.007877098396420479, 0.023358702659606934, -0.12015574425458908, -0.04092352092266083, 0.053498148918151855, 0.09491460025310516, -0.060069579631090164, 0.18100538849830627, -0.16168765723705292, -0.020975207909941673, 0.030173353850841522, -0.019834673032164574, 0.04240437597036362, 0.014735506847500801, 0.04884396120905876, -0.13155025243759155, -0.037675920873880386, 0.0163456741720438, -0.013651524670422077, -0.06703027337789536, 0.05677202716469765, -0.06693022698163986, -0.07033014297485352, -0.08201046288013458, -0.003498905338346958, 0.006890883669257164, 0.17361851036548615, 0.017633160576224327, -0.06654192507266998, -0.01823173090815544, 0.09366928040981293, 0.02194373682141304, -0.2344917505979538, -0.14181962609291077, 0.3148316442966461, 0.01497019175440073, 0.044251225888729095, -0.05854544788599014, 0.054376598447561264, -0.0972670167684555, 0.3220164179801941, 0.3822595477104187, -0.10919993370771408, -0.012421404011547565, -0.054995931684970856, 0.026529822498559952, -0.06727895140647888, 0.1466422975063324, 0.027039680629968643, 0.19114597141742706, -0.09544327855110168, 0.009551765397191048, -0.16460226476192474, -0.027596892789006233, -0.18674972653388977, -0.12213287502527237, 0.10238183289766312, -0.02843421883881092, 0.00970312487334013, 0.1377798616886139, -0.1451220065355301, 0.20307034254074097, -0.0973057970404625, -0.11064338684082031, -0.0838763639330864, -0.06061172857880592, -0.05831148102879524, 0.07210878282785416, 0.11835763603448868, -0.03170957416296005, 0.017184114083647728, 0.006176309660077095, 0.05678124353289604, -0.13912534713745117, -0.046293534338474274, 0.17125077545642853, 0.053332291543483734, 0.02395331673324108, -0.002228943631052971, 0.07690773159265518, 0.06129154562950134, 0.044584207236766815, -0.022030437365174294, 0.16790436208248138, -0.054892707616090775, 0.11412505805492401, 0.04063874110579491, 0.06762026250362396, -0.07903081178665161, -0.10690672695636749, -0.02685241959989071, -0.010762439109385014, 0.07963492721319199, -0.16988950967788696, -0.008411156944930553, 0.011996966786682606, 0.11914410442113876, -0.08325222879648209, 0.08733963966369629, 0.1527208834886551, -0.0163130946457386, -0.0600128136575222, -0.017280222848057747, 0.0455138124525547, -0.0498654805123806, -0.1399865597486496, -0.13342638313770294, -0.1232331246137619, -0.05624934658408165, -0.11700336635112762, 0.015404727309942245, -0.1885865330696106, -0.0600074902176857, -0.08385811001062393, -0.03402049466967583, -0.04960833862423897, 0.05852606147527695, 0.10655941069126129, 0.03948606923222542, -0.022349946200847626, -0.06415057182312012, 0.08333013206720352, 0.05577815696597099, -0.20567737519741058, -0.14850792288780212 ]
null
null
transformers
## About the Model An English sequence classification model, trained on MBAD Dataset to detect bias and fairness in sentences (news articles). This model was built on top of distilbert-base-uncased model and trained for 30 epochs with a batch size of 16, a learning rate of 5e-5, and a maximum sequence length of 512. - Dataset : MBAD Data - Carbon emission 0.319355 Kg | Train Accuracy | Validation Accuracy | Train loss | Test loss | |---------------:| -------------------:| ----------:|----------:| | 76.97 | 62.00 | 0.45 | 0.96 | ## Usage The easiest way is to load the inference api from huggingface and second method is through the pipeline object offered by transformers library. ```python from transformers import AutoTokenizer, TFAutoModelForSequenceClassification from transformers import pipeline tokenizer = AutoTokenizer.from_pretrained("d4data/bias-detection-model") model = TFAutoModelForSequenceClassification.from_pretrained("d4data/bias-detection-model") classifier = pipeline('text-classification', model=model, tokenizer=tokenizer) # cuda = 0,1 based on gpu availability classifier("The irony, of course, is that the exhibit that invites people to throw trash at vacuuming Ivanka Trump lookalike reflects every stereotype feminists claim to stand against, oversexualizing Ivanka’s body and ignoring her hard work.") ``` ## Author This model is part of the Research topic "Bias and Fairness in AI" conducted by Deepak John Reji, Shaina Raza. If you use this work (code, model or dataset), please star at: > Bias & Fairness in AI, (2022), GitHub repository, <https://github.com/dreji18/Fairness-in-AI>
{"language": ["en"], "tags": ["Text Classification"], "co2_eq_emissions": 0.319355, "widget": [{"text": "Nevertheless, Trump and other Republicans have tarred the protests as havens for terrorists intent on destroying property.", "example_title": "Biased example 1"}, {"text": "Billie Eilish issues apology for mouthing an anti-Asian derogatory term in a resurfaced video.", "example_title": "Biased example 2"}, {"text": "Christians should make clear that the perpetuation of objectionable vaccines and the lack of alternatives is a kind of coercion.", "example_title": "Biased example 3"}, {"text": "There have been a protest by a group of people", "example_title": "Non-Biased example 1"}, {"text": "While emphasizing he\u2019s not singling out either party, Cohen warned about the danger of normalizing white supremacist ideology.", "example_title": "Non-Biased example 2"}]}
text-classification
d4data/bias-detection-model
[ "transformers", "tf", "distilbert", "text-classification", "Text Classification", "en", "co2_eq_emissions", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #tf #distilbert #text-classification #Text Classification #en #co2_eq_emissions #autotrain_compatible #endpoints_compatible #has_space #region-us
About the Model --------------- An English sequence classification model, trained on MBAD Dataset to detect bias and fairness in sentences (news articles). This model was built on top of distilbert-base-uncased model and trained for 30 epochs with a batch size of 16, a learning rate of 5e-5, and a maximum sequence length of 512. * Dataset : MBAD Data * Carbon emission 0.319355 Kg Usage ----- The easiest way is to load the inference api from huggingface and second method is through the pipeline object offered by transformers library. Author ------ This model is part of the Research topic "Bias and Fairness in AI" conducted by Deepak John Reji, Shaina Raza. If you use this work (code, model or dataset), please star at: > > Bias & Fairness in AI, (2022), GitHub repository, <URL > > >
[]
[ "TAGS\n#transformers #tf #distilbert #text-classification #Text Classification #en #co2_eq_emissions #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ 56 ]
[ "passage: TAGS\n#transformers #tf #distilbert #text-classification #Text Classification #en #co2_eq_emissions #autotrain_compatible #endpoints_compatible #has_space #region-us \n" ]
[ 0.0030447407625615597, 0.03561275824904442, -0.003967563156038523, 0.002684403443709016, 0.09756896644830704, 0.03305675834417343, 0.048824556171894073, 0.12801195681095123, 0.004523457959294319, -0.019541088491678238, 0.12156841158866882, 0.12630324065685272, -0.06759809702634811, 0.1782236546278, -0.126469686627388, -0.2627699673175812, 0.09936679899692535, 0.0029011210426688194, -0.09142903238534927, 0.09592537581920624, 0.13408708572387695, -0.08788714557886124, 0.07706662267446518, -0.02140623889863491, -0.14296159148216248, 0.04155931994318962, 0.005567420739680529, -0.1451527625322342, 0.12652148306369781, 0.14326654374599457, 0.17194248735904694, 0.09678243845701218, -0.05411718785762787, -0.15076880156993866, 0.039128534495830536, 0.012777740135788918, -0.15161707997322083, 0.07291942089796066, 0.006335156038403511, -0.10108063369989395, 0.05753665044903755, -0.021811941638588905, 0.026889801025390625, 0.04806036129593849, -0.14684799313545227, -0.04409598559141159, 0.019228627905249596, -0.05425991863012314, 0.05815134942531586, 0.049683649092912674, -0.021508971229195595, 0.1576307862997055, -0.19024507701396942, 0.07559267431497574, 0.09761729091405869, -0.18009695410728455, 0.002716513117775321, 0.08809498697519302, 0.01801876537501812, 0.031610842794179916, -0.08693847060203552, 0.09159732609987259, 0.08004726469516754, -0.023128308355808258, 0.06726734340190887, -0.050405681133270264, -0.13078929483890533, 0.06456542760133743, -0.11197783052921295, -0.011037250980734825, 0.26318371295928955, 0.011916805058717728, 0.03963487222790718, -0.020938511937856674, -0.1258339285850525, -0.10700535774230957, 0.017341673374176025, -0.006899161264300346, -0.021580180153250694, 0.04723193496465683, 0.046958763152360916, 0.028744667768478394, -0.17049813270568848, 0.05253027006983757, -0.22258786857128143, 0.1832467019557953, -0.06044076383113861, -0.00702377874404192, -0.13008683919906616, 0.07022372633218765, -0.046192120760679245, -0.10573878139257431, 0.014253904111683369, -0.09648748487234116, -0.05518491938710213, -0.03634701296687126, -0.08533494174480438, -0.008581886067986488, 0.06512907147407532, 0.149742990732193, 0.024938706308603287, 0.010544976219534874, 0.038877684623003006, 0.07558374106884003, 0.06862390786409378, 0.12674999237060547, -0.015392402186989784, -0.07372575253248215, -0.000713519228156656, -0.14053243398666382, -0.0033616607543081045, -0.1158163994550705, -0.18509183824062347, -0.06771791726350784, 0.03348708152770996, 0.038883455097675323, 0.025843270123004913, 0.057730235159397125, -0.07846831530332565, 0.020994039252400398, 0.12722396850585938, -0.03778969496488571, 0.03465494140982628, -0.013876206241548061, 0.07767874747514725, 0.045427385717630386, 0.03367210924625397, 0.03809591010212898, 0.018858063966035843, 0.11091658473014832, -0.08701818436384201, -0.04121566191315651, 0.010687236674129963, -0.10727278143167496, 0.036869049072265625, -0.1689022332429886, 0.06902115792036057, -0.22945454716682434, 0.05547450855374336, 0.0321432426571846, -0.0120092099532485, -0.024373561143875122, -0.03712425380945206, 0.00556494016200304, -0.06527160853147507, 0.09767980128526688, -0.02088155411183834, -0.03304490074515343, -0.09601449966430664, 0.013633857481181622, -0.06739090383052826, 0.08921010047197342, -0.21043157577514648, 0.03741598129272461, -0.04572655260562897, -0.01640142872929573, -0.14512185752391815, 0.052970338612794876, -0.018074296414852142, 0.12357403337955475, -0.040601711720228195, -0.032524507492780685, -0.09457866102457047, 0.04140644147992134, -0.05257413163781166, 0.1856638640165329, -0.15960383415222168, -0.08001638948917389, 0.051270488649606705, -0.06918253004550934, -0.10409118980169296, 0.08524760603904724, -0.0023367037065327168, 0.041258517652750015, 0.02013900689780712, 0.13877539336681366, 0.09745793789625168, -0.029503101482987404, 0.016821812838315964, 0.18721766769886017, -0.05442248657345772, 0.0009334777714684606, 0.06431844830513, 0.008717769756913185, -0.0342990905046463, -0.01926446333527565, 0.13445231318473816, 0.06892441213130951, -0.04398589953780174, -0.07752135396003723, -0.03521163761615753, 0.021037979051470757, 0.13122549653053284, -0.03826991096138954, 0.09934379160404205, -0.08157957345247269, -0.07137472927570343, 0.010078543797135353, -0.01136819552630186, 0.024818211793899536, 0.05282682925462723, -0.06836940348148346, 0.09304479509592056, -0.017679596319794655, -0.0044711134396493435, -0.16164633631706238, -0.11458364129066467, -0.03337197005748749, 0.1500028669834137, -0.011598884128034115, 0.21495462954044342, 0.01009464357048273, -0.0667935386300087, -0.042998433113098145, -0.007789708208292723, 0.08050685375928879, 0.07069359719753265, -0.10684817284345627, -0.140066996216774, 0.0818798691034317, -0.08685595542192459, 0.09800353646278381, -0.08677232265472412, 0.025231724604964256, 0.22852303087711334, 0.1222604364156723, 0.016582446172833443, 0.12246667593717575, 0.015323867090046406, 0.0042299251072108746, -0.12013501673936844, -0.0002012588083744049, 0.0369831882417202, -0.026491466909646988, -0.1997324675321579, 0.18632403016090393, -0.13404302299022675, 0.29689866304397583, 0.2191237062215805, -0.19935975968837738, -0.06336968392133713, 0.01684598997235298, -0.009815037250518799, 0.044170476496219635, -0.026177553460001945, -0.08391810953617096, 0.03871358186006546, -0.06773178279399872, 0.11119136214256287, -0.047754764556884766, -0.06005491316318512, 0.021524429321289062, -0.006712525617331266, -0.05856316164135933, 0.08771272003650665, 0.05071290209889412, -0.2232901006937027, 0.1669207215309143, 0.3063679039478302, 0.03229040279984474, 0.14505162835121155, 0.018494969233870506, 0.010525359772145748, 0.07153989374637604, -0.07009445875883102, -0.0227091945707798, 0.02983943559229374, -0.16274456679821014, -0.0012598182074725628, 0.08289109915494919, 0.025264602154493332, 0.05292427912354469, -0.13099448382854462, -0.04610361158847809, 0.05480000004172325, 0.037074074149131775, -0.0657232254743576, 0.09920943528413773, 0.015757234767079353, 0.13252784311771393, 0.003881894052028656, -0.10069962590932846, 0.08822001516819, 0.016515975818037987, -0.10068555176258087, 0.2102946639060974, -0.16477222740650177, -0.357083261013031, -0.11167983710765839, -0.03467564657330513, 0.04270555078983307, 0.0764370933175087, 0.1251467615365982, -0.13060159981250763, -0.08809252083301544, -0.03781290352344513, 0.04193385690450668, -0.06754764169454575, 0.14487019181251526, -0.07124979794025421, 0.023656301200389862, -0.07252122461795807, -0.09728502482175827, -0.03742346540093422, 0.020209575071930885, 0.06277142465114594, 0.17064312100410461, -0.12329518049955368, 0.1238238736987114, 0.21825431287288666, -0.03671414405107498, -0.009631489403545856, -0.03552759066224098, 0.26180997490882874, -0.15508738160133362, 0.06448381394147873, 0.10212166607379913, -0.04427342116832733, 0.02132003754377365, 0.17291635274887085, -0.0013729145284742117, -0.08028256148099899, 0.06548487395048141, -0.014059865847229958, -0.11808150261640549, -0.1972600817680359, -0.12737207114696503, -0.11811349540948868, 0.004956601653248072, 0.00017553767247591168, 0.08646691590547562, 0.15346549451351166, 0.06403986364603043, -0.004370761103928089, 0.019686032086610794, -0.0355847105383873, 0.06621620804071426, 0.1664443165063858, -0.0006922989850863814, 0.1394825130701065, -0.08571051061153412, -0.13044239580631256, 0.13673073053359985, 0.009259702637791634, 0.06393202394247055, 0.12305624037981033, 0.07061677426099777, -0.004863244481384754, -0.0020029842853546143, 0.18924979865550995, 0.10957886278629303, 0.06180952861905098, -0.045447930693626404, -0.06153848022222519, -0.028836525976657867, 0.03836716338992119, 0.108271025121212, 0.12745170295238495, -0.1560208797454834, -0.04737381264567375, -0.10080402344465256, 0.14455485343933105, -0.019931213930249214, 0.10663782805204391, -0.19318506121635437, 0.011806810274720192, 0.12121964991092682, -0.031210802495479584, -0.05931861326098442, 0.018893906846642494, 0.08111871033906937, -0.11036419868469238, 0.08813466876745224, 0.04075728356838226, 0.09476571530103683, 0.03681543096899986, 0.07068170607089996, -0.07353818416595459, -0.15256674587726593, -0.03304918855428696, 0.06361258029937744, -0.28317368030548096, 0.22242747247219086, 0.028691334649920464, -0.13894613087177277, -0.07815413177013397, -0.010876333341002464, 0.020346343517303467, 0.2730488181114197, 0.10298702865839005, 0.018780041486024857, -0.05589989200234413, -0.07048134505748749, -0.03989514335989952, 0.011784877628087997, 0.12061217427253723, -0.07507295161485672, -0.00958926323801279, -0.007205376401543617, -0.015852494165301323, 0.06555011868476868, 0.09080211818218231, 0.0016149356961250305, -0.18333517014980316, 0.10431662946939468, 0.05640332028269768, 0.01732015050947666, 0.009756878949701786, -0.07158699631690979, -0.11494877189397812, 0.16323480010032654, -0.1392003446817398, -0.034571465104818344, -0.17626701295375824, -0.07591486722230911, 0.08612381666898727, -0.04979781433939934, 0.08264794945716858, -0.03998352587223053, 0.06358560919761658, -0.06412811577320099, -0.14935876429080963, 0.19121810793876648, -0.09406676888465881, -0.08808605372905731, -0.048479728400707245, 0.11985311657190323, -0.089913010597229, 0.08472684025764465, 0.018271273002028465, 0.041836533695459366, -0.11856237798929214, -0.13168443739414215, -0.004295655060559511, -0.04938856512308121, 0.06499574333429337, 0.04349742829799652, -0.03333562985062599, -0.06113317981362343, 0.041255708783864975, -0.005855597089976072, 0.24694912135601044, 0.17752128839492798, -0.10136143863201141, 0.13553497195243835, 0.028455577790737152, -0.01752307452261448, -0.3080705404281616, 0.006983303930610418, -0.1178988441824913, 0.015510481782257557, -0.05478774011135101, -0.09479895979166031, 0.07188025116920471, 0.017448028549551964, -0.027463743463158607, 0.06794582307338715, -0.13677293062210083, -0.139754056930542, 0.14790430665016174, -0.08356745541095734, 0.3058595657348633, -0.10392998158931732, -0.03309420868754387, -0.0963994711637497, -0.1358947455883026, 0.1964462548494339, -0.1476818025112152, 0.08089374005794525, 0.04437883570790291, -0.0002093646617140621, 0.03678297623991966, 0.004117813427001238, 0.20785404741764069, 0.020674627274274826, 0.003784669330343604, -0.060718901455402374, -0.15030832588672638, 0.07562144100666046, -0.01713521219789982, -0.02440984733402729, -0.044905804097652435, 0.0037609809078276157, -0.15431824326515198, -0.01961950771510601, -0.028383711352944374, 0.09763988107442856, -0.00565706891939044, -0.08623595535755157, -0.08437298238277435, -0.009945720434188843, 0.04380376264452934, -0.04918806254863739, 0.33725568652153015, -0.07437264174222946, 0.13758453726768494, 0.08907350152730942, 0.18406818807125092, -0.14467453956604004, 0.07122263312339783, 0.015623977407813072, -0.058786433190107346, 0.06980963796377182, -0.2144833505153656, 0.07477838546037674, 0.09643726050853729, -0.06015926972031593, 0.06263947486877441, 0.12511736154556274, 0.027230622246861458, -0.03552500158548355, 0.21577046811580658, -0.18429940938949585, -0.0004304990288801491, -0.06221403554081917, -0.07365942746400833, 0.028211914002895355, 0.05859215557575226, 0.13802923262119293, 0.03212606906890869, -0.0037002742756158113, 0.02950635552406311, -0.0451701395213604, -0.053093887865543365, 0.08158683776855469, 0.0961579978466034, 0.06430510431528091, -0.11676790565252304, 0.03829602152109146, 0.06230535730719566, -0.09062439948320389, -0.019098615273833275, 0.07243547588586807, -0.13427524268627167, -0.14091700315475464, 0.005657303147017956, 0.038999609649181366, -0.1935570389032364, -0.0175799410790205, -0.03651077672839165, -0.1312839388847351, 0.059568677097558975, 0.31135818362236023, 0.1110461950302124, 0.10825495421886444, -0.012085487134754658, -0.04572417587041855, 0.03515525534749031, -0.042650721967220306, -0.0748450756072998, 0.04857912287116051, -0.10559840500354767, 0.058037545531988144, -0.03077862598001957, 0.15600989758968353, -0.11488500237464905, -0.011271881870925426, -0.1685502529144287, -0.020894385874271393, -0.023338738828897476, -0.012252120301127434, -0.032317329198122025, -0.044740986078977585, -0.0025991806760430336, -0.08191374689340591, -0.059559416025877, -0.047004468739032745, -0.11367763578891754, 0.01688387431204319, -0.03636320307850838, 0.06981465965509415, -0.04162442684173584, -0.03376287221908569, 0.06220081076025963, -0.049075596034526825, 0.12682709097862244, 0.06035744771361351, -0.048887692391872406, 0.08230532705783844, -0.11096751689910889, -0.06472818553447723, 0.15636053681373596, 0.004200085531920195, 0.09521651268005371, 0.016190115362405777, 0.0030111323576420546, 0.0023956119548529387, 0.030156200751662254, 0.07145892828702927, 0.0013793717371299863, -0.08792637288570404, 0.07917313277721405, -0.11488862335681915, -0.17333751916885376, -0.049340520054101944, -0.05015132948756218, 0.09904062747955322, -0.025797497481107712, 0.10273797065019608, 0.002979304175823927, 0.030520223081111908, -0.02644912153482437, 0.04976718872785568, -0.04649771749973297, -0.16655005514621735, -0.06230216845870018, -0.05695953220129013, 0.02447040192782879, -0.04736072197556496, 0.26963868737220764, 0.0264127254486084, 0.06341100484132767, 0.07939191162586212, 0.05949743837118149, 0.027587521821260452, 0.042532969266176224, 0.17856641113758087, 0.12926986813545227, -0.042389366775751114, -0.08185673505067825, 0.04509705677628517, 0.07694289833307266, -0.044307611882686615, 0.16385933756828308, -0.004745738115161657, -0.028212793171405792, 0.1619081199169159, -0.05489213019609451, -0.06615646928548813, -0.0660858079791069, -0.06144105643033981, -0.02606404386460781, 0.030002271756529808, 0.002137415576726198, -0.05653412267565727, 0.15304559469223022, -0.08933818340301514, 0.0289126206189394, -0.020964285358786583, -0.03956152871251106, -0.20027826726436615, -0.006761083845049143, -0.10913234204053879, -0.10122769325971603, -0.039936549961566925, -0.11713333427906036, -0.021713102236390114, 0.023173460736870766, 0.05882539227604866, -0.007694044150412083, 0.09698870778083801, -0.03018326312303543, -0.06540556997060776, 0.07066706568002701, 0.01935192197561264, 0.07416235655546188, -0.0801234245300293, -0.03490458428859711, -0.07598550617694855, 0.029604848474264145, -0.07977797091007233, 0.007197295781224966, -0.01153017207980156, 0.003562238300219178, -0.11733362823724747, -0.09972032159566879, -0.04984261095523834, 0.03845733031630516, -0.05379603058099747, 0.07126396149396896, 0.007890703156590462, -0.005018569063395262, 0.0641038790345192, 0.202226459980011, -0.09354110062122345, -0.031231531873345375, -0.08412358909845352, 0.13310369849205017, 0.03651483356952667, 0.15110063552856445, -0.009868825785815716, -0.07745393365621567, -0.1194368228316307, 0.16777972877025604, 0.2668342888355255, -0.09160645306110382, 0.09658266603946686, 0.023202989250421524, 0.03739716112613678, 0.14158862829208374, 0.053392939269542694, 0.08685697615146637, 0.15230131149291992, -0.0682278648018837, -0.08057594299316406, -0.002351611154153943, -0.0032608509063720703, -0.039066724479198456, 0.12082313001155853, 0.06041831150650978, -0.05469052121043205, -0.050314761698246, 0.05663662776350975, -0.14111703634262085, 0.02875680848956108, 0.008586859330534935, -0.26574620604515076, -0.07403090596199036, -0.00399219011887908, 0.02810146100819111, -0.03931616246700287, 0.07404740899801254, -0.03388611227273941, -0.1369231641292572, -0.01792093552649021, 0.054841794073581696, -0.2116173356771469, 0.018501151353120804, 0.07188821583986282, -0.04803735390305519, -0.02235432341694832, -0.03459685295820236, 0.009584855288267136, 0.10308630764484406, 0.08371817320585251, -0.005952516105026007, 0.0967588722705841, 0.04329187795519829, -0.03360048308968544, 0.019538938999176025, 0.008373145945370197, -0.005120305344462395, -0.020674817264080048, 0.10427214205265045, -0.17236445844173431, 0.0754123106598854, -0.012249289080500603, -0.05253284052014351, -0.022776944562792778, -0.029909860342741013, -0.10842640697956085, 0.08185844123363495, 0.09163805097341537, -0.03337012603878975, -0.0452372282743454, 0.0033983378671109676, -0.0004710262001026422, 0.016706841066479683, -0.13970206677913666, -0.1083764135837555, -0.05217093974351883, -0.01956351287662983, 0.11455786973237991, 0.02631353959441185, -0.09206452965736389, -0.021382393315434456, -0.07599169760942459, 0.08621811866760254, -0.10824082046747208, 0.08129555732011795, 0.04368048161268234, -0.007121002767235041, -0.031804971396923065, -0.1065235584974289, 0.04479262977838516, 0.04491589963436127, -0.1251758188009262, -0.09748663008213043 ]
null
null
spacy
## About the Model This model is trained on MBAD Dataset to recognize the biased word/phrases in a sentence. This model was built on top of roberta-base offered by Spacy transformers. This model is in association with https://huggingface.co/d4data/bias-detection-model | Feature | Description | | --- | --- | | **Name** | `Bias Recognizer Model` | | **Version** | `1.0` | | **spaCy** | `>=3.2.1,<3.3.0` | | **Default Pipeline** | `transformer`, `ner` | | **Components** | `transformer`, `ner` | ## Author This model is part of the Research topic "Bias and Fairness in AI" conducted by Deepak John Reji, Shaina Raza. If you use this work (code, model or dataset), please star at: > Bias & Fairness in AI, (2022), GitHub repository, <https://github.com/dreji18/Fairness-in-AI>
{"language": ["en"], "tags": ["spacy", "token-classification"], "widget": [{"text": "Billie Eilish issues apology for mouthing an anti-Asian derogatory term in a resurfaced video.", "example_title": "Biased example 1"}, {"text": "Christians should make clear that the perpetuation of objectionable vaccines and the lack of alternatives is a kind of coercion.", "example_title": "Biased example 2"}, {"text": "But, whether this switch constitutes a true win for the racist right or not, it\u2019s clear that MAGA conservatives are highly attuned to how decisions are made in the White House and which positions they want to control.", "example_title": "Biased example 3"}, {"text": "The fact that the abortion rate among American blacks is far higher than the rate for whites is routinely chronicled and mourned.", "example_title": "Biased example 4"}]}
token-classification
d4data/en_pipeline
[ "spacy", "token-classification", "en", "model-index", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #spacy #token-classification #en #model-index #region-us
About the Model --------------- This model is trained on MBAD Dataset to recognize the biased word/phrases in a sentence. This model was built on top of roberta-base offered by Spacy transformers. This model is in association with URL Author ------ This model is part of the Research topic "Bias and Fairness in AI" conducted by Deepak John Reji, Shaina Raza. If you use this work (code, model or dataset), please star at: > > Bias & Fairness in AI, (2022), GitHub repository, <URL > > >
[]
[ "TAGS\n#spacy #token-classification #en #model-index #region-us \n" ]
[ 21 ]
[ "passage: TAGS\n#spacy #token-classification #en #model-index #region-us \n" ]
[ -0.004639077931642532, 0.08970917761325836, -0.011319623328745365, -0.04456602409482002, 0.12429428845643997, 0.08450141549110413, 0.12109387665987015, 0.02920302376151085, 0.2465081661939621, -0.04450484365224838, 0.1075228676199913, 0.05212441086769104, -0.012718789279460907, 0.10401767492294312, -0.04577607288956642, -0.28677794337272644, 0.08648546040058136, -0.036799605935811996, -0.02966843545436859, 0.08118105679750443, -0.0011295424774289131, -0.09963174164295197, 0.04967942088842392, -0.05129582807421684, -0.05529729276895523, 0.05331800878047943, 0.02475965954363346, -0.05978570133447647, 0.07705297321081161, -0.05064434930682182, 0.20552772283554077, 0.03326057270169258, -0.013609850779175758, -0.23744425177574158, 0.026697730645537376, 0.013615494593977928, -0.05840002000331879, 0.07731287181377411, 0.09350038319826126, -0.06338433176279068, 0.0036208995152264833, 0.0380440317094326, 0.04783717170357704, 0.04987424239516258, -0.21513299643993378, -0.11451633274555206, -0.009212800301611423, -0.04623694717884064, 0.0597614087164402, -0.038287725299596786, 0.011695273220539093, 0.1609041839838028, -0.20021720230579376, -0.006134488619863987, 0.02089294046163559, -0.16215701401233673, 0.024579916149377823, 0.22164592146873474, -0.012725578621029854, 0.10176205635070801, -0.01484577264636755, 0.10037428885698318, 0.049146804958581924, -0.032992903143167496, -0.037102993577718735, -0.03915470466017723, 0.0082062017172575, 0.10865304619073868, -0.09575759619474411, -0.045462388545274734, 0.2971842288970947, 0.06617727875709534, 0.04645711928606033, 0.04059813916683197, -0.03266412392258644, -0.015032447874546051, 0.02249535731971264, -0.016940826550126076, 0.04289291054010391, 0.10213420540094376, 0.22040718793869019, -0.001251775654964149, -0.09906679391860962, 0.0016000795876607299, -0.19151072204113007, 0.2202644795179367, -0.015638789162039757, 0.10112698376178741, -0.25754842162132263, -0.0145059023052454, -0.12953519821166992, -0.05042881518602371, 0.06190278008580208, -0.14057298004627228, -0.061582960188388824, -0.048871591687202454, 0.03981878608465195, 0.04016279801726341, 0.07962732017040253, 0.09794528782367706, -0.047820210456848145, 0.08675193786621094, 0.00498546939343214, 0.12064965069293976, 0.1089821457862854, 0.09155845642089844, 0.0026791237760335207, 0.03822578489780426, -0.04822041466832161, -0.09568440169095993, -0.03676281496882439, -0.09267649054527283, -0.08337756991386414, 0.00406569242477417, -0.050524722784757614, 0.09551965445280075, 0.01818021945655346, -0.035483673214912415, -0.1067454144358635, -0.016831988468766212, 0.04870854690670967, -0.022051312029361725, -0.00019913226424250752, 0.020300492644309998, -0.013283330015838146, 0.0780373215675354, -0.13371358811855316, 0.042349860072135925, 0.0634743794798851, 0.06287611275911331, -0.1173018366098404, -0.011305668391287327, -0.04903577268123627, -0.024753671139478683, 0.023300256580114365, -0.14344896376132965, 0.05461607873439789, -0.0731770396232605, -0.09850872308015823, 0.033152904361486435, 0.04465179890394211, -0.04716750979423523, 0.06489233672618866, -0.0030849126633256674, -0.0019773407839238644, -0.0427132211625576, -0.027221566066145897, -0.06601788848638535, -0.11597322672605515, 0.01127252820879221, -0.08940979093313217, 0.025642456486821175, -0.11787502467632294, 0.010194002650678158, -0.08979786187410355, 0.08618950098752975, -0.13546229898929596, 0.02606111951172352, -0.05601274594664574, 0.17304077744483948, -0.025323886424303055, 0.022280439734458923, -0.13649441301822662, 0.018360543996095657, -0.09457273781299591, 0.21585491299629211, -0.15958209335803986, -0.08008117973804474, 0.08282697945833206, -0.12750215828418732, -0.06157277524471283, 0.008730960078537464, -0.008921684697270393, 0.033555518835783005, 0.06292471289634705, 0.36872267723083496, -0.02239096909761429, 0.028824705630540848, 0.05541718378663063, 0.12799608707427979, -0.15313154458999634, -0.09711054712533951, 0.09516551345586777, -0.03142009302973747, -0.12918882071971893, 0.00782070867717266, 0.016833297908306122, 0.03994018957018852, -0.0839722752571106, -0.037850163877010345, 0.0670732706785202, -0.011009217239916325, 0.13354110717773438, 0.06073354184627533, 0.09307627379894257, -0.03658507019281387, 0.06287404894828796, 0.04370417073369026, 0.0390278659760952, 0.10642722249031067, -0.07067322731018066, -0.055674344301223755, 0.10760679841041565, -0.05291156470775604, -0.016980277374386787, -0.17015394568443298, -0.1337946057319641, 0.02425357885658741, 0.08296997100114822, -0.007017239462584257, 0.13200882077217102, 0.08337831497192383, -0.09470215439796448, 0.015831025317311287, 0.004940938204526901, 0.10201597213745117, 0.056256361305713654, -0.012591895647346973, -0.14045414328575134, 0.013452124781906605, -0.08189942687749863, -0.03761918470263481, -0.13976970314979553, 0.0022865014616400003, 0.09205739945173264, 0.09351968765258789, 0.058606091886758804, 0.04304945841431618, 0.048122573643922806, 0.004158638883382082, 0.02548038959503174, -0.020121674984693527, 0.0848366916179657, -0.09043246507644653, -0.0689830556511879, 0.05973832681775093, -0.0394069068133831, 0.25615110993385315, 0.13620048761367798, -0.21579033136367798, 0.033656034618616104, -0.05090385302901268, 0.022952359169721603, 0.04566274583339691, 0.02530262991786003, 0.020270781591534615, 0.05855812877416611, 0.013455922715365887, 0.036737214773893356, -0.030585456639528275, -0.001125612878240645, -0.021057093515992165, -0.028102826327085495, -0.12819987535476685, 0.12933386862277985, 0.17221443355083466, -0.2023140788078308, 0.13022613525390625, 0.26860547065734863, 0.15136657655239105, 0.1639244109392166, -0.05466530844569206, -0.0021928204223513603, 0.029909050092101097, -0.05984814465045929, -0.07768815755844116, 0.11841558665037155, -0.16073152422904968, -0.028791308403015137, 0.034306325018405914, 0.040091656148433685, 0.06906545907258987, -0.12994487583637238, -0.11471781134605408, 0.003427002113312483, 0.0026880980003625154, -0.16119849681854248, 0.021833928301930428, -0.013583927415311337, 0.0749448612332344, 0.037585701793432236, -0.10483847558498383, 0.06550408154726028, -0.04793604835867882, -0.06524862349033356, 0.14126069843769073, -0.10965844988822937, -0.15786713361740112, -0.04928029701113701, -0.04247290641069412, 0.008099687285721302, 0.03575556352734566, -0.041765131056308746, -0.18511976301670074, -0.024054406210780144, 0.04566461965441704, 0.025784475728869438, -0.19681578874588013, 0.028612909838557243, -0.03096773475408554, 0.11105889081954956, -0.11762578785419464, -0.007232296280562878, -0.05351674184203148, -0.10171082615852356, 0.1137947291135788, 0.08427844941616058, -0.16281627118587494, 0.03662322089076042, 0.22456440329551697, 0.021948300302028656, 0.06868069618940353, 0.0017532092751935124, 0.17346027493476868, -0.12610088288784027, -0.03955264762043953, 0.11882825940847397, -0.03700146824121475, 0.039294857531785965, 0.20534248650074005, 0.0891692191362381, -0.09006188064813614, -0.06507369130849838, -0.03757299482822418, -0.11004941910505295, -0.1988985389471054, -0.10496654361486435, -0.06669024378061295, 0.06219172105193138, 0.009807383641600609, 0.030124573037028313, 0.08064200729131699, 0.03555005043745041, 0.09797866642475128, -0.0818861871957779, -0.04048384726047516, -0.013325564563274384, 0.044969938695430756, -0.02216709777712822, 0.023401616141200066, -0.05851900950074196, -0.07130870223045349, 0.0966208353638649, 0.03682345896959305, 0.12775343656539917, 0.24501833319664001, 0.009995551779866219, 0.022519417107105255, 0.03113345056772232, 0.18061673641204834, 0.08198492228984833, 0.06316947191953659, -0.026204030960798264, -0.033824242651462555, -0.012286003679037094, 0.0019295847741886973, 0.04636871814727783, 0.10154115408658981, -0.18859173357486725, -0.030302047729492188, -0.1503351628780365, 0.07390563935041428, -0.08441091328859329, 0.13698714971542358, -0.12238597124814987, 0.10013417899608612, 0.12533064186573029, 0.03783004358410835, -0.03326953575015068, 0.09254541248083115, 0.01604602299630642, -0.10436208546161652, 0.06574864685535431, 0.05468543991446495, 0.12417032569646835, -0.07954820990562439, 0.08784982562065125, -0.05414898321032524, -0.06979965418577194, -0.0021282124798744917, 0.06781022250652313, -0.03815697133541107, 0.34460777044296265, 0.03927534818649292, -0.1322224885225296, -0.0680960938334465, -0.07073813676834106, 0.04339425638318062, 0.24450771510601044, 0.13076695799827576, 0.045797597616910934, -0.1860412210226059, -0.1423356831073761, -0.011661691591143608, 0.01112963818013668, 0.09137939661741257, -0.08778875321149826, -0.09258304536342621, 0.012080238200724125, 0.038386955857276917, -0.02470325492322445, -0.02574688009917736, 0.013626595959067345, -0.010869069024920464, 0.0028962241485714912, -0.0024593377020210028, -0.09494844079017639, 0.04063175991177559, -0.008970282040536404, -0.0426364429295063, 0.027255645021796227, 0.015760518610477448, -0.0477132610976696, -0.0862787738442421, -0.11995428055524826, 0.12820406258106232, -0.05647685378789902, -0.031126320362091064, -0.04853896051645279, -0.0242247823625803, -0.008871877565979958, -0.17192263901233673, 0.1168031245470047, -0.04483760893344879, 0.06089157983660698, -0.07858055830001831, 0.08671921491622925, -0.06555803120136261, 0.03196974843740463, -0.006401605438441038, 0.06907150149345398, -0.026392577216029167, -0.10749984532594681, 0.13508334755897522, -0.0888948142528534, 0.00805789977312088, 0.15703195333480835, 0.009936179034411907, 0.01910584792494774, 0.022369487211108208, 0.01389949582517147, 0.1242658719420433, 0.3219046890735626, -0.050361569970846176, 0.10433299839496613, 0.19669955968856812, -0.055160827934741974, -0.23867636919021606, -0.004658831749111414, -0.18856526911258698, -0.026906609535217285, 0.08217744529247284, -0.207087442278862, 0.12991946935653687, 0.07129856199026108, -0.06354179233312607, 0.1380489021539688, -0.15917861461639404, -0.01845533214509487, 0.19913698732852936, -0.06235973909497261, 0.4086400270462036, -0.09474170953035355, -0.11853466928005219, -0.02225828729569912, -0.03049946017563343, 0.13720571994781494, -0.0674581304192543, 0.01762324571609497, 0.01025628112256527, -0.06496995687484741, 0.05425776168704033, -0.022221975028514862, 0.23985937237739563, 0.015817392617464066, 0.09306802600622177, -0.050412751734256744, -0.26644742488861084, 0.09111132472753525, -0.020904146134853363, -0.05758862569928169, 0.10420601814985275, 0.007044668309390545, -0.18158294260501862, 0.009055105037987232, -0.04800739884376526, -0.0020014476031064987, 0.05716131255030632, -0.07191526144742966, -0.05414356663823128, 0.045388806611299515, -0.08195486664772034, 0.0025779185816645622, 0.28824156522750854, -0.04370611533522606, 0.11045469343662262, 0.016353119164705276, 0.10725726932287216, -0.14538061618804932, 0.014544065110385418, -0.05307826027274132, -0.028697047382593155, 0.07712770253419876, -0.1837805211544037, 0.011848709546029568, 0.13038933277130127, -0.026828432455658913, 0.04118085280060768, 0.1095285639166832, -0.019497469067573547, -0.047768451273441315, 0.18089185655117035, -0.135270893573761, -0.12138840556144714, -0.04492122307419777, -0.11153297126293182, 0.08224985003471375, 0.038569483906030655, 0.04619360342621803, 0.08390957117080688, 0.008496037684381008, 0.030369482934474945, -0.016053883358836174, -0.0688047856092453, -0.007411373779177666, 0.08882346004247665, 0.030600685626268387, -0.09679323434829712, 0.14854951202869415, 0.09886171668767929, 0.043697789311409, -0.0855928435921669, 0.10585042834281921, -0.11350740492343903, -0.07684192806482315, -0.10012117028236389, 0.15077625215053558, -0.03609536588191986, -0.043035950511693954, -0.026238704100251198, -0.09582386910915375, -0.014815926551818848, 0.1275714486837387, 0.07660308480262756, 0.0581178292632103, -0.001612194231711328, -0.07449325919151306, 0.12076675146818161, -0.08139422535896301, -0.10948988050222397, -0.05866561084985733, -0.1813315600156784, -0.08627350628376007, -0.026198316365480423, 0.1304904967546463, -0.09884200990200043, -0.10558386147022247, -0.20356711745262146, 0.06893017143011093, -0.08658204972743988, -0.05228438228368759, -0.04287975654006004, -0.04149717837572098, 0.017714638262987137, -0.06983821094036102, -0.0608670711517334, -0.07395796477794647, -0.16240860521793365, 0.0937165841460228, 0.043504081666469574, 0.09737872332334518, 0.008909969590604305, -0.034158989787101746, 0.12932902574539185, -0.00030033348593860865, 0.09372983872890472, 0.07952453941106796, 0.00713045010343194, 0.12905433773994446, -0.10803424566984177, -0.06866864860057831, 0.09711814671754837, -0.001475788070820272, 0.06869221478700638, 0.06021283194422722, -0.04274054616689682, -0.023911623284220695, -0.008821733295917511, 0.10030528903007507, -0.109192855656147, -0.06137405335903168, -0.022272862493991852, -0.016095686703920364, -0.18659444153308868, 0.02772687003016472, -0.09598401933908463, 0.18725505471229553, -0.006836308166384697, 0.043538156896829605, 0.12028458714485168, 0.07629495859146118, -0.06461116671562195, 0.01902562938630581, -0.0019335830584168434, -0.13006001710891724, 0.05873112380504608, -0.015072413720190525, 0.022259483113884926, -0.04525398835539818, 0.3661469519138336, 0.08485667407512665, -0.019175685942173004, 0.033395882695913315, 0.1745140701532364, 0.035372499376535416, 0.09521686285734177, 0.1510782241821289, 0.13052740693092346, -0.05577104538679123, -0.038019370287656784, 0.04543108865618706, -0.009173748083412647, 0.008390256203711033, 0.13104186952114105, 0.11407414078712463, -0.0039073084481060505, 0.039457883685827255, 0.03247899189591408, 0.009522458538413048, 0.0145067498087883, -0.04954832047224045, 0.019522642716765404, 0.01868300326168537, 0.027216164395213127, -0.008242812938988209, 0.145968496799469, -0.06971205025911331, 0.09183305501937866, -0.046941161155700684, -0.021801888942718506, -0.17487819492816925, -0.1290074586868286, -0.022331450134515762, -0.09879794716835022, 0.038592275232076645, -0.04632844775915146, -0.0481463223695755, 0.2544984519481659, 0.04947135969996452, 0.013621468096971512, -0.01711283251643181, -0.053839243948459625, -0.03156109154224396, 0.058983925729990005, -0.006124243605881929, -0.0014668542426079512, -0.08417309075593948, -0.05869605019688606, -0.05446097254753113, -0.0335603803396225, -0.12128662317991257, -0.026858549565076828, -0.09391071647405624, -0.033354151993989944, -0.12941251695156097, -0.1083754375576973, -0.035169802606105804, 0.017676521092653275, -0.1223810613155365, 0.017707597464323044, -0.017873268574476242, 0.017050711438059807, 0.029801826924085617, 0.19557435810565948, -0.0014115041121840477, 0.06454538553953171, -0.057563018053770065, 0.12050914764404297, -0.06442353874444962, 0.1215941533446312, -0.03651757165789604, -0.06381290405988693, -0.054175443947315216, 0.2126113474369049, 0.31652209162712097, -0.09200555086135864, -0.004950030241161585, -0.019949940964579582, 0.024400673806667328, 0.09764713048934937, 0.08890847116708755, -0.035175107419490814, 0.12033504247665405, -0.08252035826444626, 0.03625717759132385, -0.00736491521820426, -0.002577448496595025, -0.0006376832607202232, 0.06541309505701065, 0.15297411382198334, -0.03292388096451759, -0.14801739156246185, 0.16778747737407684, -0.20297187566757202, 0.1639130413532257, 0.05840076506137848, -0.19709526002407074, -0.09321457147598267, -0.05363927036523819, 0.02414136379957199, -0.049636438488960266, 0.09114091843366623, -0.0883140116930008, -0.13986757397651672, -0.18235722184181213, 0.03678620606660843, -0.3016493618488312, -0.1992723047733307, 0.09568007290363312, 0.13831616938114166, 0.1138654351234436, 0.0008787679835222661, 0.028696678578853607, 0.014323273673653603, -0.0016361117595806718, -0.01864772103726864, -0.04252839460968971, 0.018433915451169014, 0.038329001516103745, -0.14162324368953705, -0.008960560895502567, 0.004876873455941677, -0.11105360835790634, 0.1013193354010582, -0.12032376974821091, 0.007551189512014389, -0.004324350506067276, -0.11722799390554428, 0.028249593451619148, 0.022816652432084084, -0.11798198521137238, 0.03353245183825493, 0.0774940699338913, 0.014099461026489735, -0.028579184785485268, -0.002242022193968296, 0.013397528789937496, 0.009455411694943905, -0.13984155654907227, -0.13611678779125214, 0.07265457510948181, -0.06825058162212372, 0.13692006468772888, -0.033100854605436325, -0.031059378758072853, 0.0257404875010252, -0.08474133163690567, 0.0929730013012886, -0.03160342574119568, 0.044930122792720795, 0.1370825320482254, 0.02326572872698307, -0.023153288289904594, -0.13521593809127808, 0.09936809539794922, 0.0066788457334041595, -0.04449629783630371, -0.07006555795669556 ]
null
null
transformers
## About the Model An Environmental due diligence classification model, trained on customized environmental Dataset to detect contamination and remediation activities (both prevailing as well as planned) as a part of site assessment process. This model can identify the source of contamination, the extent of contamination, the types of contaminants present at the site, the flow of contaminants and their interaction with ground water, surface water and other surrounding water bodies . This model was built on top of distilbert-base-uncased model and trained for 10 epochs with a batch size of 16, a learning rate of 5e-5, and a maximum sequence length of 512. - Dataset : Open Source News data + Custom data - Carbon emission 0.1069 Kg ## Usage The easiest way is to load through the pipeline object offered by transformers library. ```python from transformers import AutoTokenizer, TFAutoModelForSequenceClassification from transformers import pipeline tokenizer = AutoTokenizer.from_pretrained("d4data/environmental-due-diligence-model") model = TFAutoModelForSequenceClassification.from_pretrained("d4data/environmental-due-diligence-model") classifier = pipeline('text-classification', model=model, tokenizer=tokenizer) # cuda = 0,1 based on gpu availability classifier("At the every month post-injection monitoring event, TCE, carbon tetrachloride, and chloroform concentrations were above CBSGs in three of the wells") ``` ## Author This model is part of the Research topic "Environmental Due Diligence" conducted by Deepak John Reji, Afreen Aman. If you use this work (code, model or dataset), please cite as: > Environmental Due Diligence, (2020), https://www.sciencedirect.com/science/article/pii/S2665963822001117 ## You can support me here :) <a href="https://www.buymeacoffee.com/deepakjohnreji" target="_blank"><img src="https://cdn.buymeacoffee.com/buttons/v2/default-yellow.png" alt="Buy Me A Coffee" style="height: 60px !important;width: 217px !important;" ></a>
{"language": ["en"], "tags": ["Text Classification"], "co2_eq_emissions": 0.1069, "widget": [{"text": "At the every month post-injection monitoring event, TCE, carbon tetrachloride, and chloroform concentrations were above CBSGs in three of the wells", "example_title": "Remediation Standards"}, {"text": "TRPH exceedances were observed in the subsurface soils immediately above the water table and there are no TRPH exceedances in surface soils.", "example_title": "Extent of Contamination"}, {"text": "weathered shale was encountered below the surface area with fluvial deposits. Sediments in the coastal plain region are found above and below the bedrock with sandstones and shales that form the basement rock", "example_title": "Geology"}]}
text-classification
d4data/environmental-due-diligence-model
[ "transformers", "tf", "distilbert", "text-classification", "Text Classification", "en", "co2_eq_emissions", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #tf #distilbert #text-classification #Text Classification #en #co2_eq_emissions #autotrain_compatible #endpoints_compatible #has_space #region-us
## About the Model An Environmental due diligence classification model, trained on customized environmental Dataset to detect contamination and remediation activities (both prevailing as well as planned) as a part of site assessment process. This model can identify the source of contamination, the extent of contamination, the types of contaminants present at the site, the flow of contaminants and their interaction with ground water, surface water and other surrounding water bodies . This model was built on top of distilbert-base-uncased model and trained for 10 epochs with a batch size of 16, a learning rate of 5e-5, and a maximum sequence length of 512. - Dataset : Open Source News data + Custom data - Carbon emission 0.1069 Kg ## Usage The easiest way is to load through the pipeline object offered by transformers library. ## Author This model is part of the Research topic "Environmental Due Diligence" conducted by Deepak John Reji, Afreen Aman. If you use this work (code, model or dataset), please cite as: > Environmental Due Diligence, (2020), URL ## You can support me here :) <a href="URL target="_blank"><img src="URL alt="Buy Me A Coffee" style="height: 60px !important;width: 217px !important;" ></a>
[ "## About the Model\nAn Environmental due diligence classification model, trained on customized environmental Dataset to detect contamination and remediation activities (both prevailing as well as planned) as a part of site assessment process. This model can identify the source of contamination, the extent of contamination, the types of contaminants present at the site, the flow of contaminants and their interaction with ground water, surface water and other surrounding water bodies .\n\nThis model was built on top of distilbert-base-uncased model and trained for 10 epochs with a batch size of 16, a learning rate of 5e-5, and a maximum sequence length of 512.\n\n- Dataset : Open Source News data + Custom data\n- Carbon emission 0.1069 Kg", "## Usage\nThe easiest way is to load through the pipeline object offered by transformers library.", "## Author\nThis model is part of the Research topic \"Environmental Due Diligence\" conducted by Deepak John Reji, Afreen Aman. If you use this work (code, model or dataset), please cite as:\n> Environmental Due Diligence, (2020), URL", "## You can support me here :)\n<a href=\"URL target=\"_blank\"><img src=\"URL alt=\"Buy Me A Coffee\" style=\"height: 60px !important;width: 217px !important;\" ></a>" ]
[ "TAGS\n#transformers #tf #distilbert #text-classification #Text Classification #en #co2_eq_emissions #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "## About the Model\nAn Environmental due diligence classification model, trained on customized environmental Dataset to detect contamination and remediation activities (both prevailing as well as planned) as a part of site assessment process. This model can identify the source of contamination, the extent of contamination, the types of contaminants present at the site, the flow of contaminants and their interaction with ground water, surface water and other surrounding water bodies .\n\nThis model was built on top of distilbert-base-uncased model and trained for 10 epochs with a batch size of 16, a learning rate of 5e-5, and a maximum sequence length of 512.\n\n- Dataset : Open Source News data + Custom data\n- Carbon emission 0.1069 Kg", "## Usage\nThe easiest way is to load through the pipeline object offered by transformers library.", "## Author\nThis model is part of the Research topic \"Environmental Due Diligence\" conducted by Deepak John Reji, Afreen Aman. If you use this work (code, model or dataset), please cite as:\n> Environmental Due Diligence, (2020), URL", "## You can support me here :)\n<a href=\"URL target=\"_blank\"><img src=\"URL alt=\"Buy Me A Coffee\" style=\"height: 60px !important;width: 217px !important;\" ></a>" ]
[ 56, 170, 23, 64, 57 ]
[ "passage: TAGS\n#transformers #tf #distilbert #text-classification #Text Classification #en #co2_eq_emissions #autotrain_compatible #endpoints_compatible #has_space #region-us \n## About the Model\nAn Environmental due diligence classification model, trained on customized environmental Dataset to detect contamination and remediation activities (both prevailing as well as planned) as a part of site assessment process. This model can identify the source of contamination, the extent of contamination, the types of contaminants present at the site, the flow of contaminants and their interaction with ground water, surface water and other surrounding water bodies .\n\nThis model was built on top of distilbert-base-uncased model and trained for 10 epochs with a batch size of 16, a learning rate of 5e-5, and a maximum sequence length of 512.\n\n- Dataset : Open Source News data + Custom data\n- Carbon emission 0.1069 Kg## Usage\nThe easiest way is to load through the pipeline object offered by transformers library.## Author\nThis model is part of the Research topic \"Environmental Due Diligence\" conducted by Deepak John Reji, Afreen Aman. If you use this work (code, model or dataset), please cite as:\n> Environmental Due Diligence, (2020), URL## You can support me here :)\n<a href=\"URL target=\"_blank\"><img src=\"URL alt=\"Buy Me A Coffee\" style=\"height: 60px !important;width: 217px !important;\" ></a>" ]
[ -0.04574066027998924, 0.15251892805099487, -0.006146837025880814, -0.02539285644888878, 0.010114723816514015, -0.03171269968152046, 0.07111353427171707, 0.031913161277770996, 0.02520304173231125, 0.05894694849848747, -0.001611691084690392, 0.06166495010256767, 0.048466041684150696, 0.12457770854234695, -0.017021093517541885, -0.19002367556095123, 0.040786392986774445, -0.07400886714458466, 0.010709427297115326, 0.08649180829524994, 0.12043632566928864, -0.12918032705783844, 0.10134986788034439, -0.017200924456119537, -0.017663389444351196, 0.044435981661081314, -0.06212648004293442, -0.04561123251914978, 0.07775898277759552, 0.07616012543439865, 0.15775227546691895, 0.09954197704792023, 0.022251907736063004, -0.12729273736476898, 0.03173838555812836, 0.03867589682340622, -0.018665794283151627, 0.0692150890827179, 0.06976564228534698, -0.014325466006994247, 0.014644082635641098, -0.08255019038915634, 0.04428892955183983, 0.03770134970545769, -0.07370244711637497, -0.07344599813222885, -0.020073657855391502, 0.06924056261777878, 0.1501713991165161, 0.06084741652011871, -0.026481354609131813, 0.0690721645951271, -0.12640956044197083, 0.041200730949640274, 0.10648705810308456, -0.1054440438747406, -0.02060776948928833, 0.007930640131235123, -0.006658782251179218, 0.09703019261360168, -0.07099346071481705, 0.0025130247231572866, -0.033968452364206314, -0.02034963294863701, 0.01953071914613247, -0.06726916879415512, 0.04353869706392288, -0.004493996500968933, -0.09410225600004196, 0.0326760858297348, 0.22215057909488678, 0.08564687520265579, -0.06492218375205994, -0.06726871430873871, -0.0899399146437645, 0.06730841845273972, 0.018037809059023857, -0.06647762656211853, 0.0482199601829052, -0.014450157061219215, 0.012517480179667473, -0.028959831222891808, -0.1480647623538971, 0.027323376387357712, -0.1069309189915657, 0.13914118707180023, -0.0025813847314566374, 0.03781161829829216, -0.020754894241690636, 0.12187818437814713, -0.009065881371498108, -0.09741107374429703, -0.029819482937455177, -0.06096413731575012, -0.06979299336671829, -0.03918179124593735, 0.031328748911619186, -0.06386405974626541, 0.03465050086379051, 0.1280697137117386, 0.03689050301909447, 0.030269978567957878, 0.06449033319950104, 0.00514121912419796, 0.09584126621484756, 0.12409459799528122, -0.11737014353275299, -0.0852200910449028, 0.051549870520830154, -0.04653343930840492, -0.0005668335361406207, -0.03660234063863754, -0.05486835911870003, -0.029879910871386528, -0.01031565759330988, 0.06290937960147858, 0.06494821608066559, -0.028829192742705345, -0.014769148081541061, -0.08349083364009857, 0.27164244651794434, -0.12325001507997513, 0.0375208705663681, 0.06494524329900742, -0.03894604369997978, 0.11188089102506638, 0.0184820294380188, 0.06169426068663597, -0.04122185707092285, 0.09696130454540253, -0.04201162979006767, -0.06632892042398453, -0.06422460824251175, -0.10219153016805649, -0.016350626945495605, -0.05285278335213661, -0.00043635134352371097, -0.14088381826877594, -0.08640777319669724, -0.028470244258642197, 0.07183897495269775, -0.014800316654145718, -0.06914765387773514, 0.01483566127717495, 0.017257312312722206, -0.018448470160365105, 0.04258464276790619, -0.02698396146297455, -0.012081420980393887, -0.01993568055331707, -0.09696568548679352, 0.029056215658783913, -0.053903114050626755, 0.023735998198390007, -0.051555510610342026, 0.02959594503045082, -0.21352297067642212, 0.08912523090839386, -0.011630861088633537, 0.09032371640205383, -0.1902482956647873, -0.02115107886493206, -0.08390575647354126, -0.046786919236183167, -0.019199313595891, 0.13468122482299805, -0.10097543150186539, -0.050236184149980545, -0.022790540009737015, -0.05940592288970947, 0.010742787271738052, 0.02676469087600708, -0.0007185001741163433, 0.07457577437162399, 0.1174575462937355, -0.005151744931936264, 0.16804596781730652, -0.12917518615722656, -0.041161153465509415, 0.02503793314099312, -0.0628400668501854, 0.07715930789709091, 0.09035305678844452, -0.0664013996720314, 0.0869283601641655, -0.02635079063475132, -0.004896120168268681, -0.008622057735919952, 0.015523879788815975, -0.0836298018693924, 0.018857410177588463, -0.0075299786403775215, -0.027405578643083572, -0.0811968669295311, 0.0004702853912021965, 0.03481777757406235, -0.05366146191954613, 0.011975564062595367, 0.05836804583668709, -0.023889854550361633, 0.018601810559630394, -0.08265522867441177, 0.05628732964396477, -0.020410221070051193, -0.01545432023704052, -0.1248493567109108, -0.029396001249551773, 0.061826884746551514, -0.10618068277835846, 0.009311976842582226, 0.011413322761654854, 0.040766820311546326, 0.010221855714917183, -0.016518693417310715, -0.04491233453154564, -0.04672187939286232, -0.01390603743493557, -0.03711352497339249, -0.1567113697528839, 0.044903941452503204, 0.012681938707828522, 0.09290537983179092, -0.08235049992799759, 0.02005711756646633, 0.1566419154405594, 0.09120216965675354, 0.01958981156349182, 0.01709512248635292, 0.026851294562220573, -0.0580076240003109, 0.0062921661883592606, -0.01967637799680233, -0.004842888563871384, 0.0003874910471495241, -0.10239338874816895, 0.08150386065244675, -0.09225215017795563, -0.057316020131111145, 0.06550699472427368, 0.034971754997968674, -0.046534519642591476, 0.030860159546136856, -0.04444156587123871, -0.051112789660692215, -0.04772012308239937, -0.07291451841592789, 0.16583245992660522, 0.058491095900535583, 0.019586583599448204, -0.09697184711694717, -0.06764780730009079, 0.0030652021523565054, 0.012540707364678383, -0.008537637069821358, 0.09213162958621979, 0.07055628299713135, -0.1919882893562317, 0.008316610008478165, 0.13165847957134247, -0.012346128933131695, 0.11933065205812454, 0.008573206141591072, -0.11763907968997955, -0.033383190631866455, 0.023902250453829765, -0.017126012593507767, 0.08658153563737869, 0.004207942634820938, 0.08018642663955688, 0.05204325541853905, -0.01178748533129692, 0.02835139073431492, -0.05615681782364845, 0.04825574904680252, 0.03334176540374756, -0.02369464375078678, -0.01218377985060215, 0.019213739782571793, 0.018231110647320747, 0.08686993271112442, 0.05067873373627663, -0.0028709617909044027, -0.005374182481318712, -0.040234122425317764, -0.11363348364830017, 0.21122588217258453, -0.07736280560493469, -0.19958265125751495, -0.09226103127002716, 0.09569676965475082, 0.022098137065768242, -0.04690944403409958, 0.03901106119155884, -0.052972640842199326, -0.11511509865522385, -0.08548660576343536, 0.00008481266559101641, -0.0038988685701042414, -0.02141997590661049, -0.08132874965667725, 0.0034935097210109234, -0.04034121707081795, -0.14254197478294373, 0.017313683405518532, 0.029784230515360832, -0.07917699962854385, 0.07133261114358902, -0.027561526745557785, 0.09087400138378143, 0.19381265342235565, -0.059069402515888214, -0.03287933021783829, 0.06563223153352737, 0.23536084592342377, -0.07497134059667587, 0.2157038003206253, 0.10634487867355347, -0.008686527609825134, 0.1232362762093544, 0.20019158720970154, 0.027471203356981277, 0.009610607288777828, 0.04202979430556297, 0.07293891161680222, -0.007507536560297012, -0.2683899998664856, -0.02862696535885334, -0.0012949573574587703, -0.08959247916936874, -0.022261030972003937, 0.0677807405591011, 0.04820893332362175, 0.0442638024687767, -0.06739676743745804, -0.009233905002474785, 0.06617529690265656, 0.07289603352546692, 0.012121142819523811, 0.025808095932006836, 0.05086859315633774, -0.07525299489498138, -0.046837083995342255, 0.11182648688554764, 0.02886536903679371, 0.2442592978477478, 0.05535789579153061, 0.16330991685390472, 0.08159395307302475, -0.013076916337013245, 0.06544455885887146, 0.08031322807073593, -0.03187863901257515, 0.010486578568816185, -0.023013664409518242, -0.09519580006599426, 0.12841014564037323, 0.09065964818000793, 0.1043134555220604, -0.03936121612787247, 0.12869194149971008, 0.00521129509434104, 0.0686698704957962, 0.22204025089740753, 0.0108240507543087, -0.119538813829422, 0.0015416473615914583, 0.06489486247301102, -0.019384892657399178, -0.06016543135046959, -0.07444785535335541, 0.0900929719209671, -0.12171205133199692, 0.012087303213775158, 0.0047754873521625996, 0.08157021552324295, -0.04616151005029678, -0.03432614356279373, 0.0007393746054731309, 0.03884008154273033, -0.03382151201367378, 0.06785280257463455, -0.1569320261478424, 0.1032557412981987, 0.040970079600811005, 0.019302034750580788, -0.04555712267756462, 0.014035450294613838, -0.0045867967419326305, 0.14293953776359558, 0.0908348560333252, 0.07396339625120163, 0.020994629710912704, -0.028138628229498863, -0.2258029580116272, 0.06273671239614487, 0.03907780349254608, -0.13910338282585144, 0.017343567684292793, 0.056232523173093796, -0.05111941322684288, -0.0006445308099500835, 0.049238964915275574, -0.2156134843826294, -0.12308091670274734, 0.07585311681032181, -0.06977231055498123, 0.04393265023827553, -0.038636136800050735, -0.013456580229103565, 0.045305974781513214, 0.15741804242134094, -0.22934263944625854, -0.12460778653621674, -0.12553048133850098, -0.16021373867988586, 0.10946594923734665, -0.11719147861003876, 0.04789116233587265, -0.04924233257770538, 0.09830271452665329, 0.001972418511286378, -0.05344719812273979, 0.09598664939403534, -0.03328228369355202, -0.1299998164176941, -0.048164427280426025, 0.09086762368679047, 0.15272800624370575, 0.03565841168165207, -0.00528474897146225, 0.012492253445088863, -0.09612254053354263, -0.1478116661310196, -0.014768947847187519, 0.12430638074874878, 0.043759435415267944, 0.037143684923648834, -0.09714917093515396, -0.1340739130973816, -0.08316044509410858, -0.03370065987110138, 0.1172633096575737, 0.10212422162294388, -0.05392861366271973, 0.08161238580942154, 0.23922491073608398, -0.05886942893266678, -0.16763414442539215, 0.0014363809023052454, 0.009329890832304955, 0.04581621289253235, -0.07136193662881851, -0.2504485547542572, 0.0982336550951004, 0.10204783827066422, -0.03136717528104782, 0.09421618282794952, -0.1962154358625412, -0.16765202581882477, 0.10345232486724854, 0.02862033061683178, 0.038599979132413864, -0.10490567982196808, -0.05683719739317894, -0.07332302629947662, -0.13308298587799072, 0.08350306749343872, -0.12249304354190826, 0.01563127152621746, 0.03922140598297119, 0.06681521236896515, -0.00869763270020485, -0.03408339247107506, 0.14909771084785461, 0.06965603679418564, 0.025404363870620728, 0.00014469541201833636, -0.1426829695701599, 0.11935272067785263, -0.061062414199113846, 0.11300837248563766, 0.058016009628772736, 0.06001058593392372, -0.1116458848118782, -0.019978094846010208, -0.0549047514796257, 0.09347530454397202, -0.08643282204866409, -0.09446042776107788, -0.05264439433813095, 0.05033815652132034, 0.10081378370523453, -0.03129594400525093, 0.03620050102472305, -0.05933946371078491, -0.07424496859312057, 0.055428918451070786, 0.151398703455925, 0.057856690138578415, -0.0828607976436615, -0.023678651079535484, -0.025805428624153137, 0.059286948293447495, -0.1450657993555069, 0.08991673588752747, 0.0907568410038948, -0.000841206987388432, 0.0796288475394249, -0.02464873157441616, -0.18418927490711212, -0.03888631984591484, 0.06895077973604202, -0.09531193971633911, -0.003635724075138569, -0.0531240813434124, 0.08743557333946228, -0.1295793205499649, -0.011847114190459251, 0.0540454275906086, 0.00477920426055789, -0.029164498671889305, 0.00736280158162117, 0.024162115529179573, 0.018646080046892166, 0.07742896676063538, 0.07681622356176376, 0.04976319521665573, -0.04623602330684662, 0.08618722856044769, 0.09175354987382889, 0.003280655248090625, -0.060200612992048264, 0.05967264622449875, -0.09912072867155075, -0.06044817343354225, 0.04447346553206444, 0.0739363431930542, -0.03644746541976929, -0.044421277940273285, -0.049585338681936264, -0.08071491122245789, 0.03914749622344971, 0.14668381214141846, -0.02129938267171383, 0.0800461396574974, 0.028539586812257767, -0.05144008994102478, 0.023256175220012665, 0.06817099452018738, -0.033774007111787796, 0.0029193032532930374, -0.06847559660673141, 0.009608691558241844, 0.013508190400898457, 0.008242862299084663, -0.021357446908950806, -0.046136584132909775, -0.05297013372182846, -0.06032828986644745, -0.06087201461195946, 0.02455926686525345, -0.014890638180077076, -0.010266530327498913, -0.05102602392435074, -0.017351699993014336, -0.031503718346357346, 0.020724624395370483, -0.05731847137212753, -0.023901930078864098, -0.040154002606868744, 0.06450407952070236, -0.09365025162696838, -0.03330626338720322, 0.06853468716144562, -0.09424836188554764, 0.07697246223688126, 0.026264110580086708, -0.009858968667685986, -0.017242813482880592, -0.11005990207195282, 0.021619873121380806, 0.010850885882973671, 0.05153895914554596, 0.026651842519640923, -0.16255483031272888, -0.017425114288926125, -0.02693134732544422, 0.001291681663133204, -0.014970771968364716, 0.02322136051952839, -0.10822205245494843, 0.1170588955283165, -0.0860414057970047, -0.12131989747285843, -0.08595510572195053, 0.002372802933678031, 0.04463030770421028, 0.028173260390758514, 0.10719000548124313, -0.014050728641450405, 0.060646865516901016, -0.07790093123912811, -0.005945276468992233, 0.03739519044756889, 0.027210399508476257, -0.04505256563425064, -0.034068480134010315, 0.0009864072781056166, -0.03456839546561241, 0.12502487003803253, -0.05894922465085983, -0.03268537297844887, 0.0679025948047638, 0.005026038736104965, -0.07048060745000839, 0.05263952538371086, 0.03813190758228302, 0.021109094843268394, 0.03549251705408096, 0.010933861136436462, -0.022892769426107407, -0.07126249372959137, -0.0868583396077156, 0.12188215553760529, 0.10870911926031113, 0.14408831298351288, 0.05676070600748062, 0.07822773605585098, -0.07495313137769699, -0.037532102316617966, 0.028093649074435234, -0.06761477887630463, 0.010241417214274406, -0.01612755097448826, 0.09092718362808228, 0.16335336863994598, -0.1759549230337143, 0.04777468740940094, 0.067558154463768, -0.07082164287567139, -0.14947204291820526, -0.1456877738237381, -0.08511488139629364, 0.028979957103729248, -0.01427305955439806, -0.059240031987428665, -0.0011258147424086928, -0.050441544502973557, 0.02276228927075863, -0.002438324736431241, 0.07120219618082047, -0.009289076551795006, -0.09958529472351074, 0.08858134597539902, 0.05381350591778755, 0.04432683810591698, 0.05483195185661316, 0.057817794382572174, 0.047271691262722015, 0.07248552143573761, 0.015762703493237495, 0.05404166132211685, 0.026815542951226234, 0.00994402077049017, -0.04090019688010216, -0.08463434129953384, -0.02962556853890419, -0.020518017932772636, -0.03571147471666336, 0.021150454878807068, 0.0340644046664238, -0.03969152644276619, 0.029361680150032043, 0.09902383387088776, -0.04655895009636879, -0.05976750701665878, -0.1007181778550148, 0.0637160986661911, -0.0011146360775455832, 0.059652168303728104, 0.0683554857969284, -0.1450282484292984, -0.0006361271953210235, 0.09253506362438202, 0.09167180210351944, 0.007030678912997246, 0.03139358386397362, 0.030607158318161964, 0.012384295463562012, 0.011382769793272018, -0.0010600588284432888, 0.0325106717646122, 0.17962783575057983, -0.07232392579317093, 0.0629541203379631, -0.01855127140879631, -0.019972141832113266, -0.09885431826114655, 0.055900558829307556, -0.010112597607076168, -0.029871173202991486, -0.08556077629327774, 0.08887435495853424, -0.02783805876970291, -0.19700294733047485, 0.023700812831521034, -0.04047194495797157, -0.08066028356552124, -0.05473269522190094, -0.04844019562005997, -0.016404690220952034, 0.018493561074137688, 0.00529532739892602, -0.04030222073197365, 0.14892658591270447, 0.05441297963261604, -0.11767860502004623, -0.01008913479745388, 0.10839762538671494, -0.07563642412424088, 0.10397191345691681, 0.05216757208108902, 0.08680932223796844, 0.083571657538414, -0.014480441808700562, -0.07667176425457001, 0.11860280483961105, 0.01000902522355318, -0.09088478982448578, -0.015942227095365524, 0.06915884464979172, 0.02334674634039402, 0.08767998963594437, 0.0773192048072815, 0.02619030326604843, 0.06464049220085144, -0.011132905259728432, -0.04356740042567253, -0.11879797279834747, 0.10462812334299088, -0.11498306691646576, 0.0818413496017456, 0.11326054483652115, -0.05780686065554619, -0.012690950185060501, -0.04206295683979988, 0.013061556033790112, 0.06296265870332718, -0.036291684955358505, -0.016661055386066437, -0.06368979066610336, 0.04306502640247345, -0.07411643862724304, 0.15654104948043823, -0.08139654248952866, -0.0495588555932045, -0.0007024703081697226, -0.0032828717958182096, -0.04319816455245018, 0.07629024237394333, -0.005902636330574751, 0.03860817849636078, -0.012678115628659725, -0.08206554502248764, -0.013408195227384567, 0.048258841037750244, -0.1439141035079956, -0.04474255442619324 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base-finetuned-marc-en This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the amazon_reviews_multi dataset. It achieves the following results on the evaluation set: - Loss: 0.8976 - Mae: 0.4268 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Mae | |:-------------:|:-----:|:----:|:---------------:|:------:| | 1.092 | 1.0 | 235 | 0.9514 | 0.5122 | | 0.9509 | 2.0 | 470 | 0.8976 | 0.4268 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.9.0+cu111 - Datasets 1.14.0 - Tokenizers 0.10.3
{"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["amazon_reviews_multi"], "base_model": "xlm-roberta-base", "model-index": [{"name": "xlm-roberta-base-finetuned-marc-en", "results": []}]}
text-classification
d4niel92/xlm-roberta-base-finetuned-marc-en
[ "transformers", "pytorch", "tensorboard", "xlm-roberta", "text-classification", "generated_from_trainer", "dataset:amazon_reviews_multi", "base_model:xlm-roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #xlm-roberta #text-classification #generated_from_trainer #dataset-amazon_reviews_multi #base_model-xlm-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us
xlm-roberta-base-finetuned-marc-en ================================== This model is a fine-tuned version of xlm-roberta-base on the amazon\_reviews\_multi dataset. It achieves the following results on the evaluation set: * Loss: 0.8976 * Mae: 0.4268 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 2 ### Training results ### Framework versions * Transformers 4.11.3 * Pytorch 1.9.0+cu111 * Datasets 1.14.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.14.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #xlm-roberta #text-classification #generated_from_trainer #dataset-amazon_reviews_multi #base_model-xlm-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.14.0\n* Tokenizers 0.10.3" ]
[ 79, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #xlm-roberta #text-classification #generated_from_trainer #dataset-amazon_reviews_multi #base_model-xlm-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.14.0\n* Tokenizers 0.10.3" ]
[ -0.09812021255493164, 0.10352690517902374, -0.001907684258185327, 0.12290644645690918, 0.1674380898475647, 0.03168686106801033, 0.1575489491224289, 0.11984023451805115, -0.058105770498514175, 0.004223859403282404, 0.13002419471740723, 0.13833214342594147, 0.01715509593486786, 0.1447533220052719, -0.06018156558275223, -0.2425287365913391, -0.0029156955424696207, 0.049298930913209915, -0.03127361834049225, 0.1479398012161255, 0.10505940020084381, -0.12610356509685516, 0.10435055196285248, 0.0037258209194988012, -0.1818360835313797, -0.010994154959917068, 0.03205085173249245, -0.05964047834277153, 0.12747356295585632, 0.036169201135635376, 0.11701266467571259, 0.01050691120326519, 0.0660516768693924, -0.16916663944721222, 0.020428365096449852, 0.035147905349731445, 0.009059338830411434, 0.1040973961353302, 0.024519264698028564, -0.028808608651161194, 0.07744082808494568, -0.058070696890354156, 0.06718502193689346, 0.017978480085730553, -0.13217800855636597, -0.24712109565734863, -0.07379481941461563, 0.044259656220674515, 0.05524620786309242, 0.0917683020234108, -0.01527134794741869, 0.15148352086544037, -0.043198853731155396, 0.0969441682100296, 0.21268121898174286, -0.2841886281967163, -0.07461445778608322, 0.03160625323653221, 0.04387715458869934, 0.10989709198474884, -0.0937267318367958, -0.02421688847243786, 0.04592442885041237, 0.04450565576553345, 0.1321316808462143, -0.03711600601673126, -0.021549640223383904, 0.01321747899055481, -0.12804503738880157, -0.03180558606982231, 0.21304011344909668, 0.0591527558863163, -0.05360223352909088, -0.05201129987835884, -0.03803767263889313, -0.1456279456615448, -0.02930956706404686, 0.01058904454112053, 0.038216665387153625, -0.05502567067742348, -0.10964832454919815, -0.03312109038233757, -0.09813667088747025, -0.05130113288760185, -0.03835471719503403, 0.14890140295028687, 0.021664323285222054, 0.01881633698940277, -0.020792227238416672, 0.1086689829826355, 0.005186812020838261, -0.130142480134964, 0.019770849496126175, 0.001205006497912109, -0.01922762393951416, -0.04256071895360947, -0.059533651918172836, -0.08207199722528458, -0.005260268691927195, 0.13816562294960022, -0.006730278022587299, 0.03263462707400322, 0.029495157301425934, 0.04845919460058212, -0.07001308351755142, 0.21546289324760437, -0.04511220380663872, -0.04219766706228256, -0.0002896675141528249, 0.08963412791490555, 0.028395041823387146, -0.020405536517500877, -0.14326611161231995, 0.008495262823998928, 0.10691153258085251, 0.0003075152635574341, -0.045838259160518646, 0.05366884917020798, -0.07306834310293198, -0.07121150940656662, 0.01563962548971176, -0.07442989200353622, 0.017976202070713043, -0.024269405752420425, -0.06692063808441162, -0.04273039475083351, 0.012124831788241863, 0.028101090341806412, 0.0075922803953289986, 0.10486268997192383, -0.10410065948963165, 0.020699841901659966, -0.08116897195577621, -0.10778643935918808, -0.00022651022300124168, -0.07472724467515945, 0.03882025182247162, -0.1064079999923706, -0.18410973250865936, -0.026412202045321465, 0.06071587651968002, -0.03473951667547226, -0.08254385739564896, -0.038923557847738266, -0.04353710636496544, 0.01611519791185856, -0.012438597157597542, 0.1264721155166626, -0.06884513050317764, 0.1065962016582489, 0.05314483866095543, 0.0545429103076458, -0.057804789394140244, 0.043849628418684006, -0.09878509491682053, 0.01811285689473152, -0.1646735519170761, 0.05127790942788124, -0.032390009611845016, 0.07116829603910446, -0.09387228637933731, -0.10144060850143433, 0.012676175683736801, -0.0026084501296281815, 0.04583049938082695, 0.07100346684455872, -0.1528351604938507, -0.0758446529507637, 0.13790775835514069, -0.04834872484207153, -0.1352408230304718, 0.12820465862751007, -0.07910808175802231, 0.0651589184999466, 0.07837335020303726, 0.17916326224803925, 0.0777864009141922, -0.04574386775493622, 0.035024162381887436, -0.021069610491394997, 0.06634294241666794, -0.061811257153749466, 0.1131177693605423, 0.01081791426986456, -0.03710639476776123, 0.02510187029838562, -0.05857214331626892, 0.047613855451345444, -0.08469528704881668, -0.08893166482448578, -0.02836320735514164, -0.10854913294315338, 0.07260706275701523, 0.057163309305906296, 0.06783141195774078, -0.10898226499557495, -0.08911730349063873, 0.059539347887039185, 0.08375418931245804, -0.05782730504870415, 0.012237601913511753, -0.06102912127971649, 0.07865146547555923, -0.0875755324959755, -0.027446014806628227, -0.1719016432762146, -0.013196619227528572, 0.01573120430111885, 0.016925813630223274, 0.04338356852531433, 0.022445863112807274, 0.06309623271226883, 0.04790724813938141, -0.07823556661605835, -0.023563319817185402, -0.04988884925842285, -0.0025884294882416725, -0.11524929106235504, -0.17892742156982422, -0.04058213531970978, -0.02493293024599552, 0.15957015752792358, -0.20539827644824982, 0.027443308383226395, -0.05687323212623596, 0.06554590165615082, 0.04896894097328186, -0.0265037901699543, -0.015622194856405258, 0.07002788037061691, -0.03115350566804409, -0.04752243682742119, 0.07251476496458054, 0.01747770793735981, -0.1277834177017212, -0.014317176304757595, -0.10614078491926193, 0.18912449479103088, 0.11936116218566895, -0.04920954629778862, -0.06957253068685532, 0.017118031159043312, -0.03793710097670555, -0.029705340042710304, -0.05631444603204727, 0.01111234724521637, 0.1715090572834015, 0.011354546062648296, 0.1495170295238495, -0.0915621891617775, -0.04248065501451492, 0.02530520223081112, -0.039175376296043396, 0.019209325313568115, 0.13292847573757172, 0.07611285895109177, -0.13196781277656555, 0.13387610018253326, 0.1607392430305481, -0.0736284926533699, 0.15296362340450287, -0.029240358620882034, -0.05728021636605263, -0.0464419424533844, -0.04359692707657814, 0.0004992583999410272, 0.11432897299528122, -0.09205528348684311, 0.003642542287707329, 0.037298522889614105, 0.007559128571301699, 0.005296600982546806, -0.194528266787529, -0.04211641848087311, 0.04960338771343231, -0.03238385170698166, -0.0195712111890316, -0.005746605806052685, 0.005445985123515129, 0.09943103045225143, 0.026033973321318626, -0.06495804339647293, 0.041400786489248276, 0.006346049718558788, -0.07763025164604187, 0.20318613946437836, -0.06127871572971344, -0.18176858127117157, -0.15809854865074158, -0.0608394481241703, -0.06807000190019608, 0.017409272491931915, 0.04937880113720894, -0.05477774143218994, -0.025062300264835358, -0.08172362297773361, -0.0028053519781678915, -0.01773783005774021, 0.01095736213028431, -0.0006763212732039392, 0.0013714163796976209, 0.06442662328481674, -0.09177060425281525, -0.011626158840954304, -0.03314218297600746, -0.014734343625605106, 0.0438874289393425, 0.023591991513967514, 0.10905226320028305, 0.13539284467697144, -0.015037366189062595, 0.004823619499802589, -0.020374136045575142, 0.2483033537864685, -0.08111010491847992, -0.040294963866472244, 0.13866938650608063, -0.01982552371919155, 0.042179569602012634, 0.1332840472459793, 0.0634307712316513, -0.08422891050577164, 0.01686006225645542, 0.01620476320385933, -0.041330400854349136, -0.22779057919979095, -0.024201340973377228, -0.05223262310028076, -0.008042591623961926, 0.10177147388458252, 0.017685074359178543, 0.01614299975335598, 0.07274973392486572, 0.03827188163995743, 0.07503646612167358, -0.0291079580783844, 0.08061981946229935, 0.09651633352041245, 0.05387271195650101, 0.14345185458660126, -0.03671906143426895, -0.06028701364994049, 0.04848731309175491, 0.0108230821788311, 0.21189790964126587, 0.01326752733439207, 0.17466220259666443, 0.04744992032647133, 0.12500524520874023, 0.014298989437520504, 0.05333346128463745, 0.015036297030746937, -0.020865047350525856, -0.03146560490131378, -0.022346103563904762, -0.04086718708276749, 0.02209469862282276, -0.016497479751706123, 0.06340839713811874, -0.11195199191570282, -0.02905680611729622, 0.050002921372652054, 0.24331700801849365, 0.037391938269138336, -0.3439534902572632, -0.11366234719753265, 0.01883978210389614, -0.03631078079342842, -0.025519123300909996, 0.005277425050735474, 0.08607487380504608, -0.11822685599327087, 0.0291055329144001, -0.08441028743982315, 0.0970635935664177, -0.0894991010427475, 0.03904876112937927, 0.05950405076146126, 0.06615041941404343, -0.012868217192590237, 0.07652463018894196, -0.26785486936569214, 0.2654654383659363, 0.0038434755988419056, 0.04138108715415001, -0.0528125986456871, -0.026833364740014076, 0.018785936757922173, 0.03856952488422394, 0.05312851443886757, 0.002691675443202257, -0.031036008149385452, -0.1918506771326065, -0.054761793464422226, 0.02141708694398403, 0.058805324137210846, -0.06886237114667892, 0.10347776114940643, -0.03732399269938469, 0.002364743035286665, 0.046839240938425064, 0.007273396942764521, -0.04171981289982796, -0.1025942862033844, 0.006834252271801233, 0.022444799542427063, -0.03227124363183975, -0.0657159686088562, -0.11993240565061569, -0.05423363298177719, 0.13757771253585815, 0.013785678893327713, -0.051518190652132034, -0.10434737056493759, 0.06923169642686844, 0.08733152598142624, -0.08491255342960358, 0.025313682854175568, 0.0014104947913438082, 0.09964048862457275, 0.02253890223801136, -0.04694976285099983, 0.09993849694728851, -0.03903364762663841, -0.17351704835891724, -0.054366543889045715, 0.1256658434867859, 0.020731644704937935, 0.06427878141403198, -0.020925333723425865, 0.022260025143623352, -0.06550727039575577, -0.07281044870615005, 0.029962480068206787, -0.016384465619921684, 0.06836926192045212, 0.037778619676828384, -0.01038407627493143, 0.02294698916375637, -0.08377562463283539, -0.04059524089097977, 0.18432417511940002, 0.23314684629440308, -0.08502037823200226, 0.013911966234445572, 0.017628196626901627, -0.058133941143751144, -0.13472311198711395, 0.01633010432124138, 0.06756463646888733, 0.01927938126027584, 0.07207529246807098, -0.14412008225917816, 0.09311923384666443, 0.08463025093078613, -0.017407385632395744, 0.10958670824766159, -0.28891894221305847, -0.13729993999004364, 0.09376998990774155, 0.14318998157978058, 0.13833874464035034, -0.13397923111915588, -0.03197073936462402, -0.04658104479312897, -0.1487695574760437, 0.13397598266601562, -0.08040919154882431, 0.13804513216018677, -0.025544079020619392, 0.096004918217659, 0.01279239822179079, -0.04526284709572792, 0.1393755078315735, 0.007535177282989025, 0.08931449800729752, -0.05096963420510292, -0.04220948740839958, 0.032130055129528046, -0.04924353212118149, 0.013439087197184563, -0.09285589307546616, 0.03006870299577713, -0.11729402840137482, -0.039012353867292404, -0.07519809901714325, 0.014598245732486248, -0.02858785353600979, -0.05347359552979469, -0.03161230683326721, 0.04778921604156494, 0.025237634778022766, -0.012278441339731216, 0.16444189846515656, -0.0027924254536628723, 0.14722470939159393, 0.10666793584823608, 0.08577708154916763, -0.046120233833789825, -0.08052004873752594, -0.03603455796837807, -0.033500637859106064, 0.04649407044053078, -0.14823013544082642, 0.031162351369857788, 0.12681978940963745, 0.007066360209137201, 0.16054615378379822, 0.06166496500372887, -0.031135370954871178, 0.01827199198305607, 0.06477196514606476, -0.15788358449935913, -0.11600645631551743, -0.03486217185854912, -0.07604873180389404, -0.14656369388103485, 0.020384302362799644, 0.13379241526126862, -0.05541857331991196, -0.036075398325920105, -0.008475614711642265, 0.004952962044626474, -0.04686073958873749, 0.17704343795776367, 0.07624934613704681, 0.05941781401634216, -0.09578975290060043, 0.08243342489004135, 0.07437931001186371, -0.04332021623849869, 0.0032680847216397524, 0.042980778962373734, -0.08939282596111298, -0.04395003616809845, 0.021134288981556892, 0.17765511572360992, -0.07725949585437775, -0.029049687087535858, -0.16312871873378754, -0.10595784336328506, 0.06462317705154419, 0.11505185812711716, 0.10770709812641144, 0.005066306795924902, -0.03954630717635155, -0.0202241949737072, -0.0869995504617691, 0.11391117423772812, 0.07455001026391983, 0.07691191881895065, -0.15757641196250916, 0.08190253376960754, 0.00984792597591877, 0.052459716796875, -0.015612490475177765, 0.023781491443514824, -0.1068887934088707, 0.01018856093287468, -0.1314578652381897, -0.00713737728074193, -0.01886388659477234, 0.017276203259825706, -0.00975011009722948, -0.0737651139497757, -0.06535785645246506, 0.005364119540899992, -0.12144448608160019, -0.03146880492568016, 0.041686635464429855, 0.059771209955215454, -0.08568035811185837, -0.032555531710386276, 0.04000971466302872, -0.054124634712934494, 0.07259633392095566, 0.034383878111839294, 0.016957225278019905, 0.044215552508831024, -0.09063462167978287, 0.034417614340782166, 0.021733084693551064, 0.006982472259551287, 0.04584461823105812, -0.13501961529254913, 0.002426386345177889, -0.006235599052160978, 0.06819358468055725, 0.022565463557839394, 0.08897408097982407, -0.14996711909770966, -0.00550141092389822, 0.002148654079064727, -0.06669206917285919, -0.06453054398298264, 0.03369293734431267, 0.06227828189730644, 0.05516096204519272, 0.21624374389648438, -0.07163715362548828, 0.03457832708954811, -0.21061794459819794, 0.003025998128578067, -0.018195707350969315, -0.12717124819755554, -0.11335045844316483, -0.08173266798257828, 0.04758157581090927, -0.06282927840948105, 0.14101436734199524, 0.025031045079231262, 0.07818511128425598, 0.023037582635879517, 0.00678018294274807, 0.022690627723932266, 0.014421696774661541, 0.1840703934431076, 0.008179683238267899, -0.0434228889644146, 0.0778246745467186, 0.04037131741642952, 0.09507386386394501, 0.13757115602493286, 0.1634574979543686, 0.16562770307064056, 0.01898784749209881, 0.06936065107584, 0.03429622948169708, -0.006622704677283764, -0.13671477138996124, 0.03102089837193489, -0.008023800328373909, 0.1043572947382927, -0.0066113825887441635, 0.2086031138896942, 0.07871729135513306, -0.17107908427715302, 0.037778399884700775, -0.055767543613910675, -0.08584146201610565, -0.09015718847513199, -0.09161543101072311, -0.09997442364692688, -0.14406391978263855, -0.0019013544078916311, -0.12244853377342224, -0.00808026734739542, 0.10944617539644241, -0.0033328398130834103, -0.04050534591078758, 0.11308660358190536, 0.020677482709288597, -0.0025744130834937096, 0.07750289142131805, 0.0031741566490381956, -0.036672383546829224, -0.0736335963010788, -0.06324776262044907, -0.012521742843091488, -0.003947390243411064, 0.032419171184301376, -0.056489042937755585, -0.06048688292503357, 0.018643613904714584, -0.023840805515646935, -0.12456218898296356, 0.012879862450063229, 0.024760475382208824, 0.07430610060691833, 0.04327857121825218, 0.008459833450615406, 0.007238071411848068, -0.00772605137899518, 0.2739601135253906, -0.06460370123386383, -0.046303700655698776, -0.12886743247509003, 0.21538853645324707, 0.015485062263906002, -0.03713414818048477, 0.02819807268679142, -0.07591656595468521, 0.02228512614965439, 0.22270861268043518, 0.21857169270515442, -0.09196339547634125, -0.012035978958010674, 0.008699605241417885, -0.006352963391691446, 0.0012919970322400331, 0.1007501557469368, 0.08472951501607895, -0.014648787677288055, -0.07979421317577362, -0.006122843828052282, -0.055351871997117996, -0.007488004397600889, -0.008557664230465889, 0.06689897924661636, 0.044714320451021194, 0.008150093257427216, -0.053536467254161835, 0.07880229502916336, -0.07123448699712753, -0.11125939339399338, 0.025823740288615227, -0.21135656535625458, -0.1810917854309082, -0.03228117153048515, 0.07168364524841309, -0.0003595422313082963, 0.06783684343099594, -0.025290708988904953, 0.004122728016227484, 0.04975440353155136, -0.011912780813872814, -0.08325883001089096, -0.09511129558086395, 0.10446051508188248, -0.09058190882205963, 0.18431179225444794, -0.05186767131090164, 0.03716496005654335, 0.12413962930440903, 0.04805907607078552, -0.07977709174156189, 0.0754096657037735, 0.03829657658934593, -0.008729022927582264, 0.0597822479903698, 0.11527280509471893, -0.024547936394810677, 0.09938172250986099, 0.052748117595911026, -0.11229771375656128, 0.00046200439101085067, -0.08963160216808319, -0.030114740133285522, -0.054946474730968475, -0.0371994748711586, -0.06871096044778824, 0.1453767567873001, 0.2193399965763092, -0.05803428590297699, -0.02145204320549965, -0.05205347016453743, 0.02569247968494892, 0.08520379662513733, 0.03857872262597084, -0.06332465261220932, -0.22206559777259827, -0.006284010596573353, 0.07150046527385712, -0.01069146953523159, -0.2947491705417633, -0.07019023597240448, -0.011928042396903038, -0.06109558045864105, -0.049017444252967834, 0.104152150452137, 0.07352162152528763, 0.04259505122900009, -0.0640808492898941, -0.03738879784941673, -0.07639223337173462, 0.15060356259346008, -0.15641680359840393, -0.09290610998868942 ]
null
null
transformers
# Harry
{"tags": ["conversational"]}
text-generation
d4rk/harry
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Harry
[ "# Harry" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Harry" ]
[ 51, 2 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Harry" ]
[ 0.008150327950716019, 0.07997918874025345, -0.007178797852247953, 0.03567842021584511, 0.13429023325443268, 0.04860587418079376, 0.13002704083919525, 0.15211623907089233, 0.032173920422792435, -0.021888386458158493, 0.1481919139623642, 0.20967841148376465, 0.002699901582673192, 0.027984336018562317, -0.07098018378019333, -0.2619108259677887, 0.04580087959766388, 0.05127737671136856, -0.03388937562704086, 0.11372090876102448, 0.08055732399225235, -0.07797889411449432, 0.10050316154956818, -0.04493314027786255, -0.1697816699743271, -0.002110888483002782, 0.03408896550536156, -0.11597410589456558, 0.14719758927822113, 0.06921592354774475, 0.048300813883543015, -0.00850814301520586, -0.076728954911232, -0.15931929647922516, 0.03596772626042366, 0.009679432958364487, -0.05316237732768059, 0.07768061757087708, 0.07793281972408295, -0.06378860771656036, 0.09473442286252975, 0.12092399597167969, 0.012448522262275219, 0.05755617842078209, -0.16281144320964813, -0.03889242932200432, -0.025217758491635323, 0.05875825509428978, 0.02882077358663082, 0.08777906000614166, -0.025393502786755562, 0.1150071993470192, -0.06023528426885605, 0.10495458543300629, 0.15590788424015045, -0.37643104791641235, -0.01927122101187706, 0.08297953754663467, 0.02421913668513298, 0.07742053270339966, -0.05841423198580742, 0.05105055496096611, 0.0291367769241333, 0.010136914439499378, -0.034277502447366714, -0.08226783573627472, -0.0691092237830162, 0.056069597601890564, -0.08630520105361938, -0.03791604936122894, 0.2868892550468445, -0.06861335039138794, 0.06846558302640915, -0.04295716807246208, -0.1125124916434288, 0.010618508793413639, -0.023355290293693542, -0.0318555124104023, -0.07080025225877762, 0.07488057017326355, 0.010736515745520592, -0.05902828648686409, -0.10992886871099472, -0.00898059457540512, -0.1566953957080841, 0.18897034227848053, 0.04560272395610809, 0.053054329007864, -0.2046297937631607, 0.08976330608129501, 0.03162526339292526, -0.06762029975652695, 0.03180040046572685, -0.10093335807323456, 0.07211042940616608, 0.01263956818729639, -0.03548488765954971, -0.0617138035595417, 0.08190520852804184, 0.09648655354976654, 0.018170509487390518, 0.03031209670007229, -0.018015019595623016, 0.07860899716615677, 0.07015649229288101, 0.03260401636362076, 0.020849935710430145, -0.10083217918872833, 0.031029850244522095, -0.12741801142692566, -0.004848834127187729, -0.07212943583726883, -0.15646961331367493, -0.032632630318403244, 0.07265544682741165, 0.04057692736387253, 0.04629788547754288, 0.0918584018945694, -0.017937956377863884, -0.031683288514614105, 0.005309654865413904, -0.03881150111556053, -0.0016185828717425466, 0.010565927252173424, 0.004867391660809517, 0.17569318413734436, -0.01173572801053524, 0.005573826376348734, -0.13824132084846497, 0.06755446642637253, -0.07359027862548828, 0.003291269764304161, -0.009804779663681984, -0.043432384729385376, 0.00019013567361980677, 0.02275741659104824, 0.009434896521270275, -0.1306990087032318, -0.09456845372915268, 0.005049117375165224, 0.001683380571193993, -0.030414460226893425, -0.0880187600851059, -0.06016523391008377, 0.0005792967858724296, 0.03403691574931145, -0.03788115829229355, -0.02053765580058098, -0.05491064488887787, 0.13973355293273926, -0.039907243102788925, 0.0680219903588295, -0.08543722331523895, 0.07607127726078033, -0.08393777906894684, -0.03660587593913078, -0.08674334734678268, 0.04517173394560814, 0.023118294775485992, 0.0952034592628479, 0.009009120985865593, -0.035681288689374924, -0.02597375586628914, 0.06909751147031784, -0.08201564103364944, 0.22704581916332245, -0.06106000021100044, -0.13297651708126068, 0.27533063292503357, -0.06215481460094452, -0.1545005589723587, 0.12745322287082672, -0.0022543161176145077, 0.04219355061650276, 0.11488337814807892, 0.1701890528202057, -0.023590579628944397, -0.012874958105385303, 0.086909219622612, 0.08520808070898056, -0.08411797881126404, 0.01972302235662937, 0.021819185465574265, -0.03323185071349144, -0.18226908147335052, 0.05368303880095482, 0.06166396662592888, 0.004056228790432215, -0.03960279002785683, -0.03434339165687561, -0.008246045559644699, -0.002105156658217311, 0.11326885968446732, -0.0030808148439973593, 0.11007247120141983, -0.07832269370555878, -0.02927342616021633, -0.061117690056562424, 0.013490136712789536, -0.03389749675989151, 0.04241752624511719, -0.030486803501844406, 0.104023277759552, 0.030795544385910034, 0.06876599043607712, -0.1441798061132431, -0.018419278785586357, -0.01465595792979002, 0.1655028760433197, 0.03634936735033989, 0.11940139532089233, 0.0469493567943573, -0.015638815239071846, -0.021675750613212585, 0.012381220236420631, 0.15536922216415405, -0.032593898475170135, -0.08762139081954956, -0.1258864551782608, 0.06891770660877228, -0.04575640708208084, 0.0731445699930191, -0.10150167346000671, 0.015967585146427155, 0.01882955990731716, 0.09948527812957764, 0.008797373622655869, 0.03413883224129677, -0.02743394672870636, -0.005820179358124733, -0.07495683431625366, 0.00875744316726923, 0.0919453501701355, -0.003998344298452139, -0.04996733367443085, 0.13806571066379547, -0.17212142050266266, 0.20840395987033844, 0.20267750322818756, -0.32493656873703003, 0.009374254383146763, -0.08949611335992813, -0.031872112303972244, 0.004958184435963631, 0.05969046801328659, -0.02312314696609974, 0.13331562280654907, -0.018370969220995903, 0.17138724029064178, -0.03461623564362526, -0.07161927968263626, -0.03657282143831253, -0.047394730150699615, -0.004794360604137182, 0.0676620677113533, 0.05765331909060478, -0.13297857344150543, 0.19050554931163788, 0.1610855609178543, 0.0399690680205822, 0.1481526792049408, 0.054114822298288345, -0.012512989342212677, 0.079967200756073, -0.012349419295787811, -0.044576261192560196, -0.044705942273139954, -0.2649945914745331, -0.059904683381319046, 0.08162160217761993, -0.016532359644770622, 0.08187162131071091, -0.10637134313583374, -0.027939267456531525, -0.02658247947692871, 0.010447057895362377, 0.04430146887898445, 0.06908783316612244, 0.04134810343384743, 0.13113145530223846, 0.009810316376388073, -0.01703023351728916, 0.08961160480976105, 0.013629522174596786, -0.10744574666023254, 0.16339001059532166, -0.15301619470119476, -0.31661707162857056, -0.1222880631685257, -0.21792426705360413, -0.047991618514060974, 0.047343697398900986, 0.11204830557107925, -0.12909647822380066, -0.018080465495586395, 0.010634495876729488, 0.08049434423446655, -0.16831138730049133, 0.014104016125202179, -0.10683790594339371, 0.050810858607292175, -0.13064362108707428, -0.08536121249198914, -0.03668107092380524, -0.02152377925813198, -0.055290527641773224, 0.15841889381408691, -0.10820674896240234, 0.014260992407798767, 0.20680807530879974, 0.056190866976976395, 0.06024555489420891, -0.04419902339577675, 0.1893211007118225, -0.09929632395505905, -0.024943780153989792, 0.13006441295146942, -0.07264143228530884, 0.07614395767450333, 0.11094104498624802, 0.010037418454885483, -0.08044503629207611, 0.017071880400180817, -0.04862537235021591, -0.03726791590452194, -0.26204052567481995, -0.11292284727096558, -0.11419184505939484, 0.14296068251132965, 0.07535058259963989, 0.06098230555653572, 0.19169019162654877, 0.05194266512989998, -0.041768621653318405, 0.060846250504255295, 0.062060531228780746, 0.11330720782279968, 0.22340160608291626, -0.06009332090616226, 0.1037193313241005, -0.026680653914809227, -0.12193924933671951, 0.08004987239837646, 0.06268240511417389, 0.08123214542865753, 0.0726436972618103, 0.029904432594776154, -0.014741581864655018, 0.11558777838945389, 0.12028518319129944, 0.07602997124195099, 0.03023551218211651, 0.0044477335177361965, -0.0442391112446785, -0.027413515374064445, -0.06894954293966293, 0.05173788592219353, 0.07534655928611755, -0.13301824033260345, -0.02288123592734337, -0.08863618224859238, 0.0733361765742302, 0.08733070641756058, 0.0683031901717186, -0.15697596967220306, -0.032764140516519547, 0.08770224452018738, -0.06571709364652634, -0.14380478858947754, 0.110891193151474, 0.005665468517690897, -0.17159228026866913, 0.03885698318481445, -0.03743424639105797, 0.11335831135511398, -0.07454986870288849, 0.08642100542783737, -0.08794202655553818, -0.09411060065031052, 0.01720825955271721, 0.10373841971158981, -0.3548857271671295, 0.1734755039215088, 0.007529132533818483, -0.05844740569591522, -0.10466589778661728, -0.0047827474772930145, 0.014064757153391838, 0.0863635241985321, 0.09708419442176819, -0.0039138952270150185, 0.036004625260829926, -0.030379468575119972, 0.00469856895506382, 0.023473018780350685, 0.13513074815273285, -0.048097193241119385, -0.00026968956808559597, -0.0625312551856041, 0.0006561221671290696, -0.05067351087927818, -0.030396992340683937, 0.056280605494976044, -0.18914814293384552, 0.10904168337583542, -0.03988165035843849, 0.11060917377471924, 0.028634263202548027, -0.0037388638593256474, -0.11019498109817505, 0.20333942770957947, -0.12370184063911438, -0.07889536768198013, -0.08892582356929779, -0.045901790261268616, 0.025939425453543663, -0.04573778435587883, 0.014072186313569546, -0.05558013170957565, 0.04509544372558594, -0.07747072726488113, -0.19148176908493042, 0.10901689529418945, -0.07811402529478073, -0.06913676112890244, -0.044795822352170944, 0.2102656215429306, -0.04631190374493599, 0.046137742698192596, 0.015599299222230911, 0.02182438224554062, -0.15843187272548676, -0.06383797526359558, 0.06230679899454117, -0.0033062263391911983, 0.041726429015398026, 0.03776678442955017, -0.042808715254068375, -0.019449908286333084, -0.08892440795898438, -0.026357483118772507, 0.32922229170799255, 0.17607098817825317, -0.023973498493433, 0.17977000772953033, 0.1232294961810112, -0.0830162838101387, -0.26702624559402466, -0.12107967585325241, -0.14790202677249908, -0.07327283173799515, -0.0888335257768631, -0.22277477383613586, 0.10721941292285919, 0.009712551720440388, 0.008863099850714207, 0.15568271279335022, -0.22837719321250916, -0.07790812104940414, 0.16020281612873077, 0.020157743245363235, 0.4183369576931, -0.14134015142917633, -0.1039331778883934, -0.052355751395225525, -0.21405909955501556, 0.12258481979370117, -0.05944736301898956, 0.10431770980358124, -0.028683850541710854, 0.16617581248283386, 0.0383528396487236, -0.0013101908843964338, 0.09088904410600662, -0.0016231819754466414, -0.04292507842183113, -0.09273933619260788, -0.11997798085212708, 0.015495941042900085, 0.022062698379158974, -0.02430347166955471, -0.048273127526044846, 0.005268789827823639, -0.11974101513624191, -0.03329618647694588, -0.07172343134880066, 0.023735133931040764, 0.00563058303669095, -0.0678010806441307, -0.022930899634957314, -0.06926808506250381, -0.0010349624790251255, 0.02622888796031475, 0.22198224067687988, -0.044733624905347824, 0.20377303659915924, 0.036108262836933136, 0.13506002724170685, -0.10838258266448975, -0.035650696605443954, -0.06348668783903122, -0.07678893208503723, 0.09647718816995621, -0.09834396094083786, 0.0431787483394146, 0.11039052158594131, -0.01863744854927063, 0.07989563792943954, 0.11678334325551987, 0.004321983549743891, -0.008723610080778599, 0.07342078536748886, -0.3068614900112152, -0.0681929960846901, -0.05015511438250542, 0.02493871934711933, 0.0664868876338005, 0.07411752641201019, 0.18184885382652283, -0.003767226357012987, -0.07080593705177307, 0.018064318224787712, 0.036588672548532486, -0.03494609147310257, 0.06030922010540962, 0.01670447550714016, 0.02351362630724907, -0.15695813298225403, 0.07625272125005722, 0.012057237327098846, -0.1320752650499344, 0.015901941806077957, 0.18004119396209717, -0.10726824402809143, -0.12882904708385468, -0.061185527592897415, 0.13282352685928345, -0.08478786051273346, -0.0034728446044027805, -0.03434884920716286, -0.12939453125, 0.06734540313482285, 0.08906140923500061, 0.05634807422757149, 0.09411884844303131, -0.07482397556304932, -0.013990667648613453, -0.028864599764347076, -0.004492250271141529, 0.007072563748806715, 0.009635094553232193, -0.060111153870821, 0.11391019076108932, -0.04253385215997696, 0.13694676756858826, -0.08699533343315125, -0.10300180315971375, -0.15517710149288177, 0.03107241541147232, -0.05667778477072716, -0.07424800097942352, -0.10900390893220901, -0.045500848442316055, 0.005280633922666311, -0.01999524235725403, -0.025221945717930794, -0.06647715717554092, -0.12245137244462967, 0.034579820930957794, -0.016108276322484016, 0.02834996208548546, -0.0634443461894989, 0.02798200026154518, 0.09315855801105499, -0.05594483017921448, 0.1529751867055893, 0.17014192044734955, -0.11237859725952148, 0.12391463667154312, -0.08101852238178253, -0.11565970629453659, 0.06840144097805023, 0.01605931855738163, 0.035929542034864426, 0.08430468291044235, -0.0229821614921093, 0.023709028959274292, 0.03800255060195923, 0.055617764592170715, 0.07614754885435104, -0.0866733193397522, 0.06365855783224106, -0.07000017911195755, -0.16882477700710297, -0.05267470329999924, -0.0494660921394825, 0.051722779870033264, -0.0029551072511821985, 0.11207453906536102, -0.06083528697490692, 0.10881359875202179, -0.03841792419552803, 0.0372023805975914, 0.027518853545188904, -0.18248122930526733, -0.047389376908540726, -0.06853744387626648, 0.04167060926556587, 0.006292351987212896, 0.24125605821609497, -0.013376148417592049, -0.014788520522415638, 0.04252627119421959, 0.05376771464943886, -0.04661921411752701, 0.004615034908056259, 0.22921457886695862, 0.1206577941775322, -0.0850074514746666, -0.10838250815868378, 0.06876291334629059, 0.03255823254585266, 0.07629062235355377, 0.13475686311721802, 0.02012047916650772, 0.07085832208395004, 0.06813158839941025, -0.04799406975507736, 0.016531076282262802, -0.08833154290914536, -0.11859961599111557, 0.006483956705778837, 0.04762992635369301, -0.022459015250205994, 0.20377089083194733, 0.16443926095962524, -0.02984069101512432, 0.01711403764784336, -0.04578550159931183, -0.061951711773872375, -0.16234396398067474, -0.07885928452014923, -0.07652788609266281, -0.14542102813720703, -0.0046520838513970375, -0.1117999479174614, 0.03598939999938011, 0.03860808536410332, 0.06861793249845505, -0.07057531923055649, 0.022871319204568863, 0.09203927218914032, -0.10495421290397644, 0.07121753692626953, -0.017546001821756363, 0.05341155081987381, -0.043894652277231216, -0.01867021806538105, -0.10766304284334183, -0.01766769029200077, 0.017951572313904762, 0.06088048964738846, -0.06718584895133972, 0.025688791647553444, -0.13271553814411163, -0.11694573611021042, -0.02835421822965145, 0.06212622672319412, -0.061673637479543686, 0.14369596540927887, -0.0004113703325856477, -0.012999610044062138, 0.03280693292617798, 0.2208307683467865, -0.09422335773706436, -0.036435529589653015, -0.04681945592164993, 0.21503198146820068, 0.029408808797597885, 0.0896737203001976, 0.0074201179668307304, 0.00048709698603488505, -0.0701461061835289, 0.3298131227493286, 0.31687265634536743, -0.08254922181367874, 0.008481976576149464, 0.0518588088452816, 0.05116882547736168, 0.12785591185092926, 0.08292554318904877, 0.09929842501878738, 0.2886435091495514, -0.07870082557201385, -0.005829294677823782, -0.01997615210711956, -0.01728188246488571, -0.0901113823056221, 0.05832153558731079, 0.05863518640398979, -0.09249113500118256, -0.03095107525587082, 0.07863713800907135, -0.25041717290878296, 0.052830521017313004, -0.06248736381530762, -0.18158715963363647, -0.05754063278436661, 0.004798607900738716, 0.0962035059928894, 0.02952558547258377, 0.10111293941736221, 0.010615319944918156, -0.09381245076656342, 0.1065032109618187, 0.037793561816215515, -0.24601826071739197, -0.01933654025197029, 0.10729585587978363, -0.10565974563360214, 0.004856930114328861, -0.015523990616202354, 0.06057559698820114, 0.05267098918557167, 0.058994874358177185, -0.024261750280857086, -0.02421385981142521, 0.01983087882399559, -0.0751492828130722, -0.02050439454615116, 0.059041865170001984, 0.03624526411294937, -0.08570602536201477, 0.10611453652381897, -0.10178600996732712, 0.019064752385020256, 0.04799460619688034, -0.04521530494093895, 0.005907328333705664, 0.013932397589087486, -0.057289909571409225, 0.04465124011039734, 0.10357904434204102, -0.01479580532759428, -0.02200905978679657, -0.05266686901450157, -0.04017287865281105, -0.0017535027582198381, -0.07938819378614426, -0.11415660381317139, -0.15015143156051636, -0.12062826007604599, 0.010497049428522587, -0.016855400055646896, -0.1783480942249298, 0.0055547296069562435, -0.08804512023925781, 0.07472650706768036, -0.19379238784313202, 0.07896775752305984, 0.10117519646883011, 0.005683623719960451, -0.005208167247474194, -0.026319080963730812, 0.04194845259189606, 0.11999927461147308, -0.10750556737184525, -0.0698806643486023 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # opus-mt-zh-en-ep1-renri-zh-to-en This model is a fine-tuned version of [Helsinki-NLP/opus-mt-zh-en](https://huggingface.co/Helsinki-NLP/opus-mt-zh-en) on an unkown dataset. It achieves the following results on the evaluation set: - Loss: 2.2192 - Bleu: 18.2579 - Gen Len: 28.4817 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len | |:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:| | 2.2194 | 1.0 | 59472 | 2.2192 | 18.2579 | 28.4817 | ### Framework versions - Transformers 4.9.2 - Pytorch 1.9.0+cu102 - Datasets 1.11.0 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["bleu"], "model_index": [{"name": "opus-mt-zh-en-ep1-renri-zh-to-en", "results": [{"task": {"name": "Sequence-to-sequence Language Modeling", "type": "text2text-generation"}, "metric": {"name": "Bleu", "type": "bleu", "value": 18.2579}}]}]}
text2text-generation
dadada/opus-mt-zh-en-ep1-renri-zh-to-en
[ "transformers", "pytorch", "tensorboard", "marian", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #marian #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
opus-mt-zh-en-ep1-renri-zh-to-en ================================ This model is a fine-tuned version of Helsinki-NLP/opus-mt-zh-en on an unkown dataset. It achieves the following results on the evaluation set: * Loss: 2.2192 * Bleu: 18.2579 * Gen Len: 28.4817 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 1e-05 * train\_batch\_size: 32 * eval\_batch\_size: 32 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 1 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.9.2 * Pytorch 1.9.0+cu102 * Datasets 1.11.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.9.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #marian #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.9.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
[ 58, 113, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #marian #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.9.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
[ -0.08791595697402954, 0.06778696924448013, -0.002774764085188508, 0.10437888652086258, 0.15418370068073273, 0.012980547733604908, 0.13824374973773956, 0.13198433816432953, -0.123493492603302, 0.015357204712927341, 0.11528561264276505, 0.1572904884815216, 0.022327890619635582, 0.1280318647623062, -0.04694313183426857, -0.26569893956184387, -0.001036170986481011, 0.03903571888804436, -0.06751322001218796, 0.13706661760807037, 0.09391605108976364, -0.1256972998380661, 0.0703691616654396, 0.014480474404990673, -0.19491037726402283, 0.012739951722323895, 0.007706410717219114, -0.05720984563231468, 0.15665139257907867, 0.02630198746919632, 0.12842091917991638, 0.01715974695980549, 0.09253322333097458, -0.20467893779277802, 0.014006925746798515, 0.05626577138900757, 0.0173840019851923, 0.08599337935447693, 0.07135938853025436, 0.0044000945053994656, 0.13464853167533875, -0.06941051036119461, 0.05643445625901222, 0.020942287519574165, -0.1284196525812149, -0.24048922955989838, -0.09883879870176315, 0.01696871966123581, 0.05995500832796097, 0.10928311198949814, -0.004749820102006197, 0.13870421051979065, -0.08293918520212173, 0.09648739546537399, 0.236740380525589, -0.2777051031589508, -0.06298667192459106, 0.008424305357038975, 0.03986828401684761, 0.08094967156648636, -0.09246138483285904, -0.026077786460518837, 0.03836773708462715, 0.04884025454521179, 0.13770347833633423, -0.030582552775740623, -0.12681959569454193, 0.008900257758796215, -0.1434907168149948, -0.037285882979631424, 0.1253456473350525, 0.032955851405858994, -0.026019761338829994, -0.038010645657777786, -0.06117980182170868, -0.14217792451381683, -0.03621850907802582, -0.017577173188328743, 0.0491693839430809, -0.021336745470762253, -0.07405712455511093, -0.02605539932847023, -0.10656807571649551, -0.0637538805603981, -0.07205202430486679, 0.10730307549238205, 0.04318959265947342, 0.005975916516035795, -0.03774913027882576, 0.09780632704496384, 0.0067856283858418465, -0.1330709606409073, 0.029071807861328125, 0.029501214623451233, -0.005510215647518635, -0.04391882196068764, -0.07265711575746536, -0.07976450771093369, 0.008835804648697376, 0.11209110170602798, -0.0746183767914772, 0.05527203530073166, 0.0029112931806594133, 0.046050816774368286, -0.10201818495988846, 0.17773960530757904, -0.04022593796253204, -0.0022444119676947594, 0.006159793585538864, 0.05227687209844589, 0.009077019058167934, -0.017360802739858627, -0.10797537118196487, 0.015214648097753525, 0.1084587350487709, 0.01941302977502346, -0.05710422247648239, 0.061585813760757446, -0.048781219869852066, -0.022617818787693977, -0.01391123328357935, -0.09246427565813065, 0.03627125173807144, -0.0005189222283661366, -0.08581742644309998, -0.01034771092236042, 0.03210000693798065, 0.03004133328795433, -0.028537388890981674, 0.09852080047130585, -0.0703904777765274, 0.043697696179151535, -0.10688280314207077, -0.12720033526420593, 0.024808792397379875, -0.05858444422483444, 0.011153876781463623, -0.09752579778432846, -0.1748233586549759, -0.026145027950406075, 0.05995228886604309, -0.024947302415966988, -0.05120270699262619, -0.05053867772221565, -0.06109324470162392, 0.012851240113377571, -0.02651648037135601, 0.14930644631385803, -0.06233479827642441, 0.10883718729019165, 0.023567747324705124, 0.05881255865097046, -0.043135132640600204, 0.060773249715566635, -0.09613388031721115, 0.0010936775943264365, -0.16715404391288757, 0.049999114125967026, -0.04385833814740181, 0.061092808842659, -0.09791086614131927, -0.10250996053218842, 0.014000436291098595, -0.0025066849775612354, 0.09485126286745071, 0.08405997604131699, -0.181447371840477, -0.07206132262945175, 0.1752375066280365, -0.06884466111660004, -0.10394755005836487, 0.11733759194612503, -0.06311088800430298, 0.057391341775655746, 0.07269296795129776, 0.1771620213985443, 0.05070186033844948, -0.07557624578475952, 0.05122784897685051, -0.03185207024216652, 0.061406657099723816, -0.049615275114774704, 0.05973168835043907, -0.0029296674765646458, 0.013962152414023876, 0.02333492785692215, -0.020313812419772148, 0.07495976984500885, -0.09990449994802475, -0.09113284945487976, -0.03881802037358284, -0.09297861158847809, 0.029516909271478653, 0.06637145578861237, 0.07810681313276291, -0.10207845270633698, -0.08077336847782135, 0.07025942206382751, 0.07737163454294205, -0.06835421919822693, 0.03965305536985397, -0.055174365639686584, 0.05420743301510811, -0.019964849576354027, -0.01201464794576168, -0.18831293284893036, -0.009081282652914524, 0.013170500285923481, -0.023948630318045616, 0.04152487963438034, 0.014624431729316711, 0.07023242115974426, 0.06408316642045975, -0.04767356440424919, -0.016442248597741127, -0.03095763362944126, -0.0033107686322182417, -0.11979462951421738, -0.20715628564357758, -0.022137418389320374, -0.018478205427527428, 0.13247568905353546, -0.1996348351240158, 0.03547767922282219, -0.022009534761309624, 0.0769844651222229, 0.00915440358221531, -0.009203110821545124, -0.04112536087632179, 0.1006089597940445, -0.033683616667985916, -0.0494849719107151, 0.08460493385791779, 0.017445456236600876, -0.08974425494670868, -0.010026541538536549, -0.1260373294353485, 0.1414521187543869, 0.1333712786436081, -0.125062495470047, -0.07347418367862701, -0.012869348749518394, -0.05217713490128517, -0.03989126533269882, -0.040883537381887436, 0.04535810276865959, 0.18881762027740479, 0.0006998516037128866, 0.15854158997535706, -0.07079102098941803, -0.042756374925374985, 0.017096158117055893, -0.03485364094376564, 0.034560464322566986, 0.11619038134813309, 0.11101613193750381, -0.07945899665355682, 0.13235031068325043, 0.14979423582553864, -0.09426113218069077, 0.1329599916934967, -0.04455898329615593, -0.07883826643228531, -0.01883394829928875, -0.01687617599964142, 0.003037179820239544, 0.0787491425871849, -0.12060259282588959, -0.0006215069442987442, 0.018558118492364883, 0.030713599175214767, 0.02561996318399906, -0.23275397717952728, -0.03194938972592354, 0.03767895698547363, -0.05134912580251694, -0.019224142655730247, -0.024468233808875084, 0.01668376661837101, 0.1060754582285881, -0.005780664272606373, -0.07749290019273758, 0.025561144575476646, 0.0007406333461403847, -0.07727126777172089, 0.20280075073242188, -0.0948910340666771, -0.16438893973827362, -0.11776162683963776, -0.08450033515691757, -0.033972568809986115, 0.004737980663776398, 0.07276551425457001, -0.08331981301307678, -0.023370802402496338, -0.06590715050697327, 0.03247026354074478, -0.0031086390372365713, 0.013540412299335003, -0.0034215173218399286, 0.0018788741435855627, 0.0756073147058487, -0.11442475765943527, -0.004830104298889637, -0.04441869258880615, -0.06595362722873688, 0.054533060640096664, 0.03986917436122894, 0.1186375841498375, 0.1548691987991333, -0.024842580780386925, 0.006325635127723217, -0.031116966158151627, 0.21491365134716034, -0.06524353474378586, -0.02686886489391327, 0.14510713517665863, -0.005088303703814745, 0.059220824390649796, 0.11008308082818985, 0.06758126616477966, -0.08248697966337204, 0.011944560334086418, 0.031643837690353394, -0.02990834228694439, -0.2331797033548355, -0.04152102395892143, -0.052740249782800674, -0.025021586567163467, 0.08995120972394943, 0.021543648093938828, 0.04736454412341118, 0.05815931782126427, 0.037492845207452774, 0.05937453359365463, -0.019225820899009705, 0.06135893985629082, 0.1426125019788742, 0.04367760941386223, 0.1375243067741394, -0.040608856827020645, -0.06634660065174103, 0.04076625406742096, -0.003692042315378785, 0.2279617339372635, 0.0077956500463187695, 0.14590901136398315, 0.0642782673239708, 0.17274583876132965, 0.004441489465534687, 0.0701407641172409, 0.011776848696172237, -0.029379138723015785, -0.02203819528222084, -0.03802997246384621, -0.03693774342536926, 0.015029317699372768, -0.05740036442875862, 0.03606141358613968, -0.11669307947158813, -0.021242037415504456, 0.04223807156085968, 0.27904757857322693, 0.022345468401908875, -0.3133646547794342, -0.07816466689109802, -0.0009326481376774609, -0.04115397483110428, -0.017890676856040955, 0.021369846537709236, 0.0849958136677742, -0.0993187204003334, 0.03384238854050636, -0.07123329490423203, 0.11041678488254547, -0.048446159809827805, 0.04948434606194496, 0.05271048843860626, 0.0968284085392952, 0.011829374358057976, 0.07941898703575134, -0.3274395167827606, 0.28487005829811096, -0.001580566051416099, 0.06564512848854065, -0.07053191214799881, 0.0022476811427623034, 0.037615302950143814, 0.03592175617814064, 0.036667272448539734, -0.019895948469638824, -0.07371751219034195, -0.19453710317611694, -0.055536698549985886, 0.030776238068938255, 0.09176211804151535, -0.0007979092188179493, 0.10239112377166748, -0.03666335344314575, 0.017148464918136597, 0.07342343777418137, -0.012709137052297592, -0.08865892142057419, -0.09956417977809906, -0.00429221335798502, 0.02834678627550602, -0.015119197778403759, -0.07138216495513916, -0.11358734965324402, -0.10876870900392532, 0.15254294872283936, 0.016884198412299156, -0.02014952339231968, -0.11183442175388336, 0.0836537554860115, 0.08581627160310745, -0.08483729511499405, 0.035657159984111786, 0.011061514727771282, 0.06761960685253143, 0.02494577132165432, -0.0650143101811409, 0.11309472471475601, -0.05951014161109924, -0.15685008466243744, -0.06122062727808952, 0.09008099883794785, 0.03560839220881462, 0.0689372569322586, -0.011277029290795326, 0.018840542063117027, -0.04258321225643158, -0.08181310445070267, 0.015118648298084736, -0.012172387912869453, 0.055099546909332275, 0.017290908843278885, -0.06719359010457993, 0.016430864110589027, -0.06926757097244263, -0.05912899971008301, 0.20074689388275146, 0.24205602705478668, -0.09248695522546768, 0.03700857609510422, 0.05333452671766281, -0.07941559702157974, -0.1849460005760193, 0.029092177748680115, 0.061317164450883865, 0.0046570925042033195, 0.05488380044698715, -0.19388100504875183, 0.0865083560347557, 0.11168190836906433, -0.013554811477661133, 0.0893917977809906, -0.34763574600219727, -0.12917818129062653, 0.11600849777460098, 0.15014195442199707, 0.10085119307041168, -0.16059812903404236, -0.022164834663271904, -0.0261512640863657, -0.11945224553346634, 0.1045541912317276, -0.10116157680749893, 0.1277056336402893, -0.021617058664560318, 0.0948023647069931, 0.004748436156660318, -0.054928869009017944, 0.1132442057132721, -0.013233612291514874, 0.09570200741291046, -0.06914517283439636, 0.015387725085020065, 0.04824453219771385, -0.037530042231082916, 0.008845441974699497, -0.08513923734426498, 0.025145044550299644, -0.07850625365972519, -0.019873928278684616, -0.08095719665288925, 0.03225196525454521, -0.03423912823200226, -0.05450601130723953, -0.027476105839014053, 0.023596445098519325, 0.05833632871508598, -0.010757312178611755, 0.12349937111139297, -0.001488467794843018, 0.16653169691562653, 0.11668171733617783, 0.06793302297592163, -0.06292713433504105, -0.04252048209309578, -0.018778536468744278, -0.02046940103173256, 0.04895153269171715, -0.12951982021331787, 0.030880825594067574, 0.14470389485359192, 0.012232869863510132, 0.14289192855358124, 0.07359272241592407, -0.03637785091996193, 0.021347787231206894, 0.05786600708961487, -0.14822165668010712, -0.09396645426750183, -0.0001800787722459063, -0.014987001195549965, -0.08235423266887665, 0.02869078516960144, 0.11042387783527374, -0.06410638988018036, -0.01557688508182764, -0.006117052864283323, 0.006403953768312931, -0.052458904683589935, 0.2050091177225113, 0.04262058809399605, 0.04175184294581413, -0.09979036450386047, 0.07624100893735886, 0.07054848223924637, -0.08902089297771454, 0.01619834080338478, 0.10298233479261398, -0.07173021882772446, -0.04470965638756752, 0.10295794159173965, 0.18799014389514923, -0.0716748759150505, -0.054219335317611694, -0.14754971861839294, -0.1280144602060318, 0.08343587070703506, 0.16131415963172913, 0.09313396364450455, 0.007549448404461145, -0.05327737703919411, 0.009944615885615349, -0.1177566647529602, 0.08184123784303665, 0.05848805978894234, 0.06333864480257034, -0.11996172368526459, 0.17138898372650146, 0.017737073823809624, 0.03105694241821766, -0.01722029596567154, 0.01946418359875679, -0.09705539792776108, 0.019925499334931374, -0.15876075625419617, -0.029074162244796753, -0.020942136645317078, 0.0009028622880578041, -0.008452793583273888, -0.05427321046590805, -0.05392909422516823, 0.015804003924131393, -0.12008407711982727, -0.030495576560497284, 0.011729372665286064, 0.053578976541757584, -0.12258152663707733, -0.04335549473762512, 0.027214687317609787, -0.05901143327355385, 0.05841781198978424, 0.037109021097421646, 0.0103760976344347, 0.05331910401582718, -0.1557071954011917, -0.0023270107340067625, 0.055356234312057495, 0.015400429256260395, 0.05510358512401581, -0.10649142414331436, -0.011317585594952106, 0.01198115386068821, 0.06615757942199707, 0.011833176016807556, 0.07519668340682983, -0.1325407326221466, -0.01523352786898613, -0.019456129521131516, -0.09154514223337173, -0.05939340218901634, 0.033159635961055756, 0.0717712789773941, 0.02496091090142727, 0.18945281207561493, -0.08803152292966843, 0.049699075520038605, -0.21472428739070892, 0.0038321211468428373, -0.014470343478024006, -0.11019500344991684, -0.1173701286315918, -0.07452867925167084, 0.06759625673294067, -0.05251024663448334, 0.13292908668518066, 0.021037813276052475, 0.057140760123729706, 0.0304748322814703, -0.038406092673540115, 0.00577499158680439, 0.019265905022621155, 0.20587655901908875, 0.03435846418142319, -0.03219500556588173, 0.06764813512563705, 0.05572960898280144, 0.08600502461194992, 0.1285698413848877, 0.19800247251987457, 0.1590302437543869, 0.021843446418642998, 0.07918921858072281, 0.04061319679021835, -0.05178162455558777, -0.13923367857933044, 0.047158997505903244, -0.02539406530559063, 0.10414725542068481, -0.03412703052163124, 0.2301473766565323, 0.07381507754325867, -0.16665711998939514, 0.060820817947387695, -0.05856316536664963, -0.08320159465074539, -0.10873010754585266, -0.04351791739463806, -0.0879126712679863, -0.14928403496742249, -0.008818564005196095, -0.10968198627233505, 0.04137770086526871, 0.09679077565670013, 0.012256140820682049, -0.024670546874403954, 0.12618574500083923, 0.041815049946308136, 0.004109008237719536, 0.04980763792991638, -0.0038069842848926783, -0.017407245934009552, -0.10616111010313034, -0.07752761244773865, -0.002031184732913971, -0.00035701438901014626, 0.03581426292657852, -0.03844163566827774, -0.06307931244373322, 0.036188434809446335, -0.039466917514801025, -0.10083185881376266, 0.018350370228290558, 0.016830839216709137, 0.07902336120605469, 0.06808411329984665, 0.013597323559224606, 0.006269005127251148, -0.00786199513822794, 0.23300839960575104, -0.07073984295129776, -0.10419963300228119, -0.09793763607740402, 0.26484137773513794, 0.033428456634283066, -0.01740160398185253, 0.025313038378953934, -0.057432834059000015, -0.002584831090644002, 0.24964624643325806, 0.19588227570056915, -0.09458409249782562, -0.015881314873695374, 0.00861701462417841, -0.010505443438887596, -0.02420739084482193, 0.11767400801181793, 0.1486971229314804, 0.03899753838777542, -0.10414429008960724, -0.03259214013814926, -0.0585840567946434, -0.011406639590859413, -0.05947783589363098, 0.06826688349246979, 0.032769493758678436, 0.001064255484379828, -0.029770469292998314, 0.06272461265325546, -0.05815092474222183, -0.08344247192144394, 0.01808124966919422, -0.20400370657444, -0.16124103963375092, -0.016623472794890404, 0.11785387992858887, 0.0013436462031677365, 0.05341049283742905, -0.026841215789318085, 0.017995573580265045, 0.0729396715760231, -0.026499157771468163, -0.07101095467805862, -0.08631309121847153, 0.09626344591379166, -0.13334153592586517, 0.18573424220085144, -0.043481599539518356, 0.05581410974264145, 0.12740737199783325, 0.06772097200155258, -0.0697949156165123, 0.07902620732784271, 0.03936053067445755, -0.07416538894176483, 0.03008655644953251, 0.0955444723367691, -0.03669044002890587, 0.0620080940425396, 0.04833829775452614, -0.12803326547145844, 0.0352826789021492, -0.08036347478628159, -0.050406068563461304, -0.020906364545226097, -0.043841950595378876, -0.05225076898932457, 0.12574462592601776, 0.21636784076690674, -0.027452873066067696, 0.019237477332353592, -0.08286303281784058, 0.0032589060720056295, 0.05018598213791847, 0.04007723182439804, -0.07625171542167664, -0.2329988181591034, 0.003988604061305523, 0.06324660778045654, -0.008761893957853317, -0.24502845108509064, -0.10347574204206467, 0.0017178135458379984, -0.07943646609783173, -0.1006859615445137, 0.09682606160640717, 0.08097605407238007, 0.04936699941754341, -0.05090503767132759, -0.0955188050866127, -0.07206091284751892, 0.16388894617557526, -0.14609870314598083, -0.07747668027877808 ]
null
null
transformers
# Similarity between two sentences (fine-tuning with KoELECTRA-Small-v3 model and KorSTS dataset) ## Usage (Amazon SageMaker inference applicable) It uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint. ### inference_korsts.py ```python import json import sys import logging import torch from torch import nn from transformers import ElectraConfig from transformers import ElectraModel, AutoTokenizer, ElectraTokenizer, ElectraForSequenceClassification logging.basicConfig( level=logging.INFO, format='[{%(filename)s:%(lineno)d} %(levelname)s - %(message)s', handlers=[ logging.FileHandler(filename='tmp.log'), logging.StreamHandler(sys.stdout) ] ) logger = logging.getLogger(__name__) max_seq_length = 128 tokenizer = AutoTokenizer.from_pretrained("daekeun-ml/koelectra-small-v3-korsts") device = torch.device("cuda" if torch.cuda.is_available() else "cpu") # Huggingface pre-trained model: 'monologg/koelectra-small-v3-discriminator' def model_fn(model_path): #### # If you have your own trained model # Huggingface pre-trained model: 'monologg/koelectra-small-v3-discriminator' #### #config = ElectraConfig.from_json_file(f'{model_path}/config.json') #model = ElectraForSequenceClassification.from_pretrained(f'{model_path}/model.pth', config=config) model = ElectraForSequenceClassification.from_pretrained('daekeun-ml/koelectra-small-v3-korsts') model.to(device) return model def input_fn(input_data, content_type="application/jsonlines"): data_str = input_data.decode("utf-8") jsonlines = data_str.split("\n") transformed_inputs = [] for jsonline in jsonlines: text = json.loads(jsonline)["text"] logger.info("input text: {}".format(text)) encode_plus_token = tokenizer.encode_plus( text, max_length=max_seq_length, add_special_tokens=True, return_token_type_ids=False, padding="max_length", return_attention_mask=True, return_tensors="pt", truncation=True, ) transformed_inputs.append(encode_plus_token) return transformed_inputs def predict_fn(transformed_inputs, model): predicted_classes = [] for data in transformed_inputs: data = data.to(device) output = model(**data) prediction_dict = {} prediction_dict['score'] = output[0].squeeze().cpu().detach().numpy().tolist() jsonline = json.dumps(prediction_dict) logger.info("jsonline: {}".format(jsonline)) predicted_classes.append(jsonline) predicted_classes_jsonlines = "\n".join(predicted_classes) return predicted_classes_jsonlines def output_fn(outputs, accept="application/jsonlines"): return outputs, accept ``` ### test.py ```python >>> from inference_korsts import model_fn, input_fn, predict_fn, output_fn >>> with open('./samples/korsts.txt', mode='rb') as file: >>> model_input_data = file.read() >>> model = model_fn() >>> transformed_inputs = input_fn(model_input_data) >>> predicted_classes_jsonlines = predict_fn(transformed_inputs, model) >>> model_outputs = output_fn(predicted_classes_jsonlines) >>> print(model_outputs[0]) [{inference_korsts.py:44} INFO - input text: ['맛있는 라면을 먹고 싶어요', '후루룩 쩝쩝 후루룩 쩝쩝 맛좋은 라면'] [{inference_korsts.py:44} INFO - input text: ['뽀로로는 내친구', '머신러닝은 러닝머신이 아닙니다.'] [{inference_korsts.py:71} INFO - jsonline: {"score": 4.786738872528076} [{inference_korsts.py:71} INFO - jsonline: {"score": 0.2319069355726242} {"score": 4.786738872528076} {"score": 0.2319069355726242} ``` ### Sample data (samples/korsts.txt) ``` {"text": ["맛있는 라면을 먹고 싶어요", "후루룩 쩝쩝 후루룩 쩝쩝 맛좋은 라면"]} {"text": ["뽀로로는 내친구", "머신러닝은 러닝머신이 아닙니다."]} ``` ## References - KoELECTRA: https://github.com/monologg/KoELECTRA - KorNLI and KorSTS Dataset: https://github.com/kakaobrain/KorNLUDatasets
{"language": ["ko"], "license": "cc-by-4.0", "tags": ["sentence-similarity", "transformers"], "datasets": ["korsts"], "metrics": ["accuracy", "f1", "precision", "recall"], "pipeline_tag": "sentence-similarity"}
sentence-similarity
daekeun-ml/koelectra-small-v3-korsts
[ "transformers", "pytorch", "electra", "text-classification", "sentence-similarity", "ko", "dataset:korsts", "license:cc-by-4.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "ko" ]
TAGS #transformers #pytorch #electra #text-classification #sentence-similarity #ko #dataset-korsts #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #region-us
# Similarity between two sentences (fine-tuning with KoELECTRA-Small-v3 model and KorSTS dataset) ## Usage (Amazon SageMaker inference applicable) It uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint. ### inference_korsts.py ### URL ### Sample data (samples/URL) ## References - KoELECTRA: URL - KorNLI and KorSTS Dataset: URL
[ "# Similarity between two sentences (fine-tuning with KoELECTRA-Small-v3 model and KorSTS dataset)", "## Usage (Amazon SageMaker inference applicable)\nIt uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint.", "### inference_korsts.py", "### URL", "### Sample data (samples/URL)", "## References\n- KoELECTRA: URL\n- KorNLI and KorSTS Dataset: URL" ]
[ "TAGS\n#transformers #pytorch #electra #text-classification #sentence-similarity #ko #dataset-korsts #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# Similarity between two sentences (fine-tuning with KoELECTRA-Small-v3 model and KorSTS dataset)", "## Usage (Amazon SageMaker inference applicable)\nIt uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint.", "### inference_korsts.py", "### URL", "### Sample data (samples/URL)", "## References\n- KoELECTRA: URL\n- KorNLI and KorSTS Dataset: URL" ]
[ 62, 31, 49, 10, 3, 11, 22 ]
[ "passage: TAGS\n#transformers #pytorch #electra #text-classification #sentence-similarity #ko #dataset-korsts #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #region-us \n# Similarity between two sentences (fine-tuning with KoELECTRA-Small-v3 model and KorSTS dataset)## Usage (Amazon SageMaker inference applicable)\nIt uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint.### inference_korsts.py### URL### Sample data (samples/URL)## References\n- KoELECTRA: URL\n- KorNLI and KorSTS Dataset: URL" ]
[ -0.0047347466461360455, 0.06698337197303772, -0.0034649649169296026, 0.03054649755358696, 0.11573665589094162, 0.024143967777490616, 0.06707829236984253, 0.12878170609474182, 0.04050864651799202, 0.009739414788782597, 0.07820677757263184, 0.17552989721298218, 0.04694357514381409, 0.06502728909254074, -0.10635994374752045, -0.3591364026069641, 0.06702136248350143, 0.016598263755440712, 0.09348224848508835, 0.10162772983312607, 0.15506428480148315, -0.054790496826171875, 0.0952109768986702, 0.026718780398368835, -0.020608941093087196, -0.026242922991514206, 0.06476584821939468, -0.13329625129699707, 0.10557737946510315, 0.027673371136188507, 0.08154527842998505, 0.0254717618227005, -0.002081052865833044, -0.2279682606458664, 0.031787388026714325, -0.024102959781885147, 0.037753842771053314, 0.021855339407920837, -0.003984875977039337, -0.20847927033901215, 0.12718260288238525, -0.04636581242084503, 0.07798516750335693, -0.011021039448678493, -0.10081597417593002, 0.03733521327376366, -0.07793368399143219, 0.06660270690917969, 0.1535687893629074, 0.04583572968840599, 0.032869063317775726, 0.04321423918008804, -0.07960789650678635, 0.09384794533252716, 0.13016077876091003, -0.1840221881866455, -0.04823864996433258, 0.08110043406486511, 0.0010659644613042474, 0.030622737482190132, -0.029681464657187462, 0.04665793851017952, -0.016974058002233505, 0.03591008484363556, -0.01858513429760933, -0.09455232322216034, -0.1471092253923416, 0.05300560221076012, -0.07636331021785736, -0.10084305703639984, 0.2835908532142639, -0.03372899442911148, 0.024038629606366158, -0.02275840938091278, -0.03667881339788437, 0.057394422590732574, -0.04198109731078148, 0.030336866155266762, -0.08690455555915833, -0.012847846373915672, -0.07601921260356903, 0.01795308105647564, -0.10957812517881393, -0.022847551852464676, -0.09131006896495819, 0.16773216426372528, 0.029936296865344048, 0.0419977605342865, -0.09686123579740524, 0.08896957337856293, -0.10706678777933121, -0.1401028335094452, -0.046550218015909195, -0.11809735745191574, 0.01889953389763832, -0.040724221616983414, -0.08290157467126846, -0.027615023776888847, 0.0985589548945427, 0.24734316766262054, 0.09044591337442398, 0.04442659392952919, 0.06372855603694916, 0.03930523619055748, 0.04060497507452965, 0.14790977537631989, -0.017625711858272552, -0.09329340606927872, 0.03682626038789749, 0.0008369545103050768, -0.008333789184689522, -0.03479443117976189, -0.09258126467466354, -0.1131964847445488, 0.07100503891706467, 0.007299143821001053, 0.03272382542490959, 0.1145659014582634, -0.004818365443497896, -0.03675428032875061, 0.012663561850786209, -0.07788670808076859, -0.0026487845461815596, -0.027539394795894623, -0.048811428248882294, 0.09620541334152222, -0.0013260728446766734, 0.08587871491909027, -0.05820903927087784, 0.1318535953760147, -0.033025726675987244, 0.05327574163675308, 0.012741033919155598, -0.06498956680297852, 0.028794562444090843, -0.13031615316867828, 0.05637364089488983, -0.2128443717956543, -0.19766803085803986, -0.022274114191532135, 0.0786905512213707, -0.019139518961310387, -0.058186691254377365, -0.06133083999156952, 0.01603388600051403, 0.0692712739109993, -0.016172679141163826, -0.1240840032696724, -0.09123723208904266, 0.021221475675702095, -0.08880855143070221, 0.069794662296772, -0.044349510222673416, 0.03370276838541031, -0.1464269906282425, -0.005079986993223429, -0.06594758480787277, 0.028122151270508766, -0.0701879933476448, 0.04013703018426895, -0.13636989891529083, -0.031493909657001495, -0.03477768227458, -0.016343610361218452, 0.0021623920183628798, 0.09788310527801514, -0.1359030306339264, 0.0012926706112921238, 0.044969502836465836, -0.05896329507231712, -0.1190456673502922, 0.11680924892425537, -0.04114580526947975, 0.03707657381892204, 0.10500764101743698, 0.1058821976184845, 0.18054896593093872, -0.002578007522970438, 0.029525861144065857, 0.04436030611395836, -0.0386032834649086, 0.056355901062488556, 0.027537574991583824, 0.014164774678647518, -0.10867151618003845, 0.0594293549656868, 0.04574683681130409, 0.05342676118016243, -0.036610864102840424, -0.04707600548863411, -0.04601868614554405, 0.009590329602360725, 0.02757030539214611, -0.04460293799638748, 0.001990234712138772, -0.00813654251396656, -0.007960189133882523, 0.03975597769021988, 0.08135986328125, -0.05554728955030441, 0.03506964445114136, -0.05718698725104332, 0.14690831303596497, -0.09786899387836456, -0.008885247632861137, -0.12564340233802795, -0.015747999772429466, -0.019216695800423622, 0.1432356834411621, 0.020975954830646515, -0.035364437848329544, 0.020245637744665146, -0.050497423857450485, -0.016525186598300934, 0.01811681129038334, 0.09831929951906204, 0.06381603330373764, -0.04247910901904106, -0.06151831895112991, 0.033495888113975525, -0.011487774550914764, -0.0011394458124414086, -0.024601496756076813, -0.022517437115311623, 0.08067267388105392, 0.14624866843223572, -0.04694090411067009, 0.07153759151697159, 0.04039362445473671, 0.03008866123855114, -0.07035024464130402, 0.01813081093132496, 0.06603799015283585, -0.02687647007405758, -0.055146921426057816, 0.21593059599399567, -0.10926887392997742, -0.10090675204992294, 0.07466711103916168, -0.05146265774965286, -0.014075319282710552, -0.10754796862602234, -0.02391841448843479, -0.011808446608483791, -0.08884567022323608, -0.02479969896376133, 0.19303607940673828, 0.02603609301149845, 0.10398738831281662, -0.07134929299354553, -0.02641408145427704, 0.030620869249105453, -0.10614041984081268, 0.003890802152454853, 0.05519390106201172, 0.028778119012713432, -0.18245956301689148, 0.09254466742277145, 0.09362498670816422, -0.0808616653084755, 0.06530854851007462, -0.03128194436430931, -0.038553688675165176, -0.00008924516441766173, -0.015624449588358402, -0.015844881534576416, 0.012938220985233784, -0.04051800072193146, -0.0013284505112096667, 0.12480886280536652, 0.06790155172348022, 0.0033283792436122894, -0.11112239211797714, 0.004745683167129755, 0.02383899688720703, 0.01764102466404438, -0.012023720890283585, 0.09021028131246567, 0.08010226488113403, 0.10574827343225479, 0.015115800313651562, -0.08780558407306671, 0.02556663565337658, 0.014809314161539078, -0.05196920409798622, 0.21710854768753052, -0.06904268264770508, -0.1560797542333603, -0.1171770840883255, -0.07013901323080063, -0.06696134060621262, -0.0026401132345199585, 0.061644598841667175, 0.031209217384457588, -0.05773366615176201, -0.06615184992551804, 0.01138634979724884, 0.04746332764625549, -0.008994297124445438, -0.16440294682979584, 0.0028075878508388996, -0.09015470743179321, -0.06607478857040405, -0.0730118453502655, 0.03814960643649101, -0.008454028517007828, 0.132660374045372, -0.08826610445976257, 0.05384072661399841, 0.11824871599674225, -0.06667730957269669, 0.021299075335264206, -0.02919897437095642, 0.18274137377738953, -0.03430095687508583, 0.07659627497196198, 0.19104690849781036, -0.07248382270336151, 0.0771467387676239, 0.20081181824207306, -0.010332014411687851, -0.008540507405996323, 0.05736943706870079, -0.015466660261154175, -0.07924813032150269, -0.14250971376895905, -0.09916990250349045, -0.0741472914814949, 0.07111512869596481, 0.11465084552764893, -0.005645128898322582, 0.07468266040086746, 0.1460951268672943, -0.06931548565626144, 0.037902481853961945, 0.10206445306539536, 0.14787845313549042, 0.1948300302028656, 0.07458969205617905, 0.11542508006095886, 0.0032028353307396173, -0.08436284214258194, 0.018628671765327454, 0.11525445431470871, 0.13860784471035004, -0.024605561047792435, 0.027958540245890617, 0.0825359970331192, -0.013789626769721508, 0.12659163773059845, 0.05612264573574066, -0.02669862098991871, 0.03061514161527157, -0.025306299328804016, -0.04269169643521309, -0.049560461193323135, 0.10885930806398392, -0.035389360040426254, -0.03067903034389019, -0.032764460891485214, 0.010890682227909565, 0.07840508222579956, 0.1526631861925125, 0.12943609058856964, -0.2499615103006363, -0.06466992199420929, 0.05711686983704567, -0.05146053060889244, -0.05732480436563492, 0.08361682295799255, -0.02486833557486534, -0.16373465955257416, 0.09590458124876022, -0.044703159481287, 0.12171551585197449, -0.10072144865989685, -0.022969935089349747, -0.06808657944202423, -0.044528715312480927, -0.020767943933606148, 0.11286257952451706, -0.17903868854045868, 0.17581471800804138, 0.013670187443494797, -0.02859630435705185, -0.09505024552345276, -0.030122527852654457, 0.049085695296525955, 0.13617222011089325, 0.04900659620761871, 0.026307910680770874, -0.20467129349708557, -0.13605576753616333, -0.06242190673947334, 0.05721459165215492, 0.038968123495578766, -0.09793967753648758, 0.0773635134100914, -0.09148816764354706, -0.02434861846268177, -0.04877699539065361, -0.06265369802713394, -0.05520952120423317, -0.16063667833805084, 0.02957036904990673, 0.040427159518003464, 0.09493216872215271, 0.005810496862977743, -0.013845046982169151, -0.020243311300873756, 0.020127352327108383, -0.04833899810910225, -0.09718294441699982, -0.10873610526323318, -0.017401957884430885, 0.10564718395471573, -0.09924653172492981, 0.04439390078186989, -0.05166979879140854, 0.06383433938026428, -0.006911689881235361, -0.13142766058444977, 0.0016416975995525718, -0.06476225703954697, -0.10809382796287537, 0.0005319464835338295, 0.13058793544769287, -0.02622077614068985, 0.0315069779753685, 0.06032245233654976, 0.09964428842067719, -0.07464274764060974, -0.05698271468281746, -0.09677543491125107, 0.23291191458702087, 0.10774855315685272, 0.04343578964471817, -0.0007272930815815926, 0.023972835391759872, -0.10081593692302704, 0.015704168006777763, 0.19455617666244507, 0.05054980143904686, -0.05172601342201233, 0.08483245968818665, 0.011247311718761921, -0.062053464353084564, -0.24973522126674652, -0.11767413467168808, -0.011349168606102467, -0.00004544588955468498, -0.07851158082485199, 0.0013691209023818374, 0.12632596492767334, -0.0040120952762663364, 0.006229415535926819, -0.059041742235422134, -0.207528218626976, -0.11996662616729736, 0.10177972167730331, -0.001314297434873879, 0.1708412766456604, -0.019580908119678497, -0.025016143918037415, -0.06228463351726532, -0.18050846457481384, 0.17034892737865448, 0.05075979605317116, 0.08136950433254242, -0.03735990449786186, 0.16268689930438995, 0.04100663959980011, -0.019745321944355965, 0.12817302346229553, 0.027777165174484253, 0.03292261064052582, -0.10049614310264587, -0.09970618784427643, -0.08059626817703247, -0.07754673808813095, 0.11599671840667725, -0.05627911910414696, 0.07178729772567749, -0.10516287386417389, -0.02294999733567238, -0.08401598036289215, 0.08642921596765518, 0.04452158510684967, -0.07071185857057571, -0.09144745022058487, 0.008736158721148968, 0.10409601032733917, -0.02113913744688034, 0.2978004217147827, -0.017782825976610184, 0.023674875497817993, 0.04828701168298721, 0.12035909295082092, -0.04376678541302681, -0.0000375343079213053, -0.040719639509916306, -0.052203476428985596, 0.06972648203372955, -0.18295533955097198, 0.059771209955215454, 0.12584635615348816, -0.0007742413436062634, 0.08752132207155228, 0.030344625934958458, -0.013708295300602913, 0.028277361765503883, 0.0678737461566925, -0.1379864513874054, -0.04028946906328201, -0.0701940730214119, 0.03341656178236008, 0.002701857592910528, -0.06182738021016121, 0.2008303552865982, -0.112788125872612, -0.06249849125742912, -0.010240797884762287, 0.016884448006749153, -0.030297618359327316, 0.1038190945982933, 0.043900832533836365, 0.05235017463564873, -0.05591624602675438, 0.07384736835956573, 0.028866160660982132, -0.15106546878814697, 0.047681279480457306, -0.013696850277483463, -0.1798030585050583, -0.07290016114711761, -0.021919790655374527, 0.03867126628756523, -0.11491549015045166, -0.06272778660058975, -0.08546964079141617, -0.05440026894211769, 0.01754038594663143, 0.26310229301452637, 0.07064638286828995, 0.06616653501987457, -0.015819698572158813, -0.04610447213053703, -0.0499856099486351, 0.11068268865346909, 0.0801640972495079, 0.07269149273633957, -0.13611233234405518, -0.011610066518187523, -0.06003193184733391, 0.09697762131690979, -0.03542131558060646, -0.06113288924098015, -0.09681657701730728, -0.002543934155255556, -0.17329226434230804, 0.06583770364522934, -0.19124391674995422, -0.0146690234541893, -0.008201994001865387, -0.07730387151241302, -0.026386970654129982, -0.017039431259036064, -0.0699889287352562, 0.020559048280119896, -0.08522441983222961, 0.10399357229471207, -0.0066791316494345665, -0.05822314694523811, 0.09705992043018341, -0.03706233948469162, 0.04381147027015686, 0.08670688420534134, -0.06898224353790283, 0.03320098668336868, -0.18770034611225128, -0.047302138060331345, 0.0685649961233139, 0.037804849445819855, 0.03260720893740654, -0.14315827190876007, 0.05502789095044136, 0.01958625018596649, 0.023919573053717613, 0.006578782573342323, 0.07972851395606995, -0.1308744102716446, -0.04521145299077034, 0.01997970975935459, -0.10257163643836975, -0.09267932921648026, -0.06278030574321747, 0.0797363668680191, 0.10620792210102081, 0.18000546097755432, -0.06776890903711319, 0.08586819469928741, -0.06019099801778793, 0.00802935753017664, -0.03433075174689293, -0.1618131697177887, -0.07327158004045486, -0.10998503118753433, 0.008108248934149742, -0.025836054235696793, 0.09860865771770477, 0.048317596316337585, -0.012455254793167114, 0.02782413735985756, 0.07568619400262833, 0.1480725109577179, 0.06667952239513397, 0.08527477085590363, 0.06806197762489319, -0.056923750787973404, -0.09170914441347122, 0.050047848373651505, 0.0517127588391304, 0.001207811408676207, 0.006005277391523123, 0.12321988493204117, 0.061931923031806946, 0.07083293795585632, 0.034842077642679214, 0.060097772628068924, 0.03873613476753235, -0.20612744987010956, -0.030887609347701073, 0.07087163627147675, -0.0239862073212862, 0.023396093398332596, 0.2147020846605301, -0.1340850442647934, 0.02039594016969204, -0.06484381854534149, -0.08815586566925049, -0.1295931190252304, -0.3253213167190552, -0.136638805270195, -0.10740829259157181, 0.024640418589115143, -0.127376526594162, -0.006660169456154108, 0.14267998933792114, 0.012675908394157887, 0.002425525104627013, 0.07537668943405151, -0.04466095194220543, -0.08130687475204468, 0.09743896871805191, -0.022760892286896706, 0.02442992851138115, 0.0988377258181572, -0.041378796100616455, -0.01671222411096096, -0.09076816588640213, -0.01748596504330635, 0.03795266151428223, 0.09094393253326416, -0.03035118617117405, -0.11668092757463455, -0.09070106595754623, 0.017284706234931946, -0.02157043106853962, 0.01052123960107565, -0.03413090854883194, 0.08679433912038803, 0.029859639704227448, 0.07145636528730392, 0.32008370757102966, -0.029655376449227333, -0.19362281262874603, -0.18223567306995392, 0.15175648033618927, 0.08199354261159897, 0.04329860210418701, 0.029489092528820038, -0.05374959856271744, -0.03846396505832672, 0.16031131148338318, 0.10079900175333023, 0.06540550291538239, 0.013679973781108856, -0.04031546413898468, 0.02943997085094452, 0.06876830756664276, 0.041694946587085724, 0.0008601690060459077, 0.09930795431137085, -0.06957045197486877, -0.05109734088182449, -0.022592982277274132, -0.03198257088661194, -0.06881614774465561, 0.07278982549905777, 0.018242783844470978, -0.08260339498519897, -0.04561510309576988, 0.1345381885766983, -0.17034970223903656, 0.17451101541519165, -0.10364153236150742, -0.1771167367696762, -0.11966773122549057, -0.022080525755882263, 0.09588928520679474, -0.0002662861661519855, 0.06859806180000305, -0.017479099333286285, -0.04662610962986946, 0.012261820025742054, 0.03096942976117134, -0.12421666830778122, 0.030777614563703537, 0.056639038026332855, 0.04994822293519974, -0.02273593097925186, -0.014569342136383057, 0.12482354789972305, 0.11352427303791046, 0.061579033732414246, -0.050142403692007065, 0.042101986706256866, 0.08091483265161514, 0.03684598207473755, 0.08040231466293335, 0.0838453471660614, 0.014814468100667, 0.038095902651548386, 0.12070711702108383, -0.09384725242853165, 0.0716153085231781, -0.018951235339045525, 0.03379041701555252, -0.09398557990789413, 0.047419045120477676, -0.02369660511612892, 0.11583822220563889, 0.09435641020536423, -0.07216659188270569, -0.0008860358502715826, -0.05256715416908264, -0.006230961065739393, 0.03302246704697609, -0.07336342334747314, -0.05973894149065018, -0.09758800268173218, -0.09193554520606995, -0.0006679300568066537, 0.02428250014781952, -0.058878093957901, -0.02775461971759796, -0.047869253903627396, -0.024115782231092453, -0.01947021484375, 0.14644072949886322, 0.015592047944664955, -0.007385212928056717, -0.033241450786590576, -0.09923451393842697, 0.03018229454755783, 0.09124084562063217, -0.1459631472826004, -0.11965306103229523 ]
null
null
transformers
# Sentiment Binary Classification (fine-tuning with KoELECTRA-Small-v3 model and Naver Sentiment Movie Corpus dataset) ## Usage (Amazon SageMaker inference applicable) It uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint. ### inference_nsmc.py ```python import json import sys import logging import torch from torch import nn from transformers import ElectraConfig from transformers import ElectraModel, AutoTokenizer, ElectraTokenizer, ElectraForSequenceClassification logging.basicConfig( level=logging.INFO, format='[{%(filename)s:%(lineno)d} %(levelname)s - %(message)s', handlers=[ logging.FileHandler(filename='tmp.log'), logging.StreamHandler(sys.stdout) ] ) logger = logging.getLogger(__name__) max_seq_length = 128 classes = ['Neg', 'Pos'] tokenizer = AutoTokenizer.from_pretrained("daekeun-ml/koelectra-small-v3-nsmc") device = torch.device("cuda" if torch.cuda.is_available() else "cpu") def model_fn(model_path=None): #### # If you have your own trained model # Huggingface pre-trained model: 'monologg/koelectra-small-v3-discriminator' #### #config = ElectraConfig.from_json_file(f'{model_path}/config.json') #model = ElectraForSequenceClassification.from_pretrained(f'{model_path}/model.pth', config=config) # Download model from the Huggingface hub model = ElectraForSequenceClassification.from_pretrained('daekeun-ml/koelectra-small-v3-nsmc') model.to(device) return model def input_fn(input_data, content_type="application/jsonlines"): data_str = input_data.decode("utf-8") jsonlines = data_str.split("\n") transformed_inputs = [] for jsonline in jsonlines: text = json.loads(jsonline)["text"][0] logger.info("input text: {}".format(text)) encode_plus_token = tokenizer.encode_plus( text, max_length=max_seq_length, add_special_tokens=True, return_token_type_ids=False, padding="max_length", return_attention_mask=True, return_tensors="pt", truncation=True, ) transformed_inputs.append(encode_plus_token) return transformed_inputs def predict_fn(transformed_inputs, model): predicted_classes = [] for data in transformed_inputs: data = data.to(device) output = model(**data) softmax_fn = nn.Softmax(dim=1) softmax_output = softmax_fn(output[0]) _, prediction = torch.max(softmax_output, dim=1) predicted_class_idx = prediction.item() predicted_class = classes[predicted_class_idx] score = softmax_output[0][predicted_class_idx] logger.info("predicted_class: {}".format(predicted_class)) prediction_dict = {} prediction_dict["predicted_label"] = predicted_class prediction_dict['score'] = score.cpu().detach().numpy().tolist() jsonline = json.dumps(prediction_dict) logger.info("jsonline: {}".format(jsonline)) predicted_classes.append(jsonline) predicted_classes_jsonlines = "\n".join(predicted_classes) return predicted_classes_jsonlines def output_fn(outputs, accept="application/jsonlines"): return outputs, accept ``` ### test.py ```python >>> from inference_nsmc import model_fn, input_fn, predict_fn, output_fn >>> with open('samples/nsmc.txt', mode='rb') as file: >>> model_input_data = file.read() >>> model = model_fn() >>> transformed_inputs = input_fn(model_input_data) >>> predicted_classes_jsonlines = predict_fn(transformed_inputs, model) >>> model_outputs = output_fn(predicted_classes_jsonlines) >>> print(model_outputs[0]) [{inference_nsmc.py:47} INFO - input text: 이 영화는 최고의 영화입니다 [{inference_nsmc.py:47} INFO - input text: 최악이에요. 배우의 연기력도 좋지 않고 내용도 너무 허접합니다 [{inference_nsmc.py:77} INFO - predicted_class: Pos [{inference_nsmc.py:84} INFO - jsonline: {"predicted_label": "Pos", "score": 0.9619030952453613} [{inference_nsmc.py:77} INFO - predicted_class: Neg [{inference_nsmc.py:84} INFO - jsonline: {"predicted_label": "Neg", "score": 0.9994170665740967} {"predicted_label": "Pos", "score": 0.9619030952453613} {"predicted_label": "Neg", "score": 0.9994170665740967} ``` ### Sample data (samples/nsmc.txt) ``` {"text": ["이 영화는 최고의 영화입니다"]} {"text": ["최악이에요. 배우의 연기력도 좋지 않고 내용도 너무 허접합니다"]} ``` ## References - KoELECTRA: https://github.com/monologg/KoELECTRA - Naver Sentiment Movie Corpus Dataset: https://github.com/e9t/nsmc
{"language": ["ko"], "license": "mit", "tags": ["classification"], "datasets": ["nsmc"], "metrics": ["accuracy", "f1", "precision", "recall- accuracy"], "widget": [{"text": "\ubd88\ud6c4\uc758 \uba85\uc791\uc785\ub2c8\ub2e4! \uc774\ub807\uac8c \uac10\ub3d9\uc801\uc778 \ub0b4\uc6a9\uc740 \ucc98\uc74c\uc774\uc5d0\uc694", "example_title": "Positive"}, {"text": "\uc2dc\uac04\uc774 \uc815\ub9d0 \uc544\uae5d\uc2b5\ub2c8\ub2e4. 10\uc810 \ub9cc\uc810\uc5d0 1\uc810\ub3c4 \uc544\uae4c\uc6cc\uc694..", "example_title": "Negative"}]}
text-classification
daekeun-ml/koelectra-small-v3-nsmc
[ "transformers", "pytorch", "electra", "text-classification", "classification", "ko", "dataset:nsmc", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "ko" ]
TAGS #transformers #pytorch #electra #text-classification #classification #ko #dataset-nsmc #license-mit #autotrain_compatible #endpoints_compatible #region-us
# Sentiment Binary Classification (fine-tuning with KoELECTRA-Small-v3 model and Naver Sentiment Movie Corpus dataset) ## Usage (Amazon SageMaker inference applicable) It uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint. ### inference_nsmc.py ### URL ### Sample data (samples/URL) ## References - KoELECTRA: URL - Naver Sentiment Movie Corpus Dataset: URL
[ "# Sentiment Binary Classification (fine-tuning with KoELECTRA-Small-v3 model and Naver Sentiment Movie Corpus dataset)", "## Usage (Amazon SageMaker inference applicable)\nIt uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint.", "### inference_nsmc.py", "### URL", "### Sample data (samples/URL)", "## References\n- KoELECTRA: URL\n- Naver Sentiment Movie Corpus Dataset: URL" ]
[ "TAGS\n#transformers #pytorch #electra #text-classification #classification #ko #dataset-nsmc #license-mit #autotrain_compatible #endpoints_compatible #region-us \n", "# Sentiment Binary Classification (fine-tuning with KoELECTRA-Small-v3 model and Naver Sentiment Movie Corpus dataset)", "## Usage (Amazon SageMaker inference applicable)\nIt uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint.", "### inference_nsmc.py", "### URL", "### Sample data (samples/URL)", "## References\n- KoELECTRA: URL\n- Naver Sentiment Movie Corpus Dataset: URL" ]
[ 54, 34, 49, 10, 3, 11, 21 ]
[ "passage: TAGS\n#transformers #pytorch #electra #text-classification #classification #ko #dataset-nsmc #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# Sentiment Binary Classification (fine-tuning with KoELECTRA-Small-v3 model and Naver Sentiment Movie Corpus dataset)## Usage (Amazon SageMaker inference applicable)\nIt uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint.### inference_nsmc.py### URL### Sample data (samples/URL)## References\n- KoELECTRA: URL\n- Naver Sentiment Movie Corpus Dataset: URL" ]
[ -0.014786467887461185, 0.23098772764205933, -0.0036969019565731287, 0.043974995613098145, 0.19116652011871338, 0.026381278410553932, 0.024669088423252106, 0.1588892936706543, 0.03526761382818222, -0.005228316877037287, 0.0304886307567358, 0.14527076482772827, 0.05437978729605675, 0.08582165837287903, -0.10416846722364426, -0.3227616548538208, -0.01728934980928898, 0.006590794771909714, 0.14946317672729492, 0.11417071521282196, 0.11091817170381546, -0.07812947034835815, 0.13666318356990814, -0.03172285109758377, -0.06569137424230576, -0.009367120452225208, 0.056371502578258514, -0.10664635896682739, 0.09169881790876389, 0.012304577976465225, 0.046281736344099045, -0.015329360961914062, 0.051237553358078, -0.1811957061290741, 0.042533837258815765, -0.039392128586769104, 0.03795422613620758, 0.017011111602187157, 0.05365443229675293, -0.09442657232284546, 0.23869279026985168, -0.0022331050131469965, 0.08739997446537018, 0.0023992094211280346, -0.07222570478916168, -0.05168024078011513, -0.05181548744440079, 0.009890105575323105, 0.0014564723242074251, 0.10038107633590698, 0.0029059213120490313, 0.10647036880254745, -0.08367267996072769, 0.08004851639270782, 0.11414217203855515, -0.053483542054891586, -0.062066301703453064, 0.045293353497982025, 0.027928823605179787, 0.09809951484203339, -0.11523560434579849, 0.02320566214621067, 0.002267599105834961, 0.018671337515115738, 0.042476072907447815, -0.09217848628759384, -0.09051650762557983, -0.033928725868463516, -0.0652826577425003, -0.11386865377426147, 0.28046083450317383, 0.00933451671153307, -0.03818202391266823, -0.0196248609572649, 0.0183839313685894, -0.034560125321149826, -0.08008775860071182, 0.04839877039194107, -0.030686816200613976, -0.03935310244560242, 0.01601087674498558, 0.06253810226917267, -0.074776791036129, -0.0019074787851423025, -0.021523647010326385, 0.1210978776216507, -0.0027828537859022617, 0.07793346047401428, -0.14864975214004517, 0.056301772594451904, -0.04611250385642052, -0.12227950245141983, -0.024495583027601242, -0.0895082950592041, 0.0689246878027916, -0.02250717207789421, -0.007382403127849102, 0.04455043002963066, 0.009525391273200512, 0.1853436380624771, 0.050949517637491226, 0.050600290298461914, 0.004629630129784346, 0.061313629150390625, 0.11959389597177505, 0.11586995422840118, -0.034441206604242325, -0.1121017262339592, 0.07980175316333771, -0.014027148485183716, 0.06617607176303864, -0.01205277070403099, -0.08080190420150757, -0.03997378051280975, 0.020236141979694366, 0.04109319671988487, 0.00852673128247261, 0.08891777694225311, -0.06397167593240738, -0.07176671177148819, 0.04971950501203537, -0.07102464884519577, 0.018057415261864662, 0.007768373936414719, -0.10334452986717224, 0.1569424569606781, -0.014089702628552914, -0.006978312041610479, -0.022347241640090942, 0.10470224171876907, -0.03991285711526871, 0.056124258786439896, -0.061078015714883804, -0.06690812855958939, 0.11003544926643372, -0.13919106125831604, 0.04884723573923111, -0.19139941036701202, -0.1646198332309723, -0.010904600843787193, 0.030261967331171036, -0.056530267000198364, -0.04375704750418663, -0.05320020765066147, 0.02929370477795601, 0.06283441185951233, -0.014886650256812572, -0.09688594937324524, -0.08956416696310043, 0.02074241451919079, -0.030373310670256615, 0.10341959446668625, -0.06934195011854172, 0.03121611289680004, -0.12863564491271973, -0.00448768911883235, -0.09640952199697495, 0.04089111089706421, -0.07335107773542404, 0.08373302966356277, -0.05496721342206001, -0.06580646336078644, -0.01001786533743143, -0.015914592891931534, -0.07118751108646393, 0.1373443752527237, -0.20217573642730713, -0.046861786395311356, -0.028281623497605324, -0.10239576548337936, -0.14513595402240753, 0.13106723129749298, -0.039178162813186646, 0.01235207449644804, 0.03497537598013878, 0.11990445852279663, 0.11705966293811798, -0.03967606648802757, -0.027553966268897057, 0.042570870369672775, -0.08456195890903473, -0.025789543986320496, 0.08167219161987305, 0.14871737360954285, -0.04368972405791283, 0.03367901220917702, -0.03292591869831085, 0.10548259317874908, -0.037426333874464035, -0.05085908994078636, 0.0035873963497579098, 0.011518720537424088, -0.04178987070918083, 0.01580202579498291, 0.03008280321955681, -0.00925477035343647, -0.012116505764424801, -0.03160947933793068, 0.1221301406621933, -0.011837703175842762, 0.07135732471942902, -0.09303443133831024, 0.1651427149772644, -0.02724090963602066, -0.0036860560066998005, -0.14710354804992676, 0.017668763175606728, -0.010094530880451202, 0.07047919929027557, -0.00962329562753439, -0.07307813316583633, -0.01465658750385046, -0.017254244536161423, -0.031033474951982498, 0.0008580565336160362, 0.06378711014986038, 0.03154159337282181, 0.0540018193423748, -0.1273265928030014, -0.004456499591469765, -0.06729898601770401, 0.000811090343631804, -0.07384240627288818, -0.01874241605401039, 0.132750004529953, 0.12240689992904663, -0.006369890179485083, -0.00783210713416338, 0.049473654478788376, 0.030738942325115204, -0.05728776752948761, -0.051307063549757004, 0.064403235912323, -0.025377925485372543, -0.03319472819566727, 0.12164656817913055, -0.13363812863826752, 0.02177814580500126, 0.12502504885196686, -0.11848224699497223, -0.03276360407471657, 0.02552836947143078, -0.029884299263358116, 0.03510993346571922, -0.07254363596439362, 0.03736807033419609, 0.10348746180534363, 0.02615479938685894, 0.12066330760717392, -0.04073816165328026, -0.018006693571805954, 0.0507483147084713, -0.08097687363624573, -0.057009413838386536, 0.06834255903959274, 0.11465656012296677, -0.10597960650920868, 0.12131784856319427, 0.084998719394207, -0.08477777987718582, 0.15689244866371155, 0.030655760318040848, -0.018503382802009583, 0.014791764318943024, -0.06454315036535263, 0.011920582503080368, 0.03260134533047676, -0.13153737783432007, -0.035835497081279755, 0.07768581807613373, -0.033894408494234085, -0.03219524770975113, -0.08948591351509094, -0.0553504079580307, 0.006901813670992851, 0.04877856373786926, 0.019178541377186775, 0.07600676268339157, 0.03861811012029648, 0.13310495018959045, 0.022772567346692085, -0.08138955384492874, 0.04758000373840332, -0.031007003039121628, -0.094902902841568, 0.1557968407869339, -0.06803297251462936, -0.29631781578063965, -0.0711941346526146, -0.05073485150933266, 0.014166298322379589, 0.0035305542405694723, 0.02276644855737686, -0.03782098367810249, -0.056603677570819855, -0.08126793056726456, -0.07534082233905792, -0.0190375167876482, -0.033699121326208115, -0.09326005727052689, 0.03846671059727669, -0.046766284853219986, -0.037568021565675735, -0.050626106560230255, 0.026703765615820885, -0.05044163390994072, 0.10584945976734161, -0.04423787444829941, 0.09241622686386108, 0.1309640109539032, -0.08259017020463943, 0.03870631009340286, -0.025226671248674393, 0.16189369559288025, -0.08904612809419632, 0.03801584616303444, 0.20988872647285461, -0.03143417090177536, 0.03776179999113083, 0.19987154006958008, 0.031123625114560127, -0.03131222724914551, 0.049705274403095245, -0.036030225455760956, -0.08795103430747986, -0.12180621922016144, -0.09042128175497055, -0.0922357589006424, 0.07835268974304199, 0.0872722938656807, 0.04816441237926483, 0.05605848506093025, 0.09430871903896332, -0.04367098957300186, -0.004400657955557108, 0.041434623301029205, 0.16593541204929352, 0.1631113886833191, 0.05736257880926132, 0.044560834765434265, -0.01537561696022749, -0.05304631590843201, 0.07876890897750854, 0.07296626269817352, 0.09604833275079727, 0.005352878011763096, 0.0928976759314537, 0.010334356687963009, 0.04836827889084816, 0.1373593658208847, -0.030325090512633324, 0.016015619039535522, 0.01429983600974083, -0.05249711498618126, -0.018572034314274788, -0.11972501873970032, 0.09557142853736877, -0.022887766361236572, -0.02819870598614216, -0.03204416111111641, 0.035432539880275726, 0.0765092670917511, 0.04480331018567085, 0.016723861917853355, -0.3110915422439575, -0.029058784246444702, 0.09191053360700607, 0.011903975158929825, -0.02918325550854206, 0.06207652390003204, 0.023702183738350868, -0.18390966951847076, 0.11956043541431427, -0.03245660662651062, 0.10681954771280289, -0.15609285235404968, -0.00995816569775343, -0.02982153184711933, -0.054476674646139145, 0.01972065679728985, 0.09647791087627411, -0.15545733273029327, 0.12181592732667923, -0.03986474126577377, -0.002573393750935793, -0.047919537872076035, -0.05653129890561104, 0.08628342300653458, 0.14367470145225525, 0.16286471486091614, 0.039471160620450974, -0.053029343485832214, -0.14049936830997467, 0.005156225524842739, 0.04320294037461281, -0.010403729975223541, -0.0644695907831192, 0.03897703066468239, -0.039867885410785675, -0.027814829722046852, -0.06708173453807831, -0.04060749709606171, -0.022815313190221786, -0.13119754195213318, 0.004938212689012289, 0.07570550590753555, 0.049442317336797714, 0.009141473099589348, -0.057906415313482285, 0.03911169245839119, 0.07864822447299957, -0.09696775674819946, -0.04998773708939552, -0.13919226825237274, 0.038068223744630814, 0.001504798186942935, -0.07123038172721863, 0.05631377175450325, -0.044743649661540985, 0.12764042615890503, -0.026508672162890434, -0.13455730676651, 0.06253505498170853, -0.07400516420602798, -0.06157023087143898, -0.03412189334630966, 0.07935240119695663, -0.019478144124150276, 0.02893778868019581, -0.022508971393108368, 0.03113342821598053, -0.003281483892351389, -0.08218267560005188, -0.04332635551691055, 0.23220466077327728, 0.05083479359745979, 0.0308883897960186, 0.006852016784250736, -0.09315716475248337, -0.07964806258678436, -0.009223462082445621, 0.1824350357055664, 0.07503814250230789, -0.05245130881667137, 0.05619673430919647, -0.024100812152028084, -0.08530927449464798, -0.24879080057144165, -0.06541472673416138, 0.032622888684272766, -0.06257314234972, 0.008126972243189812, -0.079521544277668, 0.1191362664103508, 0.023088522255420685, -0.0011491029290482402, -0.04640189930796623, -0.17867591977119446, -0.09726769477128983, 0.09559180587530136, 0.11874502897262573, 0.173162579536438, -0.05153924599289894, -0.005980140995234251, -0.05099638178944588, -0.1322740912437439, 0.22506025433540344, 0.06786258518695831, 0.06760526448488235, -0.06772254407405853, 0.16385787725448608, -0.0031384630128741264, -0.029496554285287857, 0.09183067828416824, 0.010700803250074387, 0.04521052539348602, -0.07269557565450668, -0.188497856259346, -0.009021544829010963, -0.04559504985809326, 0.12529966235160828, -0.029790280386805534, -0.020125404000282288, -0.1261119246482849, -0.027478961274027824, -0.09274829924106598, 0.06811977922916412, 0.03727315366268158, -0.05704324692487717, -0.07082168757915497, 0.08009348809719086, 0.04946598410606384, 0.0025263880379498005, 0.2668389081954956, -0.030607981607317924, -0.033143866807222366, 0.10479908436536789, 0.1555032581090927, 0.035138119012117386, -0.024531781673431396, -0.03841047361493111, -0.009145020507276058, 0.0764373242855072, -0.12063004821538925, 0.022044280543923378, 0.13774782419204712, 0.0003969144308939576, 0.15088137984275818, 0.04101712629199028, 0.014169160276651382, 0.06289204955101013, 0.0947217047214508, -0.07807566970586777, -0.0571761354804039, -0.05763532966375351, 0.11010994017124176, -0.009399296715855598, -0.10032239556312561, 0.09735772758722305, -0.16167287528514862, -0.05335620790719986, -0.021309897303581238, 0.03953997418284416, 0.007975631393492222, 0.05481638014316559, 0.03743136301636696, 0.03387037664651871, -0.11173570901155472, 0.10907525569200516, -0.0036016006488353014, -0.2169782519340515, 0.038497939705848694, 0.1441386193037033, -0.16604465246200562, -0.08114401996135712, -0.03994854539632797, 0.09082093089818954, -0.1583503931760788, -0.0587514191865921, -0.03980281576514244, -0.06608689576387405, 0.047132838517427444, 0.14270298182964325, 0.04959120973944664, 0.07600488513708115, -0.1158699169754982, -0.030137954279780388, -0.056378256529569626, 0.10941847413778305, 0.0634043961763382, 0.01950450800359249, -0.1822422593832016, -0.04500950127840042, 0.029273824766278267, 0.04198943451046944, -0.0650462657213211, -0.08770410716533661, -0.03860621899366379, 0.02225322276353836, -0.13650648295879364, 0.10901317745447159, -0.17019037902355194, 0.02299000695347786, -0.0316770002245903, -0.018465714529156685, -0.03802873566746712, -0.017109045758843422, -0.08351346850395203, -0.005621641408652067, -0.05213651806116104, 0.06288959085941315, -0.02945508435368538, -0.05376019328832626, 0.05875194072723389, -0.04813392087817192, 0.050861623138189316, 0.12331602722406387, -0.0334407277405262, 0.013660218566656113, -0.21975450217723846, -0.10679036378860474, 0.13727311789989471, 0.018242355436086655, 0.015714997425675392, -0.056771282106637955, 0.043886441737413406, 0.055208127945661545, -0.04339152202010155, -0.025372449308633804, 0.07022681087255478, -0.13117288053035736, 0.006901632063090801, 0.05656321346759796, -0.10165068507194519, -0.0882052332162857, 0.015734555199742317, 0.07635773718357086, 0.04053354263305664, 0.11111979186534882, -0.0789937973022461, 0.06885860860347748, -0.10345108807086945, -0.003868721891194582, -0.02096717059612274, -0.10016138106584549, -0.2332092523574829, -0.039405278861522675, 0.028639476746320724, -0.007526098750531673, 0.10084090381860733, 0.07173269987106323, -0.06609849631786346, 0.03380194306373596, 0.07838063687086105, 0.09073581546545029, 0.023493321612477303, 0.08670688420534134, 0.08614721894264221, -0.07485716789960861, -0.00706704193726182, -0.0027126066852360964, 0.05721583589911461, 0.07097835838794708, 0.016048923134803772, 0.12245520204305649, 0.09497840702533722, 0.04178434982895851, 0.009059916250407696, 0.06986033171415329, 0.024675630033016205, -0.024894770234823227, -0.049866192042827606, 0.07921014726161957, -0.01722046174108982, 0.051554154604673386, 0.2042725682258606, -0.06198057904839516, 0.03942282497882843, -0.044406019151210785, -0.059847284108400345, -0.0870075449347496, -0.3584827780723572, -0.09696727246046066, -0.1245264858007431, 0.031727660447359085, -0.1506877988576889, 0.00655622873455286, 0.023536767810583115, 0.08007577806711197, -0.08762671798467636, 0.07082799077033997, -0.04342757537961006, -0.09662953019142151, 0.15459978580474854, 0.03746257349848747, 0.007783498615026474, 0.033360451459884644, 0.035297419875860214, 0.001332932384684682, -0.07772250473499298, 0.025722691789269447, 0.06348886340856552, 0.04566781222820282, 0.037181466817855835, -0.050998471677303314, -0.11010127514600754, 0.014773055911064148, -0.01741762086749077, -0.03971125930547714, 0.05199752002954483, 0.026144281029701233, 0.03823786973953247, 0.05858095362782478, 0.2553360164165497, -0.04472222551703453, -0.00649634562432766, -0.0969150960445404, 0.21759743988513947, 0.024053269997239113, 0.03558950126171112, 0.0013247766764834523, -0.05480372905731201, -0.047108352184295654, 0.09435366839170456, 0.17692914605140686, -0.005356321576982737, -0.02801627665758133, -0.05842481553554535, 0.028908487409353256, 0.020263560116291046, 0.0254106055945158, 0.03512173891067505, 0.13744257390499115, -0.10185080766677856, 0.021580101922154427, -0.04332271218299866, -0.02771809510886669, 0.041409555822610855, -0.032691553235054016, 0.05581994354724884, -0.039984263479709625, -0.14047420024871826, 0.08237512409687042, -0.21469727158546448, 0.02788342721760273, 0.06989103555679321, -0.1450483351945877, -0.0662655457854271, -0.0204871017485857, 0.03774838522076607, 0.07926451414823532, 0.05496862158179283, 0.0001118204090744257, -0.056788574904203415, 0.027802549302577972, 0.06220212206244469, -0.22225652635097504, -0.05413636937737465, 0.07582198828458786, -0.024560902267694473, 0.043195728212594986, -0.08023624867200851, 0.07877560704946518, 0.05837387964129448, 0.03727540746331215, -0.09071105718612671, 0.05120844021439552, 0.031071241945028305, 0.03232815861701965, 0.011723791249096394, 0.2005050927400589, -0.0038877467159181833, 0.1197550892829895, 0.07168354839086533, -0.0780734047293663, 0.004551077727228403, -0.06502371281385422, 0.010169672779738903, -0.07687671482563019, 0.029501907527446747, -0.04932163283228874, 0.10986675322055817, 0.06903962790966034, -0.04284125939011574, -0.012502774596214294, -0.07517591118812561, -0.01890263520181179, 0.08637060225009918, -0.06347278505563736, 0.027434589341282845, -0.08509484678506851, -0.05959940701723099, -0.04201098904013634, 0.024801725521683693, -0.09538348764181137, -0.012066279537975788, -0.14108064770698547, -0.005874872673302889, -0.03362392261624336, 0.052253611385822296, -0.025058800354599953, 0.04349764063954353, -0.039022840559482574, -0.0920834019780159, 0.08402001857757568, 0.11899985373020172, -0.10500457137823105, -0.07568652927875519 ]
null
null
transformers
# DALL·E Mini Model Card This model card focuses on the model associated with the DALL·E mini space on Hugging Face, available [here](https://huggingface.co/spaces/dalle-mini/dalle-mini). The app is called “dalle-mini”, but incorporates “[DALL·E Mini](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-mini-Generate-images-from-any-text-prompt--VmlldzoyMDE4NDAy)’’ and “[DALL·E Mega](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-Mega-Training-Journal--VmlldzoxODMxMDI2)” models (further details on this distinction forthcoming). The DALL·E Mega model is the largest version of DALLE Mini. For more information specific to DALL·E Mega, see the [DALL·E Mega model card](https://huggingface.co/dalle-mini/dalle-mega). ## Model Details * **Developed by:** Boris Dayma, Suraj Patil, Pedro Cuenca, Khalid Saifullah, Tanishq Abraham, Phúc Lê, Luke, Luke Melas, Ritobrata Ghosh * **Model type:** Transformer-based text-to-image generation model * **Language(s):** English * **License:** Apache 2.0 * **Model Description:** This is a model that can be used to generate images based on text prompts. As the model developers wrote in the [project report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-mini-Generate-images-from-any-text-prompt--VmlldzoyMDE4NDAy) about DALL·E mini, “OpenAI had the first impressive model for generating images with [DALL·E](https://openai.com/blog/dall-e/). DALL·E mini is an attempt at reproducing those results with an open-source model.” * **Resources for more information:** See OpenAI’s website for more information about [DALL·E](https://openai.com/blog/dall-e/), including the [DALL·E model card](https://github.com/openai/DALL-E/blob/master/model_card.md). See the [project report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-mini-Generate-images-from-any-text-prompt--VmlldzoyMDE4NDAy) for more information from the model’s developers. To learn more about DALL·E Mega, see the DALL·E Mega [training journal](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-Mega-Training--VmlldzoxODMxMDI2#training-parameters). * **Cite as:** ```bib text @misc{Dayma_DALL·E_Mini_2021, author = {Dayma, Boris and Patil, Suraj and Cuenca, Pedro and Saifullah, Khalid and Abraham, Tanishq and Lê Khắc, Phúc and Melas, Luke and Ghosh, Ritobrata}, doi = {10.5281/zenodo.5146400}, month = {7}, title = {DALL·E Mini}, url = {https://github.com/borisdayma/dalle-mini}, year = {2021} } ``` ## Uses ### Direct Use The model is intended to be used to generate images based on text prompts for research and personal consumption. Intended uses include supporting creativity, creating humorous content, and providing generations for people curious about the model’s behavior. Intended uses exclude those described in the [Misuse and Out-of-Scope Use](#misuse-malicious-use-and-out-of-scope-use) section. ### Downstream Use The model could also be used for downstream use cases, including: * Research efforts, such as probing and better understanding the limitations and biases of generative models to further improve the state of science * Development of educational or creative tools * Generation of artwork and use in design and artistic processes. * Other uses that are newly discovered by users. This currently includes poetry illustration (give a poem as prompt), fan art (putting a character in various other visual universes), visual puns, fairy tale illustrations (give a fantasy situation as prompt), concept mashups (applying a texture to something completely different), style transfers (portraits in the style of), … We hope you will find your own application! Downstream uses exclude the uses described in [Misuse and Out-of-Scope Use](#misuse-malicious-use-and-out-of-scope-use). ### Misuse, Malicious Use, and Out-of-Scope Use The model should not be used to intentionally create or disseminate images that create hostile or alienating environments for people. This includes generating images that people would foreseeably find disturbing, distressing, or offensive; or content that propagates historical or current stereotypes. #### Out-of-Scope Use The model was not trained to be factual or true representations of people or events, and therefore using the model to generate such content is out-of-scope for the abilities of this model. #### Misuse and Malicious Use Using the model to generate content that is cruel to individuals is a misuse of this model. This includes: * Generating demeaning, dehumanizing, or otherwise harmful representations of people or their environments, cultures, religions, etc. * Intentionally promoting or propagating discriminatory content or harmful stereotypes. * Impersonating individuals without their consent. * Sexual content without consent of the people who might see it. * Mis- and disinformation * Representations of egregious violence and gore * Sharing of copyrighted or licensed material in violation of its terms of use. * Sharing content that is an alteration of copyrighted or licensed material in violation of its terms of use. ## Limitations and Bias ### Limitations The model developers discuss the limitations of the model further in the DALL·E Mini [technical report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-Mini-Explained-with-Demo--Vmlldzo4NjIxODA): * Faces and people in general are not generated properly. * Animals are usually unrealistic. * It is hard to predict where the model excels or falls short…Good prompt engineering will lead to the best results. * The model has only been trained with English descriptions and will not perform as well in other languages ### Bias **CONTENT WARNING: Readers should be aware this section contains content that is disturbing, offensive, and can propagate historical and current stereotypes.** The model was trained on unfiltered data from the Internet, limited to pictures with English descriptions. Text and images from communities and cultures using other languages were not utilized. This affects all output of the model, with white and Western culture asserted as a default, and the model’s ability to generate content using non-English prompts is observably lower quality than prompts in English. While the capabilities of image generation models are impressive, they may also reinforce or exacerbate societal biases. The extent and nature of the biases of DALL·E Mini and DALL·E Mega models have yet to be fully documented, but initial testing demonstrates that they may generate images that contain negative stereotypes against minoritized groups. Work to analyze the nature and extent of the models’ biases and limitations is ongoing. Our current analyses demonstrate that: * Images generated by the model can include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups. * When the model generates images with people in them, it tends to output people who we perceive to be white, while people of color are underrepresented. * Images generated by the model can contain biased content that depicts power differentials between people of color and people who are white, with white people in positions of privilege. * The model is generally only usable for generating images based on text in English, limiting accessibility of the model for non-English speakers and potentially contributing to the biases in images generated by the model. The [technical report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-Mini-Explained-with-Demo--Vmlldzo4NjIxODA) discusses these issues in more detail, and also highlights potential sources of bias in the model development process. ### Limitations and Bias Recommendations * Users (both direct and downstream) should be made aware of the biases and limitations. * Content that is potentially problematic should be filtered out, e.g., via automated models that detect violence or pornography. * Further work on this model should include methods for balanced and just representations of people and cultures, for example, by curating the training dataset to be both diverse and inclusive. ## Training ### Training Data The model developers used 3 datasets for the model: * [Conceptual Captions Dataset](https://aclanthology.org/P18-1238/), which contains 3 million image and caption pairs. * [Conceptual 12M](https://arxiv.org/abs/2102.08981), which contains 12 million image and caption pairs. * The [OpenAI subset](https://github.com/openai/CLIP/blob/main/data/yfcc100m.md) of [YFCC100M](https://multimediacommons.wordpress.com/yfcc100m-core-dataset/), which contains about 15 million images and that we further sub-sampled to 2 million images due to limitations in storage space. They used both title and description as caption and removed html tags, new lines and extra spaces. For fine-tuning the image encoder, a subset of 2 million images were used. All images (about 15 million) were used for training the Seq2Seq model. ### Training Procedure As described further in the [technical report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-Mini-Explained-with-Demo--Vmlldzo4NjIxODA#our-dall-e-model-architecture) for DALL·E Mini, during training, images and descriptions are both available and pass through the system as follows: * Images are encoded through a [VQGAN](https://arxiv.org/abs/2012.09841) encoder, which turns images into a sequence of tokens. * Descriptions are encoded through a [BART](https://arxiv.org/abs/1910.13461) encoder. * The output of the BART encoder and encoded images are fed through the BART decoder, which is an auto-regressive model whose goal is to predict the next token. * Loss is the [softmax cross-entropy](https://wandb.ai/sauravm/Activation-Functions/reports/Activation-Functions-Softmax--VmlldzoxNDU1Njgy#%F0%9F%93%A2-softmax-+-cross-entropy-loss-(caution:-math-alert)) between the model prediction logits and the actual image encodings from the VQGAN. The simplified training procedure for DALL·E Mega is as follows: * **Hardware:** 1 pod TPU v3-256 = 32 nodes of TPU VM v3-8 (8 TPU per node) = 256 TPU v3 * **Optimizer:** Distributed Shampoo * **Model Partition Specificiations:** 8 model parallel x 32 data parallel * **Batch:** 44 samples per model x 32 data parallel x 3 gradient accumulation steps = 4224 increasing samples per update * **Learning rate:** warmup to 0.0001 for 10,000 steps and then kept constant until plateau * Gradient checkpointing used on each Encoder/Decoder layer (ie, MHA + FFN) * Distributed Shampoo + Normformer Optimizations have proved to be effective and efficiently scaling this model. * It should also be noted that the learning rate and other parameters are sometimes adjusted on the fly, and batch size increased over time as well. There is more information about the full procedure and technical material in the DALL·E Mega [training journal](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-Mega-Training--VmlldzoxODMxMDI2#training-parameters). ## Evaluation Results The model developers discuss their results extensively in their [technical report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-Mini-Explained-with-Demo--Vmlldzo4NjIxODA#the-results-of-our-dall-e-experiment) for DALL·E Mini, which provides comparisons between DALL·E Mini’s results with [DALL·E-pytorch](https://github.com/lucidrains/DALLE-pytorch), OpenAI’s [DALL·E](https://openai.com/blog/dall-e/), and models consisting of a generator coupled with the [CLIP neural network model](https://openai.com/blog/clip/). For evaluation results related to DALL·E Mega, see this [technical report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-mini-Generate-images-from-any-text-prompt--VmlldzoyMDE4NDAy). ## Environmental Impact ### DALL·E Mini Estimated Emissions *The model is 27 times smaller than the original DALL·E and was trained on a single TPU v3-8 for only 3 days.* Based on that information, we estimate the following CO2 emissions using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact. * **Hardware Type:** TPU v3-8 * **Hours used:** 72 (3 days) * **Cloud Provider:** GCP (as mentioned in the technical report) * **Compute Region:** us-east1 (provided by model developers) * **Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid):** 30.16 kg CO2 eq. ### DALL·E Mega Estimated Emissions DALL·E Mega is still training. So far, as on June 9, 2022, the model developers report that DALL·E Mega has been training for about 40-45 days on a TPU v3-256. Using those numbers, we estimate the following CO2 emissions using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact. * **Hardware Type:** TPU v3-256 * **Hours used:** 960 - 1080 hours (40-45 days) * **Cloud Provider:** Unknown * **Compute Region:** Unknown * **Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid):** Unknown ## Citation ```bibtext @misc{Dayma_DALL·E_Mini_2021, author = {Dayma, Boris and Patil, Suraj and Cuenca, Pedro and Saifullah, Khalid and Abraham, Tanishq and Lê Khắc, Phúc and Melas, Luke and Ghosh, Ritobrata}, doi = {10.5281/zenodo.5146400}, month = {7}, title = {DALL·E Mini}, url = {https://github.com/borisdayma/dalle-mini}, year = {2021} } ``` *This model card was written by: Boris Dayma, Margaret Mitchell, Ezi Ozoani, Marissa Gerchick, Irene Solaiman, Clémentine Fourrier, Sasha Luccioni, Emily Witko, Nazneen Rajani, and Julian Herrera.*
{"language": "en", "license": "apache-2.0", "tags": ["text-to-image"], "inference": false, "co2_eq_emissions": {"emissions": 7540, "source": "MLCo2 Machine Learning Impact calculator", "geographical_location": "East USA", "hardware_used": "TPU v3-8"}, "model-index": [{"name": "dalle-mini", "results": []}]}
text-to-image
dalle-mini/dalle-mini
[ "transformers", "jax", "dallebart", "text-to-image", "en", "arxiv:2102.08981", "arxiv:2012.09841", "arxiv:1910.13461", "arxiv:1910.09700", "license:apache-2.0", "co2_eq_emissions", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2102.08981", "2012.09841", "1910.13461", "1910.09700" ]
[ "en" ]
TAGS #transformers #jax #dallebart #text-to-image #en #arxiv-2102.08981 #arxiv-2012.09841 #arxiv-1910.13461 #arxiv-1910.09700 #license-apache-2.0 #co2_eq_emissions #has_space #region-us
# DALL·E Mini Model Card This model card focuses on the model associated with the DALL·E mini space on Hugging Face, available here. The app is called “dalle-mini”, but incorporates “DALL·E Mini’’ and “DALL·E Mega” models (further details on this distinction forthcoming). The DALL·E Mega model is the largest version of DALLE Mini. For more information specific to DALL·E Mega, see the DALL·E Mega model card. ## Model Details * Developed by: Boris Dayma, Suraj Patil, Pedro Cuenca, Khalid Saifullah, Tanishq Abraham, Phúc Lê, Luke, Luke Melas, Ritobrata Ghosh * Model type: Transformer-based text-to-image generation model * Language(s): English * License: Apache 2.0 * Model Description: This is a model that can be used to generate images based on text prompts. As the model developers wrote in the project report about DALL·E mini, “OpenAI had the first impressive model for generating images with DALL·E. DALL·E mini is an attempt at reproducing those results with an open-source model.” * Resources for more information: See OpenAI’s website for more information about DALL·E, including the DALL·E model card. See the project report for more information from the model’s developers. To learn more about DALL·E Mega, see the DALL·E Mega training journal. * Cite as: ## Uses ### Direct Use The model is intended to be used to generate images based on text prompts for research and personal consumption. Intended uses include supporting creativity, creating humorous content, and providing generations for people curious about the model’s behavior. Intended uses exclude those described in the Misuse and Out-of-Scope Use section. ### Downstream Use The model could also be used for downstream use cases, including: * Research efforts, such as probing and better understanding the limitations and biases of generative models to further improve the state of science * Development of educational or creative tools * Generation of artwork and use in design and artistic processes. * Other uses that are newly discovered by users. This currently includes poetry illustration (give a poem as prompt), fan art (putting a character in various other visual universes), visual puns, fairy tale illustrations (give a fantasy situation as prompt), concept mashups (applying a texture to something completely different), style transfers (portraits in the style of), … We hope you will find your own application! Downstream uses exclude the uses described in Misuse and Out-of-Scope Use. ### Misuse, Malicious Use, and Out-of-Scope Use The model should not be used to intentionally create or disseminate images that create hostile or alienating environments for people. This includes generating images that people would foreseeably find disturbing, distressing, or offensive; or content that propagates historical or current stereotypes. #### Out-of-Scope Use The model was not trained to be factual or true representations of people or events, and therefore using the model to generate such content is out-of-scope for the abilities of this model. #### Misuse and Malicious Use Using the model to generate content that is cruel to individuals is a misuse of this model. This includes: * Generating demeaning, dehumanizing, or otherwise harmful representations of people or their environments, cultures, religions, etc. * Intentionally promoting or propagating discriminatory content or harmful stereotypes. * Impersonating individuals without their consent. * Sexual content without consent of the people who might see it. * Mis- and disinformation * Representations of egregious violence and gore * Sharing of copyrighted or licensed material in violation of its terms of use. * Sharing content that is an alteration of copyrighted or licensed material in violation of its terms of use. ## Limitations and Bias ### Limitations The model developers discuss the limitations of the model further in the DALL·E Mini technical report: * Faces and people in general are not generated properly. * Animals are usually unrealistic. * It is hard to predict where the model excels or falls short…Good prompt engineering will lead to the best results. * The model has only been trained with English descriptions and will not perform as well in other languages ### Bias CONTENT WARNING: Readers should be aware this section contains content that is disturbing, offensive, and can propagate historical and current stereotypes. The model was trained on unfiltered data from the Internet, limited to pictures with English descriptions. Text and images from communities and cultures using other languages were not utilized. This affects all output of the model, with white and Western culture asserted as a default, and the model’s ability to generate content using non-English prompts is observably lower quality than prompts in English. While the capabilities of image generation models are impressive, they may also reinforce or exacerbate societal biases. The extent and nature of the biases of DALL·E Mini and DALL·E Mega models have yet to be fully documented, but initial testing demonstrates that they may generate images that contain negative stereotypes against minoritized groups. Work to analyze the nature and extent of the models’ biases and limitations is ongoing. Our current analyses demonstrate that: * Images generated by the model can include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups. * When the model generates images with people in them, it tends to output people who we perceive to be white, while people of color are underrepresented. * Images generated by the model can contain biased content that depicts power differentials between people of color and people who are white, with white people in positions of privilege. * The model is generally only usable for generating images based on text in English, limiting accessibility of the model for non-English speakers and potentially contributing to the biases in images generated by the model. The technical report discusses these issues in more detail, and also highlights potential sources of bias in the model development process. ### Limitations and Bias Recommendations * Users (both direct and downstream) should be made aware of the biases and limitations. * Content that is potentially problematic should be filtered out, e.g., via automated models that detect violence or pornography. * Further work on this model should include methods for balanced and just representations of people and cultures, for example, by curating the training dataset to be both diverse and inclusive. ## Training ### Training Data The model developers used 3 datasets for the model: * Conceptual Captions Dataset, which contains 3 million image and caption pairs. * Conceptual 12M, which contains 12 million image and caption pairs. * The OpenAI subset of YFCC100M, which contains about 15 million images and that we further sub-sampled to 2 million images due to limitations in storage space. They used both title and description as caption and removed html tags, new lines and extra spaces. For fine-tuning the image encoder, a subset of 2 million images were used. All images (about 15 million) were used for training the Seq2Seq model. ### Training Procedure As described further in the technical report for DALL·E Mini, during training, images and descriptions are both available and pass through the system as follows: * Images are encoded through a VQGAN encoder, which turns images into a sequence of tokens. * Descriptions are encoded through a BART encoder. * The output of the BART encoder and encoded images are fed through the BART decoder, which is an auto-regressive model whose goal is to predict the next token. * Loss is the softmax cross-entropy) between the model prediction logits and the actual image encodings from the VQGAN. The simplified training procedure for DALL·E Mega is as follows: * Hardware: 1 pod TPU v3-256 = 32 nodes of TPU VM v3-8 (8 TPU per node) = 256 TPU v3 * Optimizer: Distributed Shampoo * Model Partition Specificiations: 8 model parallel x 32 data parallel * Batch: 44 samples per model x 32 data parallel x 3 gradient accumulation steps = 4224 increasing samples per update * Learning rate: warmup to 0.0001 for 10,000 steps and then kept constant until plateau * Gradient checkpointing used on each Encoder/Decoder layer (ie, MHA + FFN) * Distributed Shampoo + Normformer Optimizations have proved to be effective and efficiently scaling this model. * It should also be noted that the learning rate and other parameters are sometimes adjusted on the fly, and batch size increased over time as well. There is more information about the full procedure and technical material in the DALL·E Mega training journal. ## Evaluation Results The model developers discuss their results extensively in their technical report for DALL·E Mini, which provides comparisons between DALL·E Mini’s results with DALL·E-pytorch, OpenAI’s DALL·E, and models consisting of a generator coupled with the CLIP neural network model. For evaluation results related to DALL·E Mega, see this technical report. ## Environmental Impact ### DALL·E Mini Estimated Emissions *The model is 27 times smaller than the original DALL·E and was trained on a single TPU v3-8 for only 3 days.* Based on that information, we estimate the following CO2 emissions using the Machine Learning Impact calculator presented in Lacoste et al. (2019). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact. * Hardware Type: TPU v3-8 * Hours used: 72 (3 days) * Cloud Provider: GCP (as mentioned in the technical report) * Compute Region: us-east1 (provided by model developers) * Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid): 30.16 kg CO2 eq. ### DALL·E Mega Estimated Emissions DALL·E Mega is still training. So far, as on June 9, 2022, the model developers report that DALL·E Mega has been training for about 40-45 days on a TPU v3-256. Using those numbers, we estimate the following CO2 emissions using the Machine Learning Impact calculator presented in Lacoste et al. (2019). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact. * Hardware Type: TPU v3-256 * Hours used: 960 - 1080 hours (40-45 days) * Cloud Provider: Unknown * Compute Region: Unknown * Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid): Unknown *This model card was written by: Boris Dayma, Margaret Mitchell, Ezi Ozoani, Marissa Gerchick, Irene Solaiman, Clémentine Fourrier, Sasha Luccioni, Emily Witko, Nazneen Rajani, and Julian Herrera.*
[ "# DALL·E Mini Model Card\n\nThis model card focuses on the model associated with the DALL·E mini space on Hugging Face, available here. The app is called “dalle-mini”, but incorporates “DALL·E Mini’’ and “DALL·E Mega” models (further details on this distinction forthcoming).\n\nThe DALL·E Mega model is the largest version of DALLE Mini. For more information specific to DALL·E Mega, see the DALL·E Mega model card.", "## Model Details\n\n* Developed by: Boris Dayma, Suraj Patil, Pedro Cuenca, Khalid Saifullah, Tanishq Abraham, Phúc Lê, Luke, Luke Melas, Ritobrata Ghosh\n* Model type: Transformer-based text-to-image generation model\n* Language(s): English\n* License: Apache 2.0\n* Model Description: This is a model that can be used to generate images based on text prompts. As the model developers wrote in the project report about DALL·E mini, “OpenAI had the first impressive model for generating images with DALL·E. DALL·E mini is an attempt at reproducing those results with an open-source model.”\n* Resources for more information: See OpenAI’s website for more information about DALL·E, including the DALL·E model card. See the project report for more information from the model’s developers. To learn more about DALL·E Mega, see the DALL·E Mega training journal.\n* Cite as:", "## Uses", "### Direct Use\n\nThe model is intended to be used to generate images based on text prompts for research and personal consumption. Intended uses include supporting creativity, creating humorous content, and providing generations for people curious about the model’s behavior. Intended uses exclude those described in the Misuse and Out-of-Scope Use section.", "### Downstream Use\n\nThe model could also be used for downstream use cases, including:\n* Research efforts, such as probing and better understanding the limitations and biases of generative models to further improve the state of science\n* Development of educational or creative tools\n* Generation of artwork and use in design and artistic processes. \n* Other uses that are newly discovered by users. This currently includes poetry illustration (give a poem as prompt), fan art (putting a character in various other visual universes), visual puns, fairy tale illustrations (give a fantasy situation as prompt), concept mashups (applying a texture to something completely different), style transfers (portraits in the style of), … We hope you will find your own application!\n\nDownstream uses exclude the uses described in Misuse and Out-of-Scope Use.", "### Misuse, Malicious Use, and Out-of-Scope Use\n\nThe model should not be used to intentionally create or disseminate images that create hostile or alienating environments for people. This includes generating images that people would foreseeably find disturbing, distressing, or offensive; or content that propagates historical or current stereotypes.", "#### Out-of-Scope Use\n\nThe model was not trained to be factual or true representations of people or events, and therefore using the model to generate such content is out-of-scope for the abilities of this model.", "#### Misuse and Malicious Use \n\nUsing the model to generate content that is cruel to individuals is a misuse of this model. This includes:\n* Generating demeaning, dehumanizing, or otherwise harmful representations of people or their environments, cultures, religions, etc.\n* Intentionally promoting or propagating discriminatory content or harmful stereotypes.\n* Impersonating individuals without their consent.\n* Sexual content without consent of the people who might see it.\n* Mis- and disinformation\n* Representations of egregious violence and gore\n* Sharing of copyrighted or licensed material in violation of its terms of use.\n* Sharing content that is an alteration of copyrighted or licensed material in violation of its terms of use.", "## Limitations and Bias", "### Limitations\n\nThe model developers discuss the limitations of the model further in the DALL·E Mini technical report:\n* Faces and people in general are not generated properly.\n* Animals are usually unrealistic.\n* It is hard to predict where the model excels or falls short…Good prompt engineering will lead to the best results.\n* The model has only been trained with English descriptions and will not perform as well in other languages", "### Bias \n\nCONTENT WARNING: Readers should be aware this section contains content that is disturbing, offensive, and can propagate historical and current stereotypes.\n\nThe model was trained on unfiltered data from the Internet, limited to pictures with English descriptions. Text and images from communities and cultures using other languages were not utilized. This affects all output of the model, with white and Western culture asserted as a default, and the model’s ability to generate content using non-English prompts is observably lower quality than prompts in English.\n\nWhile the capabilities of image generation models are impressive, they may also reinforce or exacerbate societal biases. The extent and nature of the biases of DALL·E Mini and DALL·E Mega models have yet to be fully documented, but initial testing demonstrates that they may generate images that contain negative stereotypes against minoritized groups. Work to analyze the nature and extent of the models’ biases and limitations is ongoing.\n\nOur current analyses demonstrate that:\n* Images generated by the model can include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.\n* When the model generates images with people in them, it tends to output people who we perceive to be white, while people of color are underrepresented. \n* Images generated by the model can contain biased content that depicts power differentials between people of color and people who are white, with white people in positions of privilege.\n* The model is generally only usable for generating images based on text in English, limiting accessibility of the model for non-English speakers and potentially contributing to the biases in images generated by the model.\n\nThe technical report discusses these issues in more detail, and also highlights potential sources of bias in the model development process.", "### Limitations and Bias Recommendations\n\n* Users (both direct and downstream) should be made aware of the biases and limitations.\n* Content that is potentially problematic should be filtered out, e.g., via automated models that detect violence or pornography.\n* Further work on this model should include methods for balanced and just representations of people and cultures, for example, by curating the training dataset to be both diverse and inclusive.", "## Training", "### Training Data\n\nThe model developers used 3 datasets for the model:\n* Conceptual Captions Dataset, which contains 3 million image and caption pairs.\n* Conceptual 12M, which contains 12 million image and caption pairs.\n* The OpenAI subset of YFCC100M, which contains about 15 million images and that we further sub-sampled to 2 million images due to limitations in storage space. They used both title and description as caption and removed html tags, new lines and extra spaces.\n\nFor fine-tuning the image encoder, a subset of 2 million images were used.\nAll images (about 15 million) were used for training the Seq2Seq model.", "### Training Procedure\n\nAs described further in the technical report for DALL·E Mini, during training, images and descriptions are both available and pass through the system as follows:\n* Images are encoded through a VQGAN encoder, which turns images into a sequence of tokens.\n* Descriptions are encoded through a BART encoder.\n* The output of the BART encoder and encoded images are fed through the BART decoder, which is an auto-regressive model whose goal is to predict the next token.\n* Loss is the softmax cross-entropy) between the model prediction logits and the actual image encodings from the VQGAN.\n\nThe simplified training procedure for DALL·E Mega is as follows: \n\n* Hardware: 1 pod TPU v3-256 = 32 nodes of TPU VM v3-8 (8 TPU per node) = 256 TPU v3\n* Optimizer: Distributed Shampoo\n* Model Partition Specificiations: 8 model parallel x 32 data parallel\n* Batch: 44 samples per model x 32 data parallel x 3 gradient accumulation steps = 4224 increasing samples per update\n* Learning rate: warmup to 0.0001 for 10,000 steps and then kept constant until plateau\n* Gradient checkpointing used on each Encoder/Decoder layer (ie, MHA + FFN)\n* Distributed Shampoo + Normformer Optimizations have proved to be effective and efficiently scaling this model. \n* It should also be noted that the learning rate and other parameters are sometimes adjusted on the fly, and batch size increased over time as well.\n\nThere is more information about the full procedure and technical material in the DALL·E Mega training journal.", "## Evaluation Results\n\nThe model developers discuss their results extensively in their technical report for DALL·E Mini, which provides comparisons between DALL·E Mini’s results with DALL·E-pytorch, OpenAI’s DALL·E, and models consisting of a generator coupled with the CLIP neural network model. \n\nFor evaluation results related to DALL·E Mega, see this technical report.", "## Environmental Impact", "### DALL·E Mini Estimated Emissions\n\n*The model is 27 times smaller than the original DALL·E and was trained on a single TPU v3-8 for only 3 days.*\n\nBased on that information, we estimate the following CO2 emissions using the Machine Learning Impact calculator presented in Lacoste et al. (2019). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact.\n\n* Hardware Type: TPU v3-8\n* Hours used: 72 (3 days)\n* Cloud Provider: GCP (as mentioned in the technical report)\n* Compute Region: us-east1 (provided by model developers)\n* Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid): 30.16 kg CO2 eq.", "### DALL·E Mega Estimated Emissions\n\nDALL·E Mega is still training. So far, as on June 9, 2022, the model developers report that DALL·E Mega has been training for about 40-45 days on a TPU v3-256. Using those numbers, we estimate the following CO2 emissions using the Machine Learning Impact calculator presented in Lacoste et al. (2019). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact.\n\n* Hardware Type: TPU v3-256\n* Hours used: 960 - 1080 hours (40-45 days)\n* Cloud Provider: Unknown\n* Compute Region: Unknown\n* Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid): Unknown\n\n*This model card was written by: Boris Dayma, Margaret Mitchell, Ezi Ozoani, Marissa Gerchick, Irene Solaiman, Clémentine Fourrier, Sasha Luccioni, Emily Witko, Nazneen Rajani, and Julian Herrera.*" ]
[ "TAGS\n#transformers #jax #dallebart #text-to-image #en #arxiv-2102.08981 #arxiv-2012.09841 #arxiv-1910.13461 #arxiv-1910.09700 #license-apache-2.0 #co2_eq_emissions #has_space #region-us \n", "# DALL·E Mini Model Card\n\nThis model card focuses on the model associated with the DALL·E mini space on Hugging Face, available here. The app is called “dalle-mini”, but incorporates “DALL·E Mini’’ and “DALL·E Mega” models (further details on this distinction forthcoming).\n\nThe DALL·E Mega model is the largest version of DALLE Mini. For more information specific to DALL·E Mega, see the DALL·E Mega model card.", "## Model Details\n\n* Developed by: Boris Dayma, Suraj Patil, Pedro Cuenca, Khalid Saifullah, Tanishq Abraham, Phúc Lê, Luke, Luke Melas, Ritobrata Ghosh\n* Model type: Transformer-based text-to-image generation model\n* Language(s): English\n* License: Apache 2.0\n* Model Description: This is a model that can be used to generate images based on text prompts. As the model developers wrote in the project report about DALL·E mini, “OpenAI had the first impressive model for generating images with DALL·E. DALL·E mini is an attempt at reproducing those results with an open-source model.”\n* Resources for more information: See OpenAI’s website for more information about DALL·E, including the DALL·E model card. See the project report for more information from the model’s developers. To learn more about DALL·E Mega, see the DALL·E Mega training journal.\n* Cite as:", "## Uses", "### Direct Use\n\nThe model is intended to be used to generate images based on text prompts for research and personal consumption. Intended uses include supporting creativity, creating humorous content, and providing generations for people curious about the model’s behavior. Intended uses exclude those described in the Misuse and Out-of-Scope Use section.", "### Downstream Use\n\nThe model could also be used for downstream use cases, including:\n* Research efforts, such as probing and better understanding the limitations and biases of generative models to further improve the state of science\n* Development of educational or creative tools\n* Generation of artwork and use in design and artistic processes. \n* Other uses that are newly discovered by users. This currently includes poetry illustration (give a poem as prompt), fan art (putting a character in various other visual universes), visual puns, fairy tale illustrations (give a fantasy situation as prompt), concept mashups (applying a texture to something completely different), style transfers (portraits in the style of), … We hope you will find your own application!\n\nDownstream uses exclude the uses described in Misuse and Out-of-Scope Use.", "### Misuse, Malicious Use, and Out-of-Scope Use\n\nThe model should not be used to intentionally create or disseminate images that create hostile or alienating environments for people. This includes generating images that people would foreseeably find disturbing, distressing, or offensive; or content that propagates historical or current stereotypes.", "#### Out-of-Scope Use\n\nThe model was not trained to be factual or true representations of people or events, and therefore using the model to generate such content is out-of-scope for the abilities of this model.", "#### Misuse and Malicious Use \n\nUsing the model to generate content that is cruel to individuals is a misuse of this model. This includes:\n* Generating demeaning, dehumanizing, or otherwise harmful representations of people or their environments, cultures, religions, etc.\n* Intentionally promoting or propagating discriminatory content or harmful stereotypes.\n* Impersonating individuals without their consent.\n* Sexual content without consent of the people who might see it.\n* Mis- and disinformation\n* Representations of egregious violence and gore\n* Sharing of copyrighted or licensed material in violation of its terms of use.\n* Sharing content that is an alteration of copyrighted or licensed material in violation of its terms of use.", "## Limitations and Bias", "### Limitations\n\nThe model developers discuss the limitations of the model further in the DALL·E Mini technical report:\n* Faces and people in general are not generated properly.\n* Animals are usually unrealistic.\n* It is hard to predict where the model excels or falls short…Good prompt engineering will lead to the best results.\n* The model has only been trained with English descriptions and will not perform as well in other languages", "### Bias \n\nCONTENT WARNING: Readers should be aware this section contains content that is disturbing, offensive, and can propagate historical and current stereotypes.\n\nThe model was trained on unfiltered data from the Internet, limited to pictures with English descriptions. Text and images from communities and cultures using other languages were not utilized. This affects all output of the model, with white and Western culture asserted as a default, and the model’s ability to generate content using non-English prompts is observably lower quality than prompts in English.\n\nWhile the capabilities of image generation models are impressive, they may also reinforce or exacerbate societal biases. The extent and nature of the biases of DALL·E Mini and DALL·E Mega models have yet to be fully documented, but initial testing demonstrates that they may generate images that contain negative stereotypes against minoritized groups. Work to analyze the nature and extent of the models’ biases and limitations is ongoing.\n\nOur current analyses demonstrate that:\n* Images generated by the model can include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.\n* When the model generates images with people in them, it tends to output people who we perceive to be white, while people of color are underrepresented. \n* Images generated by the model can contain biased content that depicts power differentials between people of color and people who are white, with white people in positions of privilege.\n* The model is generally only usable for generating images based on text in English, limiting accessibility of the model for non-English speakers and potentially contributing to the biases in images generated by the model.\n\nThe technical report discusses these issues in more detail, and also highlights potential sources of bias in the model development process.", "### Limitations and Bias Recommendations\n\n* Users (both direct and downstream) should be made aware of the biases and limitations.\n* Content that is potentially problematic should be filtered out, e.g., via automated models that detect violence or pornography.\n* Further work on this model should include methods for balanced and just representations of people and cultures, for example, by curating the training dataset to be both diverse and inclusive.", "## Training", "### Training Data\n\nThe model developers used 3 datasets for the model:\n* Conceptual Captions Dataset, which contains 3 million image and caption pairs.\n* Conceptual 12M, which contains 12 million image and caption pairs.\n* The OpenAI subset of YFCC100M, which contains about 15 million images and that we further sub-sampled to 2 million images due to limitations in storage space. They used both title and description as caption and removed html tags, new lines and extra spaces.\n\nFor fine-tuning the image encoder, a subset of 2 million images were used.\nAll images (about 15 million) were used for training the Seq2Seq model.", "### Training Procedure\n\nAs described further in the technical report for DALL·E Mini, during training, images and descriptions are both available and pass through the system as follows:\n* Images are encoded through a VQGAN encoder, which turns images into a sequence of tokens.\n* Descriptions are encoded through a BART encoder.\n* The output of the BART encoder and encoded images are fed through the BART decoder, which is an auto-regressive model whose goal is to predict the next token.\n* Loss is the softmax cross-entropy) between the model prediction logits and the actual image encodings from the VQGAN.\n\nThe simplified training procedure for DALL·E Mega is as follows: \n\n* Hardware: 1 pod TPU v3-256 = 32 nodes of TPU VM v3-8 (8 TPU per node) = 256 TPU v3\n* Optimizer: Distributed Shampoo\n* Model Partition Specificiations: 8 model parallel x 32 data parallel\n* Batch: 44 samples per model x 32 data parallel x 3 gradient accumulation steps = 4224 increasing samples per update\n* Learning rate: warmup to 0.0001 for 10,000 steps and then kept constant until plateau\n* Gradient checkpointing used on each Encoder/Decoder layer (ie, MHA + FFN)\n* Distributed Shampoo + Normformer Optimizations have proved to be effective and efficiently scaling this model. \n* It should also be noted that the learning rate and other parameters are sometimes adjusted on the fly, and batch size increased over time as well.\n\nThere is more information about the full procedure and technical material in the DALL·E Mega training journal.", "## Evaluation Results\n\nThe model developers discuss their results extensively in their technical report for DALL·E Mini, which provides comparisons between DALL·E Mini’s results with DALL·E-pytorch, OpenAI’s DALL·E, and models consisting of a generator coupled with the CLIP neural network model. \n\nFor evaluation results related to DALL·E Mega, see this technical report.", "## Environmental Impact", "### DALL·E Mini Estimated Emissions\n\n*The model is 27 times smaller than the original DALL·E and was trained on a single TPU v3-8 for only 3 days.*\n\nBased on that information, we estimate the following CO2 emissions using the Machine Learning Impact calculator presented in Lacoste et al. (2019). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact.\n\n* Hardware Type: TPU v3-8\n* Hours used: 72 (3 days)\n* Cloud Provider: GCP (as mentioned in the technical report)\n* Compute Region: us-east1 (provided by model developers)\n* Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid): 30.16 kg CO2 eq.", "### DALL·E Mega Estimated Emissions\n\nDALL·E Mega is still training. So far, as on June 9, 2022, the model developers report that DALL·E Mega has been training for about 40-45 days on a TPU v3-256. Using those numbers, we estimate the following CO2 emissions using the Machine Learning Impact calculator presented in Lacoste et al. (2019). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact.\n\n* Hardware Type: TPU v3-256\n* Hours used: 960 - 1080 hours (40-45 days)\n* Cloud Provider: Unknown\n* Compute Region: Unknown\n* Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid): Unknown\n\n*This model card was written by: Boris Dayma, Margaret Mitchell, Ezi Ozoani, Marissa Gerchick, Irene Solaiman, Clémentine Fourrier, Sasha Luccioni, Emily Witko, Nazneen Rajani, and Julian Herrera.*" ]
[ 80, 113, 217, 3, 81, 183, 79, 51, 163, 6, 95, 421, 104, 2, 155, 380, 92, 3, 181, 236 ]
[ "passage: TAGS\n#transformers #jax #dallebart #text-to-image #en #arxiv-2102.08981 #arxiv-2012.09841 #arxiv-1910.13461 #arxiv-1910.09700 #license-apache-2.0 #co2_eq_emissions #has_space #region-us \n# DALL·E Mini Model Card\n\nThis model card focuses on the model associated with the DALL·E mini space on Hugging Face, available here. The app is called “dalle-mini”, but incorporates “DALL·E Mini’’ and “DALL·E Mega” models (further details on this distinction forthcoming).\n\nThe DALL·E Mega model is the largest version of DALLE Mini. For more information specific to DALL·E Mega, see the DALL·E Mega model card.## Model Details\n\n* Developed by: Boris Dayma, Suraj Patil, Pedro Cuenca, Khalid Saifullah, Tanishq Abraham, Phúc Lê, Luke, Luke Melas, Ritobrata Ghosh\n* Model type: Transformer-based text-to-image generation model\n* Language(s): English\n* License: Apache 2.0\n* Model Description: This is a model that can be used to generate images based on text prompts. As the model developers wrote in the project report about DALL·E mini, “OpenAI had the first impressive model for generating images with DALL·E. DALL·E mini is an attempt at reproducing those results with an open-source model.”\n* Resources for more information: See OpenAI’s website for more information about DALL·E, including the DALL·E model card. See the project report for more information from the model’s developers. To learn more about DALL·E Mega, see the DALL·E Mega training journal.\n* Cite as:## Uses### Direct Use\n\nThe model is intended to be used to generate images based on text prompts for research and personal consumption. Intended uses include supporting creativity, creating humorous content, and providing generations for people curious about the model’s behavior. Intended uses exclude those described in the Misuse and Out-of-Scope Use section.", "passage: ### Downstream Use\n\nThe model could also be used for downstream use cases, including:\n* Research efforts, such as probing and better understanding the limitations and biases of generative models to further improve the state of science\n* Development of educational or creative tools\n* Generation of artwork and use in design and artistic processes. \n* Other uses that are newly discovered by users. This currently includes poetry illustration (give a poem as prompt), fan art (putting a character in various other visual universes), visual puns, fairy tale illustrations (give a fantasy situation as prompt), concept mashups (applying a texture to something completely different), style transfers (portraits in the style of), … We hope you will find your own application!\n\nDownstream uses exclude the uses described in Misuse and Out-of-Scope Use.### Misuse, Malicious Use, and Out-of-Scope Use\n\nThe model should not be used to intentionally create or disseminate images that create hostile or alienating environments for people. This includes generating images that people would foreseeably find disturbing, distressing, or offensive; or content that propagates historical or current stereotypes.#### Out-of-Scope Use\n\nThe model was not trained to be factual or true representations of people or events, and therefore using the model to generate such content is out-of-scope for the abilities of this model.#### Misuse and Malicious Use \n\nUsing the model to generate content that is cruel to individuals is a misuse of this model. This includes:\n* Generating demeaning, dehumanizing, or otherwise harmful representations of people or their environments, cultures, religions, etc.\n* Intentionally promoting or propagating discriminatory content or harmful stereotypes.\n* Impersonating individuals without their consent.\n* Sexual content without consent of the people who might see it.\n* Mis- and disinformation\n* Representations of egregious violence and gore\n* Sharing of copyrighted or licensed material in violation of its terms of use.\n* Sharing content that is an alteration of copyrighted or licensed material in violation of its terms of use.## Limitations and Bias### Limitations\n\nThe model developers discuss the limitations of the model further in the DALL·E Mini technical report:\n* Faces and people in general are not generated properly.\n* Animals are usually unrealistic.\n* It is hard to predict where the model excels or falls short…Good prompt engineering will lead to the best results.\n* The model has only been trained with English descriptions and will not perform as well in other languages", "passage: ### Bias \n\nCONTENT WARNING: Readers should be aware this section contains content that is disturbing, offensive, and can propagate historical and current stereotypes.\n\nThe model was trained on unfiltered data from the Internet, limited to pictures with English descriptions. Text and images from communities and cultures using other languages were not utilized. This affects all output of the model, with white and Western culture asserted as a default, and the model’s ability to generate content using non-English prompts is observably lower quality than prompts in English.\n\nWhile the capabilities of image generation models are impressive, they may also reinforce or exacerbate societal biases. The extent and nature of the biases of DALL·E Mini and DALL·E Mega models have yet to be fully documented, but initial testing demonstrates that they may generate images that contain negative stereotypes against minoritized groups. Work to analyze the nature and extent of the models’ biases and limitations is ongoing.\n\nOur current analyses demonstrate that:\n* Images generated by the model can include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.\n* When the model generates images with people in them, it tends to output people who we perceive to be white, while people of color are underrepresented. \n* Images generated by the model can contain biased content that depicts power differentials between people of color and people who are white, with white people in positions of privilege.\n* The model is generally only usable for generating images based on text in English, limiting accessibility of the model for non-English speakers and potentially contributing to the biases in images generated by the model.\n\nThe technical report discusses these issues in more detail, and also highlights potential sources of bias in the model development process.### Limitations and Bias Recommendations\n\n* Users (both direct and downstream) should be made aware of the biases and limitations.\n* Content that is potentially problematic should be filtered out, e.g., via automated models that detect violence or pornography.\n* Further work on this model should include methods for balanced and just representations of people and cultures, for example, by curating the training dataset to be both diverse and inclusive.## Training### Training Data\n\nThe model developers used 3 datasets for the model:\n* Conceptual Captions Dataset, which contains 3 million image and caption pairs.\n* Conceptual 12M, which contains 12 million image and caption pairs.\n* The OpenAI subset of YFCC100M, which contains about 15 million images and that we further sub-sampled to 2 million images due to limitations in storage space. They used both title and description as caption and removed html tags, new lines and extra spaces.\n\nFor fine-tuning the image encoder, a subset of 2 million images were used.\nAll images (about 15 million) were used for training the Seq2Seq model.", "passage: ### Training Procedure\n\nAs described further in the technical report for DALL·E Mini, during training, images and descriptions are both available and pass through the system as follows:\n* Images are encoded through a VQGAN encoder, which turns images into a sequence of tokens.\n* Descriptions are encoded through a BART encoder.\n* The output of the BART encoder and encoded images are fed through the BART decoder, which is an auto-regressive model whose goal is to predict the next token.\n* Loss is the softmax cross-entropy) between the model prediction logits and the actual image encodings from the VQGAN.\n\nThe simplified training procedure for DALL·E Mega is as follows: \n\n* Hardware: 1 pod TPU v3-256 = 32 nodes of TPU VM v3-8 (8 TPU per node) = 256 TPU v3\n* Optimizer: Distributed Shampoo\n* Model Partition Specificiations: 8 model parallel x 32 data parallel\n* Batch: 44 samples per model x 32 data parallel x 3 gradient accumulation steps = 4224 increasing samples per update\n* Learning rate: warmup to 0.0001 for 10,000 steps and then kept constant until plateau\n* Gradient checkpointing used on each Encoder/Decoder layer (ie, MHA + FFN)\n* Distributed Shampoo + Normformer Optimizations have proved to be effective and efficiently scaling this model. \n* It should also be noted that the learning rate and other parameters are sometimes adjusted on the fly, and batch size increased over time as well.\n\nThere is more information about the full procedure and technical material in the DALL·E Mega training journal.## Evaluation Results\n\nThe model developers discuss their results extensively in their technical report for DALL·E Mini, which provides comparisons between DALL·E Mini’s results with DALL·E-pytorch, OpenAI’s DALL·E, and models consisting of a generator coupled with the CLIP neural network model. \n\nFor evaluation results related to DALL·E Mega, see this technical report.## Environmental Impact### DALL·E Mini Estimated Emissions\n\n*The model is 27 times smaller than the original DALL·E and was trained on a single TPU v3-8 for only 3 days.*\n\nBased on that information, we estimate the following CO2 emissions using the Machine Learning Impact calculator presented in Lacoste et al. (2019). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact.\n\n* Hardware Type: TPU v3-8\n* Hours used: 72 (3 days)\n* Cloud Provider: GCP (as mentioned in the technical report)\n* Compute Region: us-east1 (provided by model developers)\n* Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid): 30.16 kg CO2 eq." ]
[ -0.028254972770810127, 0.03553052246570587, -0.004067465662956238, 0.049623727798461914, 0.06288625299930573, -0.008076719008386135, 0.08081356436014175, 0.034917984157800674, -0.02770288847386837, 0.043668024241924286, 0.0014277361333370209, -0.07090413570404053, 0.0627814382314682, -0.02154352515935898, 0.062331415712833405, -0.22897672653198242, 0.027930893003940582, -0.056684572249650955, 0.0005141850560903549, 0.06901216506958008, 0.10026480257511139, -0.0770077258348465, 0.08852638304233551, 0.01946571096777916, -0.042475152760744095, -0.04042445868253708, -0.03032928891479969, -0.015530581586062908, 0.04920753091573715, 0.08684322237968445, 0.05095826834440231, -0.005609058775007725, 0.0387713648378849, -0.1473565697669983, 0.011568363755941391, 0.0707334578037262, -0.019062168896198273, 0.023666251450777054, 0.07249437272548676, 0.03721291571855545, 0.2478632628917694, -0.04619394987821579, 0.055587947368621826, 0.05173077434301376, -0.10144958645105362, -0.10645191371440887, -0.06079705059528351, 0.04926029592752457, 0.10882747918367386, 0.07318136096000671, -0.034097954630851746, 0.09291748702526093, -0.03491460159420967, 0.029399586841464043, 0.08232337981462479, -0.07411438971757889, -0.01771201193332672, 0.013882206752896309, 0.06669515371322632, 0.039162635803222656, -0.06075152009725571, 0.00817346666008234, 0.0069837504997849464, 0.033988527953624725, 0.05505349859595299, -0.05015191435813904, 0.2091364711523056, -0.07313799113035202, -0.1134466826915741, -0.04481673985719681, 0.10675753653049469, 0.03127077966928482, -0.10067102313041687, -0.1812151074409485, -0.037602417171001434, 0.09761438518762589, 0.00042806053534150124, -0.023337190970778465, 0.008568367920815945, -0.011262997053563595, 0.059470467269420624, -0.06772677600383759, -0.09050371497869492, -0.004825189709663391, -0.04385785013437271, 0.11731384694576263, 0.021325889974832535, 0.022427991032600403, -0.07191671431064606, 0.04677552729845047, 0.029860379174351692, -0.08833827078342438, -0.06031983345746994, -0.06868395209312439, -0.059143178164958954, -0.02178228460252285, -0.03351263701915741, -0.05831075832247734, -0.03292738273739815, 0.10511258244514465, -0.05111349746584892, 0.027061428874731064, 0.0024063875898718834, 0.04261112958192825, 0.14077980816364288, 0.09049506485462189, -0.024822315201163292, -0.01004154421389103, 0.015713516622781754, 0.023020923137664795, 0.0514226071536541, -0.0317879319190979, -0.028848174959421158, 0.021274447441101074, -0.025584641844034195, 0.04199764132499695, 0.06183002516627312, 0.03949528932571411, -0.040642499923706055, -0.05593950301408768, 0.14764374494552612, -0.07204604893922806, -0.006800378672778606, -0.014963438734412193, -0.02434675768017769, 0.050483979284763336, 0.014719635248184204, -0.0008798157796263695, -0.07039676606655121, 0.04281219467520714, -0.055165309458971024, -0.029596570879220963, -0.10530353337526321, -0.07354016602039337, 0.002370290458202362, 0.04515904188156128, -0.024029551073908806, -0.12368092685937881, -0.14114633202552795, -0.06599659472703934, 0.0413743332028389, -0.04968554526567459, -0.012931989505887032, 0.020917214453220367, -0.026565685868263245, -0.035824522376060486, 0.041345492005348206, -0.035178959369659424, -0.01451888121664524, 0.01812676340341568, -0.055077023804187775, 0.04741940647363663, 0.004489791579544544, 0.021693943068385124, -0.07072308659553528, 0.06197230890393257, -0.15173320472240448, 0.0929078608751297, -0.026153195649385452, -0.024275565519928932, -0.0450371615588665, -0.03753272444009781, -0.032213978469371796, 0.07454928755760193, -0.022696223109960556, 0.116361103951931, -0.20337817072868347, -0.007742009125649929, 0.06977975368499756, -0.1483323723077774, -0.022403698414564133, 0.10106468945741653, -0.036246828734874725, 0.07688304036855698, 0.08305046707391739, 0.036925122141838074, -0.0020695943385362625, -0.025012217462062836, -0.01153053529560566, -0.04870731383562088, -0.062164828181266785, 0.17251208424568176, 0.06693437695503235, -0.04911139979958534, 0.06920278072357178, -0.018502071499824524, -0.009777560830116272, -0.02227686159312725, 0.008205408230423927, -0.022113990038633347, 0.03777501732110977, -0.0027716532349586487, 0.007867496460676193, -0.016110830008983612, -0.057321034371852875, 0.007892405614256859, -0.09775183349847794, -0.02275952510535717, 0.08690518140792847, -0.027565091848373413, 0.03292963281273842, -0.09944845736026764, 0.05250699073076248, 0.020833153277635574, -0.005148601718246937, -0.14252260327339172, -0.02017667517066002, 0.03046880103647709, -0.08808696269989014, 0.03742404282093048, 0.059281185269355774, 0.0067289830185472965, 0.07056766748428345, -0.011954881250858307, 0.0026678312569856644, -0.07645005732774734, -0.0060551525093615055, -0.058328427374362946, -0.09309162199497223, 0.0014808042906224728, -0.028311999514698982, 0.05253634974360466, -0.1274254471063614, 0.0010249641491100192, 0.06342525035142899, 0.05306694284081459, 0.05651799589395523, -0.03859721124172211, 0.02156982384622097, 0.01850355789065361, -0.03478720784187317, -0.05338340252637863, 0.0368497408926487, -0.012614390812814236, -0.017246492207050323, 0.08421136438846588, -0.16252359747886658, -0.10591991245746613, 0.062445297837257385, -0.026578843593597412, -0.08782987296581268, -0.015330689027905464, 0.006832335144281387, -0.011905509978532791, -0.045760687440633774, -0.06502213329076767, 0.20284192264080048, 0.011342409998178482, 0.04700389876961708, -0.08265016973018646, -0.05355772003531456, 0.01115522999316454, -0.027509599924087524, -0.05203269422054291, 0.03487518057227135, 0.08564162999391556, -0.18022382259368896, 0.04727400466799736, 0.07712538540363312, 0.02383784018456936, 0.1252809315919876, 0.0259387344121933, -0.07890556752681732, -0.03962235152721405, -0.011019477620720863, 0.01223394088447094, 0.11317340284585953, 0.01215303409844637, 0.017476949840784073, 0.016512854024767876, 0.01794200949370861, 0.0327441468834877, -0.04873259365558624, 0.035352662205696106, 0.0667865127325058, 0.03654003515839577, -0.03536738455295563, -0.009528130292892456, -0.03636325150728226, 0.06997880339622498, 0.012568825855851173, 0.049166239798069, 0.00017326511442661285, -0.05551569163799286, -0.10173077136278152, 0.1352764070034027, -0.058849312365055084, -0.22323130071163177, -0.09861410409212112, 0.03589071333408356, -0.015951819717884064, 0.026546331122517586, 0.014165805652737617, -0.03896123543381691, -0.07032185792922974, -0.11490955203771591, 0.06354472786188126, -0.031060636043548584, -0.05945001542568207, 0.0015988685190677643, 0.025862351059913635, -0.004441087134182453, -0.08530192077159882, -0.005646780598908663, 0.011179406195878983, -0.058748986572027206, 0.03651847317814827, 0.01139405183494091, 0.09869913756847382, 0.059338077902793884, 0.020532462745904922, -0.054640598595142365, -0.011880309320986271, 0.22668468952178955, -0.08934766054153442, 0.09054826200008392, 0.17836976051330566, -0.10897806286811829, 0.07825756072998047, 0.08518539369106293, -0.004636690486222506, -0.06597138941287994, 0.037353768944740295, 0.04858146607875824, -0.053951188921928406, -0.10588184744119644, -0.06042233109474182, -0.021527068689465523, -0.07084983587265015, 0.029493208974599838, 0.04446162283420563, 0.010762091726064682, 0.0269143208861351, -0.0894990935921669, 0.01688317395746708, 0.04011239856481552, 0.08663161844015121, 0.04454725980758667, -0.02348225563764572, 0.04329780861735344, -0.03495113551616669, 0.000022210588213056326, 0.07992084324359894, -0.06985311210155487, 0.26336443424224854, -0.012084297835826874, 0.1265924572944641, 0.08194877952337265, -0.036599621176719666, 0.06675823032855988, 0.0014846008270978928, -0.04959314316511154, 0.0007254760712385178, -0.05866572633385658, -0.05423877015709877, -0.008063754998147488, 0.06840036064386368, 0.025193601846694946, -0.03208202123641968, -0.039588361978530884, -0.0018940456211566925, 0.04813583195209503, 0.11805759370326996, -0.03873157873749733, -0.08269913494586945, -0.04491004720330238, 0.07159694284200668, -0.033708326518535614, -0.06928415596485138, 0.010913601145148277, 0.14027845859527588, -0.11137884855270386, 0.007270073983818293, -0.0032869079150259495, 0.07233675569295883, -0.13551637530326843, 0.0071943155489861965, 0.0011359783820807934, 0.021979784592986107, -0.008933153934776783, 0.04316481202840805, -0.13815392553806305, 0.09047337621450424, 0.008687290363013744, 0.056146007031202316, -0.0850694477558136, 0.007569180801510811, 0.0006364388391375542, 0.04126792773604393, 0.11525162309408188, 0.04153529927134514, -0.07946962118148804, -0.056585464626550674, 0.0032431408762931824, 0.028699912130832672, 0.07434144616127014, -0.05389831215143204, 0.09108325839042664, -0.01798831857740879, 0.023831937462091446, -0.02646324597299099, 0.012910842895507812, -0.13566648960113525, -0.176319420337677, 0.04301288723945618, -0.12183986604213715, -0.03633871674537659, -0.07857294380664825, 0.0010197460651397705, -0.004415360279381275, 0.09473318606615067, -0.12726230919361115, -0.09807215631008148, -0.10268238931894302, -0.074149951338768, 0.05964420363306999, -0.06631974130868912, 0.043368808925151825, 0.02423635497689247, 0.14521607756614685, -0.040510646998882294, -0.061803922057151794, 0.017293665558099747, -0.07009810209274292, -0.1672673225402832, -0.05681026726961136, 0.0615042969584465, 0.13682357966899872, 0.05068415403366089, 0.011426086537539959, 0.03253782168030739, 0.0072747101075947285, -0.10684018582105637, -0.010879680514335632, 0.16073298454284668, 0.03374543786048889, -0.003926137927919626, 0.0009258128702640533, -0.023555630818009377, -0.06896185874938965, -0.05595635250210762, 0.047547418624162674, 0.1831439733505249, -0.008493962697684765, 0.09611589461565018, 0.17328378558158875, -0.08459242433309555, -0.186487078666687, -0.019887955859303474, 0.02539883367717266, 0.011104135774075985, 0.11181462556123734, -0.16512854397296906, -0.010743679478764534, 0.04467933624982834, 0.027011357247829437, 0.06291723251342773, -0.2347528040409088, -0.09806735068559647, 0.029611662030220032, 0.06993131339550018, 0.03655482828617096, -0.0708540827035904, -0.01600913517177105, -0.05113302916288376, -0.06620535999536514, 0.09347490221261978, -0.03533082827925682, 0.03588863089680672, 0.00538367684930563, 0.01825021021068096, 0.02594309113919735, -0.006132916547358036, 0.15769094228744507, -0.05109522491693497, 0.0729566439986229, -0.11589480936527252, -0.0852176696062088, 0.02390149235725403, -0.027406781911849976, 0.03099888749420643, -0.03756943717598915, 0.014072074554860592, -0.06434018909931183, -0.06413149833679199, -0.05263620242476463, 0.006362117361277342, -0.06739330291748047, -0.07019312679767609, -0.07778807729482651, 0.08532857149839401, 0.0671629011631012, -0.010375987738370895, -0.037961240857839584, -0.0735565721988678, -0.05837365612387657, 0.045114077627658844, 0.17235277593135834, 0.05043085664510727, -0.12761260569095612, -0.029124373570084572, 0.006755275651812553, 0.1039365828037262, -0.09552700072526932, 0.009892976842820644, 0.04434242472052574, -0.017265576869249344, 0.10081908106803894, 0.009294233284890652, -0.1340622454881668, 0.0487697497010231, 0.04260619729757309, -0.023452933877706528, -0.15278008580207825, -0.02933608554303646, 0.07298611849546432, -0.07325984537601471, -0.09022702276706696, 0.08628445863723755, -0.04538871347904205, -0.004823296330869198, -0.009522996842861176, 0.06333762407302856, 0.023358602076768875, 0.024976138025522232, 0.0028527677059173584, 0.0302369873970747, -0.03408993035554886, 0.08393940329551697, 0.054362453520298004, -0.042776189744472504, 0.04139868542551994, 0.07357312738895416, -0.07243712991476059, -0.064208023250103, -0.07890606671571732, 0.04653486981987953, -0.005285482853651047, -0.05212779715657234, 0.04838936775922775, -0.07191093266010284, 0.0017336253076791763, 0.0533301942050457, 0.012732119299471378, 0.023290159180760384, -0.03371123969554901, 0.015377596952021122, -0.03579647094011307, 0.03916158527135849, -0.020676905289292336, 0.002309344708919525, -0.04743212088942528, 0.032694797962903976, 0.05079290643334389, -0.0027627190575003624, -0.02251817099750042, -0.04504823684692383, -0.08618436008691788, -0.0047812615521252155, -0.13337156176567078, -0.007625746540725231, -0.09619545191526413, -0.02372058853507042, -0.0005544261075556278, 0.018877064809203148, -0.007224884815514088, 0.01956903375685215, -0.027327407151460648, -0.02873135358095169, -0.00702556362375617, 0.044896624982357025, -0.05840778350830078, 0.011147970333695412, 0.07285720109939575, -0.05937982350587845, 0.07405173033475876, -0.059571970254182816, -0.042355410754680634, -0.011622674763202667, -0.07043963670730591, 0.052913691848516464, -0.03563066944479942, 0.009132534265518188, -0.0056062377989292145, -0.12580133974552155, -0.00755306426435709, -0.03438004106283188, -0.043141745030879974, -0.01551198773086071, 0.05328216776251793, -0.06578578054904938, 0.06367267668247223, 0.04630739986896515, -0.03355802595615387, -0.07394486665725708, 0.01726502552628517, 0.03311347961425781, 0.035739585757255554, 0.1184537410736084, -0.0021206100936979055, 0.041592374444007874, -0.10370165854692459, -0.0005483492277562618, 0.04291285201907158, 0.053148772567510605, 0.04874550178647041, -0.054888710379600525, 0.0162661112844944, -0.03041190654039383, 0.12889668345451355, 0.007716603577136993, -0.03280327469110489, 0.04245828464627266, 0.021763024851679802, -0.049131445586681366, 0.028598498553037643, 0.004202160984277725, -0.01807636208832264, -0.004112003371119499, -0.04956802725791931, -0.015185982920229435, -0.0360383614897728, -0.05825777351856232, 0.10740920901298523, 0.08184842765331268, 0.09030456840991974, 0.03267945349216461, 0.0066353571601212025, -0.0509047731757164, -0.04715346544981003, 0.0016539925709366798, 0.030112003907561302, 0.011715339496731758, -0.058775369077920914, 0.027856627479195595, 0.1601923406124115, -0.08058802038431168, 0.08788246661424637, -0.013036543503403664, -0.028034061193466187, -0.018286975100636482, -0.23193016648292542, -0.006117557175457478, 0.040568433701992035, -0.012143414467573166, -0.05696731060743332, 0.028236743062734604, 0.06608431041240692, -0.00397786358371377, -0.038200631737709045, 0.06065433472394943, -0.11099310219287872, -0.1268375962972641, 0.026680264621973038, 0.006194673478603363, 0.03170626610517502, 0.05601657181978226, 0.04785269498825073, 0.018475983291864395, 0.036732178181409836, 0.04482865333557129, 0.07156772911548615, 0.005309471860527992, 0.01584353856742382, -0.05377034470438957, -0.06344524025917053, 0.00503256730735302, -0.001949157565832138, 0.01793571561574936, 0.19221460819244385, 0.044266775250434875, -0.007715856656432152, -0.0005565970204770565, 0.10910259932279587, 0.0050235651433467865, -0.07127586007118225, -0.10156755149364471, 0.11592833697795868, -0.009269963949918747, 0.007596705108880997, -0.004099417477846146, -0.09853032231330872, 0.035631634294986725, 0.14595124125480652, 0.11793772876262665, -0.06497637927532196, -0.010544723831117153, -0.024871796369552612, 0.01234507467597723, -0.04233910143375397, 0.08465882390737534, -0.0002979077398777008, 0.27663376927375793, -0.042093705385923386, 0.08295948803424835, -0.048427827656269073, -0.014403104782104492, -0.020525701344013214, 0.09552521258592606, 0.025097861886024475, 0.01702967658638954, -0.05797044560313225, 0.07815112173557281, -0.043247804045677185, -0.18040023744106293, 0.02749481052160263, 0.017609108239412308, -0.0531073622405529, 0.0482812263071537, -0.00957708340138197, 0.004043959081172943, 0.0781099796295166, 0.010732724331319332, -0.007306508719921112, 0.10703416913747787, 0.023312224075198174, -0.023071173578500748, 0.015670571476221085, 0.08158473670482635, 0.00038789771497249603, 0.17934975028038025, 0.018467240035533905, 0.11872272193431854, 0.07088959962129593, 0.023334555327892303, -0.09702812135219574, 0.04668901488184929, 0.019254200160503387, -0.047453783452510834, 0.013965137302875519, 0.16018974781036377, 0.0009419107809662819, 0.04975389689207077, 0.08270584791898727, 0.03995238617062569, 0.08036544173955917, -0.02389020100235939, -0.007218390237540007, -0.03987351059913635, 0.09136070311069489, -0.09668055176734924, 0.14175905287265778, 0.0936768427491188, 0.00007086072582751513, -0.03070146217942238, -0.018281973898410797, 0.02458261512219906, 0.010043719783425331, 0.0709032416343689, -0.008239459246397018, -0.07810752093791962, 0.00790075771510601, -0.0020623058080673218, 0.033364538103342056, -0.12251053005456924, -0.049385227262973785, 0.035902246832847595, -0.029955897480249405, 0.02067720890045166, 0.051837995648384094, 0.049863290041685104, 0.00032302155159413815, -0.03971381485462189, -0.010713146068155766, -0.020958060398697853, 0.05297427996993065, -0.06335750222206116, -0.02293112501502037 ]
null
null
transformers
## VQGAN-f16-16384 ### Model Description This is a Flax/JAX implementation of VQGAN, which learns a codebook of context-rich visual parts by leveraging both the use of convolutional methods and transformers. It was introduced in [Taming Transformers for High-Resolution Image Synthesis](https://compvis.github.io/taming-transformers/) ([CVPR paper](https://openaccess.thecvf.com/content/CVPR2021/html/Esser_Taming_Transformers_for_High-Resolution_Image_Synthesis_CVPR_2021_paper.html)). The model allows the encoding of images as a fixed-length sequence of tokens taken from the codebook. This version of the model uses a reduction factor `f=16` and a vocabulary of `16,384` tokens. As an example of how the reduction factor works, images of size `256x256` are encoded to sequences of `256` tokens: `256/16 * 256/16`. Images of `512x512` would result in sequences of `1024` tokens. This model was ported to JAX using [a checkpoint trained on ImageNet](https://heibox.uni-heidelberg.de/d/a7530b09fed84f80a887/). ### How to Use The checkpoint can be loaded using [Suraj Patil's implementation](https://github.com/patil-suraj/vqgan-jax) of `VQModel`. ### Other This model can be used as part of the implementation of [DALL·E mini](https://github.com/borisdayma/dalle-mini). Our [report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-mini--Vmlldzo4NjIxODA) contains more details on how to leverage it in an image encoding / generation pipeline.
{}
null
dalle-mini/vqgan_imagenet_f16_16384
[ "transformers", "jax", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #jax #endpoints_compatible #has_space #region-us
## VQGAN-f16-16384 ### Model Description This is a Flax/JAX implementation of VQGAN, which learns a codebook of context-rich visual parts by leveraging both the use of convolutional methods and transformers. It was introduced in Taming Transformers for High-Resolution Image Synthesis (CVPR paper). The model allows the encoding of images as a fixed-length sequence of tokens taken from the codebook. This version of the model uses a reduction factor 'f=16' and a vocabulary of '16,384' tokens. As an example of how the reduction factor works, images of size '256x256' are encoded to sequences of '256' tokens: '256/16 * 256/16'. Images of '512x512' would result in sequences of '1024' tokens. This model was ported to JAX using a checkpoint trained on ImageNet. ### How to Use The checkpoint can be loaded using Suraj Patil's implementation of 'VQModel'. ### Other This model can be used as part of the implementation of DALL·E mini. Our report contains more details on how to leverage it in an image encoding / generation pipeline.
[ "## VQGAN-f16-16384", "### Model Description\n\nThis is a Flax/JAX implementation of VQGAN, which learns a codebook of context-rich visual parts by leveraging both the use of convolutional methods and transformers. It was introduced in Taming Transformers for High-Resolution Image Synthesis (CVPR paper).\n\nThe model allows the encoding of images as a fixed-length sequence of tokens taken from the codebook.\n\nThis version of the model uses a reduction factor 'f=16' and a vocabulary of '16,384' tokens.\n\nAs an example of how the reduction factor works, images of size '256x256' are encoded to sequences of '256' tokens: '256/16 * 256/16'. Images of '512x512' would result in sequences of '1024' tokens.\n\nThis model was ported to JAX using a checkpoint trained on ImageNet.", "### How to Use\n\nThe checkpoint can be loaded using Suraj Patil's implementation of 'VQModel'.", "### Other\n\nThis model can be used as part of the implementation of DALL·E mini. Our report contains more details on how to leverage it in an image encoding / generation pipeline." ]
[ "TAGS\n#transformers #jax #endpoints_compatible #has_space #region-us \n", "## VQGAN-f16-16384", "### Model Description\n\nThis is a Flax/JAX implementation of VQGAN, which learns a codebook of context-rich visual parts by leveraging both the use of convolutional methods and transformers. It was introduced in Taming Transformers for High-Resolution Image Synthesis (CVPR paper).\n\nThe model allows the encoding of images as a fixed-length sequence of tokens taken from the codebook.\n\nThis version of the model uses a reduction factor 'f=16' and a vocabulary of '16,384' tokens.\n\nAs an example of how the reduction factor works, images of size '256x256' are encoded to sequences of '256' tokens: '256/16 * 256/16'. Images of '512x512' would result in sequences of '1024' tokens.\n\nThis model was ported to JAX using a checkpoint trained on ImageNet.", "### How to Use\n\nThe checkpoint can be loaded using Suraj Patil's implementation of 'VQModel'.", "### Other\n\nThis model can be used as part of the implementation of DALL·E mini. Our report contains more details on how to leverage it in an image encoding / generation pipeline." ]
[ 24, 9, 208, 27, 43 ]
[ "passage: TAGS\n#transformers #jax #endpoints_compatible #has_space #region-us \n## VQGAN-f16-16384### Model Description\n\nThis is a Flax/JAX implementation of VQGAN, which learns a codebook of context-rich visual parts by leveraging both the use of convolutional methods and transformers. It was introduced in Taming Transformers for High-Resolution Image Synthesis (CVPR paper).\n\nThe model allows the encoding of images as a fixed-length sequence of tokens taken from the codebook.\n\nThis version of the model uses a reduction factor 'f=16' and a vocabulary of '16,384' tokens.\n\nAs an example of how the reduction factor works, images of size '256x256' are encoded to sequences of '256' tokens: '256/16 * 256/16'. Images of '512x512' would result in sequences of '1024' tokens.\n\nThis model was ported to JAX using a checkpoint trained on ImageNet.### How to Use\n\nThe checkpoint can be loaded using Suraj Patil's implementation of 'VQModel'.### Other\n\nThis model can be used as part of the implementation of DALL·E mini. Our report contains more details on how to leverage it in an image encoding / generation pipeline." ]
[ -0.006597475148737431, 0.06075231358408928, -0.004456229507923126, 0.04626632481813431, 0.13018949329853058, -0.0017908165464177728, 0.10405097901821136, 0.12290865182876587, -0.14265818893909454, 0.0680757537484169, 0.05414735898375511, 0.021042343229055405, 0.06811831146478653, 0.10318948328495026, 0.08109906315803528, -0.24791069328784943, 0.04128504544496536, -0.02032165229320526, -0.0578073225915432, 0.04168806970119476, 0.09300445765256882, -0.1249479278922081, 0.09109122306108475, 0.03338020294904709, -0.16884322464466095, 0.09058205038309097, -0.004405213985592127, -0.04228891059756279, 0.08763128519058228, 0.105554960668087, 0.04212513938546181, -0.03425802290439606, 0.0791950672864914, -0.16392023861408234, 0.010320218279957771, 0.10594967007637024, -0.013712236657738686, 0.053026873618364334, 0.019013753160834312, 0.10305416584014893, 0.14872924983501434, -0.07393950968980789, 0.025447476655244827, 0.01763094589114189, -0.047375500202178955, -0.2571936249732971, -0.05998358502984047, 0.0817716047167778, 0.05016034096479416, 0.04075494781136513, 0.014871301129460335, 0.12323004007339478, 0.06085049733519554, 0.0879000648856163, 0.18393535912036896, -0.258868008852005, -0.01534380204975605, 0.12966981530189514, 0.05181879177689552, 0.06185908243060112, -0.028509553521871567, 0.020341485738754272, 0.020377730950713158, 0.030974650755524635, 0.14051689207553864, -0.05164024606347084, 0.014032313600182533, -0.01648372784256935, -0.08758914470672607, -0.11567975580692291, 0.07135719805955887, -0.026942329481244087, -0.059363286942243576, -0.13615205883979797, -0.07512452453374863, -0.030233802273869514, -0.023930590599775314, -0.1090991199016571, 0.032140638679265976, 0.022404339164495468, 0.026487985625863075, -0.10340812802314758, -0.11289744824171066, -0.03791520372033119, -0.028171759098768234, 0.042469050735235214, 0.03576323762536049, 0.06014057248830795, -0.14054431021213531, 0.042105112224817276, -0.07394247502088547, -0.04886261001229286, -0.056407660245895386, -0.06822925806045532, -0.018442349508404732, 0.03717406466603279, 0.009981624782085419, -0.11268740147352219, -0.024045506492257118, 0.05856532230973244, -0.05484084412455559, 0.0833575576543808, -0.008255742490291595, 0.020815057680010796, 0.04935887083411217, 0.15444493293762207, -0.02487511932849884, 0.06773091852664948, 0.026983026415109634, -0.034935593605041504, 0.0204461682587862, -0.08993024379014969, -0.10091729462146759, -0.0000751412080717273, 0.04272305965423584, 0.04678112268447876, -0.07348236441612244, 0.015170237980782986, 0.03322199732065201, -0.019192127510905266, 0.2181636542081833, -0.10131433606147766, 0.03327342867851257, -0.011691084131598473, -0.0007019525510258973, 0.1201300248503685, 0.04993820935487747, -0.05283498018980026, -0.09633058309555054, -0.042896125465631485, -0.07582768052816391, -0.04715600237250328, -0.10648223757743835, -0.1572473794221878, -0.02822878025472164, -0.08625566214323044, 0.014249906875193119, -0.1635131537914276, -0.09891855716705322, 0.021731648594141006, 0.018696719780564308, -0.020777670666575432, 0.04252914711833, 0.014576028101146221, -0.04842452332377434, 0.0018831976922228932, 0.02426300384104252, -0.029093675315380096, 0.0022141761146485806, 0.0033505542669445276, 0.04526962712407112, 0.08727675676345825, -0.026030369102954865, -0.017985988408327103, -0.046722494065761566, 0.07640711218118668, -0.0996888130903244, 0.08612365275621414, -0.02746998891234398, -0.008499612100422382, -0.08770707994699478, -0.06252073496580124, 0.009874527342617512, 0.005788901820778847, 0.06697894632816315, 0.0740143209695816, -0.18879136443138123, -0.012639903463423252, 0.21226376295089722, -0.14473704993724823, -0.04311424493789673, 0.08136896044015884, -0.02050139009952545, 0.00577427726238966, 0.06739732623100281, 0.11653833091259003, 0.13528983294963837, -0.11646295338869095, -0.038024045526981354, 0.08854041248559952, -0.006200303323566914, -0.010076467879116535, 0.04961676523089409, 0.004704251419752836, 0.0714205875992775, 0.009145119227468967, 0.05470747873187065, -0.05887226387858391, -0.022057505324482918, -0.032411523163318634, -0.025771135464310646, 0.009696109220385551, 0.0036381424870342016, -0.004393335897475481, 0.009960193186998367, -0.021631227806210518, -0.08955947309732437, -0.00195832597091794, 0.09833954274654388, -0.1251336932182312, 0.04001585766673088, -0.09963786602020264, 0.05185620114207268, -0.1132979616522789, 0.026814347133040428, -0.16703198850154877, -0.005378156900405884, 0.043721772730350494, -0.08561054617166519, -0.030888544395565987, 0.06687929481267929, 0.040966808795928955, 0.03634021803736687, 0.06186843663454056, 0.039850231260061264, -0.06892046332359314, -0.03140026330947876, -0.0678844153881073, -0.024940405040979385, -0.11805476993322372, -0.06634949892759323, -0.05006800591945648, -0.14766615629196167, 0.01235931646078825, 0.12429475784301758, 0.031140774488449097, 0.04423638433218002, -0.02016276866197586, -0.012890681624412537, -0.024266621097922325, -0.014412510208785534, -0.08429989218711853, 0.014282654039561749, 0.04672420769929886, 0.040734656155109406, 0.04981568455696106, -0.20533928275108337, -0.004526039585471153, 0.0733381137251854, 0.07315460592508316, -0.007910169661045074, -0.016951825469732285, -0.007829505950212479, 0.02091672644019127, -0.027646712958812714, -0.02911241166293621, 0.15080216526985168, 0.00466998340561986, 0.12674598395824432, -0.09203733503818512, -0.038568466901779175, 0.11060196161270142, -0.0323791429400444, -0.05549416318535805, 0.011546962894499302, -0.020403606817126274, -0.09077993780374527, 0.013790734112262726, 0.025763211771845818, 0.043240610510110855, 0.12315165996551514, 0.01666579209268093, -0.08409300446510315, -0.061058081686496735, -0.05177260562777519, 0.02807004377245903, 0.09256446361541748, -0.03530491888523102, -0.05356869846582413, 0.03532037511467934, 0.054990701377391815, 0.04022848606109619, -0.09337830543518066, 0.06275992840528488, 0.10194756090641022, -0.041075292974710464, -0.0023498975206166506, 0.007090691477060318, -0.0664815679192543, 0.02529526688158512, 0.06247000768780708, 0.11698020994663239, 0.0023803780786693096, -0.049999549984931946, -0.09919466823339462, 0.16966071724891663, -0.12142793089151382, -0.2606508433818817, -0.1457156091928482, 0.005495865363627672, 0.01880567893385887, 0.005322332493960857, 0.03676815703511238, -0.0627497136592865, -0.09325451403856277, -0.10142833739519119, -0.013901646248996258, -0.13209189474582672, -0.013541982509195805, 0.06323250383138657, -0.020253533497452736, 0.029269423335790634, -0.15135237574577332, -0.003396119922399521, 0.060459356755018234, -0.03442729264497757, -0.014017732813954353, 0.013907987624406815, 0.08593756705522537, 0.05869848281145096, -0.09817526489496231, 0.04889814183115959, -0.012063420377671719, 0.25734779238700867, -0.04158835485577583, 0.03772962465882301, 0.19707587361335754, -0.0643192008137703, 0.05201348289847374, 0.01932741515338421, -0.008330519311130047, -0.04560014232993126, 0.009362342767417431, -0.00662068510428071, -0.09569109231233597, -0.1392700970172882, 0.015375569462776184, -0.05292893573641777, 0.01994434930384159, 0.07249003648757935, 0.045524463057518005, 0.005793232470750809, 0.05838911980390549, 0.006608649156987667, 0.13250994682312012, -0.06137358397245407, 0.08221586048603058, 0.16668537259101868, 0.027526775375008583, 0.028263969346880913, -0.09543631970882416, 0.00855342485010624, 0.07282500714063644, 0.08445308357477188, 0.15255744755268097, -0.02411399967968464, 0.15503886342048645, 0.04605429619550705, 0.09253711253404617, 0.07632394880056381, 0.09708815068006516, -0.09284227341413498, 0.0043264818377792835, -0.031210321933031082, -0.06493503600358963, -0.053058214485645294, 0.026106497272849083, -0.04852234572172165, -0.02268843725323677, -0.0718969851732254, 0.0593363493680954, 0.005808097776025534, 0.12530788779258728, 0.01945807971060276, -0.19315417110919952, -0.09573047608137131, 0.024855302646756172, 0.014949682168662548, -0.09338223189115524, 0.02405383437871933, 0.20658162236213684, -0.11168184131383896, -0.029361600056290627, -0.05385259911417961, 0.07018579542636871, -0.12810710072517395, 0.018515532836318016, -0.005681218113750219, 0.10133697837591171, 0.032182879745960236, 0.1119547039270401, -0.212228462100029, 0.06661240011453629, 0.023691875860095024, 0.08967912197113037, -0.0917946994304657, 0.09227713197469711, -0.034604329615831375, 0.007272906601428986, 0.13490265607833862, -0.012805657461285591, -0.11157333850860596, -0.09284047037363052, -0.009131581522524357, 0.044310424476861954, 0.0792350172996521, 0.003048093756660819, 0.09695817530155182, -0.04963739216327667, -0.00020549895998556167, -0.020748862996697426, 0.06001126393675804, -0.08139175921678543, -0.22103972733020782, 0.026732521131634712, -0.016249913722276688, -0.07797889411449432, -0.04177949205040932, -0.0050312429666519165, 0.0004254453524481505, 0.10854334384202957, -0.018195275217294693, -0.08033880591392517, -0.11139972507953644, -0.001542100333608687, 0.055060893297195435, -0.06872950494289398, 0.0857735425233841, -0.0012561575276777148, 0.22845135629177094, -0.028693117201328278, -0.14935676753520966, 0.04444728046655655, -0.086384616792202, -0.06287693977355957, 0.011671624146401882, 0.033289022743701935, 0.0040389783680438995, 0.0053517804481089115, 0.05004841089248657, 0.02189982496201992, -0.01815970055758953, -0.09972747415304184, 0.005550415720790625, 0.09419801831245422, 0.02809162065386772, -0.08013885468244553, 0.030777225270867348, 0.022404231131076813, -0.003428179770708084, 0.025596708059310913, 0.13273324072360992, 0.18982960283756256, -0.10918436199426651, 0.07709953933954239, 0.14891357719898224, -0.11561719328165054, -0.1807917207479477, -0.04807161167263985, -0.00025641193496994674, 0.07081091403961182, 0.07354515045881271, -0.14198927581310272, 0.005407281219959259, 0.03167736902832985, -0.0030958608258515596, 0.016827527433633804, -0.3455412983894348, -0.06075385957956314, 0.006283250171691179, 0.12176771461963654, 0.007726958021521568, -0.09619499742984772, -0.027506904676556587, -0.027833834290504456, -0.16414155066013336, 0.12659214437007904, -0.12525399029254913, 0.026031112298369408, 0.0020261602476239204, -0.01911516673862934, 0.032257769256830215, -0.03817703574895859, 0.06458020955324173, -0.07834397256374359, 0.07468651235103607, -0.0668681189417839, -0.014223086647689342, 0.15718825161457062, -0.05069514364004135, 0.10355378687381744, -0.03157993033528328, 0.10741634666919708, -0.13223323225975037, -0.04418587684631348, -0.0068824090994894505, -0.016070617362856865, -0.06778328120708466, -0.06885010749101639, -0.09452197700738907, -0.004750771913677454, 0.0843408927321434, -0.010388900525867939, 0.08726832270622253, -0.009407827630639076, 0.061569783836603165, 0.24702170491218567, 0.07448070496320724, 0.003772465046495199, -0.0968116894364357, 0.009650221094489098, 0.008484577760100365, 0.09505777806043625, -0.17656753957271576, 0.05040058121085167, 0.07832375168800354, 0.012504837475717068, 0.04480051249265671, 0.03847331926226616, -0.11980484426021576, 0.05803735554218292, 0.027435265481472015, -0.11127851903438568, -0.11785715073347092, -0.015646275132894516, 0.06931867450475693, -0.053927477449178696, 0.02001003362238407, 0.1374770849943161, -0.08543037623167038, -0.029841013252735138, -0.012725061737000942, 0.0015431044157594442, -0.01536548137664795, 0.09658888727426529, 0.09057372808456421, 0.06130669638514519, -0.0581083782017231, 0.0878656879067421, 0.06505430489778519, -0.12993492186069489, 0.0748833417892456, 0.11734998971223831, -0.09341247379779816, -0.04442073777318001, -0.014730908907949924, 0.09598558396100998, -0.004352628253400326, -0.06942406296730042, -0.03553401678800583, -0.05442747101187706, 0.006994308438152075, 0.013820101507008076, 0.02109253779053688, 0.02499021776020527, -0.0893305242061615, 0.03774488717317581, -0.17469793558120728, 0.10036244988441467, -0.05669589713215828, 0.018628878518939018, -0.15796402096748352, 0.11058587580919266, 0.013718505389988422, 0.12749408185482025, -0.026879139244556427, -0.03250686451792717, -0.06825758516788483, -0.02350478246808052, -0.09902729094028473, 0.03509525582194328, -0.055099762976169586, -0.021361948922276497, -0.009451759979128838, 0.0593080036342144, -0.03397646173834801, 0.061375875025987625, -0.03203617408871651, -0.03825535625219345, -0.03777103126049042, -0.007951532490551472, -0.0702953115105629, 0.017413616180419922, -0.014654004015028477, -0.05552372708916664, 0.07954970002174377, -0.041637446731328964, -0.026718569919466972, 0.03670267388224602, 0.08424276858568192, 0.03195495903491974, 0.02637110836803913, 0.02550472691655159, 0.04410451278090477, 0.0027906291652470827, 0.001845145015977323, 0.008112234994769096, -0.034366805106401443, -0.04744882136583328, 0.013782696798443794, -0.06374825537204742, -0.003988170996308327, -0.0711117535829544, 0.042406100779771805, -0.0447426363825798, 0.08148878067731857, 0.04901253432035446, 0.035465240478515625, 0.055809393525123596, -0.058375775814056396, 0.06057257577776909, -0.12482760846614838, -0.05911221355199814, 0.0314796045422554, -0.030342796817421913, -0.010087205097079277, -0.06973439455032349, 0.035862065851688385, -0.07887274026870728, 0.15780657529830933, 0.04261070489883423, 0.11324318498373032, 0.014319085516035557, -0.1634550392627716, -0.08779416978359222, 0.007092856336385012, 0.18563377857208252, 0.03199489414691925, -0.010736756026744843, 0.03260720148682594, 0.033558931201696396, -0.0020708914380520582, 0.11565103381872177, 0.1250579059123993, -0.03536168113350868, 0.02845638617873192, 0.12242060154676437, 0.026716085150837898, -0.035879477858543396, -0.13783526420593262, 0.01163066178560257, -0.05714265629649162, 0.11408662796020508, -0.06558507680892944, -0.027836283668875694, 0.21363282203674316, -0.0946296826004982, 0.10407939553260803, 0.002776293782517314, -0.07000042498111725, -0.08789616078138351, -0.15721938014030457, -0.04738462716341019, -0.12313863635063171, -0.030244335532188416, -0.13286641240119934, 0.06727684289216995, 0.07897362858057022, 0.004011448938399553, -0.031941525638103485, 0.11826067417860031, 0.013859299942851067, -0.06373712420463562, 0.034589286893606186, -0.009034894406795502, 0.026351328939199448, 0.03934279829263687, 0.08221487700939178, 0.06607542186975479, 0.018375547602772713, 0.06798091530799866, 0.05738638713955879, 0.12599711120128632, 0.041117943823337555, -0.03357372060418129, -0.05379308760166168, -0.0221824012696743, 0.010575284250080585, -0.004123662132769823, 0.06506295502185822, 0.04956422746181488, -0.07565885782241821, -0.034462738782167435, 0.09981994330883026, -0.06497890502214432, -0.02921859733760357, -0.09181968122720718, 0.11621768772602081, 0.02331257238984108, 0.07822538912296295, 0.01085728220641613, -0.09844338148832321, -0.04450950399041176, 0.22150084376335144, 0.11270467191934586, 0.030686188489198685, 0.011965623125433922, 0.05013643950223923, 0.0018154382705688477, -0.049349017441272736, 0.170135498046875, 0.03877163678407669, 0.23721516132354736, -0.04799175262451172, -0.001933518098667264, -0.023950425907969475, -0.007442439906299114, -0.06679996103048325, 0.07730937749147415, -0.013000392355024815, -0.012245827354490757, 0.015115676447749138, 0.006956877652555704, -0.030422406271100044, -0.16831789910793304, 0.033524464815855026, -0.06351705640554428, -0.043520230799913406, 0.03693699091672897, 0.002903989516198635, 0.008957106620073318, 0.07417849451303482, -0.025139357894659042, 0.020773641765117645, 0.24488048255443573, 0.01650667004287243, -0.07464957982301712, -0.033127978444099426, 0.041611023247241974, -0.047933436930179596, 0.16033388674259186, 0.0018052976811304688, 0.05183035880327225, 0.07483185082674026, 0.013185892254114151, -0.1254570335149765, -0.012934296391904354, 0.02052290551364422, 0.00007073237065924332, -0.0352497361600399, 0.08517999202013016, -0.013035633601248264, -0.028518186882138252, 0.022692952305078506, -0.08160614967346191, -0.0008438150398433208, -0.012770135886967182, 0.06884321570396423, -0.047352395951747894, -0.025135094299912453, -0.08898429572582245, 0.127938911318779, 0.08836307376623154, -0.008296037092804909, -0.02270093932747841, -0.006957540288567543, 0.0747096836566925, 0.0004607860173564404, 0.12509967386722565, -0.00608002208173275, -0.06737674027681351, -0.013234317302703857, -0.05201323702931404, 0.0066835917532444, -0.1706818789243698, -0.029993275180459023, 0.046318527311086655, -0.031917065382003784, -0.034376516938209534, 0.07591889798641205, 0.10142166912555695, 0.04014519229531288, -0.07242535054683685, -0.015774894505739212, -0.03859666734933853, 0.1052083820104599, -0.09445430338382721, -0.07853350043296814 ]
null
null
transformers
# HIV_BERT model ## Table of Contents - [Summary](#model-summary) - [Model Description](#model-description) - [Intended Uses & Limitations](#intended-uses-&-limitations) - [How to Use](#how-to-use) - [Training Data](#training-data) - [Training Procedure](#training-procedure) - [Preprocessing](#preprocessing) - [Training](#training) - [Evaluation Results](#evaluation-results) - [BibTeX Entry and Citation Info](#bibtex-entry-and-citation-info) ## Summary The HIV-BERT model was trained as a refinement of the [ProtBert-BFD model](https://huggingface.co/Rostlab/prot_bert_bfd) for HIV centric tasks. It was refined with whole viral genomes from the [Los Alamos HIV Sequence Database](https://www.hiv.lanl.gov/content/sequence/HIV/mainpage.html). This pretraining is important for HIV related tasks as the original BFD database contains few viral proteins making it sub-optimal when used as the basis for transfer learning tasks. This model and other related HIV prediction tasks have been published (link). ## Model Description Like the original [ProtBert-BFD model](https://huggingface.co/Rostlab/prot_bert_bfd), this model encodes each amino acid as an individual token. This model was trained using Masked Language Modeling: a process in which a random set of tokens are masked with the model trained on their prediction. This model was trained using the damlab/hiv-flt dataset with 256 amino acid chunks and a 15% mask rate. ## Intended Uses & Limitations As a masked language model this tool can be used to predict expected mutations using a masking approach. This could be used to identify highly mutated sequences, sequencing artifacts, or other contexts. As a BERT model, this tool can also be used as the base for transfer learning. This pretrained model could be used as the base when developing HIV-specific classification tasks. ## How to use As this is a BERT-style Masked Language learner, it can be used to determine the most likely amino acid at a masked position. ```python from transformers import pipeline unmasker = pipeline("fill-mask", model="damlab/HIV_FLT") unmasker(f"C T R P N [MASK] N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C") [ { "score": 0.9581968188285828, "token": 17, "token_str": "N", "sequence": "C T R P N N N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C" }, { "score": 0.022986575961112976, "token": 12, "token_str": "K", "sequence": "C T R P N K N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C" }, { "score": 0.003997281193733215, "token": 14, "token_str": "D", "sequence": "C T R P N D N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C" }, { "score": 0.003636382520198822, "token": 15, "token_str": "T", "sequence": "C T R P N T N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C" }, { "score": 0.002701344434171915, "token": 10, "token_str": "S", "sequence": "C T R P N S N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C" } ] ``` ## Training Data The dataset [damlab/HIV_FLT](https://huggingface.co/datasets/damlab/HIV_FLT) was used to refine the original [rostlab/Prot-bert-bfd](https://huggingface.co/Rostlab/prot_bert_bfd). This dataset contains 1790 full HIV genomes from across the globe. When translated, these genomes contain approximately 3.9 million amino-acid tokens. ## Training Procedure ### Preprocessing As with the [rostlab/Prot-bert-bfd](https://huggingface.co/Rostlab/prot_bert_bfd) model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation. ### Training Training was performed with the HuggingFace training module using the MaskedLM data loader with a 15% masking rate. The learning rate was set at E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. ## BibTeX Entry and Citation Info [More Information Needed]
{"license": "mit", "datasets": ["damlab/HIV_FLT"], "metrics": ["accuracy"], "widget": [{"text": "C T R P N N N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C", "example_title": "V3"}, {"text": "M E P V D P R L E P W K H P G S Q P K T A C T N C Y C K K C C F H C Q V C F I T K A L G I S Y G R K K R R Q R R R A H Q N S Q T H Q A S L S K Q P T S Q P R G D P T G P K E S K K K V E R E T E T D P F D", "example_title": "Tat"}, {"text": "P Q I T L W Q R P L V T I K I G G Q L K E A L L D T G A D D T V L E E M N L P G R W K P K M I G G I G G F I K V R Q Y D Q I L I E I C G H K A I G T V L V G P T P V N I I G R N L L T Q I G C T L N F", "example_title": "PR"}]}
fill-mask
damlab/HIV_BERT
[ "transformers", "pytorch", "bert", "fill-mask", "dataset:damlab/HIV_FLT", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #bert #fill-mask #dataset-damlab/HIV_FLT #license-mit #autotrain_compatible #endpoints_compatible #region-us
# HIV_BERT model ## Table of Contents - Summary - Model Description - Intended Uses & Limitations - How to Use - Training Data - Training Procedure - Preprocessing - Training - Evaluation Results - BibTeX Entry and Citation Info ## Summary The HIV-BERT model was trained as a refinement of the ProtBert-BFD model for HIV centric tasks. It was refined with whole viral genomes from the Los Alamos HIV Sequence Database. This pretraining is important for HIV related tasks as the original BFD database contains few viral proteins making it sub-optimal when used as the basis for transfer learning tasks. This model and other related HIV prediction tasks have been published (link). ## Model Description Like the original ProtBert-BFD model, this model encodes each amino acid as an individual token. This model was trained using Masked Language Modeling: a process in which a random set of tokens are masked with the model trained on their prediction. This model was trained using the damlab/hiv-flt dataset with 256 amino acid chunks and a 15% mask rate. ## Intended Uses & Limitations As a masked language model this tool can be used to predict expected mutations using a masking approach. This could be used to identify highly mutated sequences, sequencing artifacts, or other contexts. As a BERT model, this tool can also be used as the base for transfer learning. This pretrained model could be used as the base when developing HIV-specific classification tasks. ## How to use As this is a BERT-style Masked Language learner, it can be used to determine the most likely amino acid at a masked position. ## Training Data The dataset damlab/HIV_FLT was used to refine the original rostlab/Prot-bert-bfd. This dataset contains 1790 full HIV genomes from across the globe. When translated, these genomes contain approximately 3.9 million amino-acid tokens. ## Training Procedure ### Preprocessing As with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation. ### Training Training was performed with the HuggingFace training module using the MaskedLM data loader with a 15% masking rate. The learning rate was set at E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. ## BibTeX Entry and Citation Info
[ "# HIV_BERT model", "## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info", "## Summary\r\n\r\nThe HIV-BERT model was trained as a refinement of the ProtBert-BFD model for HIV centric tasks. It was refined with whole viral genomes from the Los Alamos HIV Sequence Database. This pretraining is important for HIV related tasks as the original BFD database contains few viral proteins making it sub-optimal when used as the basis for transfer learning tasks. This model and other related HIV prediction tasks have been published (link).", "## Model Description\r\n\r\nLike the original ProtBert-BFD model, this model encodes each amino acid as an individual token. This model was trained using Masked Language Modeling: a process in which a random set of tokens are masked with the model trained on their prediction. This model was trained using the damlab/hiv-flt dataset with 256 amino acid chunks and a 15% mask rate.", "## Intended Uses & Limitations\r\n\r\nAs a masked language model this tool can be used to predict expected mutations using a masking approach. This could be used to identify highly mutated sequences, sequencing artifacts, or other contexts. As a BERT model, this tool can also be used as the base for transfer learning. This pretrained model could be used as the base when developing HIV-specific classification tasks.", "## How to use\r\n\r\nAs this is a BERT-style Masked Language learner, it can be used to determine the most likely amino acid at a masked position.", "## Training Data\r\n\r\nThe dataset damlab/HIV_FLT was used to refine the original rostlab/Prot-bert-bfd. This dataset contains 1790 full HIV genomes from across the globe. When translated, these genomes contain approximately 3.9 million amino-acid tokens.", "## Training Procedure", "### Preprocessing\r\n\r\nAs with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.", "### Training\r\n\r\nTraining was performed with the HuggingFace training module using the MaskedLM data loader with a 15% masking rate. The learning rate was set at E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset.", "## BibTeX Entry and Citation Info" ]
[ "TAGS\n#transformers #pytorch #bert #fill-mask #dataset-damlab/HIV_FLT #license-mit #autotrain_compatible #endpoints_compatible #region-us \n", "# HIV_BERT model", "## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info", "## Summary\r\n\r\nThe HIV-BERT model was trained as a refinement of the ProtBert-BFD model for HIV centric tasks. It was refined with whole viral genomes from the Los Alamos HIV Sequence Database. This pretraining is important for HIV related tasks as the original BFD database contains few viral proteins making it sub-optimal when used as the basis for transfer learning tasks. This model and other related HIV prediction tasks have been published (link).", "## Model Description\r\n\r\nLike the original ProtBert-BFD model, this model encodes each amino acid as an individual token. This model was trained using Masked Language Modeling: a process in which a random set of tokens are masked with the model trained on their prediction. This model was trained using the damlab/hiv-flt dataset with 256 amino acid chunks and a 15% mask rate.", "## Intended Uses & Limitations\r\n\r\nAs a masked language model this tool can be used to predict expected mutations using a masking approach. This could be used to identify highly mutated sequences, sequencing artifacts, or other contexts. As a BERT model, this tool can also be used as the base for transfer learning. This pretrained model could be used as the base when developing HIV-specific classification tasks.", "## How to use\r\n\r\nAs this is a BERT-style Masked Language learner, it can be used to determine the most likely amino acid at a masked position.", "## Training Data\r\n\r\nThe dataset damlab/HIV_FLT was used to refine the original rostlab/Prot-bert-bfd. This dataset contains 1790 full HIV genomes from across the globe. When translated, these genomes contain approximately 3.9 million amino-acid tokens.", "## Training Procedure", "### Preprocessing\r\n\r\nAs with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.", "### Training\r\n\r\nTraining was performed with the HuggingFace training module using the MaskedLM data loader with a 15% masking rate. The learning rate was set at E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset.", "## BibTeX Entry and Citation Info" ]
[ 53, 6, 52, 109, 93, 96, 35, 69, 4, 80, 82, 11 ]
[ "passage: TAGS\n#transformers #pytorch #bert #fill-mask #dataset-damlab/HIV_FLT #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# HIV_BERT model## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info## Summary\r\n\r\nThe HIV-BERT model was trained as a refinement of the ProtBert-BFD model for HIV centric tasks. It was refined with whole viral genomes from the Los Alamos HIV Sequence Database. This pretraining is important for HIV related tasks as the original BFD database contains few viral proteins making it sub-optimal when used as the basis for transfer learning tasks. This model and other related HIV prediction tasks have been published (link).## Model Description\r\n\r\nLike the original ProtBert-BFD model, this model encodes each amino acid as an individual token. This model was trained using Masked Language Modeling: a process in which a random set of tokens are masked with the model trained on their prediction. This model was trained using the damlab/hiv-flt dataset with 256 amino acid chunks and a 15% mask rate.## Intended Uses & Limitations\r\n\r\nAs a masked language model this tool can be used to predict expected mutations using a masking approach. This could be used to identify highly mutated sequences, sequencing artifacts, or other contexts. As a BERT model, this tool can also be used as the base for transfer learning. This pretrained model could be used as the base when developing HIV-specific classification tasks.## How to use\r\n\r\nAs this is a BERT-style Masked Language learner, it can be used to determine the most likely amino acid at a masked position." ]
[ -0.03160731494426727, 0.08184020221233368, -0.0024370141327381134, -0.0017657733988016844, 0.1079944297671318, 0.006299123167991638, 0.11540769785642624, 0.10264714062213898, 0.07042404264211655, 0.08010247349739075, 0.013361725956201553, -0.08970969170331955, 0.07429111748933792, 0.17727544903755188, 0.10472066700458527, -0.2174062728881836, 0.06442227214574814, -0.05051722005009651, -0.04334679991006851, 0.06222013011574745, 0.08324696868658066, -0.0671345591545105, 0.032846543937921524, 0.0032635671086609364, 0.03448178246617317, 0.013491138815879822, -0.03732994571328163, -0.039532002061605453, 0.0732407197356224, 0.0759970024228096, 0.07817909866571426, 0.07251995801925659, 0.10258915275335312, -0.24010802805423737, 0.005486767273396254, 0.07223475724458694, -0.019493550062179565, 0.027336424216628075, 0.08258278667926788, 0.011854149401187897, 0.15314346551895142, -0.030468327924609184, 0.09030371904373169, 0.08182527124881744, -0.11814821511507034, -0.1487777978181839, -0.06202559545636177, 0.11534034460783005, -0.007193629164248705, 0.08529489487409592, -0.0072442227974534035, 0.03316739201545715, 0.03165663778781891, 0.032895952463150024, 0.07277265191078186, -0.13950832188129425, 0.0044417791068553925, 0.18499457836151123, 0.07112042605876923, 0.032196447253227234, -0.06537136435508728, 0.0008379845530726016, -0.01113959588110447, 0.06624946743249893, 0.18331147730350494, -0.04116291552782059, 0.014031779021024704, -0.0869210734963417, -0.16838794946670532, -0.09083826839923859, 0.074562206864357, -0.012571455910801888, -0.07326241582632065, -0.11875617504119873, -0.007513594813644886, 0.08611667901277542, -0.02506449446082115, -0.04385041072964668, -0.030140668153762817, 0.054710716009140015, -0.006755960639566183, -0.10260245203971863, -0.015962377190589905, -0.09109465777873993, 0.049168769270181656, 0.0693390816450119, 0.015662826597690582, 0.02303341031074524, -0.06443852186203003, 0.04439757019281387, -0.13740015029907227, -0.06974027305841446, -0.03589406609535217, -0.05470375344157219, -0.05780988559126854, -0.029501045122742653, 0.0006595068261958659, -0.23916099965572357, 0.009552656672894955, 0.07025258243083954, -0.03208720311522484, 0.013530377298593521, -0.0480194017291069, 0.008104556240141392, 0.07704471796751022, 0.10271189361810684, -0.09971120208501816, 0.007560933008790016, 0.019337201490998268, -0.0014732856070622802, -0.01306725014001131, 0.0016453496646136045, -0.021655114367604256, 0.01043171901255846, 0.040306538343429565, 0.0770094096660614, -0.01743638701736927, 0.05049917474389076, 0.03415177762508392, -0.0168002862483263, 0.1082007884979248, -0.07711682468652725, -0.03906136006116867, -0.014817952178418636, -0.015302197076380253, -0.013202599249780178, 0.11944244801998138, -0.07413462549448013, -0.03427067771553993, 0.06302405893802643, -0.054763101041316986, -0.0794227123260498, -0.0979824811220169, -0.1480003446340561, -0.014773267321288586, 0.06731698662042618, -0.06953515112400055, -0.07586141675710678, -0.16603928804397583, -0.08370038866996765, -0.0031973354052752256, 0.011809129267930984, 0.010481659322977066, -0.002352563664317131, -0.002000609878450632, 0.007760423235595226, 0.011679306626319885, -0.00511739170178771, -0.017887115478515625, 0.009769155643880367, -0.1315736621618271, 0.08200034499168396, 0.0767325833439827, 0.023149242624640465, -0.09360368549823761, 0.08319010585546494, -0.25641965866088867, 0.09263823181390762, -0.14606024324893951, 0.013222502544522285, -0.13882675766944885, -0.06467430293560028, -0.05345821753144264, -0.023891186341643333, 0.0452410951256752, 0.10663195699453354, -0.10864375531673431, -0.025458596646785736, 0.3334011137485504, -0.13566696643829346, -0.008723828941583633, 0.06359130889177322, -0.03992639482021332, 0.14630448818206787, 0.08286789059638977, 0.07404768466949463, 0.10201387107372284, -0.029866553843021393, -0.05626515671610832, -0.0029493430629372597, -0.021125486120581627, 0.1304551362991333, 0.003975489642471075, -0.07363470643758774, -0.016918087378144264, -0.004502100870013237, -0.028936713933944702, 0.013756237924098969, -0.029400087893009186, 0.0018798175733536482, 0.00503338361158967, -0.0561564601957798, 0.023376064375042915, -0.018961075693368912, -0.04946143180131912, 0.023546185344457626, -0.0778389573097229, 0.04167269542813301, 0.13663369417190552, -0.111199289560318, 0.03823857009410858, -0.068715900182724, 0.022078987210989, -0.014179069548845291, -0.0015426711179316044, -0.1750757396221161, -0.08655544370412827, 0.04300263896584511, -0.09966649115085602, 0.07571746408939362, -0.11990191042423248, 0.014994307421147823, 0.049916334450244904, -0.06765154749155045, 0.007487924303859472, -0.07730565965175629, -0.013647693209350109, -0.07051810622215271, -0.129505917429924, -0.08810960501432419, -0.05179751664400101, 0.10057520866394043, -0.12918749451637268, -0.0020136674866080284, -0.05913009122014046, 0.05756545811891556, 0.07480531185865402, -0.05125879868865013, 0.12508521974086761, 0.04679304361343384, -0.00485282763838768, -0.04968849569559097, -0.007092348299920559, 0.017676912248134613, -0.06722059845924377, -0.004071415867656469, -0.054200902581214905, -0.22943422198295593, -0.002058328129351139, 0.05710954591631889, -0.10711275041103363, 0.02745118737220764, 0.007747008930891752, 0.0004480364441405982, -0.0932864099740982, -0.015478672459721565, 0.12693412601947784, 0.0074186441488564014, 0.07410623133182526, -0.02657615765929222, 0.04684929549694061, 0.026029640808701515, 0.017007220536470413, -0.009218432940542698, 0.03400113806128502, 0.09209150820970535, -0.15974603593349457, 0.05429435521364212, -0.04989403113722801, 0.03709767758846283, 0.11756640672683716, 0.015987396240234375, -0.09369159489870071, 0.0410664901137352, -0.024875590577721596, 0.044307123869657516, 0.07563639432191849, -0.06674271821975708, 0.004805553238838911, 0.03249132260680199, -0.005684540141373873, 0.011619137600064278, 0.010314682498574257, 0.0743369609117508, 0.026360981166362762, -0.0753566101193428, -0.07903289794921875, -0.05104093253612518, -0.03921645134687424, 0.07496732473373413, 0.04149901494383812, -0.0035801606718450785, -0.03663461655378342, -0.030806943774223328, -0.13376981019973755, 0.13642382621765137, -0.08264297991991043, -0.12520423531532288, -0.16399821639060974, 0.0393928661942482, -0.02282705530524254, 0.012948415242135525, -0.006458152085542679, 0.03005753643810749, -0.07008805125951767, -0.11811563372612, 0.11437851935625076, -0.06783881783485413, -0.04900658130645752, -0.050251949578523636, 0.03875108063220978, 0.06965424120426178, -0.08301272988319397, 0.030640767887234688, -0.006817636080086231, -0.028605429455637932, 0.02017347887158394, -0.04948161542415619, 0.1089276373386383, 0.10561530292034149, 0.06171383708715439, -0.02672586217522621, -0.06609233468770981, 0.19854214787483215, -0.06727918237447739, 0.11289804428815842, 0.10889790207147598, -0.03343617543578148, 0.03395366296172142, 0.08421114832162857, 0.0023107193410396576, -0.023346057161688805, 0.07187540829181671, 0.04792638123035431, -0.06971272826194763, -0.18097008764743805, -0.0976305678486824, -0.020187528803944588, -0.010687149129807949, 0.042016346007585526, 0.03795215114951134, 0.06717193126678467, 0.025557825341820717, -0.08396553248167038, 0.017423806712031364, 0.0007197275408543646, 0.06256408989429474, -0.016830705106258392, 0.019698625430464745, 0.04722112789750099, -0.03720737248659134, 0.013955305330455303, 0.082737497985363, -0.0956791415810585, 0.20778913795948029, 0.006416297983378172, 0.1454060971736908, 0.07481274753808975, 0.08124256134033203, 0.05509515106678009, 0.07210695743560791, -0.03684117645025253, 0.033437591046094894, 0.006032864097505808, -0.055803775787353516, 0.020915597677230835, 0.03014427237212658, -0.028910603374242783, 0.0010487191611900926, 0.0033973711542785168, -0.10603979229927063, -0.019858913496136665, 0.19137801229953766, 0.06879045814275742, -0.13722455501556396, -0.09666093438863754, 0.023248231038451195, -0.044597748667001724, -0.13211849331855774, -0.008966930210590363, 0.08480949699878693, -0.16601990163326263, 0.05876415595412254, -0.02526981383562088, 0.1181800588965416, -0.0379619374871254, -0.03588990494608879, -0.07319354265928268, 0.0526764802634716, -0.09817440062761307, 0.09795933961868286, -0.23473189771175385, 0.09108690917491913, 0.009039143100380898, 0.07738828659057617, -0.07091183960437775, 0.007512554991990328, 0.07903625816106796, 0.1357707977294922, 0.1511835753917694, 0.01604856364428997, -0.0828404575586319, -0.060985926538705826, -0.12917175889015198, -0.006699876859784126, 0.0529341958463192, -0.10671232640743256, 0.1350230723619461, 0.027585729956626892, 0.019079705700278282, -0.02510324865579605, 0.10989196598529816, -0.20259182155132294, -0.08384209126234055, 0.02859257161617279, -0.041456107050180435, 0.13330771028995514, -0.03827917203307152, -0.038814838975667953, -0.0118956184014678, 0.07642661035060883, -0.08309301733970642, -0.09962732344865799, -0.14849746227264404, 0.020433329045772552, 0.10089107602834702, -0.08184989541769028, 0.041065577417612076, 0.015195704065263271, 0.16021110117435455, -0.05906122177839279, -0.1313001811504364, 0.04946618899703026, -0.08779840171337128, -0.15352584421634674, -0.0683356299996376, 0.04100200906395912, 0.15021930634975433, 0.10053034126758575, 0.031091948971152306, 0.05869051814079285, 0.08793776482343674, -0.07756544649600983, 0.05476096272468567, 0.1467912346124649, -0.010593698360025883, 0.06717412918806076, -0.11423752456903458, -0.0022176317870616913, -0.1154656857252121, -0.0679335743188858, 0.10547921061515808, 0.07683044672012329, -0.06638002395629883, 0.07459528744220734, 0.17174145579338074, -0.14396347105503082, -0.26237061619758606, -0.021390384063124657, 0.07517693191766739, 0.05139570310711861, 0.04520454630255699, -0.31111499667167664, -0.013472672551870346, 0.012048818171024323, -0.04515145719051361, 0.030407149344682693, -0.30475470423698425, -0.11408619582653046, 0.09906263649463654, 0.021627308800816536, -0.04134950041770935, -0.10673472285270691, -0.05366339161992073, -0.05232163891196251, 0.005613545887172222, 0.09186619520187378, -0.003874091897159815, 0.0290667787194252, -0.0011168023338541389, 0.020208969712257385, 0.022507863119244576, 0.007088452577590942, 0.06697382032871246, -0.004702885635197163, 0.05286511778831482, -0.05073634162545204, 0.09919659793376923, 0.08743195235729218, -0.06430305540561676, 0.09296002984046936, 0.09006441384553909, 0.02909923531115055, -0.14122337102890015, -0.06018795818090439, -0.07830724865198135, 0.04160580411553383, -0.0311388298869133, -0.07857856899499893, -0.10017159581184387, 0.07978589832782745, 0.07483614236116409, -0.014357876032590866, 0.028417523950338364, -0.10843921452760696, 0.051863063126802444, 0.14290058612823486, 0.10864768177270889, 0.09554078429937363, -0.11520737409591675, 0.06326237320899963, -0.02823018841445446, 0.06830411404371262, -0.04200564697384834, 0.08296621590852737, 0.04463620111346245, 0.044546496123075485, 0.13018985092639923, 0.024470621719956398, -0.19115090370178223, 0.00866314209997654, 0.038987815380096436, -0.11250537633895874, -0.15485775470733643, -0.012188976630568504, 0.01697447896003723, -0.144207164645195, -0.06209023296833038, 0.11877616494894028, -0.09773363173007965, -0.02378847263753414, -0.06136301904916763, 0.07287470251321793, -0.006168958730995655, 0.11332488805055618, 0.0769299641251564, 0.010792993009090424, -0.033668018877506256, 0.11311119794845581, 0.10772878676652908, -0.023058723658323288, 0.017621925100684166, 0.103319451212883, -0.12781469523906708, -0.04093518480658531, 0.013308694586157799, 0.17064911127090454, -0.031489789485931396, -0.08550266176462173, 0.010269268415868282, -0.0732814148068428, 0.03020956926047802, 0.2873147130012512, -0.012263068929314613, -0.010687115602195263, -0.04333498328924179, -0.005429246928542852, -0.1097390204668045, 0.047697339206933975, -0.024259664118289948, 0.03277336433529854, -0.00228871894069016, 0.12751010060310364, 0.03488048538565636, 0.0032585631124675274, -0.05147862434387207, -0.045694150030612946, -0.07938101887702942, -0.0007907294202595949, -0.07521766424179077, 0.0083607267588377, -0.05946226045489311, -0.06398316472768784, -0.021153632551431656, 0.014610878191888332, 0.03417413309216499, 0.037653207778930664, -0.05514027923345566, -0.06387418508529663, -0.0534047931432724, 0.08635072410106659, -0.104946568608284, -0.024999504908919334, 0.026352720335125923, -0.06258925050497055, 0.07679574936628342, 0.022163791581988335, 0.03532474488019943, 0.004434044472873211, 0.013533213175833225, 0.028438910841941833, -0.044944312423467636, 0.009288948960602283, -0.019555632025003433, -0.13018125295639038, -0.04048657417297363, -0.09097623825073242, -0.08578509837388992, 0.012447060085833073, 0.028310224413871765, -0.022109124809503555, 0.041516974568367004, 0.006526718381792307, 0.07814847677946091, -0.0106589924544096, 0.0719728097319603, 0.04976297914981842, 0.06030280888080597, 0.0757729634642601, -0.052615225315093994, 0.05220349133014679, -0.16826477646827698, -0.01869014836847782, -0.011319675482809544, 0.031503405421972275, -0.0016281139105558395, -0.004496106877923012, 0.055898621678352356, 0.019526282325387, 0.1037278026342392, 0.015969248488545418, 0.05092817172408104, -0.0072476621717214584, -0.04169609025120735, 0.02931950055062771, -0.02954006753861904, 0.07024990022182465, 0.06208176538348198, -0.01353778038173914, 0.07201679050922394, 0.03634333238005638, -0.03969399258494377, 0.11280103772878647, 0.12077973037958145, 0.027768276631832123, 0.17382313311100006, 0.02325740084052086, -0.0022627986036241055, -0.05406462773680687, -0.041376009583473206, -0.13184191286563873, 0.07027874886989594, 0.06977260857820511, 0.0015727955615147948, 0.0796724185347557, 0.091609425842762, -0.16743120551109314, 0.17163099348545074, 0.04555178061127663, -0.08702503144741058, -0.0664086788892746, -0.17916418612003326, -0.0180982518941164, -0.0658874437212944, -0.018318844959139824, -0.1344824731349945, -0.004736015107482672, 0.23141685128211975, -0.016429012641310692, 0.02095138654112816, 0.04612371698021889, -0.11564882099628448, -0.13097380101680756, 0.07957284152507782, 0.010798080824315548, 0.013363288715481758, -0.06348557025194168, 0.05876903235912323, 0.061607252806425095, 0.04946037009358406, 0.03263687714934349, 0.05961637198925018, 0.0638686940073967, 0.028516510501503944, -0.0000816223764559254, -0.04278930649161339, 0.0031329705379903316, -0.007779599167406559, -0.014379450120031834, 0.2180221974849701, 0.05310342460870743, -0.023029213771224022, 0.0006738431402482092, 0.19974569976329803, -0.0627751350402832, 0.001479700906202197, -0.14140281081199646, 0.285643607378006, 0.02874072454869747, -0.011257763020694256, -0.022850453853607178, -0.10173040628433228, -0.007690343074500561, 0.2491970956325531, 0.05874577537178993, 0.007352062501013279, -0.013507021591067314, -0.039145778864622116, -0.00336787779815495, 0.05735045671463013, 0.07761713862419128, 0.00730143254622817, 0.1860969364643097, -0.0791102796792984, 0.1288548707962036, -0.07933977246284485, -0.011456173844635487, -0.09948314726352692, 0.06998540461063385, -0.034847237169742584, -0.023341134190559387, -0.04567922279238701, 0.10097071528434753, -0.09982161223888397, -0.25055545568466187, 0.0066065252758562565, -0.02080710604786873, -0.0699877142906189, -0.03633860498666763, -0.15775062143802643, 0.07641047239303589, 0.03961900621652603, 0.009179677814245224, 0.08007105439901352, 0.16303613781929016, 0.06013350561261177, 0.027892109006643295, -0.14789946377277374, 0.06910606473684311, -0.10954555124044418, 0.258358895778656, 0.010404652915894985, 0.004275194369256496, 0.05665592476725578, -0.03424082696437836, -0.11772129684686661, 0.05173727497458458, 0.027354545891284943, 0.05828693509101868, -0.0024237572215497494, 0.19032509624958038, -0.02321215346455574, 0.13119223713874817, 0.012200595811009407, -0.0574658066034317, 0.06812827289104462, -0.029567541554570198, -0.07447396963834763, -0.03271806985139847, 0.05477922037243843, -0.08938704431056976, 0.11554192006587982, 0.16245220601558685, -0.08159726858139038, -0.016407953575253487, -0.044470228254795074, 0.005118480417877436, -0.002644595690071583, 0.05763623118400574, -0.0526757538318634, -0.08004289120435715, 0.016442440450191498, -0.09890700876712799, -0.0070437039248645306, -0.29775190353393555, -0.013063786551356316, 0.0060995640233159065, -0.015924198552966118, 0.02456398867070675, 0.07617245614528656, -0.0017113048816099763, -0.016144322231411934, -0.037635937333106995, -0.014578529633581638, 0.013219890184700489, 0.0966174378991127, -0.12027067691087723, -0.04590489715337753 ]
null
null
transformers
# HIV_PR_resist model ## Table of Contents - [Summary](#model-summary) - [Model Description](#model-description) - [Intended Uses & Limitations](#intended-uses-&-limitations) - [How to Use](#how-to-use) - [Training Data](#training-data) - [Training Procedure](#training-procedure) - [Preprocessing](#preprocessing) - [Training](#training) - [Evaluation Results](#evaluation-results) - [BibTeX Entry and Citation Info](#bibtex-entry-and-citation-info) ## Summary The HIV-BERT-Protease-Resistance model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict whether an HIV protease sequence will be resistant to certain protease inhibitors. HIV-BERT is a model refined from the [ProtBert-BFD model](https://huggingface.co/Rostlab/prot_bert_bfd) to better fulfill HIV-centric tasks. This model was then trained using HIV protease sequences from the [Stanford HIV Genotype-Phenotype Database](https://hivdb.stanford.edu/pages/genotype-phenotype.html), allowing even more precise prediction protease inhibitor resistance than the HIV-BERT model can provide. ## Model Description The HIV-BERT-Protease-Resistance model is intended to predict the likelihood that an HIV protease sequence will be resistant to protease inhibitors. The protease gene is responsible for cleaving viral proteins into their active states, and as such is an ideal target for antiretroviral therapy. Annotation programs designed to predict and identify protease resistance using known mutations already exist, however with varied results. The HIV-BERT-Protease-Resistance model is designed to provide an alternative, NLP-based mechanism for predicting resistance mutations when provided with an HIV protease sequence. ## Intended Uses & Limitations This tool can be used as a predictor of protease resistance mutations within an HIV genomic sequence. It should not be considered a clinical diagnostic tool. ## How to use *Prediction example of protease sequences* ## Training Data This model was trained using the [damlab/HIV-PI dataset](https://huggingface.co/datasets/damlab/HIV_PI) using the 0th fold. The dataset consists of 1959 sequences (approximately 99 tokens each) extracted from the Stanford HIV Genotype-Phenotype Database. ## Training Procedure ### Preprocessing As with the [rostlab/Prot-bert-bfd model](https://huggingface.co/Rostlab/prot_bert_bfd), the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation. ### Training The [damlab/HIV-BERT model](https://huggingface.co/damlab/HIV_BERT) was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be resistant to multiple drugs) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance. ## Evaluation Results *Need to add* ## BibTeX Entry and Citation Info [More Information Needed]
{"license": "mit"}
text-classification
damlab/HIV_PR_resist
[ "transformers", "pytorch", "bert", "text-classification", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #bert #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us
# HIV_PR_resist model ## Table of Contents - Summary - Model Description - Intended Uses & Limitations - How to Use - Training Data - Training Procedure - Preprocessing - Training - Evaluation Results - BibTeX Entry and Citation Info ## Summary The HIV-BERT-Protease-Resistance model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict whether an HIV protease sequence will be resistant to certain protease inhibitors. HIV-BERT is a model refined from the ProtBert-BFD model to better fulfill HIV-centric tasks. This model was then trained using HIV protease sequences from the Stanford HIV Genotype-Phenotype Database, allowing even more precise prediction protease inhibitor resistance than the HIV-BERT model can provide. ## Model Description The HIV-BERT-Protease-Resistance model is intended to predict the likelihood that an HIV protease sequence will be resistant to protease inhibitors. The protease gene is responsible for cleaving viral proteins into their active states, and as such is an ideal target for antiretroviral therapy. Annotation programs designed to predict and identify protease resistance using known mutations already exist, however with varied results. The HIV-BERT-Protease-Resistance model is designed to provide an alternative, NLP-based mechanism for predicting resistance mutations when provided with an HIV protease sequence. ## Intended Uses & Limitations This tool can be used as a predictor of protease resistance mutations within an HIV genomic sequence. It should not be considered a clinical diagnostic tool. ## How to use *Prediction example of protease sequences* ## Training Data This model was trained using the damlab/HIV-PI dataset using the 0th fold. The dataset consists of 1959 sequences (approximately 99 tokens each) extracted from the Stanford HIV Genotype-Phenotype Database. ## Training Procedure ### Preprocessing As with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation. ### Training The damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be resistant to multiple drugs) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance. ## Evaluation Results *Need to add* ## BibTeX Entry and Citation Info
[ "# HIV_PR_resist model", "## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info", "## Summary\r\n\r\nThe HIV-BERT-Protease-Resistance model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict whether an HIV protease sequence will be resistant to certain protease inhibitors. HIV-BERT is a model refined from the ProtBert-BFD model to better fulfill HIV-centric tasks. This model was then trained using HIV protease sequences from the Stanford HIV Genotype-Phenotype Database, allowing even more precise prediction protease inhibitor resistance than the HIV-BERT model can provide.", "## Model Description\r\n\r\nThe HIV-BERT-Protease-Resistance model is intended to predict the likelihood that an HIV protease sequence will be resistant to protease inhibitors. The protease gene is responsible for cleaving viral proteins into their active states, and as such is an ideal target for antiretroviral therapy. Annotation programs designed to predict and identify protease resistance using known mutations already exist, however with varied results. The HIV-BERT-Protease-Resistance model is designed to provide an alternative, NLP-based mechanism for predicting resistance mutations when provided with an HIV protease sequence.", "## Intended Uses & Limitations\r\n\r\nThis tool can be used as a predictor of protease resistance mutations within an HIV genomic sequence. It should not be considered a clinical diagnostic tool.", "## How to use\r\n\r\n*Prediction example of protease sequences*", "## Training Data\r\n\r\nThis model was trained using the damlab/HIV-PI dataset using the 0th fold. The dataset consists of 1959 sequences (approximately 99 tokens each) extracted from the Stanford HIV Genotype-Phenotype Database.", "## Training Procedure", "### Preprocessing\r\n\r\nAs with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.", "### Training\r\n\r\nThe damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be resistant to multiple drugs) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.", "## Evaluation Results\r\n\r\n*Need to add*", "## BibTeX Entry and Citation Info" ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us \n", "# HIV_PR_resist model", "## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info", "## Summary\r\n\r\nThe HIV-BERT-Protease-Resistance model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict whether an HIV protease sequence will be resistant to certain protease inhibitors. HIV-BERT is a model refined from the ProtBert-BFD model to better fulfill HIV-centric tasks. This model was then trained using HIV protease sequences from the Stanford HIV Genotype-Phenotype Database, allowing even more precise prediction protease inhibitor resistance than the HIV-BERT model can provide.", "## Model Description\r\n\r\nThe HIV-BERT-Protease-Resistance model is intended to predict the likelihood that an HIV protease sequence will be resistant to protease inhibitors. The protease gene is responsible for cleaving viral proteins into their active states, and as such is an ideal target for antiretroviral therapy. Annotation programs designed to predict and identify protease resistance using known mutations already exist, however with varied results. The HIV-BERT-Protease-Resistance model is designed to provide an alternative, NLP-based mechanism for predicting resistance mutations when provided with an HIV protease sequence.", "## Intended Uses & Limitations\r\n\r\nThis tool can be used as a predictor of protease resistance mutations within an HIV genomic sequence. It should not be considered a clinical diagnostic tool.", "## How to use\r\n\r\n*Prediction example of protease sequences*", "## Training Data\r\n\r\nThis model was trained using the damlab/HIV-PI dataset using the 0th fold. The dataset consists of 1959 sequences (approximately 99 tokens each) extracted from the Stanford HIV Genotype-Phenotype Database.", "## Training Procedure", "### Preprocessing\r\n\r\nAs with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.", "### Training\r\n\r\nThe damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be resistant to multiple drugs) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.", "## Evaluation Results\r\n\r\n*Need to add*", "## BibTeX Entry and Citation Info" ]
[ 41, 8, 52, 140, 145, 45, 16, 60, 4, 80, 147, 10, 11 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# HIV_PR_resist model## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info## Summary\r\n\r\nThe HIV-BERT-Protease-Resistance model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict whether an HIV protease sequence will be resistant to certain protease inhibitors. HIV-BERT is a model refined from the ProtBert-BFD model to better fulfill HIV-centric tasks. This model was then trained using HIV protease sequences from the Stanford HIV Genotype-Phenotype Database, allowing even more precise prediction protease inhibitor resistance than the HIV-BERT model can provide.## Model Description\r\n\r\nThe HIV-BERT-Protease-Resistance model is intended to predict the likelihood that an HIV protease sequence will be resistant to protease inhibitors. The protease gene is responsible for cleaving viral proteins into their active states, and as such is an ideal target for antiretroviral therapy. Annotation programs designed to predict and identify protease resistance using known mutations already exist, however with varied results. The HIV-BERT-Protease-Resistance model is designed to provide an alternative, NLP-based mechanism for predicting resistance mutations when provided with an HIV protease sequence.## Intended Uses & Limitations\r\n\r\nThis tool can be used as a predictor of protease resistance mutations within an HIV genomic sequence. It should not be considered a clinical diagnostic tool.## How to use\r\n\r\n*Prediction example of protease sequences*## Training Data\r\n\r\nThis model was trained using the damlab/HIV-PI dataset using the 0th fold. The dataset consists of 1959 sequences (approximately 99 tokens each) extracted from the Stanford HIV Genotype-Phenotype Database." ]
[ 0.03759939223527908, 0.1811004877090454, -0.005275247152894735, -0.0016653244383633137, 0.06647336483001709, -0.02797546423971653, 0.07238733768463135, 0.09313631057739258, 0.08026246726512909, 0.09454163908958435, -0.022272001951932907, -0.053498901426792145, 0.07700856775045395, 0.14030224084854126, 0.11999727785587311, -0.1583612561225891, 0.021882042288780212, -0.07187645137310028, 0.08002239465713501, 0.05448950082063675, 0.05009039491415024, -0.0752778947353363, 0.0388336181640625, -0.01012490689754486, 0.17756116390228271, 0.08721281588077545, -0.032658182084560394, -0.0684254914522171, 0.06963898986577988, 0.029028626158833504, 0.0370267853140831, 0.0573788620531559, 0.0966794416308403, -0.31986719369888306, -0.01749889738857746, 0.05986238643527031, 0.023922357708215714, 0.01625056378543377, 0.05478396266698837, -0.01305504608899355, -0.058195777237415314, -0.10853391885757446, 0.0759270042181015, 0.02460586652159691, -0.09214690327644348, -0.049027614295482635, -0.09103091806173325, 0.061085738241672516, -0.02033831551671028, 0.10036751627922058, 0.008781551383435726, 0.06017056852579117, 0.0696747675538063, -0.02498628944158554, 0.12560752034187317, -0.19922679662704468, 0.024764124304056168, 0.07729950547218323, 0.033481746912002563, 0.07083631306886673, -0.05794446915388107, -0.011227642185986042, -0.04463041201233864, 0.014577838592231274, 0.19673016667366028, -0.03273582458496094, 0.09773971885442734, -0.05113109201192856, -0.1341041475534439, -0.11683572828769684, 0.12823784351348877, -0.05819873511791229, -0.06788526475429535, -0.17247320711612701, -0.0010948420967906713, 0.08283565938472748, -0.0028273488860577345, -0.07561152428388596, 0.05030634254217148, 0.03513815999031067, -0.03077141009271145, -0.13899745047092438, -0.03295093774795532, -0.009500078856945038, -0.041425447911024094, 0.06321065127849579, 0.03332485258579254, -0.010962124913930893, -0.009678233414888382, 0.10489029437303543, -0.13822005689144135, -0.04670137166976929, -0.06343226134777069, -0.06876541674137115, -0.06350399553775787, -0.045724645256996155, 0.017421245574951172, -0.15425941348075867, -0.001517034717835486, 0.14616365730762482, 0.0031839213334023952, 0.028689850121736526, -0.009004035964608192, -0.03304310142993927, 0.08482928574085236, 0.17576980590820312, -0.06782957911491394, -0.08227598667144775, -0.005507292225956917, -0.009776064194738865, 0.07551927864551544, 0.00890791229903698, 0.03580969572067261, -0.03999567776918411, 0.019512062892317772, 0.08640681207180023, -0.03331591188907623, 0.021170027554035187, -0.0037219570949673653, 0.0006380330305546522, 0.031259745359420776, -0.10808227956295013, -0.009877776727080345, -0.06388071179389954, 0.009706777520477772, 0.042752064764499664, 0.03746505081653595, -0.04442311078310013, -0.07136335968971252, 0.0412011593580246, -0.022126296535134315, -0.060908928513526917, -0.10125567018985748, -0.06278789043426514, -0.005686821881681681, 0.06507955491542816, -0.07500436902046204, -0.05149383470416069, -0.08016815036535263, -0.03620055317878723, -0.06018335744738579, -0.020619908347725868, -0.0057435669004917145, 0.029504727572202682, 0.04645809158682823, -0.014163045212626457, -0.05207303166389465, -0.13491271436214447, -0.008516674861311913, -0.023376157507300377, -0.03553251549601555, 0.08672550320625305, 0.14190787076950073, 0.0015477915294468403, -0.12378039211034775, 0.07251696288585663, -0.312050998210907, 0.08705675601959229, -0.13963238894939423, -0.03205489367246628, -0.13100355863571167, -0.02807345613837242, -0.007766786962747574, -0.06336680054664612, 0.03532043844461441, 0.01419304870069027, -0.1384613811969757, 0.00836973637342453, 0.27244964241981506, -0.09564736485481262, 0.006154497154057026, -0.0070581091567873955, -0.03632959723472595, 0.09271364659070969, 0.08387318253517151, 0.0694236159324646, 0.06896570324897766, -0.07470788061618805, -0.13361191749572754, -0.040645260363817215, -0.043732449412345886, 0.19100266695022583, 0.018884822726249695, -0.0801161378622055, 0.04775583744049072, 0.04335174709558487, -0.060410164296627045, -0.06926874816417694, 0.016057610511779785, -0.008313636295497417, -0.017341557890176773, -0.0050749084912240505, -0.03172372654080391, -0.07324813306331635, -0.09037307649850845, 0.03587707504630089, -0.09270653873682022, 0.04492701590061188, 0.10338139533996582, -0.097435861825943, 0.05097056180238724, -0.09203905612230301, -0.03374597057700157, 0.00649010855704546, -0.03952035307884216, -0.14833825826644897, -0.0851527601480484, 0.0947282686829567, -0.19074124097824097, 0.07122239470481873, -0.10220426321029663, 0.026657545939087868, 0.030098918825387955, -0.029873989522457123, 0.006681392900645733, -0.05475631728768349, 0.000668159918859601, -0.044532015919685364, -0.035881832242012024, -0.0659172385931015, -0.006512963213026524, 0.12988300621509552, 0.00792597234249115, -0.0030877566896378994, -0.10502265393733978, 0.10357020050287247, 0.04787364602088928, -0.014810997992753983, 0.08366970717906952, 0.02277916669845581, -0.012709617614746094, -0.014826282858848572, -0.008842221461236477, -0.05102645605802536, 0.0006726623978465796, 0.03935408592224121, -0.07918965816497803, -0.2221839725971222, 0.016947265714406967, 0.11526995152235031, -0.057258911430835724, -0.03404441475868225, 0.016513053327798843, 0.0074379644356667995, -0.10275450348854065, 0.01089879684150219, 0.18504828214645386, 0.06840232014656067, 0.07905574887990952, -0.027851048856973648, -0.05271309241652489, -0.01441711001098156, 0.0139690563082695, -0.03298703208565712, -0.003460625885054469, 0.124714195728302, -0.06294621527194977, 0.03458625078201294, 0.012200135737657547, 0.09262026101350784, 0.02443649247288704, 0.030827391892671585, -0.13604114949703217, -0.012728985399007797, -0.07606378197669983, 0.029514286667108536, 0.012785453349351883, 0.06849613785743713, 0.046066246926784515, 0.020651958882808685, -0.04153621941804886, -0.028905903920531273, -0.015369297005236149, 0.064766526222229, 0.04846344143152237, -0.04698481410741806, -0.033767785876989365, -0.08754422515630722, 0.023052990436553955, 0.0541781410574913, 0.024164851754903793, 0.08100020885467529, -0.07884149253368378, -0.008584176190197468, -0.0635836273431778, 0.039979238063097, -0.07380315661430359, -0.2706591486930847, -0.1736638844013214, 0.02013574168086052, -0.002409926615655422, 0.005408197175711393, -0.03597337007522583, 0.029153980314731598, -0.07886345684528351, -0.1492493450641632, 0.08341369032859802, -0.013178545981645584, -0.12553085386753082, -0.02026027999818325, 0.10003422945737839, 0.08914697170257568, -0.06733259558677673, 0.009895531460642815, -0.005878643598407507, -0.07189464569091797, 0.021841909736394882, 0.05086406320333481, 0.10593097656965256, -0.01905815303325653, 0.03232419863343239, -0.0713084414601326, -0.05403566360473633, 0.11207576096057892, -0.04962191358208656, 0.08423410356044769, 0.1685478538274765, -0.04984462633728981, 0.06545472145080566, 0.07483260333538055, -0.030343621969223022, -0.00303398328833282, 0.06539268791675568, 0.063563272356987, 0.008387415669858456, -0.1520681381225586, -0.0407644584774971, 0.026288751512765884, -0.0022619592491537333, 0.08520010113716125, 0.020384911447763443, 0.016820218414068222, 0.04035505652427673, -0.08212796598672867, 0.028108257800340652, 0.026777375489473343, 0.07921186089515686, 0.18859553337097168, -0.03249955177307129, 0.07380087673664093, -0.04385247826576233, -0.07986612617969513, 0.08481666445732117, 0.06506460160017014, 0.1401170790195465, -0.011412478052079678, 0.1483108103275299, 0.021862756460905075, 0.12464991211891174, 0.042023781687021255, 0.0314248725771904, 0.02337219938635826, 0.03774510696530342, -0.006389522459357977, -0.07185009121894836, -0.052570175379514694, 0.008461512625217438, -0.003190064337104559, -0.0074364920146763325, 0.07872216403484344, -0.05678480491042137, -0.0061725773848593235, 0.16314932703971863, 0.014008614234626293, -0.02947455830872059, -0.026647068560123444, 0.05906763672828674, -0.11058709025382996, -0.09346568584442139, 0.035319019109010696, 0.06496059894561768, -0.16121035814285278, 0.07342901825904846, 0.032305583357810974, 0.09486091136932373, -0.15488353371620178, -0.02228248491883278, 0.0025402952451258898, 0.057733092457056046, -0.03541799634695053, 0.11765548586845398, -0.15791404247283936, -0.06339266896247864, 0.004578631836920977, 0.05556744337081909, -0.05548364669084549, 0.04722217842936516, -0.0370306521654129, 0.021431710571050644, 0.1475292444229126, 0.0033288474660366774, 0.031824659556150436, -0.06494658440351486, -0.19097405672073364, -0.019413381814956665, 0.06244737654924393, -0.16081535816192627, 0.17074409127235413, -0.05187441408634186, -0.004669396672397852, -0.06761005520820618, 0.026811186224222183, -0.15285727381706238, -0.20637941360473633, 0.04397650435566902, -0.056215040385723114, 0.1372886300086975, -0.03868239372968674, -0.008260798640549183, 0.005108225625008345, 0.06520271301269531, -0.15427598357200623, -0.08898352086544037, -0.12270700931549072, -0.031082777306437492, 0.17381176352500916, -0.05810571834445, 0.02394566312432289, 0.034921564161777496, 0.20335090160369873, -0.04102395474910736, -0.15232068300247192, -0.02884870208799839, -0.03996500372886658, -0.12436136603355408, -0.06956590712070465, 0.11134812980890274, 0.03589320927858353, 0.10841906070709229, 0.02668655477464199, 0.08365277945995331, 0.0629497766494751, -0.057779014110565186, 0.03994737192988396, 0.22221556305885315, 0.009785555303096771, 0.01050223782658577, -0.1030258983373642, -0.03467562794685364, -0.12832535803318024, -0.06445877254009247, 0.1747114062309265, 0.13407352566719055, -0.05581485852599144, 0.1269439458847046, 0.06776411831378937, -0.13774912059307098, -0.20570212602615356, -0.08163122832775116, 0.0879523754119873, -0.02847713977098465, 0.09350083768367767, -0.3371870815753937, 0.03778991848230362, 0.06769061088562012, 0.0006754352943971753, -0.07756920903921127, -0.1849459707736969, -0.12298566848039627, 0.018717538565397263, 0.006286133546382189, -0.11018836498260498, -0.03562384471297264, -0.03705140948295593, -0.05089316889643669, -0.032666660845279694, 0.1327630877494812, 0.022651463747024536, 0.010037020780146122, -0.005573086440563202, 0.10131679475307465, 0.035189077258110046, 0.0034852323587983847, 0.02730105072259903, -0.05083545297384262, 0.03323967754840851, -0.044979095458984375, 0.09806713461875916, 0.09106766432523727, -0.028752246871590614, 0.0945800319314003, 0.13296785950660706, 0.0061467853374779224, -0.059247761964797974, -0.09029071033000946, -0.005557273980230093, -0.025198932737112045, -0.03789623826742172, -0.07522541284561157, -0.08176995813846588, 0.06209201365709305, 0.04889515042304993, -0.07742760330438614, 0.03534708917140961, -0.06802807748317719, -0.05819077789783478, 0.11488523334264755, 0.1370917111635208, 0.16688239574432373, -0.0789678692817688, 0.03504548594355583, -0.015266964212059975, -0.021080389618873596, 0.0298110730946064, 0.097935751080513, 0.043939635157585144, 0.015425147488713264, 0.05543963983654976, -0.00025826715864241123, -0.16025932133197784, -0.017884396016597748, 0.02569904364645481, -0.12419571727514267, -0.18786364793777466, 0.00536001892760396, 0.07122281193733215, -0.1181166023015976, -0.082621268928051, 0.1459631770849228, -0.06228185072541237, -0.06968487799167633, -0.03565364331007004, 0.06677500158548355, 0.08385327458381653, 0.0804404690861702, -0.0030789305455982685, -0.04232468456029892, -0.06873565912246704, 0.12887097895145416, 0.129071444272995, -0.05559881031513214, 0.004469600506126881, 0.06196723133325577, -0.09761014580726624, -0.017652496695518494, -0.07085452973842621, 0.07986007630825043, -0.09518279880285263, -0.048345476388931274, 0.03516027331352234, -0.058493807911872864, 0.047562241554260254, 0.19755050539970398, -0.03967907279729843, -0.01644454523921013, -0.017222687602043152, -0.010730108246207237, -0.09492789208889008, 0.06691862642765045, -0.08225513994693756, 0.06157761067152023, -0.005548554938286543, 0.08087248355150223, 0.032446905970573425, -0.0357397124171257, -0.030051447451114655, -0.06825295090675354, -0.040125392377376556, -0.03623053804039955, -0.08979237079620361, -0.006579010747373104, -0.046404413878917694, -0.07661603391170502, 0.024807341396808624, 0.028376556932926178, 0.027941405773162842, 0.02333184890449047, -0.011040539480745792, -0.06538629531860352, -0.05979723855853081, 0.09290378540754318, -0.1571292281150818, 0.03457926958799362, 0.08653062582015991, -0.07237344980239868, 0.08966808021068573, 0.03050336241722107, 0.07629293948411942, -0.031047234311699867, 0.08682555705308914, -0.017902277410030365, -0.04124719649553299, 0.09740127623081207, -0.00970284640789032, -0.13471093773841858, -0.008250370621681213, -0.028333835303783417, -0.14069518446922302, -0.017407046630978584, 0.04344906657934189, -0.05070403218269348, 0.006401121616363525, 0.06992943584918976, 0.03042948991060257, -0.039441049098968506, 0.02695251628756523, 0.10999053716659546, -0.0154475849121809, 0.07093559950590134, -0.02988836169242859, 0.029247445985674858, -0.1580517590045929, -0.024078261107206345, 0.001577793387696147, 0.030457347631454468, -0.018856598064303398, -0.05827171728014946, 0.046446919441223145, 0.04718555510044098, 0.19375720620155334, 0.027657831087708473, 0.09311176091432571, 0.01098061352968216, -0.05215401202440262, -0.01187965553253889, 0.0027061705477535725, 0.04091552272439003, 0.09907020628452301, -0.025152741000056267, 0.026153119280934334, -0.0030253189615905285, -0.060988135635852814, -0.03702419251203537, -0.01833326183259487, 0.01889619044959545, 0.12874072790145874, -0.06339117139577866, 0.005924062803387642, -0.02882041223347187, -0.05775367096066475, 0.004353818949311972, -0.018385134637355804, 0.003812011331319809, -0.00735698314383626, 0.1382600963115692, 0.035366982221603394, -0.12821809947490692, 0.16777533292770386, -0.011427806690335274, -0.05010339617729187, -0.07769612222909927, -0.1536966860294342, -0.04494483768939972, -0.0382576659321785, -0.00758164469152689, -0.14694911241531372, 0.025452380999922752, 0.152153879404068, -0.017203103750944138, -0.012809069827198982, 0.0057948557659983635, -0.10290820896625519, -0.1644412875175476, 0.010107362642884254, 0.044301606714725494, 0.049468718469142914, -0.047109536826610565, 0.0798170194029808, 0.043427012860774994, 0.0658397227525711, 0.050030417740345, 0.06626547873020172, 0.11816535890102386, -0.01583992876112461, -0.021196018904447556, -0.01810024119913578, -0.0047297696582973, -0.057024359703063965, -0.029013697057962418, 0.13414441049098969, 0.04826921597123146, 0.029293011873960495, 0.0027151750400662422, 0.31061697006225586, -0.04727781563997269, -0.017880911007523537, -0.09975673258304596, 0.24604973196983337, 0.06904648244380951, 0.061678461730480194, 0.01742108352482319, -0.0796351507306099, -0.01785106211900711, 0.1940334439277649, -0.03901670128107071, -0.010308751836419106, 0.0038644627202302217, -0.01250983402132988, 0.010589761659502983, 0.05745153874158859, 0.045966751873493195, 0.03238805755972862, 0.17104360461235046, -0.0903128832578659, 0.13837242126464844, -0.061429329216480255, 0.018846873193979263, -0.09426000714302063, 0.13641102612018585, -0.08204039931297302, 0.014797335490584373, -0.038325317203998566, 0.04486700892448425, -0.012459568679332733, -0.26624271273612976, 0.010427931323647499, -0.050739601254463196, -0.06692800670862198, 0.022182591259479523, 0.001386547926813364, -0.016342846676707268, -0.003273446811363101, 0.09379615634679794, 0.07814252376556396, 0.2623385488986969, 0.06187966838479042, -0.013552546501159668, 0.01925036683678627, 0.046031299978494644, -0.07766826450824738, 0.1698407232761383, 0.01158224232494831, 0.004491983447223902, 0.036291927099227905, -0.05160602927207947, -0.13909533619880676, 0.08550453186035156, 0.0442168191075325, -0.02016105130314827, 0.01336743589490652, 0.15764692425727844, 0.03137124329805374, 0.19532187283039093, 0.060700953006744385, -0.006959350313991308, 0.0877094715833664, 0.02735014446079731, -0.04875956475734711, -0.01390833593904972, 0.0586174838244915, -0.09777700901031494, 0.09242677688598633, 0.10537789016962051, -0.05395539104938507, 0.02350202202796936, -0.08035305142402649, 0.005768085364252329, -0.014955663122236729, 0.07118266820907593, 0.00983533076941967, 0.0033639161847531796, -0.019128110259771347, -0.08082140982151031, 0.025024354457855225, -0.26988399028778076, -0.0382113978266716, 0.07463887333869934, -0.011230136267840862, -0.002007109811529517, 0.09312419593334198, 0.030261240899562836, 0.026292182505130768, -0.024744225665926933, -0.021597329527139664, 0.053426846861839294, 0.03971724584698677, -0.07576552033424377, 0.018197868019342422 ]
null
null
transformers
# HIV_V3_coreceptor model ## Table of Contents - [Summary](#model-summary) - [Model Description](#model-description) - [Intended Uses & Limitations](#intended-uses-&-limitations) - [How to Use](#how-to-use) - [Training Data](#training-data) - [Training Procedure](#training-procedure) - [Preprocessing](#preprocessing) - [Training](#training) - [Evaluation Results](#evaluation-results) - [BibTeX Entry and Citation Info](#bibtex-entry-and-citation-info) ## Summary The HIV-BERT-Coreceptor model was trained as a refinement of the [HIV-BERT model](https://huggingface.co/damlab/HIV_BERT) and serves to better predict HIV V3 coreceptor tropism. HIV-BERT is a model refined from the [ProtBert-BFD model](https://huggingface.co/Rostlab/prot_bert_bfd) to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the [Los Alamos HIV Sequence Database](https://www.hiv.lanl.gov/content/sequence/HIV/mainpage.html), allowing even more precise prediction of V3 coreceptor tropism than the HIV-BERT model can provide. ## Model Description The HIV-BERT-Coreceptor model is intended to predict the Co-receptor tropism of HIV from a segment of the envelope protein. These envelope proteins encapsulate the virus and interact with the host cell through the human CD4 receptor. HIV then requires the interaction of one, of two, co-receptors: CCR5 or CXCR4. The availability of these co-receptors on different cell types allows the virus to invade different areas of the body and evade antiretroviral therapy. The 3rd variable loop of the envelope protein, the V3 loop, is responsible for this interaction. Given a V3 loop sequence, the HIV-BERT-Coreceptor model will predict the likelihood of binding to each of these co-receptors. ## Intended Uses & Limitations This tool can be used as a predictor of HIV tropism from the Env-V3 loop. It can recognize both R5, X4, and dual tropic viruses natively. It should not be considered a clinical diagnostic tool. This tool was trained using the [Los Alamos HIV sequence dataset](https://www.hiv.lanl.gov/content/sequence/HIV/mainpage.html). Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences. ## How to use *Need to add* ## Training Data This model was trained using the [damlab/HIV_V3_coreceptor dataset](https://huggingface.co/datasets/damlab/HIV_V3_coreceptor) using the 0th fold. The dataset consists of 2935 V3 sequences (approximately 35 tokens each) extracted from the [Los Alamos HIV Sequence database](https://www.hiv.lanl.gov/content/sequence/HIV/mainpage.html). ## Training Procedure ### Preprocessing As with the [rostlab/Prot-bert-bfd model](https://huggingface.co/Rostlab/prot_bert_bfd), the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation. ### Training The [damlab/HIV-BERT model](https://huggingface.co/damlab/HIV_BERT) was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can bind to CCR5, CXCR4, neither, or both) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance. ## Evaluation Results *Need to add* ## BibTeX Entry and Citation Info [More Information Needed]
{"license": "mit", "widget": [{"text": "C T R P N N N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C"}, {"text": "C T R P N N N T R K S I H I G P G R A F Y T T G Q I I G D I R Q A Y C"}, {"text": "C T R P N N N T R R S I R I G P G Q A F Y A T G D I I G D I R Q A H C"}, {"text": "C G R P N N H R I K G L R I G P G R A F F A M G A I G G G E I R Q A H C"}]}
text-classification
damlab/HIV_V3_Coreceptor
[ "transformers", "pytorch", "bert", "text-classification", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #bert #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us
# HIV_V3_coreceptor model ## Table of Contents - Summary - Model Description - Intended Uses & Limitations - How to Use - Training Data - Training Procedure - Preprocessing - Training - Evaluation Results - BibTeX Entry and Citation Info ## Summary The HIV-BERT-Coreceptor model was trained as a refinement of the HIV-BERT model and serves to better predict HIV V3 coreceptor tropism. HIV-BERT is a model refined from the ProtBert-BFD model to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database, allowing even more precise prediction of V3 coreceptor tropism than the HIV-BERT model can provide. ## Model Description The HIV-BERT-Coreceptor model is intended to predict the Co-receptor tropism of HIV from a segment of the envelope protein. These envelope proteins encapsulate the virus and interact with the host cell through the human CD4 receptor. HIV then requires the interaction of one, of two, co-receptors: CCR5 or CXCR4. The availability of these co-receptors on different cell types allows the virus to invade different areas of the body and evade antiretroviral therapy. The 3rd variable loop of the envelope protein, the V3 loop, is responsible for this interaction. Given a V3 loop sequence, the HIV-BERT-Coreceptor model will predict the likelihood of binding to each of these co-receptors. ## Intended Uses & Limitations This tool can be used as a predictor of HIV tropism from the Env-V3 loop. It can recognize both R5, X4, and dual tropic viruses natively. It should not be considered a clinical diagnostic tool. This tool was trained using the Los Alamos HIV sequence dataset. Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences. ## How to use *Need to add* ## Training Data This model was trained using the damlab/HIV_V3_coreceptor dataset using the 0th fold. The dataset consists of 2935 V3 sequences (approximately 35 tokens each) extracted from the Los Alamos HIV Sequence database. ## Training Procedure ### Preprocessing As with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation. ### Training The damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can bind to CCR5, CXCR4, neither, or both) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance. ## Evaluation Results *Need to add* ## BibTeX Entry and Citation Info
[ "# HIV_V3_coreceptor model", "## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info", "## Summary\r\n\r\nThe HIV-BERT-Coreceptor model was trained as a refinement of the HIV-BERT model and serves to better predict HIV V3 coreceptor tropism. HIV-BERT is a model refined from the ProtBert-BFD model to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database, allowing even more precise prediction of V3 coreceptor tropism than the HIV-BERT model can provide.", "## Model Description\r\n\r\nThe HIV-BERT-Coreceptor model is intended to predict the Co-receptor tropism of HIV from a segment of the envelope protein. These envelope proteins encapsulate the virus and interact with the host cell through the human CD4 receptor. HIV then requires the interaction of one, of two, co-receptors: CCR5 or CXCR4. The availability of these co-receptors on different cell types allows the virus to invade different areas of the body and evade antiretroviral therapy. The 3rd variable loop of the envelope protein, the V3 loop, is responsible for this interaction. Given a V3 loop sequence, the HIV-BERT-Coreceptor model will predict the likelihood of binding to each of these co-receptors.", "## Intended Uses & Limitations\r\n\r\nThis tool can be used as a predictor of HIV tropism from the Env-V3 loop. It can recognize both R5, X4, and dual tropic viruses natively. It should not be considered a clinical diagnostic tool. \r\n \r\nThis tool was trained using the Los Alamos HIV sequence dataset. Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences.", "## How to use\r\n\r\n*Need to add*", "## Training Data\r\n\r\nThis model was trained using the damlab/HIV_V3_coreceptor dataset using the 0th fold. The dataset consists of 2935 V3 sequences (approximately 35 tokens each) extracted from the Los Alamos HIV Sequence database.", "## Training Procedure", "### Preprocessing\r\n\r\nAs with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.", "### Training\r\n\r\nThe damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can bind to CCR5, CXCR4, neither, or both) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.", "## Evaluation Results\r\n\r\n*Need to add*", "## BibTeX Entry and Citation Info" ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us \n", "# HIV_V3_coreceptor model", "## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info", "## Summary\r\n\r\nThe HIV-BERT-Coreceptor model was trained as a refinement of the HIV-BERT model and serves to better predict HIV V3 coreceptor tropism. HIV-BERT is a model refined from the ProtBert-BFD model to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database, allowing even more precise prediction of V3 coreceptor tropism than the HIV-BERT model can provide.", "## Model Description\r\n\r\nThe HIV-BERT-Coreceptor model is intended to predict the Co-receptor tropism of HIV from a segment of the envelope protein. These envelope proteins encapsulate the virus and interact with the host cell through the human CD4 receptor. HIV then requires the interaction of one, of two, co-receptors: CCR5 or CXCR4. The availability of these co-receptors on different cell types allows the virus to invade different areas of the body and evade antiretroviral therapy. The 3rd variable loop of the envelope protein, the V3 loop, is responsible for this interaction. Given a V3 loop sequence, the HIV-BERT-Coreceptor model will predict the likelihood of binding to each of these co-receptors.", "## Intended Uses & Limitations\r\n\r\nThis tool can be used as a predictor of HIV tropism from the Env-V3 loop. It can recognize both R5, X4, and dual tropic viruses natively. It should not be considered a clinical diagnostic tool. \r\n \r\nThis tool was trained using the Los Alamos HIV sequence dataset. Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences.", "## How to use\r\n\r\n*Need to add*", "## Training Data\r\n\r\nThis model was trained using the damlab/HIV_V3_coreceptor dataset using the 0th fold. The dataset consists of 2935 V3 sequences (approximately 35 tokens each) extracted from the Los Alamos HIV Sequence database.", "## Training Procedure", "### Preprocessing\r\n\r\nAs with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.", "### Training\r\n\r\nThe damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can bind to CCR5, CXCR4, neither, or both) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.", "## Evaluation Results\r\n\r\n*Need to add*", "## BibTeX Entry and Citation Info" ]
[ 41, 10, 52, 122, 178, 166, 10, 67, 4, 80, 157, 10, 11 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# HIV_V3_coreceptor model## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info## Summary\r\n\r\nThe HIV-BERT-Coreceptor model was trained as a refinement of the HIV-BERT model and serves to better predict HIV V3 coreceptor tropism. HIV-BERT is a model refined from the ProtBert-BFD model to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database, allowing even more precise prediction of V3 coreceptor tropism than the HIV-BERT model can provide.## Model Description\r\n\r\nThe HIV-BERT-Coreceptor model is intended to predict the Co-receptor tropism of HIV from a segment of the envelope protein. These envelope proteins encapsulate the virus and interact with the host cell through the human CD4 receptor. HIV then requires the interaction of one, of two, co-receptors: CCR5 or CXCR4. The availability of these co-receptors on different cell types allows the virus to invade different areas of the body and evade antiretroviral therapy. The 3rd variable loop of the envelope protein, the V3 loop, is responsible for this interaction. Given a V3 loop sequence, the HIV-BERT-Coreceptor model will predict the likelihood of binding to each of these co-receptors." ]
[ 0.06304185092449188, 0.038277871906757355, -0.005255136173218489, 0.011456751264631748, 0.05832764133810997, 0.01965138129889965, 0.06130440533161163, 0.059777140617370605, 0.07188352942466736, 0.08445162326097488, 0.02141999453306198, -0.13655461370944977, 0.11485499143600464, 0.11875556409358978, 0.11328098922967911, -0.17541760206222534, 0.017055882140994072, -0.0301374364644289, -0.07306236773729324, 0.047906726598739624, 0.05424671620130539, -0.0775853618979454, 0.016127677634358406, 0.012562540359795094, 0.14036977291107178, 0.04261400178074837, -0.06231396272778511, -0.025280790403485298, 0.0826248973608017, 0.03791312128305435, 0.07303430885076523, 0.11607526242733002, 0.13070784509181976, -0.2738986015319824, -0.013374436646699905, 0.04605238884687424, -0.03385396674275398, 0.0027804195415228605, 0.06675755232572556, -0.045513514429330826, -0.024566659703850746, -0.06545241922140121, 0.11289360374212265, 0.02610521763563156, -0.078685462474823, -0.10580453276634216, -0.09225580096244812, 0.10832826793193817, 0.07367023825645447, 0.08334372937679291, -0.003148880787193775, 0.06594443321228027, 0.060694094747304916, -0.011583025567233562, 0.11060794442892075, -0.2684257924556732, 0.0417213998734951, 0.13195350766181946, 0.10475316643714905, 0.0325782373547554, -0.0344257615506649, -0.005889280699193478, -0.005764179863035679, 0.04434531554579735, 0.23558297753334045, -0.03120933659374714, 0.1304090917110443, -0.06521786749362946, -0.18202099204063416, -0.13945253193378448, 0.09942317008972168, 0.0029369089752435684, -0.0411800742149353, -0.018591908738017082, 0.011966670863330364, -0.0017762277275323868, 0.02636871114373207, -0.12912605702877045, 0.014965733513236046, 0.037421129643917084, -0.03313160315155983, -0.1453290730714798, -0.06925689429044724, -0.034851182252168655, -0.06483589112758636, -0.014762414619326591, 0.032990697771310806, -0.0030665576923638582, -0.04735502973198891, 0.1114123985171318, -0.13629299402236938, -0.0730876624584198, -0.03158801421523094, -0.06353745609521866, -0.11720213294029236, -0.0205636415630579, 0.03513287380337715, -0.12049135565757751, 0.015110020525753498, 0.07598813623189926, -0.06479258835315704, -0.0058450279757380486, -0.044877681881189346, -0.011358481831848621, 0.0758550763130188, 0.07122500985860825, -0.10512079298496246, -0.019925301894545555, -0.025464538484811783, -0.05369582399725914, 0.013921063393354416, -0.008144865743815899, -0.0165707990527153, 0.0012589811813086271, 0.001352486084215343, 0.09475207328796387, -0.020759284496307373, 0.010765650309622288, 0.056437619030475616, 0.013524743728339672, 0.02619141899049282, -0.050621725618839264, -0.004649277310818434, -0.05001932382583618, 0.04791996628046036, 0.07179481536149979, 0.07484856247901917, -0.011866617947816849, -0.048270974308252335, 0.02597421035170555, -0.014187011867761612, -0.08982949703931808, -0.02509119361639023, -0.1434115767478943, 0.04200616106390953, 0.001062069903127849, -0.07476682215929031, -0.0981450080871582, 0.039714351296424866, -0.05778617784380913, -0.04408406466245651, -0.030377229675650597, -0.010950113646686077, 0.003428337862715125, 0.05056367442011833, 0.012823468074202538, -0.030861124396324158, -0.03982600197196007, -0.005648248828947544, -0.04418274387717247, -0.11474456638097763, 0.11387906223535538, 0.14653600752353668, 0.016151832416653633, -0.019368644803762436, 0.08030670136213303, -0.15906256437301636, -0.02269372157752514, -0.141703262925148, -0.0316464826464653, -0.16859391331672668, -0.0683484673500061, 0.03348802402615547, -0.09503242373466492, 0.04929175600409508, 0.04025646671652794, -0.05260516703128815, -0.021593736484646797, 0.32618772983551025, -0.0729360282421112, 0.004837557673454285, -0.019868839532136917, 0.019788723438978195, 0.08223002403974533, 0.03054015524685383, -0.0018649415578693151, 0.14938129484653473, -0.09599261730909348, -0.07157807797193527, -0.02511782944202423, -0.05676668509840965, 0.1672838032245636, 0.07907973229885101, -0.08840636163949966, 0.034039195626974106, -0.006876127794384956, -0.02262250706553459, -0.02177908830344677, -0.027521761134266853, 0.021832099184393883, 0.004905417561531067, 0.016735803335905075, 0.08150217682123184, -0.10590943694114685, -0.07635287940502167, 0.0004211771010886878, -0.15232639014720917, -0.0482960119843483, 0.10598732531070709, -0.11164949089288712, 0.07850310951471329, -0.09995913505554199, -0.02309507131576538, -0.037483882158994675, 0.005992199759930372, -0.07179397344589233, -0.05110282078385353, 0.04660904407501221, -0.08148818463087082, 0.06149788945913315, 0.02868681214749813, 0.03719834238290787, 0.04128427430987358, -0.0259479321539402, 0.024950483813881874, -0.021803267300128937, 0.011342105455696583, -0.020574655383825302, -0.09605486690998077, -0.10301151126623154, -0.057718340307474136, 0.22426657378673553, -0.02834719605743885, -0.021654529497027397, -0.014526769518852234, 0.09868931025266647, 0.0021537637803703547, 0.04237409308552742, 0.09961269050836563, 0.03243880718946457, -0.005038838367909193, -0.010977767407894135, 0.001113676349632442, -0.038338176906108856, -0.018898451700806618, 0.03995632007718086, -0.07929535955190659, -0.28655126690864563, 0.009290214627981186, 0.07118827849626541, -0.08192501217126846, -0.0379057377576828, 0.0551086887717247, 0.007499190047383308, -0.06613204628229141, -0.0313890241086483, 0.20230650901794434, 0.08687224984169006, 0.06374077498912811, 0.028014304116368294, -0.03572671860456467, 0.05107852816581726, -0.004121359437704086, -0.048383209854364395, -0.0025889407843351364, 0.09656529128551483, -0.07325196266174316, -0.014419058337807655, 0.12971480190753937, 0.059261586517095566, -0.007417609449476004, 0.033718500286340714, -0.09270873665809631, -0.023544564843177795, -0.06898431479930878, 0.032838787883520126, -0.047804977744817734, -0.0572294183075428, 0.024990664795041084, 0.05409792438149452, -0.0029779369942843914, -0.0548243373632431, -0.015034960582852364, 0.056535571813583374, 0.09979942440986633, -0.04659482464194298, -0.06763483583927155, -0.07434526830911636, 0.0020132220815867186, 0.03781338036060333, 0.04328995570540428, -0.029249366372823715, -0.04640140011906624, 0.004631000570952892, -0.12200570106506348, 0.05236726254224777, -0.09863657504320145, -0.2765951454639435, -0.1120847538113594, -0.10583920031785965, 0.012647552415728569, 0.009440762922167778, 0.048752039670944214, -0.04113860800862312, -0.05399787798523903, -0.10969893634319305, 0.12205632776021957, -0.05777646601200104, -0.0347459577023983, 0.0496504046022892, 0.08351383358240128, 0.026896104216575623, -0.05024004355072975, 0.014179032295942307, 0.005243465304374695, -0.11105494946241379, 0.09598727524280548, -0.02685021422803402, 0.08549114316701889, 0.14944300055503845, 0.045886971056461334, -0.021046141162514687, -0.08322375267744064, 0.04890952259302139, -0.029875174164772034, 0.10896429419517517, 0.1566115915775299, -0.026005355641245842, 0.03198747709393501, 0.08636834472417831, -0.04137278348207474, 0.05372130125761032, 0.05555278807878494, 0.01089609507471323, -0.045836761593818665, -0.20544400811195374, -0.04724711924791336, 0.04146306589245796, -0.0031864899210631847, 0.058058883994817734, -0.0026699507143348455, 0.03713122755289078, 0.06226210296154022, -0.04082668945193291, -0.04384096711874008, 0.04495798051357269, 0.07372234016656876, 0.1654563695192337, -0.059835284948349, 0.0831209123134613, 0.0027721773367375135, -0.0257403701543808, 0.08603367954492569, 0.05738729238510132, 0.1873473823070526, 0.06160099804401398, 0.0646575540304184, 0.06981179863214493, 0.0785248801112175, 0.014019660651683807, 0.1237616315484047, 0.05058739706873894, 0.026031184941530228, -0.016380442306399345, -0.05235745385289192, 0.045278631150722504, -0.016949832439422607, 0.06565877050161362, -0.08839518576860428, -0.030551359057426453, -0.04619378224015236, -0.06370609253644943, 0.09487385302782059, 0.1242203339934349, -0.046852875500917435, -0.01088802795857191, 0.049867983907461166, -0.06811561435461044, -0.08465846627950668, 0.004262722097337246, 0.056823037564754486, -0.18137530982494354, 0.08067908138036728, 0.032502759248018265, 0.09726283699274063, -0.0886302962899208, -0.016787763684988022, -0.07661639153957367, -0.04813670739531517, -0.05851779133081436, 0.10212942212820053, -0.05801958218216896, 0.09469825774431229, -0.010536075569689274, -0.03665851801633835, -0.04785652086138725, 0.02855571359395981, -0.053140223026275635, 0.1541147232055664, 0.14416712522506714, 0.005398011766374111, 0.08278685063123703, -0.05517059937119484, -0.16197219491004944, -0.011432445608079433, 0.05011100694537163, -0.15613631904125214, 0.15725071728229523, -0.01887868344783783, -0.01784617453813553, -0.04338131472468376, 0.12509246170520782, -0.12660053372383118, -0.10954976081848145, 0.0802057534456253, -0.07003729045391083, 0.15521842241287231, -0.011015117168426514, -0.05835800990462303, 0.020828772336244583, 0.04451818764209747, -0.1413705199956894, -0.03828517347574234, -0.12314609438180923, 0.004024547524750233, 0.1496194303035736, -0.0934191644191742, 0.019255924969911575, 0.003828669199720025, 0.17677581310272217, -0.08922654390335083, -0.11051526665687561, -0.012163834646344185, -0.06383049488067627, -0.11407368630170822, -0.05333055555820465, 0.1633380651473999, -0.0009648537961766124, 0.1245366707444191, 0.052181169390678406, 0.03906460106372833, 0.04601620137691498, -0.040530189871788025, 0.08610653132200241, 0.17122265696525574, 0.021884877234697342, -0.0367249995470047, -0.04043189808726311, -0.016488738358020782, -0.10849010944366455, -0.07525812834501266, 0.25780677795410156, 0.20693102478981018, -0.07523921132087708, 0.1333887130022049, 0.05168937146663666, -0.13050861656665802, -0.18076229095458984, -0.12448012083768845, 0.10860200971364975, -0.019582344219088554, 0.07587889581918716, -0.2439144402742386, -0.05081578716635704, 0.014813202433288097, -0.0012409311020746827, -0.10266922414302826, -0.16276507079601288, -0.08745688199996948, 0.08941500633955002, -0.05698670074343681, -0.03694053366780281, -0.00953042134642601, -0.020108159631490707, -0.07911936938762665, -0.20620031654834747, 0.08701407164335251, -0.040902767330408096, 0.014856589958071709, 0.013640847988426685, 0.07488173246383667, 0.026014532893896103, 0.02450283244252205, 0.058709513396024704, -0.037056658416986465, -0.04650451987981796, -0.06184700131416321, 0.06162923946976662, 0.04463960602879524, -0.04459527134895325, 0.05470937490463257, 0.13293029367923737, -0.031214579939842224, -0.09644337743520737, -0.0720161497592926, -0.0038323646876960993, -0.03335823863744736, -0.03758377581834793, -0.07516448944807053, -0.0681750625371933, 0.008401933126151562, 0.09172461926937103, -0.0755753219127655, 0.02262652851641178, -0.0704009011387825, -0.042772326618433, 0.2065475583076477, 0.10661358386278152, 0.05373998358845711, -0.12913952767848969, 0.05661953613162041, -0.04890184476971626, -0.03763178735971451, -0.07610036432743073, 0.09916724264621735, 0.04724298417568207, -0.00048201228491961956, 0.04658383131027222, 0.029229160398244858, -0.15197037160396576, -0.0008786373073235154, 0.05023971199989319, -0.09994117170572281, -0.11050596833229065, -0.00620501721277833, 0.04906553775072098, -0.13080033659934998, -0.055501751601696014, 0.18776562809944153, -0.0773831233382225, -0.046732157468795776, -0.04193071648478508, 0.06726586073637009, -0.0018346133874729276, 0.10329953581094742, 0.011979230679571629, 0.00871298648416996, -0.038688432425260544, 0.06962879747152328, 0.14304883778095245, -0.026209479197859764, 0.02003627084195614, 0.042441099882125854, -0.10191446542739868, -0.08200469613075256, -0.02938578836619854, 0.048647332936525345, -0.13232110440731049, -0.07231497764587402, 0.11401233822107315, -0.08589986711740494, 0.03567283973097801, 0.2359609454870224, -0.0576138012111187, -0.030248748138546944, -0.0009561756742186844, -0.04373728111386299, -0.11285912245512009, 0.04005805775523186, -0.038870103657245636, 0.04827633500099182, -0.017275074496865273, 0.16283348202705383, 0.04089611396193504, 0.08429885655641556, -0.05325634777545929, -0.07728512585163116, -0.049083709716796875, -0.025054845958948135, -0.14265528321266174, -0.05127342790365219, -0.05496468394994736, -0.10414478182792664, 0.03820716589689255, 0.02523139864206314, 0.03242756798863411, 0.022152388468384743, -0.027517085894942284, -0.022041451185941696, -0.07923436164855957, 0.09213250130414963, -0.10168766230344772, -0.0007751139928586781, 0.08668641000986099, -0.04746590927243233, 0.10150681436061859, -0.027933571487665176, 0.08749864250421524, -0.014271017163991928, 0.05583091825246811, -0.016906317323446274, -0.03234570473432541, 0.056694645434617996, 0.03376379609107971, -0.18050982058048248, -0.03237343579530716, -0.08500757068395615, -0.17610931396484375, 0.005044750869274139, 0.052656397223472595, 0.020985621958971024, -0.0030218723695725203, -0.03244859352707863, 0.018944961950182915, 0.0019031489500775933, 0.007913932204246521, 0.04810767620801926, 0.01788301020860672, 0.04541456699371338, -0.022417431697249413, 0.10082297027111053, -0.15900923311710358, -0.030022811144590378, -0.03335358574986458, 0.041458435356616974, 0.06307192891836166, -0.07363659888505936, 0.045667655766010284, 0.045111119747161865, 0.05913393944501877, 0.026986001059412956, 0.175282284617424, -0.024613594636321068, -0.07280116528272629, 0.05333014577627182, -0.018102627247571945, -0.03421558812260628, 0.09990638494491577, -0.044693849980831146, -0.020567605271935463, 0.001525377156212926, -0.008814873173832893, -0.0517512746155262, 0.02796175889670849, -0.10379165410995483, 0.11819389462471008, -0.03892535716295242, 0.041308771818876266, -0.07702023535966873, -0.0940595269203186, -0.158336341381073, 0.08430630713701248, 0.045421820133924484, -0.020004330202937126, 0.12952034175395966, 0.05257087200880051, -0.13810716569423676, 0.11156881600618362, -0.006082148756831884, -0.050174612551927567, -0.07194142788648605, -0.1375245451927185, -0.047754235565662384, -0.16307185590267181, -0.0109166344627738, -0.140285462141037, 0.004929568152874708, 0.14871491491794586, 0.0275628250092268, 0.041758839040994644, -0.046950992196798325, -0.1348588466644287, -0.12898586690425873, -0.015132687985897064, 0.047615841031074524, 0.0745978131890297, -0.02981341816484928, 0.11998172104358673, 0.0399327352643013, 0.08236002922058105, 0.021400844678282738, 0.060576051473617554, 0.10936515778303146, 0.001344833755865693, -0.02010592445731163, -0.0479782335460186, -0.03850357607007027, -0.014355931431055069, -0.0148639976978302, 0.20363420248031616, 0.05810800939798355, 0.027991382405161858, -0.010875318199396133, 0.2232297956943512, -0.04765631631016731, -0.041434139013290405, -0.11074668914079666, 0.2421865165233612, 0.06861910223960876, 0.07472074031829834, -0.001283062039874494, -0.08484840393066406, -0.09068533033132553, 0.20498539507389069, 0.02784121036529541, 0.06517013162374496, -0.001685642870143056, 0.018674984574317932, 0.019500769674777985, 0.07384303212165833, 0.025525497272610664, 0.07930118590593338, 0.14074470102787018, -0.14401187002658844, 0.1033245176076889, -0.06155822053551674, 0.0347275547683239, -0.10232886672019958, 0.013716918416321278, -0.044855717569589615, -0.004370691254734993, 0.0756082609295845, 0.1079501286149025, 0.025035370141267776, -0.19509050250053406, -0.01826980896294117, -0.06230061128735542, -0.032054975628852844, 0.028220098465681076, -0.010632585734128952, 0.0244760625064373, 0.0378497838973999, 0.046843092888593674, 0.028017748147249222, 0.27924153208732605, 0.06676096469163895, 0.012742509134113789, -0.02761792205274105, 0.10039977729320526, -0.13041366636753082, 0.17278729379177094, 0.01243556383997202, 0.06207701936364174, 0.01920948550105095, -0.00760053563863039, -0.14095449447631836, 0.06774874776601791, 0.0627657100558281, 0.0063465009443461895, 0.037174418568611145, 0.12625809013843536, -0.018857285380363464, 0.08823982626199722, 0.04358869418501854, -0.09473904967308044, 0.0791286900639534, 0.03211073577404022, 0.051379963755607605, 0.07334025949239731, 0.05727555975317955, -0.09651822596788406, 0.08398137241601944, 0.11305423080921173, -0.08340760320425034, -0.03028222918510437, -0.07216130942106247, 0.0020380974747240543, -0.008609727956354618, 0.1337393969297409, 0.0031060189940035343, -0.027202390134334564, 0.0197328832000494, -0.08333741128444672, 0.03189343214035034, -0.23835958540439606, -0.05579927936196327, 0.03661378100514412, 0.014116104692220688, -0.000055254087783396244, 0.11352834105491638, 0.03810524195432663, 0.027585996314883232, -0.03415660560131073, 0.01794571615755558, 0.03528251126408577, 0.05738217011094093, -0.14821724593639374, 0.013086756691336632 ]
null
null
transformers
# Model Card for [HIV_V3_bodysite] ## Table of Contents - [Table of Contents](#table-of-contents) - [Summary](#model-summary) - [Model Description](#model-description) - [Intended Uses & Limitations](#intended-uses-&-limitations) - [How to Use](#how-to-use) - [Training Data](#training-data) - [Training Procedure](#training-procedure) - [Preprocessing](#preprocessing) - [Training](#training) - [Evaluation Results](#evaluation-results) - [BibTeX Entry and Citation Info](#bibtex-entry-and-citation-info) ## Summary The HIV-BERT-Bodysite-Identification model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict the location that an HIV V3 loop sample was derived from. HIV-BERT is a model refined from the ProtBert-BFD model (https://huggingface.co/Rostlab/prot_bert_bfd) to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database (https://www.hiv.lanl.gov/content/sequence/HIV/mainpage.html), allowing even more precise prediction of body site location than the HIV-BERT model can provide. ## Model Description The HIV-BERT-Bodysite-Identification model is intended to predict the location as to where an HIV sequence was most likely derived from. Because HIV infects immune cells, it uses these as a means of rapidly spreading throughout the body. Thus, body site identification can help determine where exactly these HIV particles ultimately end up. This would be helpful when attempting to study HIV treatment strategies. When provided with an HIV genomic sequence, the HIV-BERT-Bodysite-Identification model can predict which tissue it was derived from. ## Intended Uses & Limitations This tool can be used as a predictor of which body site an HIV sample was derived from based on its genomic sequence. It should not be considered a clinical diagnostic tool. This tool was trained using the Los Alamos HIV sequence dataset (https://www.hiv.lanl.gov/content/sequence/HIV/mainpage.html). Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences. ## How to use This model is able to predict the likely bodysite from a V3 sequence. This may be use for surveillance of cells that are emerging from latent reservoirs. Remember, a sequence can come from multiple sites, they are not mutually exclusive. ```python from transformers import pipeline predictor = pipeline("text-classification", model="damlab/HIV_V3_bodysite") predictor(f"C T R P N N N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C") [ [ { "label": "periphery-tcell", "score": 0.29097115993499756 }, { "label": "periphery-monocyte", "score": 0.014322502538561821 }, { "label": "CNS", "score": 0.06870711594820023 }, { "label": "breast-milk", "score": 0.002785981632769108 }, { "label": "female-genitals", "score": 0.024997007101774216 }, { "label": "male-genitals", "score": 0.01040483545511961 }, { "label": "gastric", "score": 0.06872137635946274 }, { "label": "lung", "score": 0.04432062804698944 }, { "label": "organ", "score": 0.47476938366889954 } ] ] ``` ## Training Data This model was trained using the damlab/HIV_V3_bodysite dataset using the 0th fold. The dataset consists of 5510 sequences (approximately 35 tokens each) extracted from the Los Alamos HIV Sequence database. ## Training Procedure ### Preprocessing As with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation. ### Training The damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be found in multiple sites) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance. ## Evaluation Results *Need to add* ## BibTeX Entry and Citation Info [More Information Needed]
{"datasets": ["damlab/HIV_V3_bodysite"], "metrics": ["accuracy"], "licence": "mit", "widget": [{"text": "T R P N N N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C", "example_title": "V3 Macrophage"}, {"text": "C T R P N N N T R K S I H I G P G R A F Y T T G Q I I G D I R Q A Y C", "example_title": "V3 T-cell"}]}
text-classification
damlab/HIV_V3_bodysite
[ "transformers", "pytorch", "bert", "text-classification", "dataset:damlab/HIV_V3_bodysite", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #bert #text-classification #dataset-damlab/HIV_V3_bodysite #autotrain_compatible #endpoints_compatible #region-us
# Model Card for [HIV_V3_bodysite] ## Table of Contents - Table of Contents - Summary - Model Description - Intended Uses & Limitations - How to Use - Training Data - Training Procedure - Preprocessing - Training - Evaluation Results - BibTeX Entry and Citation Info ## Summary The HIV-BERT-Bodysite-Identification model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict the location that an HIV V3 loop sample was derived from. HIV-BERT is a model refined from the ProtBert-BFD model (URL to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database (URL allowing even more precise prediction of body site location than the HIV-BERT model can provide. ## Model Description The HIV-BERT-Bodysite-Identification model is intended to predict the location as to where an HIV sequence was most likely derived from. Because HIV infects immune cells, it uses these as a means of rapidly spreading throughout the body. Thus, body site identification can help determine where exactly these HIV particles ultimately end up. This would be helpful when attempting to study HIV treatment strategies. When provided with an HIV genomic sequence, the HIV-BERT-Bodysite-Identification model can predict which tissue it was derived from. ## Intended Uses & Limitations This tool can be used as a predictor of which body site an HIV sample was derived from based on its genomic sequence. It should not be considered a clinical diagnostic tool. This tool was trained using the Los Alamos HIV sequence dataset (URL Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences. ## How to use This model is able to predict the likely bodysite from a V3 sequence. This may be use for surveillance of cells that are emerging from latent reservoirs. Remember, a sequence can come from multiple sites, they are not mutually exclusive. ## Training Data This model was trained using the damlab/HIV_V3_bodysite dataset using the 0th fold. The dataset consists of 5510 sequences (approximately 35 tokens each) extracted from the Los Alamos HIV Sequence database. ## Training Procedure ### Preprocessing As with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation. ### Training The damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be found in multiple sites) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance. ## Evaluation Results *Need to add* ## BibTeX Entry and Citation Info
[ "# Model Card for [HIV_V3_bodysite]", "## Table of Contents\r\n- Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info", "## Summary\r\n\r\nThe HIV-BERT-Bodysite-Identification model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict the location that an HIV V3 loop sample was derived from. HIV-BERT is a model refined from the ProtBert-BFD model (URL to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database (URL allowing even more precise prediction of body site location than the HIV-BERT model can provide.", "## Model Description\r\n\r\nThe HIV-BERT-Bodysite-Identification model is intended to predict the location as to where an HIV sequence was most likely derived from. Because HIV infects immune cells, it uses these as a means of rapidly spreading throughout the body. Thus, body site identification can help determine where exactly these HIV particles ultimately end up. This would be helpful when attempting to study HIV treatment strategies. When provided with an HIV genomic sequence, the HIV-BERT-Bodysite-Identification model can predict which tissue it was derived from.", "## Intended Uses & Limitations\r\n\r\nThis tool can be used as a predictor of which body site an HIV sample was derived from based on its genomic sequence. It should not be considered a clinical diagnostic tool. \r\n \r\nThis tool was trained using the Los Alamos HIV sequence dataset (URL Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences.", "## How to use\r\n\r\nThis model is able to predict the likely bodysite from a V3 sequence.\r\nThis may be use for surveillance of cells that are emerging from latent reservoirs.\r\nRemember, a sequence can come from multiple sites, they are not mutually exclusive.", "## Training Data\r\n\r\nThis model was trained using the damlab/HIV_V3_bodysite dataset using the 0th fold. The dataset consists of 5510 sequences (approximately 35 tokens each) extracted from the Los Alamos HIV Sequence database.", "## Training Procedure", "### Preprocessing\r\n\r\nAs with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.", "### Training\r\n\r\nThe damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be found in multiple sites) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.", "## Evaluation Results\r\n\r\n*Need to add*", "## BibTeX Entry and Citation Info" ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #dataset-damlab/HIV_V3_bodysite #autotrain_compatible #endpoints_compatible #region-us \n", "# Model Card for [HIV_V3_bodysite]", "## Table of Contents\r\n- Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info", "## Summary\r\n\r\nThe HIV-BERT-Bodysite-Identification model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict the location that an HIV V3 loop sample was derived from. HIV-BERT is a model refined from the ProtBert-BFD model (URL to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database (URL allowing even more precise prediction of body site location than the HIV-BERT model can provide.", "## Model Description\r\n\r\nThe HIV-BERT-Bodysite-Identification model is intended to predict the location as to where an HIV sequence was most likely derived from. Because HIV infects immune cells, it uses these as a means of rapidly spreading throughout the body. Thus, body site identification can help determine where exactly these HIV particles ultimately end up. This would be helpful when attempting to study HIV treatment strategies. When provided with an HIV genomic sequence, the HIV-BERT-Bodysite-Identification model can predict which tissue it was derived from.", "## Intended Uses & Limitations\r\n\r\nThis tool can be used as a predictor of which body site an HIV sample was derived from based on its genomic sequence. It should not be considered a clinical diagnostic tool. \r\n \r\nThis tool was trained using the Los Alamos HIV sequence dataset (URL Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences.", "## How to use\r\n\r\nThis model is able to predict the likely bodysite from a V3 sequence.\r\nThis may be use for surveillance of cells that are emerging from latent reservoirs.\r\nRemember, a sequence can come from multiple sites, they are not mutually exclusive.", "## Training Data\r\n\r\nThis model was trained using the damlab/HIV_V3_bodysite dataset using the 0th fold. The dataset consists of 5510 sequences (approximately 35 tokens each) extracted from the Los Alamos HIV Sequence database.", "## Training Procedure", "### Preprocessing\r\n\r\nAs with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.", "### Training\r\n\r\nThe damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be found in multiple sites) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.", "## Evaluation Results\r\n\r\n*Need to add*", "## BibTeX Entry and Citation Info" ]
[ 51, 14, 57, 135, 134, 154, 64, 64, 4, 80, 146, 10, 11 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #dataset-damlab/HIV_V3_bodysite #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for [HIV_V3_bodysite]## Table of Contents\r\n- Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info## Summary\r\n\r\nThe HIV-BERT-Bodysite-Identification model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict the location that an HIV V3 loop sample was derived from. HIV-BERT is a model refined from the ProtBert-BFD model (URL to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database (URL allowing even more precise prediction of body site location than the HIV-BERT model can provide.## Model Description\r\n\r\nThe HIV-BERT-Bodysite-Identification model is intended to predict the location as to where an HIV sequence was most likely derived from. Because HIV infects immune cells, it uses these as a means of rapidly spreading throughout the body. Thus, body site identification can help determine where exactly these HIV particles ultimately end up. This would be helpful when attempting to study HIV treatment strategies. When provided with an HIV genomic sequence, the HIV-BERT-Bodysite-Identification model can predict which tissue it was derived from." ]
[ 0.012887292541563511, 0.06151840463280678, -0.004258778877556324, 0.006362018175423145, 0.1302095502614975, -0.024671535938978195, -0.004384350962936878, 0.11429488658905029, 0.011422245763242245, 0.03229138255119324, -0.028772104531526566, -0.06519068777561188, 0.07248427718877792, 0.055942993611097336, 0.11257477104663849, -0.16200335323810577, 0.06368739902973175, -0.017705049365758896, -0.006136989686638117, 0.09631022065877914, 0.07635916024446487, -0.06845550984144211, 0.022369256243109703, 0.010594753548502922, 0.02978496067225933, 0.10197686403989792, -0.05623631924390793, -0.00017819467757362872, 0.06551805138587952, 0.030090991407632828, 0.10532616823911667, 0.04674935340881348, 0.11340585350990295, -0.3081340491771698, -0.025692861527204514, 0.07937849313020706, -0.01604086346924305, 0.0011388957500457764, 0.08173859119415283, -0.058101508766412735, 0.06873352825641632, -0.011925478465855122, 0.05739499628543854, 0.03818191960453987, -0.10923519730567932, -0.17877399921417236, -0.0551244355738163, 0.07483562082052231, 0.01519174687564373, 0.131434366106987, -0.04492952302098274, 0.016592277213931084, 0.005909188184887171, 0.020080238580703735, 0.11836729943752289, -0.17855849862098694, 0.05196421965956688, 0.18489761650562286, 0.07832688838243484, 0.05058107525110245, -0.056761402636766434, -0.0059006838127970695, -0.021662579849362373, 0.04394694045186043, 0.23263123631477356, 0.0018252730369567871, 0.06676644086837769, -0.024941299110651016, -0.18229180574417114, -0.15425047278404236, 0.10358088463544846, 0.0033026011660695076, -0.032439131289720535, -0.10920988768339157, 0.010087123140692711, -0.025654785335063934, -0.01183370128273964, -0.13304631412029266, 0.014272408559918404, 0.03525824099779129, 0.012636978179216385, -0.06621433794498444, -0.05822742357850075, -0.026632189750671387, -0.0086868517100811, 0.030498631298542023, 0.033026158809661865, 0.03190593793988228, -0.08317345380783081, 0.12327539175748825, -0.06723096966743469, -0.06601660698652267, -0.01748684048652649, -0.06785068660974503, -0.07212547957897186, -0.02455214597284794, 0.04695058986544609, -0.08990133553743362, -0.03787358105182648, 0.09198123961687088, 0.01898838020861149, -0.01421631034463644, 0.0482005774974823, -0.03928138688206673, 0.06909767538309097, 0.10551634430885315, -0.09267488121986389, -0.006819534581154585, -0.02529069036245346, -0.014049369841814041, -0.02892865613102913, 0.03596708923578262, -0.02421029657125473, 0.004155692644417286, 0.09699009358882904, 0.09782527387142181, -0.043391916900873184, 0.03700830787420273, 0.04033190384507179, -0.04208134859800339, 0.016642510890960693, -0.06719732284545898, -0.03839370980858803, -0.05494130775332451, -0.026601940393447876, 0.04783088341355324, 0.1254914402961731, -0.0036604467313736677, -0.0830947756767273, 0.05232146754860878, 0.01617281697690487, -0.0775766521692276, -0.11344961076974869, -0.1401938498020172, -0.034438036382198334, 0.03904158994555473, -0.06022157147526741, -0.12735344469547272, -0.1033669114112854, -0.10631021857261658, -0.023626644164323807, -0.0006380007253028452, 0.021520348265767097, -0.004056849516928196, -0.006118765566498041, -0.010859561152756214, -0.01939123310148716, 0.005538190249353647, -0.019273169338703156, -0.01122094877064228, -0.09847062826156616, 0.06950119882822037, 0.15504896640777588, 0.05901160463690758, -0.10700967907905579, 0.05687057226896286, -0.17613112926483154, 0.08629896491765976, -0.12810148298740387, 0.009037183597683907, -0.186017706990242, -0.057714540511369705, -0.028785046190023422, -0.017580820247530937, 0.019781963899731636, 0.06242162734270096, -0.07480897009372711, -0.01670507900416851, 0.26055651903152466, -0.10492757707834244, -0.03641529753804207, 0.011913743801414967, -0.04415324702858925, 0.17137327790260315, 0.08708783239126205, 0.06571190804243088, 0.08601225167512894, -0.07179263979196548, -0.06348806619644165, -0.015124248340725899, -0.009700821712613106, 0.17370325326919556, 0.01970459148287773, -0.02283773384988308, 0.013318299315869808, 0.01953735388815403, -0.05121449753642082, 0.03341323137283325, -0.01104023028165102, -0.012370200827717781, -0.017936265096068382, -0.009650399908423424, -0.03275647386908531, -0.006768854334950447, -0.02116566337645054, 0.051190461963415146, -0.10584553331136703, 0.12361074984073639, 0.09477990120649338, -0.12892483174800873, 0.04084309935569763, -0.08253917843103409, -0.014403436332941055, -0.02766610123217106, -0.009005319327116013, -0.17579568922519684, -0.11831831932067871, 0.0559016652405262, -0.11227898299694061, -0.012260925956070423, -0.04504537954926491, 0.04670209810137749, 0.00149160111322999, -0.07747117429971695, -0.036389730870723724, -0.03766532614827156, 0.007084935903549194, -0.042842548340559006, -0.057931117713451385, -0.12029367685317993, -0.03248000517487526, 0.052269063889980316, -0.10253547877073288, -0.015847498551011086, -0.0721929743885994, 0.12910272181034088, 0.04262322187423706, -0.009643194265663624, 0.09273183345794678, 0.04947248473763466, -0.017650719732046127, -0.05320250988006592, 0.02441239170730114, -0.026526102796196938, 0.0003528133674990386, 0.028487198054790497, -0.05207078903913498, -0.19223234057426453, 0.030003782361745834, 0.03475531190633774, -0.06790245324373245, -0.060199517756700516, -0.023165108636021614, 0.013188173063099384, -0.07556750625371933, -0.07282821834087372, 0.10334975272417068, 0.016429772600531578, 0.07627879828214645, -0.029541490599513054, -0.05829354003071785, -0.001734513440169394, 0.04460212588310242, -0.02732153795659542, 0.04426313191652298, 0.1370672881603241, -0.11487479507923126, 0.04899001121520996, 0.0075276135466992855, 0.08519256114959717, 0.03238862007856369, -0.009216784499585629, -0.10901125520467758, 0.0440395288169384, -0.044030338525772095, 0.012349446304142475, -0.03579283133149147, -0.08520973473787308, -0.02304900623857975, 0.03592671826481819, -0.000748879392631352, 0.006143375765532255, -0.01275670062750578, 0.06575703620910645, 0.07190787047147751, -0.05295172333717346, -0.02355881780385971, -0.06206244230270386, -0.04233216866850853, 0.06893905252218246, 0.0028262303676456213, -0.0016713906079530716, -0.03335881233215332, -0.010941363871097565, -0.13292723894119263, 0.1067304015159607, -0.03768392279744148, -0.22099265456199646, -0.12338364124298096, -0.0629907175898552, 0.011318654753267765, -0.004643134772777557, 0.022639263421297073, -0.03684712573885918, -0.07750391960144043, -0.13269321620464325, 0.12285836040973663, -0.09458009153604507, -0.0565369576215744, -0.10928758978843689, 0.09623570740222931, 0.058799196034669876, -0.06304925680160522, 0.00017854604811873287, -0.020906833931803703, -0.02614372782409191, 0.03272782266139984, -0.08536405116319656, 0.13291022181510925, 0.0662907138466835, 0.06842458993196487, -0.02777400240302086, -0.06689327955245972, 0.18134362995624542, -0.057362694293260574, 0.08085675537586212, 0.19916005432605743, -0.017099332064390182, 0.01981503702700138, 0.12397634983062744, -0.017399000003933907, -0.007241521030664444, 0.08217058330774307, 0.06209579482674599, -0.031000999733805656, -0.20324131846427917, -0.06931722164154053, 0.0049181291833519936, -0.16222703456878662, 0.09595222771167755, 0.032315827906131744, -0.006610723212361336, 0.03306223452091217, -0.06776181608438492, 0.00699925422668457, 0.013976983726024628, 0.05983612313866615, 0.0893184021115303, -0.00276610953733325, 0.08657602965831757, 0.0019032452255487442, -0.06248388811945915, 0.05629395693540573, 0.030935153365135193, 0.20945386588573456, -0.001673980848863721, 0.07227261364459991, 0.12206456810235977, 0.15035219490528107, 0.04996349290013313, 0.05584040284156799, 0.014599867165088654, 0.026299644261598587, -0.002185997786000371, -0.0478147491812706, 0.06596379727125168, 0.01321526151150465, -0.03144485503435135, -0.01850949600338936, 0.013846066780388355, -0.09899923205375671, -0.003309585154056549, 0.17481017112731934, 0.03241235390305519, -0.12584620714187622, -0.05763930082321167, 0.06275447458028793, -0.04912459850311279, -0.1055913195014, 0.0050491453148424625, 0.04196362942457199, -0.1433035433292389, 0.04547049105167389, -0.004558056592941284, 0.11272744834423065, -0.07493476569652557, -0.01478924322873354, 0.03175162523984909, -0.012731343507766724, -0.047391269356012344, 0.14609898626804352, -0.22803665697574615, 0.09898733347654343, 0.00033433354110457003, 0.03176498785614967, -0.05315471068024635, 0.024522388353943825, -0.02653498202562332, 0.11462824791669846, 0.2059725970029831, 0.019247684627771378, 0.023520953953266144, -0.10064761340618134, -0.06939798593521118, 0.021397346630692482, 0.049318280071020126, -0.18099994957447052, 0.17395354807376862, -0.007085139863193035, 0.017034703865647316, -0.03359585255384445, 0.028349166736006737, -0.1249225065112114, -0.1403205841779709, 0.03333389386534691, -0.12390125542879105, 0.20293672382831573, -0.031239043921232224, -0.028091520071029663, -0.044373735785484314, 0.10086078941822052, -0.05470593273639679, -0.09342528879642487, -0.15313206613063812, 0.04486581310629845, 0.14021526277065277, -0.0571739487349987, 0.045809660106897354, 0.05553135275840759, 0.1785009205341339, -0.03245517238974571, -0.15005147457122803, 0.022948548197746277, -0.0818212702870369, -0.10615754127502441, -0.056420598179101944, 0.10059263557195663, 0.09197677671909332, 0.05252281203866005, 0.042138662189245224, 0.03685314208269119, 0.04096671938896179, -0.05430780351161957, 0.1147325411438942, 0.22266031801700592, -0.014379465021193027, 0.008723573759198189, -0.11295894533395767, -0.03396187722682953, -0.14748001098632812, -0.08626767247915268, 0.12662777304649353, 0.042685989290475845, -0.05854545906186104, 0.07588420808315277, 0.10798167437314987, -0.16585427522659302, -0.20671062171459198, -0.09265704452991486, 0.13029901683330536, 0.02723987214267254, 0.028375715017318726, -0.26500633358955383, 0.0595407709479332, 0.01729048602283001, -0.016695799306035042, -0.1207960769534111, -0.24037370085716248, -0.1237783208489418, 0.08034287393093109, 0.04341192543506622, -0.09733373671770096, -0.03666049987077713, -0.027385011315345764, -0.06260029226541519, -0.03809675946831703, 0.12410107254981995, 0.02254951186478138, 0.004474301356822252, -0.013225039467215538, 0.06845452636480331, 0.03478122502565384, 0.011129920370876789, 0.07252702116966248, 0.03305421397089958, 0.049718499183654785, -0.04216274246573448, 0.09056010097265244, 0.06302536278963089, -0.05838059261441231, 0.10484781861305237, 0.16260194778442383, 0.005040259100496769, -0.1991230696439743, -0.08767138421535492, -0.051085952669382095, -0.002138921059668064, -0.02608432061970234, -0.10336363315582275, -0.06830218434333801, 0.04514848440885544, 0.1220749095082283, -0.05401338264346123, 0.020827215164899826, -0.09426812082529068, -0.00478715542703867, 0.13374412059783936, 0.12320049852132797, 0.0859149694442749, -0.16011343896389008, 0.049271415919065475, -0.012033580802381039, 0.04514162614941597, -0.07821802794933319, 0.08531176298856735, 0.09915575385093689, 0.002508297562599182, 0.10409419238567352, 0.011874858289957047, -0.1845703125, -0.029974836856126785, 0.05471672862768173, -0.10237817466259003, -0.11054009199142456, -0.058045707643032074, 0.033616598695516586, -0.1297440230846405, -0.08827165514230728, 0.108632393181324, -0.10154467076063156, -0.0580969974398613, -0.05878300964832306, 0.07619377970695496, 0.02401014044880867, 0.1056516170501709, -0.003983214497566223, 0.03334534168243408, -0.051888566464185715, 0.10907109826803207, 0.09734705835580826, -0.06694486737251282, -0.021579276770353317, 0.04510325938463211, -0.14465919137001038, -0.041397109627723694, -0.009048184379935265, 0.10939307510852814, -0.08192956447601318, -0.07514877617359161, 0.049719586968421936, -0.09085695445537567, 0.04781816527247429, 0.20221541821956635, -0.005057462956756353, -0.025093775242567062, -0.06609842926263809, -0.04964565858244896, -0.10709129273891449, 0.06694014370441437, 0.028527235612273216, 0.024805227294564247, -0.05772258713841438, 0.0926315188407898, 0.058225229382514954, -0.024552948772907257, -0.045788440853357315, -0.020091241225600243, -0.08378448337316513, 0.028142817318439484, -0.11609061062335968, -0.03419467434287071, -0.01969808340072632, -0.06242380291223526, 0.0031408267095685005, 0.01789969950914383, 0.014275576919317245, 0.0537700355052948, -0.07911156117916107, -0.05483726039528847, -0.026508398354053497, 0.04984379932284355, -0.06793858110904694, -0.02009834162890911, 0.04744263365864754, -0.08470276743173599, 0.07587844878435135, 0.026316415518522263, 0.08204741775989532, 0.011742666363716125, -0.047058794647455215, -0.023241743445396423, 0.05985129624605179, 0.06168239936232567, -0.03680925816297531, -0.14892318844795227, -0.01217821054160595, -0.0188972856849432, -0.1631450206041336, -0.014167898334562778, -0.02359636127948761, -0.03122730739414692, -0.05007247254252434, 0.03510918468236923, 0.10810108482837677, -0.03274676203727722, 0.07678423821926117, 0.05437168478965759, 0.0346166156232357, 0.09152014553546906, -0.027301503345370293, 0.0387318953871727, -0.1836395114660263, -0.037817928940057755, -0.04206538200378418, 0.01727692037820816, -0.07361917942762375, -0.07898217439651489, 0.04874015599489212, 0.004404996056109667, 0.13077573478221893, 0.07732725143432617, 0.12684805691242218, -0.004598189145326614, -0.059280868619680405, 0.05126990005373955, -0.02342306077480316, 0.04850829392671585, 0.09297559410333633, -0.0003294401103630662, 0.06699632108211517, 0.004898841492831707, -0.03690188750624657, 0.13181185722351074, -0.040133558213710785, 0.06722390651702881, 0.13568946719169617, -0.0156870037317276, 0.02448122762143612, -0.0793728455901146, -0.07304597645998001, -0.10195977240800858, 0.008597454987466335, 0.06382250785827637, -0.028246955946087837, 0.10790937393903732, 0.05368416756391525, -0.14168837666511536, 0.1382240206003189, 0.02025742270052433, -0.06004124507308006, -0.07230665534734726, -0.19975151121616364, -0.01872202940285206, -0.10035151243209839, -0.011819781735539436, -0.14206039905548096, -0.03383460268378258, 0.15635524690151215, 0.015712503343820572, 0.01657671108841896, -0.008531395345926285, -0.05436290428042412, -0.1921509951353073, 0.04118581861257553, -0.021324949339032173, 0.03407832607626915, -0.037715714424848557, 0.12888126075267792, 0.10040091723203659, 0.027437033131718636, -0.0017360521014779806, 0.03923061117529869, 0.06351657211780548, 0.018736934289336205, 0.011803303845226765, -0.02619599550962448, -0.0016353518003597856, -0.010711491107940674, -0.018739717081189156, 0.2128135710954666, 0.08643458038568497, 0.010859278962016106, -0.013327307067811489, 0.2719590961933136, -0.013744792900979519, 0.00038731511449441314, -0.10017619282007217, 0.2847927510738373, 0.051864150911569595, 0.05363902449607849, -0.024702399969100952, -0.06802988052368164, -0.00497882766649127, 0.2589443027973175, -0.04052933678030968, 0.05868498235940933, -0.03245920315384865, 0.010165945626795292, -0.00025562828523106873, 0.08680397272109985, 0.05918252468109131, 0.0788346454501152, 0.16360385715961456, -0.12562677264213562, 0.1111917644739151, -0.06995449215173721, -0.04773583635687828, 0.010547695681452751, 0.13325881958007812, -0.01633138582110405, 0.012655765749514103, -0.015622206032276154, 0.10090653598308563, -0.055712636560201645, -0.21698783338069916, 0.033841684460639954, -0.00856885313987732, -0.03553998842835426, -0.01651792973279953, -0.07828269898891449, 0.07462149113416672, -0.005076433066278696, 0.013068592175841331, 0.04322120174765587, 0.24629251658916473, 0.09069105237722397, -0.0426977202296257, -0.07326193153858185, 0.08207601308822632, -0.05020296573638916, 0.13780580461025238, -0.0005999592249281704, 0.008184017613530159, 0.04595213383436203, -0.03715120628476143, -0.16037903726100922, 0.06034229323267937, 0.020304184406995773, 0.012375769205391407, 0.0346093513071537, 0.16013646125793457, 0.030511880293488503, 0.1392693668603897, 0.023589445278048515, -0.04924188181757927, 0.0775790587067604, -0.03543220832943916, -0.07890307903289795, 0.009469885379076004, 0.028317265212535858, -0.050621990114450455, 0.06648676842451096, 0.17520247399806976, -0.06480361521244049, -0.00023298927408177406, -0.10357184708118439, 0.04084021598100662, -0.011726520024240017, 0.04468474164605141, 0.04073478654026985, -0.05806577950716019, 0.018614808097481728, 0.0017798752523958683, -0.03336693346500397, -0.18242084980010986, -0.021990565583109856, 0.09258150309324265, -0.02295256406068802, -0.0020227113272994757, 0.1114959716796875, 0.03674369677901268, 0.041227374225854874, -0.04708845540881157, -0.016768645495176315, 0.03697868436574936, 0.07144790887832642, -0.12090978026390076, -0.006146871950477362 ]
null
null
transformers
#dialogue
{"tags": ["text-generation"]}
text-generation
danchang11/GPT2-TraditionalChat
[ "transformers", "pytorch", "gpt2", "text-generation", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #endpoints_compatible #text-generation-inference #region-us
#dialogue
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 39 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0177617147564888, 0.00413000863045454, -0.008743596263229847, -0.014220969751477242, 0.164886936545372, 0.03891453891992569, 0.046693552285432816, 0.13523150980472565, 0.015874169766902924, -0.004716119728982449, 0.1357993483543396, 0.1721617579460144, 0.006274322979152203, 0.005896460264921188, -0.05455324426293373, -0.2677989900112152, 0.07517701387405396, 0.08140098303556442, -0.042099032551050186, 0.12790833413600922, 0.048993777483701706, -0.04671265929937363, 0.08353178203105927, -0.018136927857995033, -0.14306217432022095, 0.012910505756735802, 0.0358387716114521, -0.0987507775425911, 0.11372382193803787, 0.02934994176030159, 0.10152171552181244, 0.0027315847110003233, -0.11149000376462936, -0.16700994968414307, 0.029015103355050087, 0.05075612664222717, -0.07201661914587021, 0.05401792749762535, 0.08914339542388916, -0.12092448025941849, 0.12759658694267273, 0.029667116701602936, -0.051347605884075165, 0.049732182174921036, -0.1655282825231552, -0.09687618911266327, -0.02854047901928425, 0.025650938972830772, 0.03822527453303337, 0.10048377513885498, -0.016911430284380913, 0.06471045315265656, -0.09987933188676834, 0.08688510209321976, 0.22739006578922272, -0.32783472537994385, 0.00570092024281621, 0.11302487552165985, 0.05605051666498184, 0.053175803273916245, -0.02404518984258175, 0.05229756981134415, 0.01313769444823265, 0.01957249827682972, 0.00137265061493963, -0.07302742451429367, -0.0565350204706192, 0.08518718928098679, -0.10417694598436356, -0.10550206154584885, 0.2108500748872757, -0.06404201686382294, 0.06052133068442345, -0.024971894919872284, -0.10919266194105148, -0.058717526495456696, -0.0136891920119524, 0.047699883580207825, -0.05458793044090271, 0.09626783430576324, 0.04053015634417534, -0.08627340197563171, -0.13180461525917053, -0.06005355343222618, -0.17273692786693573, 0.20520804822444916, 0.024591796100139618, 0.07910293340682983, -0.2197970300912857, 0.11058394610881805, -0.038688383996486664, -0.06404577195644379, 0.002036143559962511, -0.10641399770975113, 0.05131516978144646, 0.03840160742402077, -0.07174071669578552, -0.05681565776467323, 0.09885678440332413, 0.0772608071565628, -0.08083102107048035, 0.02651338279247284, -0.03876886144280434, 0.10964981466531754, 0.013784559443593025, 0.07848335802555084, -0.028710849583148956, 0.03977745771408081, 0.01966063678264618, -0.17050744593143463, -0.0029470184817910194, -0.057985570281744, -0.11522450298070908, -0.0807778462767601, 0.03479090705513954, 0.09325509518384933, 0.010769971646368504, 0.0810101106762886, -0.03773309290409088, -0.025342658162117004, 0.0269398745149374, -0.05964063107967377, -0.030446602031588554, 0.032796863466501236, 0.012962838634848595, 0.1671743243932724, -0.0137644587084651, -0.0017080423422157764, -0.15164265036582947, 0.05012911558151245, -0.07729051262140274, 0.017933771014213562, -0.041089978069067, -0.03279638662934303, 0.022797340527176857, -0.11531727015972137, 0.004681426100432873, -0.11883518099784851, -0.17856267094612122, 0.027148963883519173, -0.00009361533011542633, -0.015542703680694103, -0.0018064614851027727, -0.028014300391077995, -0.052028052508831024, 0.024566810578107834, -0.06845179945230484, 0.001042075571604073, -0.06374721974134445, 0.1192232072353363, -0.06242113187909126, 0.06873898208141327, -0.11195021122694016, 0.07659032940864563, -0.10372360050678253, 0.01668623648583889, -0.12242420017719269, 0.0736471563577652, -0.005726009141653776, 0.09933587908744812, -0.0381762757897377, -0.07277093082666397, -0.11802409589290619, 0.052166618406772614, -0.052433934062719345, 0.19637678563594818, -0.05141520872712135, -0.12538684904575348, 0.3230677843093872, -0.07854989171028137, -0.12571749091148376, 0.08480606228113174, 0.017596984282135963, -0.009740262292325497, 0.08438948541879654, 0.22746822237968445, 0.013743794523179531, 0.005568898748606443, 0.05719493329524994, 0.1417950987815857, -0.13118228316307068, -0.0950319841504097, 0.04732563719153404, -0.054628971964120865, -0.10563048720359802, 0.030682945623993874, -0.007876500487327576, 0.0630379170179367, -0.0447998009622097, -0.012814326211810112, -0.04133640602231026, 0.02648276835680008, 0.08373766392469406, 0.001436202903278172, 0.10575487464666367, -0.043722447007894516, -0.04275013506412506, -0.01118161529302597, -0.01528346911072731, -0.06143542379140854, 0.06491949409246445, -0.014812358655035496, 0.14447778463363647, -0.03690091893076897, 0.04783803969621658, -0.18692760169506073, -0.08071625232696533, -0.023726923391222954, 0.10554835200309753, -0.03628280386328697, 0.07664146274328232, 0.06979329884052277, -0.06121150404214859, -0.011977508664131165, -0.00036841665860265493, 0.10460905730724335, -0.04247349128127098, -0.024623505771160126, -0.040476251393556595, 0.04367184638977051, -0.050804685801267624, -0.03764241188764572, -0.043807633221149445, 0.026319099590182304, 0.12130927294492722, 0.08805370330810547, -0.011063527315855026, 0.01108147669583559, 0.0008715765434317291, 0.002451015170663595, -0.04143768176436424, -0.01616770774126053, 0.1087322011590004, -0.01118565909564495, -0.09968315064907074, 0.20184315741062164, -0.12686163187026978, 0.18463411927223206, 0.1917361617088318, -0.28694963455200195, 0.044258002191782, -0.08406686037778854, -0.04173731803894043, 0.03883618488907814, 0.0610942617058754, -0.054276976734399796, 0.1691325157880783, 0.01757095381617546, 0.1580827832221985, -0.07692365348339081, -0.06435651332139969, -0.009723078459501266, -0.027760373428463936, 0.008332083001732826, 0.08378587663173676, 0.08490931242704391, -0.09190400689840317, 0.18039464950561523, 0.17116224765777588, 0.05264123156666756, 0.1655624955892563, -0.02682124264538288, -0.04416370019316673, 0.08961658924818039, 0.05326202139258385, -0.04856514930725098, -0.07251864671707153, -0.2637195289134979, -0.02071956731379032, 0.08327561616897583, 0.07756919413805008, 0.14509862661361694, -0.14547564089298248, -0.05993083491921425, -0.02641592174768448, -0.05574978142976761, -0.05129653960466385, 0.0821082666516304, 0.061170388013124466, 0.1182895228266716, 0.011390134692192078, 0.04968147352337837, 0.10705231875181198, 0.0072124917060136795, -0.0920310989022255, 0.22349299490451813, -0.11133892834186554, -0.34016597270965576, -0.14436322450637817, -0.11788070946931839, -0.030494745820760727, 0.04577368497848511, 0.10457808524370193, -0.136713445186615, 0.003679262939840555, 0.02427663654088974, 0.14769063889980316, -0.13262353837490082, 0.017893647775053978, -0.041815612465143204, 0.04790092632174492, -0.10780538618564606, -0.09514795243740082, -0.05651114135980606, -0.036214686930179596, -0.0907103642821312, 0.13784471154212952, -0.12632136046886444, 0.03253214433789253, 0.1913776844739914, 0.06095456704497337, 0.06541487574577332, -0.040304239839315414, 0.17827682197093964, -0.1092667505145073, -0.034401800483465195, 0.21323581039905548, -0.02538575418293476, 0.09009216725826263, 0.05674639344215393, 0.0030026263557374477, -0.08552411198616028, -0.005033917725086212, -0.023024432361125946, -0.11021243780851364, -0.26796042919158936, -0.10801071673631668, -0.12731461226940155, 0.0727757066488266, 0.022434573620557785, 0.06192222982645035, 0.1212952733039856, 0.0660119503736496, -0.02397187612950802, -0.007011115085333586, 0.027425037696957588, 0.07903468608856201, 0.1664983183145523, -0.029877230525016785, 0.09351065754890442, -0.05715025216341019, -0.09589941799640656, 0.08371074497699738, 0.07919108867645264, 0.20126491785049438, 0.03149360418319702, 0.07982219010591507, 0.045319534838199615, 0.058093857020139694, 0.13299427926540375, 0.1143735721707344, -0.02366695925593376, -0.00476102065294981, -0.02725902758538723, -0.023289548233151436, -0.04717805236577988, 0.0179450660943985, 0.007347718812525272, -0.1569279283285141, -0.058918919414281845, -0.1493380069732666, 0.11772475391626358, 0.0877038910984993, 0.05193497985601425, -0.18433049321174622, -0.0027360362000763416, 0.09183459728956223, -0.016695870086550713, -0.11040494590997696, 0.10688318312168121, 0.025785457342863083, -0.13458241522312164, 0.05990788713097572, -0.062118303030729294, 0.12250654399394989, -0.060343630611896515, 0.08080611377954483, -0.04021589085459709, -0.09235090017318726, 0.02796080708503723, 0.1174207478761673, -0.2523665726184845, 0.22567634284496307, -0.004532721359282732, -0.05051229149103165, -0.09785845875740051, -0.004584147594869137, -0.006866491865366697, 0.10252957046031952, 0.15214580297470093, 0.026027202606201172, -0.014887488447129726, -0.0593542717397213, -0.0008516228408552706, 0.03857841715216637, 0.1469660848379135, -0.06455042958259583, -0.01702277362346649, -0.021815843880176544, 0.005529934074729681, -0.0365385003387928, -0.03495530039072037, 0.0736987516283989, -0.1503359079360962, 0.06277604401111603, 0.008920188993215561, 0.09208177030086517, 0.001672430313192308, 0.009417003951966763, -0.07299439609050751, 0.20857995748519897, -0.13122130930423737, -0.13205280900001526, -0.09646400064229965, -0.0619744136929512, 0.08571894466876984, -0.05828427895903587, 0.0448489636182785, -0.06629502773284912, 0.007595032919198275, -0.05393918231129646, -0.21739354729652405, 0.10785721987485886, -0.06963614374399185, -0.023533938452601433, 0.0036003789864480495, 0.246476411819458, -0.06690777838230133, 0.006640743464231491, 0.01160342339426279, 0.020129498094320297, -0.10629308968782425, -0.12935012578964233, 0.026942672207951546, 0.016331583261489868, 0.08289532363414764, 0.08407847583293915, -0.04428640007972717, 0.055529527366161346, -0.020418979227542877, 0.02320215106010437, 0.31711724400520325, 0.10947677493095398, -0.03180767595767975, 0.18130561709403992, 0.08401691168546677, -0.08147493004798889, -0.2822689414024353, -0.06397486478090286, -0.1189005896449089, -0.05097424238920212, -0.09442377835512161, -0.2242605984210968, 0.08713490515947342, 0.06292454153299332, 0.008994982577860355, 0.15863020718097687, -0.31459149718284607, -0.05044165253639221, 0.08400531858205795, -0.004015425220131874, 0.4111485481262207, -0.14122027158737183, -0.1208438128232956, -0.02601119875907898, -0.21869809925556183, 0.15852734446525574, -0.0737798884510994, 0.107994444668293, -0.03963931277394295, 0.10422861576080322, 0.03994893655180931, -0.0722108855843544, 0.11042612791061401, 0.05572624132037163, 0.0006772224442102015, -0.07959822565317154, -0.03933039307594299, 0.07553410530090332, 0.01759473606944084, 0.01572279818356037, -0.027233067899942398, 0.031986165791749954, -0.15855683386325836, -0.04007037356495857, -0.11417832225561142, 0.04086349159479141, 0.047451384365558624, -0.05472583696246147, -0.022158317267894745, -0.06267859041690826, 0.0036529593635350466, 0.02783399075269699, 0.2299688458442688, -0.055845387279987335, 0.14678794145584106, 0.0181744322180748, 0.08333985507488251, -0.12178298830986023, -0.07275176048278809, -0.06875777989625931, -0.028424803167581558, 0.0988553985953331, -0.16970506310462952, 0.053557995706796646, 0.1198544055223465, -0.024564003571867943, 0.046731337904930115, 0.13614340126514435, 0.005411188583821058, 0.008843154646456242, 0.11319591850042343, -0.24978069961071014, -0.06546348333358765, -0.06806614249944687, -0.06450063735246658, 0.09459757059812546, 0.11480807512998581, 0.1617022007703781, 0.044600822031497955, -0.034621790051460266, -0.015353205613791943, 0.019293226301670074, -0.06277140974998474, 0.03454780951142311, 0.001771117327734828, 0.029499240219593048, -0.14590995013713837, 0.08422111719846725, -0.01743590086698532, -0.15481045842170715, 0.0021370314061641693, 0.1372750699520111, -0.141866534948349, -0.10142909735441208, -0.062354519963264465, 0.05168379843235016, -0.11357450485229492, -0.023666704073548317, -0.03506386652588844, -0.12296982854604721, 0.09337733685970306, 0.15561185777187347, 0.06566707044839859, 0.11923837661743164, -0.026205087080597878, -0.024233929812908173, -0.00927067268639803, -0.06355437636375427, -0.04055394232273102, -0.006395004689693451, -0.06335654109716415, 0.06671498715877533, -0.021831858903169632, 0.14458994567394257, -0.07311990857124329, -0.07258106768131256, -0.15852877497673035, 0.05074010789394379, -0.11197909712791443, -0.0547679141163826, -0.10145534574985504, -0.061023611575365067, -0.02561028115451336, -0.003428260562941432, -0.04455279931426048, -0.03193459287285805, -0.11993534862995148, 0.03322478011250496, -0.05789078772068024, 0.013840819709002972, -0.07794932276010513, 0.013500452041625977, 0.09164173901081085, -0.043710220605134964, 0.15036717057228088, 0.16742666065692902, -0.09713615477085114, 0.14140373468399048, -0.15944255888462067, -0.09699242562055588, 0.1065526008605957, -0.004710288718342781, 0.01055878959596157, 0.08062771707773209, 0.03611044958233833, 0.04957300424575806, 0.00015067339700181037, 0.0604633130133152, -0.036007050424814224, -0.12734907865524292, 0.03170613944530487, -0.042643286287784576, -0.10766422003507614, -0.06020660325884819, -0.049973342567682266, 0.06957629323005676, 0.05035904794931412, 0.06711549311876297, -0.011089643463492393, 0.11557795107364655, -0.050442613661289215, 0.018985580652952194, 0.026638763025403023, -0.16677306592464447, 0.008527749218046665, -0.061151642352342606, 0.03204059973359108, 0.010763505473732948, 0.28714069724082947, -0.00013240271073300391, -0.007132190745323896, 0.014635466039180756, 0.0900757908821106, 0.04493294656276703, 0.017144059762358665, 0.2578983008861542, 0.10303528606891632, -0.06955546885728836, -0.08077843487262726, 0.06792601943016052, 0.012195531278848648, 0.02784431166946888, 0.17747752368450165, 0.08463454991579056, 0.02749769017100334, 0.09605445712804794, -0.040658168494701385, 0.021314077079296112, -0.10011026263237, -0.11307407170534134, 0.013244640082120895, 0.04699188470840454, 0.0012895361287519336, 0.1374855935573578, 0.14933034777641296, -0.040645238012075424, 0.06967397034168243, 0.016282886266708374, -0.05199767276644707, -0.15802720189094543, -0.14379066228866577, -0.04298508167266846, -0.14894913136959076, 0.021038243547081947, -0.13705667853355408, 0.03882404416799545, 0.13195720314979553, 0.06392589956521988, -0.03431929647922516, 0.1295498013496399, 0.07832565158605576, -0.11752352863550186, 0.08415429294109344, -0.032199811190366745, 0.08165641874074936, 0.03971846029162407, -0.017476707696914673, -0.03917689621448517, -0.04564012959599495, 0.011646753177046776, 0.06824102252721786, -0.030940555036067963, 0.012069804593920708, -0.16889329254627228, -0.08907386660575867, -0.0529489628970623, 0.09230942279100418, -0.05592195317149162, 0.13175910711288452, 0.0032278846483677626, -0.04776560515165329, 0.0334765650331974, 0.2324390858411789, -0.0619046650826931, -0.018577080219984055, -0.018306629732251167, 0.17200686037540436, 0.05329083278775215, 0.08580727875232697, -0.00465694535523653, -0.006655924487859011, -0.07144376635551453, 0.3501553535461426, 0.2633618414402008, -0.03886903077363968, 0.012465479783713818, 0.041434697806835175, 0.04495251551270485, 0.17578192055225372, 0.1147269532084465, 0.11233104765415192, 0.29954424500465393, -0.07917296886444092, -0.0670166090130806, -0.009023794904351234, -0.004121111705899239, -0.10456884652376175, 0.10345932841300964, 0.039856355637311935, -0.09799449145793915, -0.04581701010465622, 0.11436376720666885, -0.25572752952575684, 0.09593817591667175, -0.048779647797346115, -0.1697508841753006, -0.047806475311517715, -0.013967693783342838, 0.10703971982002258, 0.010679148137569427, 0.10505855083465576, -0.006022031884640455, -0.13563650846481323, 0.04811090603470802, 0.06031722202897072, -0.26589345932006836, -0.01062775868922472, 0.05762946978211403, -0.02625512331724167, -0.010761603713035583, -0.020649395883083344, 0.04956243932247162, 0.06090697646141052, 0.03886969015002251, -0.02475808933377266, 0.01703202910721302, -0.004924232140183449, -0.040182795375585556, -0.02096102386713028, 0.05230327695608139, 0.006438924930989742, -0.15112970769405365, 0.06094548851251602, -0.135308176279068, 0.02098478563129902, 0.028334662318229675, -0.032334886491298676, -0.019326696172356606, -0.044271863996982574, -0.08218339085578918, 0.022670846432447433, 0.08878248929977417, -0.0012725733686238527, -0.01129397377371788, -0.08182375878095627, -0.022717300802469254, -0.019387206062674522, -0.05968303605914116, -0.09699249267578125, -0.09970560669898987, -0.10814160108566284, 0.1260090470314026, -0.020027613267302513, -0.1781155914068222, 0.031491469591856, -0.05561300739645958, 0.07841652631759644, -0.1629980504512787, 0.0618479959666729, 0.04849570244550705, 0.02506282366812229, -0.00006693792238365859, -0.01461757067590952, 0.06004534289240837, 0.09022362530231476, -0.08918830007314682, -0.08250809460878372 ]