sha
null
last_modified
null
library_name
stringclasses
154 values
text
stringlengths
1
900k
metadata
stringlengths
2
348k
pipeline_tag
stringclasses
45 values
id
stringlengths
5
122
tags
listlengths
1
1.84k
created_at
stringlengths
25
25
arxiv
listlengths
0
201
languages
listlengths
0
1.83k
tags_str
stringlengths
17
9.34k
text_str
stringlengths
0
389k
text_lists
listlengths
0
722
processed_texts
listlengths
1
723
tokens_length
listlengths
1
723
input_texts
listlengths
1
61
embeddings
listlengths
768
768
null
null
null
https://community.afpglobal.org/network/members/profile?UserKey=fb4fdcef-dde4-4258-a423-2159545d84c1 https://community.afpglobal.org/network/members/profile?UserKey=e6ccc088-b709-45ec-b61e-4d56088acbda https://community.afpglobal.org/network/members/profile?UserKey=ba280059-0890-4510-81d0-a79522b75ac8 https://community.afpglobal.org/network/members/profile?UserKey=799ba769-6e99-4a6a-a173-4f1b817e978c https://community.afpglobal.org/network/members/profile?UserKey=babb84d7-e91a-4972-b26a-51067c66d793 https://community.afpglobal.org/network/members/profile?UserKey=8e4656bc-8d0d-44e1-b280-e68a2ace9353 https://community.afpglobal.org/network/members/profile?UserKey=8e7b41a8-9bed-4cb0-9021-a164b0aa6dd3 https://community.afpglobal.org/network/members/profile?UserKey=e4f38596-d772-4fbe-9e93-9aef5618f26e https://community.afpglobal.org/network/members/profile?UserKey=18221e49-74ba-4155-ac1e-6f184bfb2398 https://community.afpglobal.org/network/members/profile?UserKey=ef4391e8-03df-467f-bf3f-4a45087817eb https://community.afpglobal.org/network/members/profile?UserKey=832774fd-a035-421a-8236-61cf45a7747d https://community.afpglobal.org/network/members/profile?UserKey=9f05cd73-b75c-4820-b60a-5df6357b2af9 https://community.afpglobal.org/network/members/profile?UserKey=c1727992-5024-4321-b0c9-ecc6f51e6532 https://www.hybrid-analysis.com/sample/255948e335dd9f873d11bf0224f8d180cd097509d23d27506292c22443fa92b8 https://www.facebook.com/PS5Giveaways2021 https://cgvmovie.cookpad-blog.jp/articles/589986 https://myanimelist.net/blog.php?eid=850892 https://comicvine.gamespot.com/profile/full-tv-free/about-me/ https://pantip.com/topic/40658194
{}
null
fullshowbox/full-tv-free
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
URL URL URL URL URL URL URL URL URL URL URL URL URL URL URL URL URL URL URL
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
null
https://volunteer.alz.org/network/members/profile?UserKey=f4774542-39b3-4cfd-8c21-7b834795f7d7 https://volunteer.alz.org/network/members/profile?UserKey=05a00b90-f854-45fb-9a3a-7420144d290c https://volunteer.alz.org/network/members/profile?UserKey=45cceddd-29b9-4c6c-8612-e2a16aaa391a https://volunteer.alz.org/network/members/profile?UserKey=ae3c28f9-72a3-4af5-bd50-3b2ea2c0d3a3 https://volunteer.alz.org/network/members/profile?UserKey=7ab8e28e-e31f-4906-ab06-84b9ea3a880f https://volunteer.alz.org/network/members/profile?UserKey=1b31fc90-e18e-4ef6-81f0-5c0b55fb95a3 https://volunteer.alz.org/network/members/profile?UserKey=23971b11-04ad-4eb4-abc5-6e659c6b071c 123movies-watch-online-movie-full-free-2021 https://myanimelist.net/blog.php?eid=849353 https://comicvine.gamespot.com/profile/nacenetwork21/about-me/ https://pantip.com/topic/40639721
{}
null
fullshowbox/nacenetwork21
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
URL URL URL URL URL URL URL 123movies-watch-online-movie-full-free-2021 URL URL URL
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
null
https://www.nace.org/network/members/profile?UserKey=461a690a-bff6-4e4c-be63-ea8e39264459 https://www.nace.org/network/members/profile?UserKey=b4a6a66a-fb8a-4f2b-8af9-04f003ad9d46 https://www.nace.org/network/members/profile?UserKey=24544ab2-551d-42aa-adbe-7a1c1d68fd9c https://www.nace.org/network/members/profile?UserKey=3e8035d5-056a-482d-9010-9883e5990f4a https://www.nace.org/network/members/profile?UserKey=d7241c69-28c4-4146-a077-a00cc2c9ccf5 https://www.nace.org/network/members/profile?UserKey=2c58c2fb-13a4-4e5a-b044-f467bb295d83 https://www.nace.org/network/members/profile?UserKey=dd8a290c-e53a-4b56-9a17-d35dbcb6b8bd https://www.nace.org/network/members/profile?UserKey=0e96a1af-91f4-496a-af02-6d753a1bbded
{}
null
fullshowbox/networkprofile
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
URL URL URL URL URL URL URL URL
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
null
https://ragbrai.com/groups/hd-movie-watch-french-exit-2021-full-movie-online-for-free/ https://ragbrai.com/groups/hd-movie-watch-nobody-2021-full-movie-online-for-free/ https://ragbrai.com/groups/hd-movie-watch-voyagers-2021-full-movie-online-for-free/ https://ragbrai.com/groups/hd-movie-watch-godzilla-vs-kong-2021-full-movie-online-for-free/ https://ragbrai.com/groups/hd-movie-watch-raya-and-the-last-dragon-2021-full-movie-online-for-free/ https://ragbrai.com/groups/hd-movie-watch-mortal-kombat-2021-full-movie-online-for-free/ https://ragbrai.com/groups/hd-movie-watch-the-father-2021-full-movie-online-for-free/
{}
null
fullshowbox/ragbrai
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
URL URL URL URL URL URL URL
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
transformers
# Funnel Transformer intermediate model (B6-6-6 without decoder) Pretrained model on English language using a similar objective objective as [ELECTRA](https://huggingface.co/transformers/model_doc/electra.html). It was introduced in [this paper](https://arxiv.org/pdf/2006.03236.pdf) and first released in [this repository](https://github.com/laiguokun/Funnel-Transformer). This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. **Note:** This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth of the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if you need one input per initial token. You should use the `intermediate` model in that case. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=funnel-transformer) to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import FunnelTokenizer, FunnelBaseModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/intermediate-base") model = FunnelBaseModel.from_pretrained("funnel-transformer/intermediate-base") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import FunnelTokenizer, TFFunnelBaseModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/intermediate-base") model = TFFunnelBaseModel.from_pretrained("funnel-transformer/intermediate-base") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` ## Training data The BERT model was pretrained on: - [BookCorpus](https://yknzhu.wixsite.com/mbweb), a dataset consisting of 11,038 unpublished books, - [English Wikipedia](https://en.wikipedia.org/wiki/English_Wikipedia) (excluding lists, tables and headers), - [Clue Web](https://lemurproject.org/clueweb12/), a dataset of 733,019,372 English web pages, - [GigaWord](https://catalog.ldc.upenn.edu/LDC2011T07), an archive of newswire text data, - [Common Crawl](https://commoncrawl.org/), a dataset of raw web pages. ### BibTeX entry and citation info ```bibtex @misc{dai2020funneltransformer, title={Funnel-Transformer: Filtering out Sequential Redundancy for Efficient Language Processing}, author={Zihang Dai and Guokun Lai and Yiming Yang and Quoc V. Le}, year={2020}, eprint={2006.03236}, archivePrefix={arXiv}, primaryClass={cs.LG} } ```
{"language": "en", "license": "apache-2.0", "datasets": ["bookcorpus", "wikipedia", "gigaword"]}
feature-extraction
funnel-transformer/intermediate-base
[ "transformers", "pytorch", "tf", "funnel", "feature-extraction", "en", "dataset:bookcorpus", "dataset:wikipedia", "dataset:gigaword", "arxiv:2006.03236", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2006.03236" ]
[ "en" ]
TAGS #transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us
# Funnel Transformer intermediate model (B6-6-6 without decoder) Pretrained model on English language using a similar objective objective as ELECTRA. It was introduced in this paper and first released in this repository. This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. Note: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth of the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if you need one input per initial token. You should use the 'intermediate' model in that case. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: and in TensorFlow: ## Training data The BERT model was pretrained on: - BookCorpus, a dataset consisting of 11,038 unpublished books, - English Wikipedia (excluding lists, tables and headers), - Clue Web, a dataset of 733,019,372 English web pages, - GigaWord, an archive of newswire text data, - Common Crawl, a dataset of raw web pages. ### BibTeX entry and citation info
[ "# Funnel Transformer intermediate model (B6-6-6 without decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nNote: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth\nof the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if\nyou need one input per initial token. You should use the 'intermediate' model in that case.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ "TAGS\n#transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us \n", "# Funnel Transformer intermediate model (B6-6-6 without decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nNote: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth\nof the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if\nyou need one input per initial token. You should use the 'intermediate' model in that case.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ 72, 103, 289, 133, 32, 95, 11 ]
[ "passage: TAGS\n#transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us \n# Funnel Transformer intermediate model (B6-6-6 without decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nNote: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth\nof the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if\nyou need one input per initial token. You should use the 'intermediate' model in that case." ]
[ -0.08720822632312775, 0.13673396408557892, -0.004108911845833063, 0.0489746555685997, 0.10349368304014206, -0.0015920527512207627, 0.03418269008398056, 0.060560036450624466, -0.09277887642383575, 0.10485635697841644, -0.025255752727389336, -0.049272362142801285, 0.09565799683332443, 0.05283638834953308, 0.07918329536914825, -0.20990310609340668, 0.06475108861923218, -0.04593084380030632, 0.13395768404006958, 0.09233882278203964, 0.05439838021993637, -0.09617021679878235, 0.0691683292388916, -0.02259259670972824, -0.05693874880671501, -0.03375393524765968, -0.00703376205638051, -0.019729837775230408, 0.027898987755179405, 0.053484659641981125, 0.11137280613183975, 0.0034756301902234554, 0.035197220742702484, -0.07327409833669662, 0.015732116997241974, 0.08856358379125595, 0.008083944208920002, 0.048323046416044235, 0.030341098085045815, 0.010727260261774063, 0.07862145453691483, -0.027580786496400833, 0.059140268713235855, 0.056676916778087616, -0.15410763025283813, -0.14883793890476227, -0.06354764848947525, 0.12343835085630417, 0.057169102132320404, 0.11150859296321869, -0.037510357797145844, 0.11116575449705124, -0.009387088939547539, 0.0695895329117775, 0.12401249259710312, -0.19335035979747772, -0.0038728693034499884, -0.01360295433551073, -0.02136051468551159, 0.10680485516786575, -0.06947559118270874, -0.0664990022778511, -0.013680286705493927, 0.031072285026311874, 0.04184078797698021, -0.027591336518526077, -0.043129801750183105, -0.08015836030244827, -0.1316748410463333, -0.05648968741297722, 0.15820439159870148, -0.004239402711391449, -0.13837002217769623, -0.04922005534172058, -0.05481768026947975, -0.14435559511184692, 0.013149605132639408, -0.011849838308990002, 0.003146487521007657, 0.013149742968380451, 0.025641899555921555, -0.011392727494239807, -0.11058026552200317, -0.031314291059970856, -0.05000666156411171, 0.11062031239271164, 0.022216035053133965, 0.04724281281232834, -0.0661555826663971, 0.11540139466524124, -0.06240539625287056, -0.055045511573553085, -0.0841476172208786, -0.038458142429590225, -0.12847177684307098, -0.052715808153152466, -0.03122149221599102, -0.12356983870267868, -0.07791973650455475, 0.041985854506492615, -0.10120090842247009, 0.050200384110212326, -0.017746655270457268, 0.0408901646733284, 0.10895685106515884, 0.10556401312351227, -0.013035662472248077, 0.07371609658002853, -0.014117266982793808, -0.0846620574593544, 0.0455034039914608, -0.04019617289304733, -0.010195761919021606, -0.002067578723654151, 0.049866363406181335, 0.05518952012062073, -0.004279361572116613, 0.06604710221290588, -0.056787341833114624, -0.05250666290521622, 0.12704309821128845, -0.09899132698774338, 0.005839226767420769, 0.026035251095891, -0.024846648797392845, 0.08802798390388489, 0.06112935394048691, -0.050732165575027466, -0.12834863364696503, 0.022644802927970886, -0.07076971232891083, -0.01149824820458889, -0.0833648294210434, -0.11049388349056244, 0.00907260924577713, -0.047179970890283585, -0.09414803236722946, -0.10061337053775787, -0.18461094796657562, -0.07595019787549973, 0.016597673296928406, -0.005480315536260605, 0.014699602499604225, 0.006810646038502455, -0.000002153011337213684, -0.020851897075772285, -0.007896769791841507, -0.13535623252391815, -0.012298873625695705, 0.04051767289638519, -0.04692593216896057, 0.059143662452697754, -0.01872301660478115, 0.027359778061509132, -0.16805782914161682, 0.009019589051604271, -0.19411751627922058, 0.092104472219944, -0.002607777714729309, 0.034167446196079254, -0.05094394087791443, -0.008947515860199928, -0.07941710203886032, 0.01227178331464529, -0.009641456417739391, 0.12141614407300949, -0.1206267774105072, -0.02940273843705654, 0.21456262469291687, -0.19963151216506958, 0.0037524071522057056, 0.08274208754301071, -0.013699131086468697, 0.09255952388048172, 0.16497701406478882, 0.04212379828095436, 0.13653689622879028, -0.04911814257502556, -0.08861054480075836, 0.000989542342722416, -0.09165287762880325, 0.07329905033111572, 0.03484426066279411, -0.06666473299264908, -0.007085772696882486, -0.005859557073563337, -0.020612407475709915, -0.03244755044579506, -0.013599338941276073, -0.023149538785219193, 0.0001982789544854313, -0.0166273545473814, -0.021063584834337234, 0.020443769171833992, 0.003965398296713829, 0.011658074334263802, -0.1031242161989212, -0.04542168229818344, 0.08622842282056808, -0.06987183541059494, 0.048840899020433426, -0.07760217040777206, 0.022125061601400375, -0.06578650325536728, -0.010971310548484325, -0.20360179245471954, -0.04289877414703369, 0.03670673444867134, -0.05019289255142212, 0.06837590038776398, 0.0508904755115509, 0.022545984014868736, 0.08772188425064087, -0.010456554591655731, -0.01850094459950924, -0.006283401045948267, -0.010208920575678349, -0.05834413692355156, -0.09543342888355255, -0.05288849025964737, -0.04560856893658638, 0.03134234622120857, -0.06112121790647507, 0.017759595066308975, 0.06061996892094612, -0.005643843207508326, 0.06101309135556221, -0.09006831794977188, 0.03141511604189873, 0.00443275086581707, -0.0014154657255858183, -0.03567962720990181, 0.03645949065685272, 0.05668117105960846, -0.03906779736280441, 0.027275027707219124, -0.18629911541938782, -0.12221936881542206, 0.06319569796323776, 0.004660504870116711, -0.16912294924259186, 0.02038245089352131, 0.005852261558175087, -0.011416464112699032, -0.057149261236190796, -0.07026511430740356, 0.1946265697479248, 0.004467998631298542, 0.09008118510246277, -0.04925358295440674, -0.009454086422920227, 0.031242823228240013, -0.03210488334298134, -0.0031244747806340456, 0.05200798064470291, 0.0005439313827082515, -0.15450836718082428, 0.06423503905534744, 0.004395703319460154, 0.01785152032971382, 0.1417408585548401, 0.03325621783733368, -0.0843934640288353, -0.02952333725988865, 0.034340932965278625, 0.02237013913691044, 0.022144611924886703, -0.05297551304101944, 0.0011478739324957132, 0.025459155440330505, 0.10170090198516846, 0.00989725161343813, -0.053374920040369034, 0.06699469685554504, 0.07153100520372391, -0.004389651119709015, -0.06895419955253601, -0.06259769946336746, -0.04891546070575714, 0.0790366530418396, 0.03605163097381592, 0.10585296899080276, 0.04058271646499634, -0.031815506517887115, -0.1524791419506073, 0.1501787006855011, -0.08999856561422348, -0.20171959698200226, -0.11323580145835876, -0.029958227649331093, 0.0346851646900177, 0.04765794053673744, 0.05109371617436409, -0.05138663202524185, -0.06766336411237717, -0.10653387755155563, 0.07603698968887329, -0.05629872530698776, -0.04880121350288391, -0.02269679307937622, -0.05581085756421089, -0.012269821017980576, -0.07655573636293411, 0.010962443426251411, -0.00048542555305175483, -0.10316085815429688, 0.023213563486933708, -0.04992373660206795, 0.06911959499120712, 0.16882486641407013, -0.010740355588495731, -0.018420884385704994, -0.00020395847968757153, 0.19130480289459229, -0.03968143090605736, 0.09190952777862549, 0.12357641756534576, -0.0746065005660057, 0.04931945353746414, 0.079559326171875, 0.004999913740903139, -0.022913342341780663, 0.04426591843366623, 0.010295701213181019, -0.08143842965364456, -0.12867344915866852, -0.03429488092660904, -0.044126957654953, 0.014809408225119114, 0.041371773928403854, 0.03868798539042473, 0.0322665274143219, 0.047807224094867706, -0.05942576751112938, -0.006975505035370588, 0.04464973881840706, 0.11809788644313812, -0.026486804708838463, -0.02428327687084675, 0.043879494071006775, -0.07838220149278641, 0.039814528077840805, 0.10044966638088226, -0.043350283056497574, 0.20909476280212402, -0.0290280282497406, 0.1923631727695465, 0.06402953714132309, 0.0024041219148784876, 0.093107208609581, 0.06099283695220947, -0.038052793592214584, 0.012073992751538754, -0.026593852788209915, -0.06759770959615707, -0.059939734637737274, -0.0028860894963145256, -0.06440668553113937, 0.030206803232431412, -0.12259376049041748, -0.04690324142575264, 0.019937701523303986, 0.18320639431476593, 0.027452655136585236, -0.15258412063121796, -0.1293516606092453, 0.02951151691377163, -0.004888654686510563, -0.08788600564002991, 0.014396497048437595, 0.07275895029306412, -0.09896501898765564, -0.012637863866984844, -0.011738710105419159, 0.08188703656196594, -0.14790695905685425, 0.015112643130123615, -0.027563918381929398, 0.030750697478652, -0.03907115384936333, 0.05782834440469742, -0.09298761188983917, 0.020872287452220917, 0.013061555102467537, 0.1148606538772583, -0.024818846955895424, -0.002578325569629669, -0.04483339935541153, 0.12865407764911652, 0.11058912426233292, 0.0431830957531929, -0.05849761143326759, -0.09322010725736618, -0.03949807211756706, 0.009033040143549442, 0.060251809656620026, -0.04523945599794388, 0.10578764230012894, -0.0005714800790883601, 0.03196774423122406, -0.026299629360437393, -0.006031502969563007, -0.08552104234695435, -0.11273477971553802, 0.052518635988235474, -0.0055497027933597565, 0.1069466695189476, -0.021290289238095284, -0.001609373721294105, 0.018941184505820274, 0.1795884370803833, -0.17860962450504303, -0.09379293024539948, -0.11843258142471313, -0.017955446615815163, 0.015231456607580185, -0.06048694998025894, -0.006482475437223911, -0.028554346412420273, 0.10834133625030518, 0.027132460847496986, -0.056812480092048645, 0.050488635897636414, -0.06254146993160248, -0.15650425851345062, -0.05541952699422836, 0.049885619431734085, 0.14149758219718933, 0.039108797907829285, -0.01949893869459629, 0.04004552215337753, -0.012397498823702335, -0.11244619637727737, -0.010135860182344913, 0.12755651772022247, 0.03594135493040085, 0.09491150826215744, -0.04685850813984871, -0.12190568447113037, -0.03232380747795105, -0.01658463291823864, 0.11691806465387344, 0.14453460276126862, -0.05871756747364998, 0.13025310635566711, 0.2649490237236023, -0.10729817301034927, -0.19360758364200592, 0.010643326677381992, 0.0061299726366996765, 0.0188958328217268, 0.0033949704375118017, -0.18178920447826385, 0.07717869430780411, 0.06782566010951996, 0.007232817355543375, 0.007590859197080135, -0.26613253355026245, -0.06733734160661697, 0.05649760365486145, 0.09489047527313232, 0.09888705611228943, -0.10372861474752426, -0.024666180834174156, -0.009402130730450153, -0.1341104507446289, 0.1384093314409256, -0.10752195119857788, 0.07272546738386154, 0.015329848974943161, -0.029202621430158615, 0.029145121574401855, -0.044565003365278244, 0.06682443618774414, 0.03353410214185715, 0.06285936385393143, -0.07301070541143417, 0.029553117230534554, 0.11589957028627396, -0.036223940551280975, 0.1317211389541626, 0.06748421490192413, 0.05256224051117897, -0.09078563749790192, -0.048322826623916626, -0.08735582232475281, 0.01548714004456997, -0.06280534714460373, -0.040054142475128174, -0.051278021186590195, 0.07947436720132828, 0.07226724922657013, -0.006103191524744034, 0.00851644016802311, -0.10162439942359924, 0.052692584693431854, 0.11841414123773575, 0.1282070279121399, 0.07742055505514145, -0.15537703037261963, -0.04369165375828743, -0.028508728370070457, 0.10344090312719345, -0.0651608482003212, 0.03755488619208336, 0.07178975641727448, 0.02791557088494301, 0.11240556836128235, 0.04749145731329918, -0.13588222861289978, 0.037396859377622604, 0.01615297794342041, -0.11619797348976135, -0.05642109736800194, 0.02413163334131241, 0.016995975747704506, -0.12152066826820374, 0.024082455784082413, 0.10681425034999847, -0.0635867714881897, 0.013295610435307026, -0.00670105405151844, 0.044655732810497284, 0.004165414720773697, 0.06812004745006561, 0.0071419053710997105, 0.01603415049612522, -0.0580202080309391, 0.11868447065353394, 0.10236486792564392, -0.09314025938510895, 0.039808791130781174, 0.10989047586917877, -0.12213359028100967, -0.08853495866060257, -0.06846896559000015, 0.08577515184879303, -0.011402628384530544, -0.06266690790653229, 0.018202926963567734, -0.08483488112688065, 0.06945431232452393, 0.0885210782289505, -0.02287193574011326, 0.08480077981948853, -0.056976981461048126, 0.030161581933498383, -0.09150256216526031, 0.03443494811654091, -0.03269883990287781, 0.009375369176268578, -0.061859939247369766, 0.16609816253185272, 0.041164133697748184, 0.014847992919385433, -0.014308629557490349, -0.10013077408075333, -0.09881220757961273, 0.004386632237583399, -0.05953716114163399, -0.005038957577198744, -0.03168472275137901, -0.025502707809209824, -0.019102027639746666, 0.034625094383955, 0.016591008752584457, 0.016380568966269493, -0.02924320474267006, -0.0004131196765229106, -0.021342279389500618, 0.018557902425527573, -0.08419565856456757, 0.029699724167585373, 0.03160632401704788, -0.03924137353897095, 0.08316349983215332, 0.02611183002591133, -0.03934931010007858, 0.02357451617717743, -0.081295907497406, 0.08992289006710052, -0.0471879281103611, -0.04869502782821655, -0.005235752090811729, -0.06945008784532547, -0.026010194793343544, 0.002955146599560976, -0.04802514612674713, 0.007070381194353104, 0.097808837890625, -0.07653790712356567, 0.10968402773141861, 0.030264029279351234, 0.010852226056158543, -0.12199801951646805, 0.055628933012485504, 0.016076555475592613, 0.033181894570589066, 0.09393054991960526, -0.036722127348184586, 0.08062737435102463, -0.12459800392389297, -0.013789252378046513, 0.07310207933187485, 0.030726158991456032, -0.021569669246673584, -0.026777775958180428, 0.043153077363967896, -0.024825507774949074, 0.020287662744522095, -0.015593762509524822, 0.018382441252470016, 0.02542419545352459, -0.037007614970207214, -0.08302121609449387, 0.009640620090067387, 0.04610806703567505, -0.012068087235093117, -0.057012178003787994, 0.01651059277355671, 0.027509670704603195, -0.06477928906679153, -0.013198421336710453, 0.18047906458377838, 0.036878637969493866, 0.06305929273366928, 0.02587207779288292, -0.07251999527215958, -0.029281625524163246, -0.10079176723957062, -0.0037770546041429043, -0.0046500759199261665, 0.032017216086387634, -0.028884977102279663, 0.07746922969818115, 0.1431746482849121, -0.022970471531152725, 0.10573678463697433, -0.008951821364462376, -0.06390636414289474, -0.06434512883424759, -0.2204858958721161, 0.025792552158236504, -0.0021036367397755384, -0.03503746911883354, -0.11225217580795288, 0.036510638892650604, 0.04690343141555786, 0.00843519065529108, -0.023609787225723267, 0.14830376207828522, -0.07852374762296677, -0.09117071330547333, 0.013406007550656796, -0.025194674730300903, 0.05163058266043663, 0.024029647931456566, 0.062296584248542786, 0.07180789113044739, 0.06538783013820648, 0.0722818523645401, 0.10800837725400925, 0.07844580709934235, 0.007584805600345135, -0.027683883905410767, -0.09270866960287094, -0.003943813499063253, 0.019332360476255417, 0.027151428163051605, 0.16795511543750763, 0.03214298561215401, -0.0374600812792778, -0.006576416548341513, 0.16651935875415802, -0.08111496269702911, -0.020925987511873245, -0.12603506445884705, 0.21349090337753296, 0.015614082105457783, -0.03239060938358307, 0.02092622220516205, -0.1269250214099884, 0.048199959099292755, 0.17646721005439758, 0.07940921187400818, 0.013050254434347153, 0.02756865881383419, -0.014133267104625702, 0.002095867646858096, 0.02431810460984707, 0.1173439621925354, 0.015391585417091846, 0.32499319314956665, -0.06646983325481415, 0.1849062293767929, -0.026502836495637894, 0.010358515195548534, -0.061947066336870193, 0.06032123044133186, -0.03189493715763092, 0.056204140186309814, -0.09026232361793518, 0.06553894281387329, -0.09533907473087311, -0.2331204116344452, 0.03962903469800949, 0.01863224059343338, -0.020513268187642097, -0.003710505086928606, -0.006665102671831846, 0.023325204849243164, 0.08192325383424759, 0.002363622421398759, 0.010062665678560734, 0.17985664308071136, 0.022860432043671608, -0.07156620919704437, -0.0775955468416214, 0.07765986025333405, -0.06693810224533081, 0.13646939396858215, 0.010486001148819923, 0.09318152815103531, 0.08411945402622223, -0.009364176541566849, -0.11709713190793991, 0.07445552945137024, -0.06303013861179352, -0.031207691878080368, -0.011620789766311646, 0.15169757604599, -0.004141881596297026, 0.13461372256278992, 0.0498359277844429, -0.08513712882995605, 0.012910147197544575, 0.01622478850185871, 0.02264299802482128, -0.08837918192148209, 0.05038238689303398, -0.030326740816235542, 0.15815937519073486, 0.16111715137958527, -0.009934988804161549, -0.019837886095046997, -0.04054621234536171, 0.014715026132762432, -0.01943824253976345, 0.053939275443553925, 0.0024755573831498623, -0.10393606871366501, 0.01235715951770544, 0.06427904218435287, 0.061906587332487106, -0.24948813021183014, -0.04159945249557495, 0.013469717465341091, -0.022035928443074226, -0.01368871983140707, 0.05697489529848099, 0.007251076400279999, 0.04534943774342537, -0.051117077469825745, 0.07974301278591156, 0.012608920224010944, 0.1257849484682083, -0.0846102237701416, -0.08220173418521881 ]
null
null
transformers
# Funnel Transformer intermediate model (B6-6-6 with decoder) Pretrained model on English language using a similar objective objective as [ELECTRA](https://huggingface.co/transformers/model_doc/electra.html). It was introduced in [this paper](https://arxiv.org/pdf/2006.03236.pdf) and first released in [this repository](https://github.com/laiguokun/Funnel-Transformer). This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=funnel-transformer) to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import FunnelTokenizer, FunnelModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/intermediate") model = FunneModel.from_pretrained("funnel-transformer/intermediate") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import FunnelTokenizer, TFFunnelModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/intermediate") model = TFFunnelModel.from_pretrained("funnel-transformer/intermediatesmall") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` ## Training data The BERT model was pretrained on: - [BookCorpus](https://yknzhu.wixsite.com/mbweb), a dataset consisting of 11,038 unpublished books, - [English Wikipedia](https://en.wikipedia.org/wiki/English_Wikipedia) (excluding lists, tables and headers), - [Clue Web](https://lemurproject.org/clueweb12/), a dataset of 733,019,372 English web pages, - [GigaWord](https://catalog.ldc.upenn.edu/LDC2011T07), an archive of newswire text data, - [Common Crawl](https://commoncrawl.org/), a dataset of raw web pages. ### BibTeX entry and citation info ```bibtex @misc{dai2020funneltransformer, title={Funnel-Transformer: Filtering out Sequential Redundancy for Efficient Language Processing}, author={Zihang Dai and Guokun Lai and Yiming Yang and Quoc V. Le}, year={2020}, eprint={2006.03236}, archivePrefix={arXiv}, primaryClass={cs.LG} } ```
{"language": "en", "license": "apache-2.0", "datasets": ["bookcorpus", "wikipedia", "gigaword"]}
feature-extraction
funnel-transformer/intermediate
[ "transformers", "pytorch", "tf", "safetensors", "funnel", "feature-extraction", "en", "dataset:bookcorpus", "dataset:wikipedia", "dataset:gigaword", "arxiv:2006.03236", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2006.03236" ]
[ "en" ]
TAGS #transformers #pytorch #tf #safetensors #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us
# Funnel Transformer intermediate model (B6-6-6 with decoder) Pretrained model on English language using a similar objective objective as ELECTRA. It was introduced in this paper and first released in this repository. This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: and in TensorFlow: ## Training data The BERT model was pretrained on: - BookCorpus, a dataset consisting of 11,038 unpublished books, - English Wikipedia (excluding lists, tables and headers), - Clue Web, a dataset of 733,019,372 English web pages, - GigaWord, an archive of newswire text data, - Common Crawl, a dataset of raw web pages. ### BibTeX entry and citation info
[ "# Funnel Transformer intermediate model (B6-6-6 with decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ "TAGS\n#transformers #pytorch #tf #safetensors #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us \n", "# Funnel Transformer intermediate model (B6-6-6 with decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ 77, 103, 206, 133, 32, 95, 11 ]
[ "passage: TAGS\n#transformers #pytorch #tf #safetensors #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us \n# Funnel Transformer intermediate model (B6-6-6 with decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs." ]
[ -0.07877761870622635, 0.0806143656373024, -0.0039480822160840034, 0.051602814346551895, 0.10760805010795593, -0.00195114326197654, 0.048399098217487335, 0.07863440364599228, -0.14395105838775635, 0.07349174469709396, -0.005998771637678146, -0.029404466971755028, 0.09100716561079025, 0.05465076118707657, 0.08332588523626328, -0.24436283111572266, 0.08314777910709381, -0.03970705717802048, 0.11030774563550949, 0.09407324343919754, 0.08927347511053085, -0.09456053376197815, 0.07160373032093048, 0.013903917744755745, -0.06653225421905518, -0.019826233386993408, 0.00007003499922575429, -0.0409378781914711, 0.04906412959098816, 0.08917345851659775, 0.09829742461442947, 0.004122805781662464, 0.04303142800927162, -0.06105384975671768, 0.025144003331661224, 0.08172793686389923, 0.01355485338717699, 0.04342939332127571, 0.021715547889471054, 0.010155211202800274, 0.09396157413721085, -0.004768379498273134, 0.06029914319515228, 0.06179165840148926, -0.11271068453788757, -0.2064758688211441, -0.0416613444685936, 0.12240204215049744, 0.011070936918258667, 0.11775818467140198, -0.03983364254236221, 0.15518218278884888, -0.03992033749818802, 0.05269993469119072, 0.14061011373996735, -0.1800365000963211, 0.0010166051797568798, 0.001217230106703937, -0.04234732687473297, 0.11597635596990585, -0.0290615726262331, -0.05739668011665344, -0.008207166567444801, 0.0442349873483181, 0.03885537385940552, 0.005339971277862787, -0.024205341935157776, -0.06054821237921715, -0.1327497810125351, -0.0674661174416542, 0.12563051283359528, -0.040714606642723083, -0.1178407222032547, -0.1041848286986351, -0.045532964169979095, -0.052774857729673386, 0.000007419939720421098, -0.0493960864841938, -0.007505671586841345, 0.02120753936469555, 0.03616003692150116, -0.020607788115739822, -0.10560434311628342, -0.01916203275322914, -0.08135909587144852, 0.0928734764456749, 0.04742584004998207, 0.05994155630469322, -0.10850376635789871, 0.10272365808486938, -0.09465128928422928, -0.05838455632328987, -0.07867015898227692, -0.0625331699848175, -0.04935922101140022, -0.02568216621875763, -0.04708482325077057, -0.11201110482215881, -0.04791045933961868, 0.03503277152776718, -0.08820293098688126, 0.039454761892557144, -0.014503940008580685, 0.028053829446434975, 0.11380881071090698, 0.11635667830705643, -0.045405223965644836, 0.13498961925506592, 0.010475910268723965, -0.05516373738646507, 0.04170877858996391, -0.04422668367624283, -0.04956439137458801, -0.02264924719929695, 0.04244659096002579, 0.0581611804664135, 0.004277688916772604, 0.060235414654016495, -0.03761103376746178, -0.06443043798208237, 0.12827306985855103, -0.11674609035253525, -0.010564823634922504, 0.04194650053977966, -0.03725990280508995, 0.1340785026550293, 0.05079817771911621, -0.03160514310002327, -0.13478930294513702, -0.01696043834090233, -0.06275871396064758, -0.02213135175406933, -0.07977177947759628, -0.10453265905380249, -0.0008839680813252926, -0.023870987817645073, -0.07930366694927216, -0.12500251829624176, -0.19148261845111847, -0.0415823720395565, 0.00507690291851759, -0.006690027192234993, 0.0027713181916624308, 0.01440039835870266, -0.00004301001172279939, -0.011369181796908379, 0.0024141937028616667, -0.12428513914346695, -0.005687591154128313, 0.04863932728767395, -0.05286875739693642, 0.035585250705480576, -0.03330022841691971, 0.04834022372961044, -0.16731329262256622, -0.014040707610547543, -0.19175094366073608, 0.12899087369441986, -0.0009715465712361038, -0.009043198078870773, -0.057371921837329865, -0.03829760476946831, -0.09961620718240738, 0.008427562192082405, -0.01792241632938385, 0.13304124772548676, -0.13771408796310425, -0.06497570127248764, 0.20399515330791473, -0.23628781735897064, 0.009694433771073818, 0.0928274467587471, -0.03256233409047127, 0.10600943863391876, 0.1765095293521881, 0.0370217002928257, 0.16711415350437164, -0.03400592878460884, -0.07769253849983215, 0.03349572420120239, -0.06224098056554794, 0.07607565075159073, 0.04356664419174194, -0.06359512358903885, -0.026599416509270668, 0.003340129042044282, -0.04626106843352318, -0.01196138933300972, -0.021901186555624008, -0.02660350874066353, -0.016459189355373383, -0.0016343491151928902, 0.019872291013598442, 0.030852988362312317, 0.02056793123483658, 0.029883064329624176, -0.1151280626654625, 0.012105545029044151, 0.10319354385137558, -0.08508969843387604, 0.040548138320446014, -0.11511552333831787, 0.04515878111124039, -0.046131934970617294, -0.008431513793766499, -0.19139012694358826, -0.022157378494739532, 0.03889116644859314, -0.06997769325971603, 0.06768413633108139, 0.04218859598040581, 0.0336272306740284, 0.064994677901268, -0.003051555948331952, -0.008291793055832386, -0.026511428877711296, -0.013494948856532574, -0.02835896611213684, -0.09529904276132584, -0.06008762866258621, -0.04566986858844757, 0.042669203132390976, -0.06985511630773544, 0.034926850348711014, 0.07012568414211273, -0.007291343994438648, 0.03561142832040787, -0.04942053556442261, 0.0001651768252486363, 0.00193815550301224, -0.01155765913426876, -0.038865022361278534, 0.039244286715984344, 0.03826034069061279, -0.06391102820634842, 0.05739147961139679, -0.19801244139671326, -0.08999192714691162, 0.09119604527950287, 0.031608644872903824, -0.13239924609661102, 0.022001076489686966, 0.006209350191056728, 0.00851654913276434, -0.028122613206505775, -0.08143250644207001, 0.20972932875156403, 0.00020004637190140784, 0.08955667912960052, -0.09776709973812103, -0.0184515081346035, 0.026217231526970863, -0.031509879976511, -0.011273723095655441, 0.04563842713832855, -0.013126634061336517, -0.10220005363225937, 0.0618528388440609, 0.051911525428295135, 0.025348855182528496, 0.15813829004764557, -0.0009414005326107144, -0.07363452762365341, 0.0013527105329558253, 0.012613150291144848, 0.01030734647065401, -0.02680770494043827, 0.018747709691524506, 0.02450406923890114, 0.036192961037158966, 0.09292805194854736, 0.03820523992180824, -0.0547981932759285, 0.06215113028883934, 0.09779386967420578, -0.019921831786632538, -0.04354739189147949, -0.03914438560605049, -0.022253289818763733, 0.08218882977962494, 0.03766729310154915, 0.11525111645460129, 0.04927486926317215, -0.011869067326188087, -0.15292704105377197, 0.16178163886070251, -0.10262108594179153, -0.19460074603557587, -0.12488368898630142, 0.009093946777284145, 0.03975459188222885, 0.040447428822517395, 0.05302465334534645, -0.10811308026313782, -0.06570567190647125, -0.11563806980848312, 0.13621889054775238, -0.04930702596902847, -0.04398969188332558, -0.016562558710575104, -0.062263693660497665, -0.019368350505828857, -0.11754021793603897, 0.006724075879901648, 0.00825490988790989, -0.12728281319141388, 0.00639354856684804, -0.07194250822067261, 0.012229339219629765, 0.18131287395954132, -0.009592565707862377, -0.004047028254717588, -0.016304219141602516, 0.23402772843837738, -0.011393556371331215, 0.08387637138366699, 0.21665677428245544, -0.08494437485933304, 0.04613150656223297, 0.058067262172698975, 0.00691273994743824, -0.008010434918105602, 0.0375506691634655, -0.009128094650804996, -0.07746238261461258, -0.16042324900627136, -0.06379365175962448, -0.012907528318464756, -0.016555728390812874, 0.014754477888345718, 0.04493894800543785, 0.020004281774163246, 0.06286103278398514, -0.06326743960380554, -0.024420328438282013, 0.07080725580453873, 0.08550848066806793, 0.043395496904850006, -0.021104538813233376, 0.06947436183691025, -0.08123879134654999, 0.02657410316169262, 0.10019301623106003, -0.08406123518943787, 0.19719070196151733, -0.03405071049928665, 0.1991186887025833, 0.06530934572219849, 0.026677921414375305, 0.12382327765226364, 0.057646963745355606, -0.043492887169122696, 0.007338196039199829, -0.024687418714165688, -0.07869341224431992, -0.07976514101028442, -0.014315765351057053, -0.05684591084718704, 0.05578088387846947, -0.1355118751525879, -0.019952598959207535, 0.0037211354356259108, 0.1737080216407776, -0.016285095363855362, -0.19573701918125153, -0.1397743672132492, 0.02677765302360058, -0.0047227623872458935, -0.06727927178144455, 0.01764320209622383, 0.05930826812982559, -0.1255616545677185, -0.012785281985998154, 0.00710182124748826, 0.07498664408922195, -0.11931999027729034, 0.027045516297221184, -0.04796432703733444, 0.030795373022556305, -0.01796494796872139, 0.07205067574977875, -0.1032867506146431, 0.019474511966109276, 0.01166493073105812, 0.0955338254570961, -0.04664841666817665, 0.01940789259970188, -0.030091678723692894, 0.14759130775928497, 0.09556488692760468, 0.02258547581732273, -0.03327365219593048, -0.09742096066474915, -0.041499413549900055, 0.036939818412065506, 0.05241627246141434, -0.0418131947517395, 0.0903313159942627, -0.048642054200172424, 0.054315295070409775, -0.0034903038758784533, 0.021085988730192184, -0.10287299752235413, -0.1333637535572052, 0.04238764941692352, -0.02527397871017456, 0.11233863979578018, -0.04721301421523094, -0.04786735773086548, -0.026041418313980103, 0.140121191740036, -0.15574103593826294, -0.12124151736497879, -0.12170691043138504, -0.029534969478845596, 0.04472329467535019, -0.07847215980291367, 0.0208702702075243, -0.03606663644313812, 0.12105462700128555, 0.020044617354869843, -0.10678759962320328, 0.0201432928442955, -0.05276349559426308, -0.150821715593338, -0.012178200297057629, 0.04025867208838463, 0.16491779685020447, 0.0465066097676754, 0.006915218196809292, 0.02434735931456089, -0.01648503914475441, -0.10797909647226334, -0.05188344419002533, 0.17135410010814667, 0.005278053693473339, 0.08667036890983582, -0.041387349367141724, -0.14208820462226868, -0.03317435458302498, 0.00575308408588171, 0.14300651848316193, 0.12248224020004272, -0.06755327433347702, 0.13022208213806152, 0.2571095824241638, -0.09954754263162613, -0.20946873724460602, -0.003383548464626074, 0.01631392538547516, 0.0303184874355793, 0.0140986954793334, -0.192316934466362, 0.08718635886907578, 0.06399072706699371, 0.00003841496436507441, -0.07080738991498947, -0.31683865189552307, -0.07145683467388153, 0.11239447444677353, 0.07547812163829803, 0.1120183989405632, -0.08813236653804779, -0.01772613637149334, -0.007348302286118269, -0.09555529057979584, 0.15768323838710785, -0.1395648717880249, 0.04817594960331917, 0.013690534047782421, -0.01441079843789339, 0.03239903971552849, -0.025014828890562057, 0.0644214004278183, -0.004860250744968653, 0.056159891188144684, -0.08642715960741043, 0.03100195713341236, 0.10399805009365082, -0.013453667983412743, 0.1315017193555832, 0.05978235602378845, 0.06791650503873825, -0.07267840206623077, -0.06252463907003403, -0.09169252961874008, 0.0493796169757843, -0.056121766567230225, -0.0475178137421608, -0.07170050591230392, 0.06252305209636688, 0.04998159781098366, -0.008806440979242325, 0.00481946999207139, -0.09530491381883621, 0.0953063890337944, 0.11374221742153168, 0.16937458515167236, 0.07142326980829239, -0.12450242787599564, -0.02981981821358204, -0.0212713610380888, 0.11753806471824646, -0.0556982085108757, 0.05359413102269173, 0.05970029905438423, 0.011458219029009342, 0.09349441528320312, 0.04821068421006203, -0.14783217012882233, 0.02287214994430542, 0.00021811893384438008, -0.09668665379285812, -0.11007590591907501, 0.01492543239146471, 0.04336349293589592, -0.11645133048295975, -0.0006386800669133663, 0.12253675609827042, -0.07269944995641708, -0.017362920567393303, -0.013521228916943073, 0.028707191348075867, 0.011915981769561768, 0.06718331575393677, 0.030705060809850693, 0.0114463334903121, -0.059775497764348984, 0.09650910645723343, 0.09908328205347061, -0.10846804827451706, 0.06912200897932053, 0.0645529180765152, -0.10684717446565628, -0.08361247926950455, -0.021806053817272186, 0.08005813509225845, -0.02309286966919899, -0.08755229413509369, 0.02603718638420105, -0.11566676944494247, 0.06196556240320206, 0.08799532055854797, 0.005224588327109814, 0.07806836813688278, -0.07717390358448029, 0.023690538480877876, -0.07409567385911942, 0.04123755171895027, -0.0241374671459198, -0.014501371420919895, -0.07867737114429474, 0.1572933942079544, 0.06616944819688797, 0.033846426755189896, -0.030389804393053055, -0.0907023623585701, -0.09414790570735931, 0.00036255401209928095, -0.09470874816179276, 0.018267599865794182, -0.017756683751940727, -0.020688427612185478, -0.020368702709674835, 0.02105959877371788, 0.022359997034072876, 0.025866534560918808, -0.04251721128821373, 0.007248986978083849, -0.029549993574619293, 0.029962176457047462, -0.08942288160324097, 0.06580277532339096, 0.03657876327633858, -0.018490564078092575, 0.10356120765209198, 0.04175340756773949, -0.06868337839841843, 0.045299481600522995, -0.07798069715499878, 0.04453841596841812, -0.053525809198617935, -0.029688801616430283, -0.015612629242241383, -0.07347431778907776, -0.006912174168974161, 0.02466987632215023, -0.06629112362861633, 0.009896648116409779, 0.12909092009067535, -0.07852789759635925, 0.12390433251857758, 0.04730331525206566, 0.030956313014030457, -0.09748335182666779, 0.04737190157175064, -0.005458551459014416, 0.029565809294581413, 0.10753124207258224, -0.04310838133096695, 0.0707373097538948, -0.13670934736728668, -0.03088844195008278, 0.02987620234489441, 0.04314963147044182, -0.06104457750916481, -0.04083269461989403, 0.025574639439582825, -0.020107408985495567, 0.04379298910498619, -0.009413783438503742, 0.012364991009235382, 0.02173522301018238, -0.009257583878934383, -0.04114038497209549, 0.018971310928463936, 0.10482019931077957, 0.0016009940300136805, -0.03837290406227112, 0.033992987126111984, 0.006554692052304745, -0.0754368007183075, -0.028754469007253647, 0.17212198674678802, 0.07256139814853668, 0.05272740498185158, 0.02230611816048622, -0.033185601234436035, -0.010150219313800335, -0.06859449297189713, -0.04446328803896904, -0.0072888839058578014, 0.0007410926045849919, -0.022865010425448418, 0.09584178030490875, 0.1708192378282547, -0.02205795980989933, 0.10650470852851868, -0.009540568105876446, -0.05035362392663956, -0.12569580972194672, -0.23920275270938873, -0.00033797722426243126, -0.047121427953243256, -0.03476502001285553, -0.10536716878414154, 0.005694267340004444, 0.07663992047309875, 0.030210772529244423, -0.045577485114336014, 0.14494837820529938, -0.09109645336866379, -0.1160946786403656, -0.007147370837628841, -0.024624459445476532, 0.04792733117938042, 0.011718282476067543, 0.052380338311195374, 0.0939282774925232, 0.07657065242528915, 0.0680563747882843, 0.1086750477552414, 0.07533884048461914, 0.01490518357604742, -0.06600230187177658, -0.07331506907939911, -0.002275362377986312, 0.019252588972449303, 0.007840760983526707, 0.14137843251228333, 0.01926802098751068, -0.05736459419131279, -0.014819708652794361, 0.1539592295885086, -0.05884777009487152, -0.07111764699220657, -0.11374541372060776, 0.24626915156841278, -0.004496793262660503, 0.0007895802846178412, 0.019525330513715744, -0.09960716962814331, 0.00730931106954813, 0.1804536134004593, 0.13573184609413147, 0.0039020548574626446, 0.030834734439849854, -0.00975059624761343, 0.003765581175684929, 0.007557238452136517, 0.15571939945220947, 0.033918801695108414, 0.32191532850265503, -0.07813803106546402, 0.18584245443344116, -0.042576611042022705, 0.004592848476022482, -0.07802468538284302, 0.08220738917589188, -0.05289503186941147, 0.028863707557320595, -0.06439299881458282, 0.06050535663962364, -0.1050591990351677, -0.17515060305595398, -0.021133022382855415, -0.0019022875931113958, -0.022303229197859764, -0.006878265645354986, -0.005281501915305853, 0.01856229640543461, 0.0625925362110138, -0.00018453079974278808, 0.0015780520625412464, 0.14670929312705994, 0.026936665177345276, -0.07702787220478058, -0.0642232894897461, 0.11030203849077225, -0.07508910447359085, 0.1223684623837471, 0.019059520214796066, 0.11994994431734085, 0.08112931251525879, -0.0013429379323497415, -0.11433180421590805, 0.06164315715432167, -0.04716015234589577, 0.003741904627531767, -0.010384873487055302, 0.09305566549301147, 0.01937536709010601, 0.116378054022789, 0.05329480394721031, -0.11284970492124557, 0.030527731403708458, -0.009509333409368992, -0.0111605916172266, -0.07148156315088272, 0.04762551560997963, -0.04199153929948807, 0.13963286578655243, 0.1517711579799652, -0.014582475647330284, -0.03379027545452118, -0.05842336267232895, 0.0339813232421875, -0.018957803025841713, 0.04949142038822174, 0.00500476686283946, -0.11835405230522156, -0.0035484605468809605, 0.04519765451550484, 0.05922836437821388, -0.23018376529216766, -0.04057270288467407, 0.007285906467586756, -0.021999917924404144, -0.023098450154066086, 0.05310387536883354, 0.021289974451065063, 0.053577251732349396, -0.036404650658369064, 0.09970943629741669, -0.012200947850942612, 0.10379459708929062, -0.09557478129863739, -0.0689840316772461 ]
null
null
transformers
# Funnel Transformer large model (B8-8-8 without decoder) Pretrained model on English language using a similar objective objective as [ELECTRA](https://huggingface.co/transformers/model_doc/electra.html). It was introduced in [this paper](https://arxiv.org/pdf/2006.03236.pdf) and first released in [this repository](https://github.com/laiguokun/Funnel-Transformer). This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. **Note:** This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth of the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if you need one input per initial token. You should use the `large` model in that case. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=funnel-transformer) to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import FunnelTokenizer, FunnelBaseModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/large-base") model = FunnelBaseModel.from_pretrained("funnel-transformer/large-base") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import FunnelTokenizer, TFFunnelBaseModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/large-base") model = TFFunnelBaseModel.from_pretrained("funnel-transformer/large-base") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` ## Training data The BERT model was pretrained on: - [BookCorpus](https://yknzhu.wixsite.com/mbweb), a dataset consisting of 11,038 unpublished books, - [English Wikipedia](https://en.wikipedia.org/wiki/English_Wikipedia) (excluding lists, tables and headers), - [Clue Web](https://lemurproject.org/clueweb12/), a dataset of 733,019,372 English web pages, - [GigaWord](https://catalog.ldc.upenn.edu/LDC2011T07), an archive of newswire text data, - [Common Crawl](https://commoncrawl.org/), a dataset of raw web pages. ### BibTeX entry and citation info ```bibtex @misc{dai2020funneltransformer, title={Funnel-Transformer: Filtering out Sequential Redundancy for Efficient Language Processing}, author={Zihang Dai and Guokun Lai and Yiming Yang and Quoc V. Le}, year={2020}, eprint={2006.03236}, archivePrefix={arXiv}, primaryClass={cs.LG} } ```
{"language": "en", "license": "apache-2.0", "datasets": ["bookcorpus", "wikipedia", "gigaword"]}
feature-extraction
funnel-transformer/large-base
[ "transformers", "pytorch", "tf", "funnel", "feature-extraction", "en", "dataset:bookcorpus", "dataset:wikipedia", "dataset:gigaword", "arxiv:2006.03236", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2006.03236" ]
[ "en" ]
TAGS #transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us
# Funnel Transformer large model (B8-8-8 without decoder) Pretrained model on English language using a similar objective objective as ELECTRA. It was introduced in this paper and first released in this repository. This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. Note: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth of the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if you need one input per initial token. You should use the 'large' model in that case. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: and in TensorFlow: ## Training data The BERT model was pretrained on: - BookCorpus, a dataset consisting of 11,038 unpublished books, - English Wikipedia (excluding lists, tables and headers), - Clue Web, a dataset of 733,019,372 English web pages, - GigaWord, an archive of newswire text data, - Common Crawl, a dataset of raw web pages. ### BibTeX entry and citation info
[ "# Funnel Transformer large model (B8-8-8 without decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nNote: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth\nof the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if\nyou need one input per initial token. You should use the 'large' model in that case.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ "TAGS\n#transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us \n", "# Funnel Transformer large model (B8-8-8 without decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nNote: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth\nof the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if\nyou need one input per initial token. You should use the 'large' model in that case.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ 72, 102, 288, 133, 32, 95, 11 ]
[ "passage: TAGS\n#transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us \n# Funnel Transformer large model (B8-8-8 without decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nNote: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth\nof the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if\nyou need one input per initial token. You should use the 'large' model in that case." ]
[ -0.08509158343076706, 0.13337621092796326, -0.003956535831093788, 0.04026637598872185, 0.11322996765375137, -0.007403940428048372, 0.04114893823862076, 0.06030670553445816, -0.07223165780305862, 0.10341677069664001, -0.02855512872338295, -0.0483456626534462, 0.09565079212188721, 0.05713433772325516, 0.07933434844017029, -0.20990756154060364, 0.06472504138946533, -0.03458240628242493, 0.12517668306827545, 0.08910307288169861, 0.049179550260305405, -0.09388680756092072, 0.06527996063232422, -0.02278853952884674, -0.06577128916978836, -0.03272102400660515, -0.009365299716591835, -0.022188439965248108, 0.02878202125430107, 0.05365830287337303, 0.11163776367902756, -0.0014013063628226519, 0.03094651736319065, -0.0944070890545845, 0.012463237158954144, 0.0861830934882164, 0.004565797280520201, 0.043074458837509155, 0.028206689283251762, 0.007121452130377293, 0.09396955370903015, -0.02890770696103573, 0.05999656021595001, 0.06194445118308067, -0.15032704174518585, -0.16540193557739258, -0.057576876133680344, 0.11874853074550629, 0.06735299527645111, 0.1142655536532402, -0.038928937166929245, 0.10065672546625137, -0.014973829500377178, 0.06935672461986542, 0.11867744475603104, -0.18378107249736786, -0.00010656953236320987, -0.02686396799981594, -0.024700956419110298, 0.10618259757757187, -0.07107681035995483, -0.06579902768135071, -0.013911291025578976, 0.02808862365782261, 0.05142360180616379, -0.02695923112332821, -0.04540777951478958, -0.08287356048822403, -0.13517145812511444, -0.05668409541249275, 0.16642358899116516, 0.002100659068673849, -0.13468705117702484, -0.04910883679986, -0.051921296864748, -0.14157474040985107, 0.013092316687107086, -0.017930077388882637, 0.007782122120261192, 0.012841914780437946, 0.01782643422484398, -0.01714312843978405, -0.11608964949846268, -0.0296722874045372, -0.04438627511262894, 0.11192280054092407, 0.0207389947026968, 0.0384335070848465, -0.05957544222474098, 0.1084815263748169, -0.05915743485093117, -0.05336165055632591, -0.08787038177251816, -0.02760864421725273, -0.1347523033618927, -0.05772474408149719, -0.036039844155311584, -0.11638811975717545, -0.08078990131616592, 0.05340394750237465, -0.10812720656394958, 0.04728083685040474, -0.013980713672935963, 0.040235359221696854, 0.10886383056640625, 0.11003540456295013, 0.003296404844149947, 0.06979943811893463, -0.01631172187626362, -0.08118505775928497, 0.049163155257701874, -0.042161546647548676, -0.014001351781189442, 0.0021059426944702864, 0.0516713410615921, 0.053287580609321594, -0.006242040544748306, 0.05506492778658867, -0.0531536266207695, -0.04931629076600075, 0.11924964189529419, -0.09799488633871078, 0.0019664994906634092, 0.027596386149525642, -0.02014489844441414, 0.08700194209814072, 0.06209941580891609, -0.0475441999733448, -0.12729507684707642, 0.015229947865009308, -0.06992128491401672, 0.0010369932278990746, -0.08274570852518082, -0.10685274004936218, 0.013931717723608017, -0.056595996022224426, -0.08911888301372528, -0.10105687379837036, -0.17914754152297974, -0.08301416784524918, 0.017996003851294518, -0.0011432492174208164, 0.013018544763326645, -0.00007342743629124016, 0.0008505680016241968, -0.025213254615664482, -0.010013985447585583, -0.14119228720664978, -0.01657770946621895, 0.038454070687294006, -0.040321916341781616, 0.05916741490364075, -0.01913384534418583, 0.03001822531223297, -0.16901670396327972, 0.0037630447186529636, -0.18049326539039612, 0.09596876800060272, -0.006589666940271854, 0.04702397808432579, -0.048943255096673965, -0.004635889548808336, -0.07305197417736053, 0.020853007212281227, -0.011646551080048084, 0.11914356797933578, -0.09821312129497528, -0.039834294468164444, 0.2094273418188095, -0.19414861500263214, 0.0002536638348829001, 0.09031642973423004, -0.012173759751021862, 0.10309990495443344, 0.16827471554279327, 0.05061706155538559, 0.13016322255134583, -0.04572379216551781, -0.08888900279998779, 0.01228309329599142, -0.08986110240221024, 0.05266726389527321, 0.03687891736626625, -0.0671294704079628, -0.004651034716516733, -0.011396767571568489, -0.013266269117593765, -0.029961688444018364, -0.010255740024149418, -0.027823299169540405, 0.0023756714072078466, -0.01578536629676819, -0.03311409428715706, 0.030508657917380333, 0.010650458745658398, 0.01170205045491457, -0.10818437486886978, -0.024743160232901573, 0.08050594478845596, -0.07324295490980148, 0.05272373557090759, -0.07354521006345749, 0.0138715710490942, -0.07488363236188889, -0.010514304973185062, -0.2067730575799942, -0.04534687474370003, 0.03459596633911133, -0.04972288757562637, 0.0726671889424324, 0.06439366191625595, 0.020682891830801964, 0.07928565889596939, -0.010180278681218624, -0.01802246831357479, -0.012734923511743546, -0.0135747529566288, -0.06910625845193863, -0.09282170236110687, -0.05622262507677078, -0.049406323581933975, 0.046807657927274704, -0.06725845485925674, 0.02247217670083046, 0.06151553988456726, -0.02162281423807144, 0.05830948054790497, -0.08810589462518692, 0.028933480381965637, 0.011786295101046562, -0.001997696002945304, -0.03455224633216858, 0.04298296198248863, 0.05982058122754097, -0.04936691373586655, 0.015396890230476856, -0.18997076153755188, -0.11683550477027893, 0.06289362162351608, -0.0033673024736344814, -0.1678343564271927, 0.00951747503131628, 0.001769279595464468, -0.01393882930278778, -0.0596446730196476, -0.06742709875106812, 0.18725436925888062, -0.0024974248372018337, 0.08594337850809097, -0.04384215176105499, -0.013333058916032314, 0.02286284975707531, -0.03270459547638893, -0.0005852586473338306, 0.05422263965010643, 0.006425811443477869, -0.15200896561145782, 0.07122824341058731, 0.0012852568179368973, 0.008395963348448277, 0.14636129140853882, 0.036022257059812546, -0.07992667704820633, -0.02588636428117752, 0.03446698561310768, 0.018599193543195724, 0.024999966844916344, -0.0740639865398407, -0.004486115649342537, 0.024721132591366768, 0.09918426722288132, 0.01014642883092165, -0.062201663851737976, 0.06853581964969635, 0.07366222143173218, -0.007358175236731768, -0.0677599236369133, -0.07304007560014725, -0.058813564479351044, 0.0748506635427475, 0.04252082109451294, 0.10828332602977753, 0.045309823006391525, -0.030562516301870346, -0.15738409757614136, 0.15301983058452606, -0.07738903164863586, -0.19003354012966156, -0.11289342492818832, -0.031957246363162994, 0.04169266298413277, 0.05344917252659798, 0.059684157371520996, -0.05299023166298866, -0.06551846116781235, -0.10575547814369202, 0.06921476125717163, -0.058502376079559326, -0.05405429005622864, -0.02141827903687954, -0.052994970232248306, -0.013519329950213432, -0.06899947673082352, 0.013355433009564877, -0.006172560155391693, -0.10928259789943695, 0.014601367525756359, -0.05345369502902031, 0.07368378341197968, 0.17473982274532318, -0.0007231972995214164, -0.017136938869953156, -0.006701607257127762, 0.18219085037708282, -0.046224091202020645, 0.08493354171514511, 0.11940577626228333, -0.06691556423902512, 0.04441778361797333, 0.07880489528179169, 0.0021761010866612196, -0.022102074697613716, 0.041417594999074936, 0.01661258563399315, -0.09311670809984207, -0.13288645446300507, -0.026533132418990135, -0.04375888407230377, 0.015145854093134403, 0.038111988455057144, 0.03657389432191849, 0.0477939248085022, 0.05666155368089676, -0.0633847787976265, -0.010235070250928402, 0.03635786846280098, 0.11720148473978043, -0.02057211473584175, -0.01805833913385868, 0.04959479719400406, -0.07949098944664001, 0.04517573490738869, 0.09696530550718307, -0.0513797402381897, 0.21195600926876068, -0.03169308230280876, 0.18677382171154022, 0.06058202311396599, -0.0003275115741416812, 0.0779808759689331, 0.053001951426267624, -0.04172612726688385, 0.011762107722461224, -0.028899431228637695, -0.06253892183303833, -0.06588740646839142, 0.008862593211233616, -0.06640227138996124, 0.024225806817412376, -0.13014690577983856, -0.05079168453812599, 0.01828441023826599, 0.183964341878891, 0.02740728110074997, -0.15184098482131958, -0.13456355035305023, 0.036677755415439606, 0.00025144926621578634, -0.08280818164348602, 0.014995191246271133, 0.08122568577528, -0.0940047949552536, -0.009189186617732048, -0.012000048533082008, 0.08092479407787323, -0.1482715606689453, 0.013604569248855114, -0.028069524094462395, 0.03546198084950447, -0.03930571302771568, 0.06208910793066025, -0.09026088565587997, 0.019635405391454697, 0.010769756510853767, 0.12663301825523376, -0.024748055264353752, 0.0026072741020470858, -0.0436776801943779, 0.11993372440338135, 0.10691747069358826, 0.042883217334747314, -0.07469170540571213, -0.09997417032718658, -0.033392664045095444, 0.01574227772653103, 0.04897686466574669, -0.04535574093461037, 0.10657811909914017, 0.0023254500702023506, 0.028394466266036034, -0.023470349609851837, -0.01436981838196516, -0.07788502424955368, -0.11506368219852448, 0.048716526478528976, 0.002736905822530389, 0.10767200589179993, -0.01451362669467926, -0.00043269674642942846, 0.036572884768247604, 0.17641346156597137, -0.17740413546562195, -0.09735537320375443, -0.11745253205299377, -0.008711392059922218, -0.0010035985615104437, -0.06217740476131439, -0.006827280391007662, -0.024773983284831047, 0.09904159605503082, 0.02478594146668911, -0.049802515655756, 0.05822312459349632, -0.06421226263046265, -0.15285000205039978, -0.04889075830578804, 0.055434875190258026, 0.14612948894500732, 0.041799478232860565, -0.019941706210374832, 0.0365951843559742, -0.0102496687322855, -0.11215461045503616, -0.003265086095780134, 0.12526476383209229, 0.027838829904794693, 0.09999795258045197, -0.05121435225009918, -0.12156373262405396, -0.03489362448453903, -0.021260686218738556, 0.11910554766654968, 0.1473701149225235, -0.053076740354299545, 0.12822505831718445, 0.2690357565879822, -0.11309657245874405, -0.18714940547943115, 0.018882717937231064, 0.0025459756143391132, 0.01642758585512638, 0.0021035184618085623, -0.189838707447052, 0.08212582767009735, 0.07042246311903, 0.006502986885607243, 0.008538288064301014, -0.26695749163627625, -0.06488244235515594, 0.05328600853681564, 0.09920623898506165, 0.1167863979935646, -0.10876437276601791, -0.022984016686677933, -0.011315981857478619, -0.12530547380447388, 0.14753951132297516, -0.11078561842441559, 0.07813744992017746, 0.016683151945471764, -0.038932789117097855, 0.028663290664553642, -0.04436046630144119, 0.06623296439647675, 0.04198122397065163, 0.062482528388500214, -0.07329827547073364, 0.036456767469644547, 0.10724927484989166, -0.041996728628873825, 0.1279149204492569, 0.06161028519272804, 0.03740396350622177, -0.10626727342605591, -0.05588182806968689, -0.09058991074562073, 0.022135648876428604, -0.06420913338661194, -0.038935814052820206, -0.05054460093379021, 0.07473494857549667, 0.0747029110789299, -0.0010190809844061732, 0.0218331478536129, -0.11118771880865097, 0.06488718837499619, 0.1126987636089325, 0.1340443342924118, 0.06644374132156372, -0.14986830949783325, -0.04626442864537239, -0.027900036424398422, 0.1031988337635994, -0.08002763241529465, 0.03125372529029846, 0.07144544273614883, 0.02953554317355156, 0.11278076469898224, 0.052147842943668365, -0.13256117701530457, 0.04719405248761177, 0.011957703158259392, -0.10628300160169601, -0.06368064135313034, 0.02331223152577877, 0.00817903969436884, -0.13036459684371948, 0.030467219650745392, 0.10910903662443161, -0.06264104694128036, 0.013159398920834064, -0.009114679880440235, 0.045962054282426834, 0.008350755088031292, 0.06893222033977509, 0.009032822214066982, 0.013006357476115227, -0.05448430776596069, 0.11041293293237686, 0.09959074854850769, -0.09970659017562866, 0.04647646099328995, 0.1184055507183075, -0.1193440705537796, -0.09133272618055344, -0.07365266978740692, 0.0867873951792717, -0.030140932649374008, -0.06459572911262512, 0.018143627792596817, -0.09040894359350204, 0.07297881692647934, 0.0938946083188057, -0.01633661612868309, 0.07769254595041275, -0.05420741066336632, 0.04329822584986687, -0.09895047545433044, 0.03040158562362194, -0.039982445538043976, 0.008249490521848202, -0.05887427181005478, 0.1675858199596405, 0.043671928346157074, 0.01903451420366764, -0.012959381565451622, -0.10498958826065063, -0.10247378051280975, 0.006210905499756336, -0.0548417791724205, -0.013701402582228184, -0.037503309547901154, -0.022580120712518692, -0.021894369274377823, 0.03360053896903992, 0.012877743691205978, 0.019851909950375557, -0.02893833816051483, -0.0004681193095166236, -0.022315112873911858, 0.01909704878926277, -0.0827997550368309, 0.030063416808843613, 0.028298620134592056, -0.04640503227710724, 0.08377212285995483, 0.027738505974411964, -0.039843399077653885, 0.02166919596493244, -0.08792311698198318, 0.09709346294403076, -0.04152809455990791, -0.05247339606285095, -0.002837574342265725, -0.06390946358442307, -0.025332216173410416, 0.002069003414362669, -0.03733474761247635, 0.005871387664228678, 0.09194781631231308, -0.07600634545087814, 0.115127794444561, 0.027960501611232758, 0.0023242298047989607, -0.1221715658903122, 0.06072698161005974, 0.013796729035675526, 0.02952287346124649, 0.09409020841121674, -0.03627517446875572, 0.0843406617641449, -0.12763477861881256, -0.013556101359426975, 0.07685259729623795, 0.028175925835967064, -0.020038140937685966, -0.032380931079387665, 0.04458818957209587, -0.0155708072707057, 0.017487874254584312, -0.019169596955180168, 0.03510414436459541, 0.028700176626443863, -0.04608737677335739, -0.05870326980948448, 0.0044769723899662495, 0.046687137335538864, -0.00541047053411603, -0.05945761874318123, 0.03308100625872612, 0.031776752322912216, -0.052724141627550125, -0.013534054160118103, 0.18563294410705566, 0.0328875295817852, 0.0542977936565876, 0.035823509097099304, -0.0780312716960907, -0.029473265632987022, -0.10281987488269806, 0.008845512755215168, 0.004957527853548527, 0.034472350031137466, -0.0279534999281168, 0.0781191810965538, 0.1446288377046585, -0.020795010030269623, 0.10191436856985092, -0.0050028539262712, -0.06398724764585495, -0.06628153473138809, -0.2061779499053955, 0.02727261371910572, -0.007543056737631559, -0.03705977648496628, -0.10850516706705093, 0.026827316731214523, 0.06225360929965973, 0.004002527799457312, -0.020348073914647102, 0.1497683823108673, -0.06764388829469681, -0.08172813802957535, 0.00805845856666565, -0.03533343970775604, 0.04760149121284485, 0.019189851358532906, 0.05424271523952484, 0.0763554647564888, 0.06821359694004059, 0.07497966289520264, 0.10626542568206787, 0.07018501311540604, 0.013906668871641159, -0.02547595277428627, -0.0899253711104393, -0.0027060015127062798, 0.024247393012046814, 0.0328228659927845, 0.1571069359779358, 0.032598529011011124, -0.040055546909570694, 0.0000888035210664384, 0.17227348685264587, -0.08030664920806885, -0.0188534464687109, -0.12345994263887405, 0.2113489955663681, 0.022247640416026115, -0.03701888024806976, 0.013983082957565784, -0.12380184233188629, 0.05566161870956421, 0.17118142545223236, 0.06178908050060272, 0.018760383129119873, 0.02199096977710724, -0.026370283216238022, 0.0009084442281164229, 0.024234486743807793, 0.10818150639533997, 0.018107660114765167, 0.3354184329509735, -0.06726019829511642, 0.1805253028869629, -0.01816728338599205, 0.008752057328820229, -0.0686226636171341, 0.050887905061244965, -0.025570373982191086, 0.06280649453401566, -0.0837278664112091, 0.06626657396554947, -0.08993234485387802, -0.22984705865383148, 0.050686631351709366, 0.025965161621570587, -0.02068129926919937, -0.0018154368735849857, 0.009161412715911865, 0.023148493841290474, 0.08417598903179169, 0.0016637343214824796, 0.007241051644086838, 0.18279804289340973, 0.025084758177399635, -0.08017490804195404, -0.07382059842348099, 0.07755158841609955, -0.05710591748356819, 0.13710330426692963, 0.003909047693014145, 0.08878455311059952, 0.08664697408676147, -0.010395017452538013, -0.11444955319166183, 0.07134287804365158, -0.06421102583408356, -0.03633454069495201, -0.00370686873793602, 0.15206092596054077, -0.012116880156099796, 0.14115774631500244, 0.05432697758078575, -0.09149175882339478, 0.012678073719143867, 0.012646356597542763, 0.015393815003335476, -0.08458124846220016, 0.03982129693031311, -0.036935631185770035, 0.15839450061321259, 0.16028492152690887, -0.010527675971388817, -0.023358773440122604, -0.03867470100522041, 0.009862053208053112, -0.022184543311595917, 0.06129137799143791, 0.0012792657362297177, -0.10535703599452972, 0.014917047694325447, 0.06675396114587784, 0.05700794979929924, -0.25565171241760254, -0.04078518971800804, 0.014379885978996754, -0.016513744369149208, -0.012846955098211765, 0.05023254454135895, 0.0032534513156861067, 0.04325869306921959, -0.0583144836127758, 0.06428106874227524, 0.012934593483805656, 0.13524436950683594, -0.0859699621796608, -0.082684226334095 ]
null
null
transformers
# Funnel Transformer large model (B8-8-8 with decoder) Pretrained model on English language using a similar objective as [ELECTRA](https://huggingface.co/transformers/model_doc/electra.html). It was introduced in [this paper](https://arxiv.org/pdf/2006.03236.pdf) and first released in [this repository](https://github.com/laiguokun/Funnel-Transformer). This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=funnel-transformer) to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import FunnelTokenizer, FunnelModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/large") model = FunneModel.from_pretrained("funnel-transformer/large") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import FunnelTokenizer, TFFunnelModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/large") model = TFFunnelModel.from_pretrained("funnel-transformer/large") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` ## Training data The BERT model was pretrained on: - [BookCorpus](https://yknzhu.wixsite.com/mbweb), a dataset consisting of 11,038 unpublished books, - [English Wikipedia](https://en.wikipedia.org/wiki/English_Wikipedia) (excluding lists, tables and headers), - [Clue Web](https://lemurproject.org/clueweb12/), a dataset of 733,019,372 English web pages, - [GigaWord](https://catalog.ldc.upenn.edu/LDC2011T07), an archive of newswire text data, - [Common Crawl](https://commoncrawl.org/), a dataset of raw web pages. ### BibTeX entry and citation info ```bibtex @misc{dai2020funneltransformer, title={Funnel-Transformer: Filtering out Sequential Redundancy for Efficient Language Processing}, author={Zihang Dai and Guokun Lai and Yiming Yang and Quoc V. Le}, year={2020}, eprint={2006.03236}, archivePrefix={arXiv}, primaryClass={cs.LG} } ```
{"language": "en", "license": "apache-2.0", "datasets": ["bookcorpus", "wikipedia", "gigaword"]}
feature-extraction
funnel-transformer/large
[ "transformers", "pytorch", "tf", "safetensors", "funnel", "feature-extraction", "en", "dataset:bookcorpus", "dataset:wikipedia", "dataset:gigaword", "arxiv:2006.03236", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2006.03236" ]
[ "en" ]
TAGS #transformers #pytorch #tf #safetensors #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us
# Funnel Transformer large model (B8-8-8 with decoder) Pretrained model on English language using a similar objective as ELECTRA. It was introduced in this paper and first released in this repository. This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: and in TensorFlow: ## Training data The BERT model was pretrained on: - BookCorpus, a dataset consisting of 11,038 unpublished books, - English Wikipedia (excluding lists, tables and headers), - Clue Web, a dataset of 733,019,372 English web pages, - GigaWord, an archive of newswire text data, - Common Crawl, a dataset of raw web pages. ### BibTeX entry and citation info
[ "# Funnel Transformer large model (B8-8-8 with decoder)\n\nPretrained model on English language using a similar objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ "TAGS\n#transformers #pytorch #tf #safetensors #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us \n", "# Funnel Transformer large model (B8-8-8 with decoder)\n\nPretrained model on English language using a similar objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ 77, 101, 206, 133, 32, 95, 11 ]
[ "passage: TAGS\n#transformers #pytorch #tf #safetensors #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us \n# Funnel Transformer large model (B8-8-8 with decoder)\n\nPretrained model on English language using a similar objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs." ]
[ -0.0871284082531929, 0.1151079535484314, -0.003065721597522497, 0.04894547909498215, 0.10655200481414795, 0.005208133719861507, 0.057479195296764374, 0.06679825484752655, -0.09776948392391205, 0.07049014419317245, -0.013438836671411991, -0.03500070422887802, 0.09534851461648941, 0.07315583527088165, 0.07434357702732086, -0.22481438517570496, 0.07243063300848007, -0.0458211712539196, 0.1079794391989708, 0.09391138702630997, 0.0799252912402153, -0.09456133097410202, 0.06179911643266678, -0.003921495750546455, -0.05568768456578255, -0.01897876150906086, -0.0028535204473882914, -0.02783369831740856, 0.04723686724901199, 0.07809282839298248, 0.10976643115282059, 0.010463125072419643, 0.06238357350230217, -0.08460585027933121, 0.024774743244051933, 0.079522505402565, 0.009372297674417496, 0.04587528854608536, 0.02145909145474434, -0.013575693592429161, 0.08694247901439667, -0.002686123363673687, 0.06544928252696991, 0.061304498463869095, -0.11744583398103714, -0.1899247169494629, -0.047068312764167786, 0.12728507816791534, 0.024917323142290115, 0.10944022983312607, -0.039551474153995514, 0.14300231635570526, -0.011708764359354973, 0.06307963281869888, 0.1216130405664444, -0.17517654597759247, -0.0016121116932481527, 0.014612528495490551, -0.025480294600129128, 0.11508426070213318, -0.037245698273181915, -0.04863312840461731, -0.0029578651301562786, 0.03560584411025047, 0.055940065532922745, -0.0021381115075200796, -0.0011954241199418902, -0.06775246560573578, -0.12924039363861084, -0.06452289968729019, 0.14703665673732758, -0.025077015161514282, -0.12449720501899719, -0.11598338186740875, -0.050355665385723114, -0.06467070430517197, 0.00222732569091022, -0.0392569899559021, -0.009377052076160908, 0.016472507268190384, 0.03722316771745682, -0.03507175296545029, -0.10330472886562347, -0.032743413001298904, -0.051356181502342224, 0.10252176225185394, 0.044309038668870926, 0.05833087116479874, -0.08994631469249725, 0.1306600719690323, -0.09600051492452621, -0.06461882591247559, -0.06691300868988037, -0.07181106507778168, -0.08030443638563156, -0.03221685811877251, -0.028773454949259758, -0.13112269341945648, -0.06213598698377609, 0.04838290438055992, -0.0847555547952652, 0.04655754193663597, -0.016508376225829124, 0.023229973390698433, 0.12711544334888458, 0.12241937965154648, -0.03744979575276375, 0.10947724431753159, 0.006802880205214024, -0.0564291886985302, 0.03210104629397392, -0.04313160106539726, -0.02792408876121044, -0.0020542244892567396, 0.03631427884101868, 0.06043005734682083, 0.004831602331250906, 0.05692208930850029, -0.03894945979118347, -0.0643664300441742, 0.14076727628707886, -0.12089410424232483, -0.015694160014390945, 0.03897733613848686, -0.03506636619567871, 0.12732498347759247, 0.062442339956760406, -0.03014371730387211, -0.12919802963733673, -0.014157393015921116, -0.0588834322988987, -0.024968331679701805, -0.08374892175197601, -0.1099313348531723, -0.004132574889808893, -0.027867745608091354, -0.08244657516479492, -0.11420373618602753, -0.21094153821468353, -0.048240456730127335, 0.0199629757553339, -0.008709399960935116, 0.012545415200293064, 0.017470790073275566, 0.0105156060308218, -0.01636241376399994, -0.0006374689401127398, -0.1410139948129654, -0.010430376045405865, 0.03937223181128502, -0.04434942454099655, 0.041899848729372025, -0.025674918666481972, 0.03946859762072563, -0.1518656462430954, -0.007559967692941427, -0.20099234580993652, 0.12686064839363098, -0.012932303361594677, -0.0033755479380488396, -0.05363943800330162, -0.027629822492599487, -0.07519125938415527, 0.006054630037397146, -0.02314905636012554, 0.13895182311534882, -0.13448786735534668, -0.04441162198781967, 0.22910232841968536, -0.22365501523017883, 0.01034400425851345, 0.08033052831888199, -0.0335044227540493, 0.1037285327911377, 0.179879829287529, 0.03027339093387127, 0.16846689581871033, -0.04872441664338112, -0.07155171036720276, 0.025326045230031013, -0.05062948167324066, 0.0812399685382843, 0.03866797685623169, -0.04665401577949524, -0.01992030255496502, 0.007524245418608189, -0.04919636249542236, -0.02040565386414528, -0.021880175918340683, -0.027994778007268906, -0.015355338342487812, -0.022142063826322556, 0.011001702398061752, 0.03016168810427189, 0.012558224610984325, 0.02467256970703602, -0.11271670460700989, -0.01282541174441576, 0.10093564540147781, -0.09996327757835388, 0.027024300768971443, -0.10320445895195007, 0.03353332355618477, -0.04667392373085022, -0.0042763217352330685, -0.19306433200836182, -0.0419895201921463, 0.04484577104449272, -0.07970479875802994, 0.06249101832509041, 0.05994996801018715, 0.030209220945835114, 0.08807853609323502, -0.01400353666394949, -0.01344282180070877, -0.01745750941336155, -0.008786797523498535, -0.020764973014593124, -0.09638810157775879, -0.05785031616687775, -0.05480904132127762, 0.06230182200670242, -0.09345605969429016, 0.04148067161440849, 0.04674016311764717, 0.005611087661236525, 0.04590770974755287, -0.06467100232839584, 0.01718384586274624, -0.005490605719387531, -0.00886590126901865, -0.04341120645403862, 0.031224504113197327, 0.031738270074129105, -0.06528948247432709, 0.05595042556524277, -0.19024430215358734, -0.09040722995996475, 0.08305484056472778, 0.03823203220963478, -0.14701296389102936, 0.044428709894418716, 0.006798328831791878, -0.002049995120614767, -0.04407394304871559, -0.08797135949134827, 0.2072296142578125, 0.009197176434099674, 0.09681294858455658, -0.0769001767039299, -0.024658404290676117, 0.02864150144159794, -0.028736546635627747, -0.0014639602741226554, 0.04864325374364853, -0.019111765548586845, -0.11067167669534683, 0.05008990690112114, 0.03262583538889885, 0.03596668317914009, 0.15382906794548035, 0.007510260678827763, -0.07676185667514801, -0.006922588218003511, 0.013182515278458595, 0.005944007076323032, 0.007198300678282976, 0.01256111916154623, 0.023561883717775345, 0.03581644594669342, 0.09698490053415298, 0.023943977430462837, -0.05615374445915222, 0.06047797575592995, 0.08982004970312119, -0.026562277227640152, -0.06152953580021858, -0.04956262186169624, -0.03392764925956726, 0.07880597561597824, 0.017881901934742928, 0.12463518232107162, 0.03699304163455963, -0.02532496117055416, -0.15818484127521515, 0.15329743921756744, -0.09405411034822464, -0.18999256193637848, -0.12658296525478363, -0.004790471866726875, 0.030983202159404755, 0.04443659633398056, 0.0516209602355957, -0.0756462886929512, -0.07319522649049759, -0.11999493837356567, 0.11239119619131088, -0.044769566506147385, -0.0608832947909832, -0.021268505603075027, -0.06860769540071487, -0.00814378447830677, -0.11388993263244629, 0.005634379107505083, 0.017051368951797485, -0.12691819667816162, 0.017295295372605324, -0.06422523409128189, 0.03440089896321297, 0.17475126683712006, -0.022153884172439575, -0.01008618250489235, -0.008028940297663212, 0.22538310289382935, -0.017349734902381897, 0.09119025617837906, 0.19113481044769287, -0.0853547751903534, 0.04635090008378029, 0.058918606489896774, 0.013823255896568298, -0.01986968331038952, 0.042263664305210114, -0.005480136256664991, -0.08709781616926193, -0.17243877053260803, -0.05989333242177963, -0.01741732843220234, -0.010340608656406403, 0.020107153803110123, 0.03873264044523239, 0.024086976423859596, 0.0603477843105793, -0.06905127316713333, -0.01915409043431282, 0.06984281539916992, 0.09057983756065369, 0.02702656202018261, -0.03307691588997841, 0.0684266984462738, -0.0858948603272438, 0.03570990264415741, 0.10292743891477585, -0.07871323823928833, 0.20501607656478882, -0.026298610493540764, 0.2012941837310791, 0.06490151584148407, 0.021864071488380432, 0.10720107704401016, 0.08262132853269577, -0.03260408341884613, 0.015794504433870316, -0.023399412631988525, -0.07611563801765442, -0.07949720323085785, -0.012673902325332165, -0.057626813650131226, 0.046387650072574615, -0.12444509565830231, -0.03276936337351799, -0.0006902652094140649, 0.17694520950317383, -0.0017607647459954023, -0.17602308094501495, -0.12504160404205322, 0.03429986909031868, -0.005775818135589361, -0.07580485194921494, 0.01913060061633587, 0.06904507428407669, -0.12653443217277527, -0.0019438242306932807, -0.001090434961952269, 0.08719204366207123, -0.15392425656318665, 0.01582234725356102, -0.05123118683695793, 0.012359175831079483, -0.02525617927312851, 0.07156973332166672, -0.06870723515748978, 0.02449505776166916, 0.015705207362771034, 0.10371632128953934, -0.03954707086086273, 0.015022792853415012, -0.03258686885237694, 0.14615006744861603, 0.09597761929035187, 0.027948325499892235, -0.03525635600090027, -0.10023147612810135, -0.05922573432326317, 0.03417714685201645, 0.03960750252008438, -0.0630909875035286, 0.09371889382600784, -0.035830333828926086, 0.04681648686528206, -0.012169625610113144, -0.009862194769084454, -0.11952580511569977, -0.12888048589229584, 0.042950455099344254, -0.026392214000225067, 0.10734465718269348, -0.04330623894929886, -0.03783746063709259, -0.031064853072166443, 0.14290930330753326, -0.1603301763534546, -0.11802761256694794, -0.12874390184879303, -0.03459417074918747, 0.03796994313597679, -0.06669259071350098, 0.011262916028499603, -0.02885415218770504, 0.11753352731466293, 0.021696390584111214, -0.0898924395442009, 0.011809729039669037, -0.06569135189056396, -0.1483246088027954, -0.02293134108185768, 0.046633027493953705, 0.16984432935714722, 0.041485920548439026, 0.0035984490532428026, 0.021191129460930824, -0.008804414421319962, -0.118376225233078, -0.04088146984577179, 0.17080733180046082, 0.014052650891244411, 0.10184195637702942, -0.03783446177840233, -0.15233740210533142, -0.03462791070342064, 0.0008213559631258249, 0.11966460198163986, 0.14646337926387787, -0.06618748605251312, 0.1288822442293167, 0.25939133763313293, -0.10267311334609985, -0.19674581289291382, -0.014436806552112103, 0.007050547748804092, 0.0326056145131588, 0.02915995940566063, -0.1862264722585678, 0.08653776347637177, 0.05196639522910118, -0.00010879088949877769, -0.03303975611925125, -0.2984239161014557, -0.07666287571191788, 0.10866270959377289, 0.09257206320762634, 0.11001003533601761, -0.08544142544269562, -0.015959331765770912, -0.007591600529849529, -0.14445587992668152, 0.1447674036026001, -0.14156188070774078, 0.053880129009485245, 0.013752294704318047, -0.014497539959847927, 0.028599219396710396, -0.03663479536771774, 0.06549733132123947, 0.003692478407174349, 0.06220468878746033, -0.08248836547136307, 0.053545091301202774, 0.1100664883852005, -0.02493627928197384, 0.14732137322425842, 0.05565604940056801, 0.07043904811143875, -0.09179364889860153, -0.062086790800094604, -0.07006494700908661, 0.0372488871216774, -0.05578169226646423, -0.04657581076025963, -0.06421314179897308, 0.05648777261376381, 0.05892753228545189, -0.009817956946790218, -0.018093688413500786, -0.09370945394039154, 0.08082829415798187, 0.12975932657718658, 0.15301631391048431, 0.0830712616443634, -0.13076405227184296, -0.04841626435518265, -0.02062363177537918, 0.09845349192619324, -0.0684959813952446, 0.05808146297931671, 0.06746714562177658, 0.02203948050737381, 0.10882192105054855, 0.02957654371857643, -0.14108094573020935, 0.026273252442479134, -0.0024886876344680786, -0.11647927761077881, -0.10717801004648209, 0.010469544678926468, 0.024932242929935455, -0.11133811622858047, 0.006656266283243895, 0.1241912916302681, -0.07643915712833405, -0.0042359791696071625, -0.019269080832600594, 0.02146683819591999, 0.01582052931189537, 0.07498867809772491, 0.02153930813074112, 0.015994783490896225, -0.06571339070796967, 0.09448052942752838, 0.09884491562843323, -0.08000509440898895, 0.054608069360256195, 0.05748087912797928, -0.11253667622804642, -0.08038298040628433, -0.03788169100880623, 0.08659033477306366, -0.01189754530787468, -0.08483649045228958, 0.031176013872027397, -0.09622402489185333, 0.04916279762983322, 0.08966733515262604, 0.0039037729147821665, 0.0804147720336914, -0.0709414854645729, 0.02402576059103012, -0.0818917453289032, 0.052606478333473206, -0.013374886475503445, -0.009915937669575214, -0.08215820044279099, 0.1304800659418106, 0.07501541078090668, 0.025137903168797493, -0.0296674482524395, -0.10022151470184326, -0.08662089705467224, 0.0008191564120352268, -0.10382449626922607, 0.006678407080471516, -0.015888746827840805, -0.01730886474251747, -0.02146111987531185, 0.015859205275774002, 0.019958743825554848, 0.0254992563277483, -0.0381510853767395, 0.0015794015489518642, -0.030544396489858627, 0.034307144582271576, -0.10214468836784363, 0.0517408512532711, 0.038281165063381195, -0.03232531622052193, 0.09325671941041946, 0.04396965727210045, -0.06708580255508423, 0.049806542694568634, -0.07452047616243362, 0.07304112613201141, -0.058495037257671356, -0.027626236900687218, -0.011207040399312973, -0.0795838013291359, -0.015705078840255737, 0.019127285107970238, -0.06161119043827057, 0.01100620161741972, 0.10997716337442398, -0.08252160251140594, 0.10719563812017441, 0.04439657926559448, 0.026822999119758606, -0.10191137343645096, 0.05578310415148735, 0.008528406731784344, 0.03455488011240959, 0.10885637998580933, -0.038796670734882355, 0.07706601172685623, -0.14565090835094452, -0.027461964637041092, 0.04748838022351265, 0.04269689694046974, -0.03905316814780235, -0.03440551459789276, 0.02280169166624546, -0.02650030516088009, 0.06896223872900009, -0.018416905775666237, 0.02714603766798973, 0.021153196692466736, -0.026117492467164993, -0.04690342769026756, 0.012809298932552338, 0.09519615024328232, -0.006294120568782091, -0.03710545226931572, 0.014992346987128258, 0.01683972403407097, -0.07163099944591522, -0.05430993437767029, 0.15984097123146057, 0.049851734191179276, 0.032494161278009415, 0.010973026975989342, -0.045942068099975586, -0.003986907657235861, -0.0769236832857132, -0.03763679042458534, -0.006351735908538103, 0.009828833863139153, -0.024456264451146126, 0.11002006381750107, 0.1509469449520111, -0.03085351176559925, 0.10800018161535263, -0.004804321564733982, -0.05684948340058327, -0.10687067359685898, -0.232297882437706, 0.00516853341832757, -0.039881717413663864, -0.034959953278303146, -0.1074390560388565, 0.018749596551060677, 0.07108020782470703, 0.025698209181427956, -0.04178740084171295, 0.14686335623264313, -0.09184644371271133, -0.09963681548833847, -0.010625865310430527, -0.008839276619255543, 0.03693561255931854, 0.010691115632653236, 0.06256840378046036, 0.08804628252983093, 0.0823940858244896, 0.06888403743505478, 0.10653261095285416, 0.06816709041595459, 0.015251733362674713, -0.0547616183757782, -0.07903402298688889, -0.002896279562264681, 0.023033028468489647, 0.014736540615558624, 0.14240294694900513, 0.022953597828745842, -0.04851304367184639, -0.016336657106876373, 0.16723477840423584, -0.06504184007644653, -0.05786322429776192, -0.12575609982013702, 0.227651908993721, 0.0005392669700086117, -0.006632321979850531, 0.027246886864304543, -0.11268523335456848, 0.027227945625782013, 0.17424729466438293, 0.13658419251441956, 0.01376582495868206, 0.025830471888184547, -0.019511090591549873, 0.002165030688047409, 0.010802168399095535, 0.1346973329782486, 0.02376578189432621, 0.30485999584198, -0.06736931949853897, 0.1883552223443985, -0.039507102221250534, -0.008220058865845203, -0.06637771427631378, 0.0760352686047554, -0.04686335474252701, 0.030234437435865402, -0.06874865293502808, 0.07727023214101791, -0.09143871814012527, -0.20239703357219696, -0.013412371277809143, 0.012600879184901714, -0.02025531604886055, -0.00496172159910202, -0.002007629256695509, 0.014588546939194202, 0.05416123941540718, -0.008348208852112293, 0.003384500741958618, 0.17614294588565826, 0.027456749230623245, -0.07121731340885162, -0.06177152693271637, 0.09361221641302109, -0.07504318654537201, 0.13954448699951172, 0.0172695554792881, 0.12375747412443161, 0.07570754736661911, -0.014097160659730434, -0.10474678128957748, 0.06188386678695679, -0.047408536076545715, -0.0010253229411318898, -0.018442003056406975, 0.13035057485103607, 0.008408797904849052, 0.11970318108797073, 0.049752216786146164, -0.10491733253002167, 0.020232217386364937, -0.0251413993537426, -0.010494414716959, -0.07106718420982361, 0.053386345505714417, -0.03205636143684387, 0.14236599206924438, 0.15176603198051453, -0.01170020829886198, -0.017087938264012337, -0.052979644387960434, 0.03247830644249916, -0.01936114951968193, 0.050555966794490814, 0.010385381989181042, -0.12392652779817581, -0.007262043654918671, 0.06006027013063431, 0.059650711715221405, -0.24738599359989166, -0.04501049593091011, 0.0035056043416261673, -0.019862178713083267, -0.006837569177150726, 0.058731477707624435, 0.02108912542462349, 0.05917820334434509, -0.041118983179330826, 0.08222977817058563, -0.017611859366297722, 0.10309085249900818, -0.09857727587223053, -0.07185879349708557 ]
null
null
transformers
# Funnel Transformer medium model (B6-3x2-3x2 without decoder) Pretrained model on English language using a similar objective objective as [ELECTRA](https://huggingface.co/transformers/model_doc/electra.html). It was introduced in [this paper](https://arxiv.org/pdf/2006.03236.pdf) and first released in [this repository](https://github.com/laiguokun/Funnel-Transformer). This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. **Note:** This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth of the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if you need one input per initial token. You should use the `medium` model in that case. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=funnel-transformer) to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import FunnelTokenizer, FunnelBaseModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/medium-base") model = FunnelBaseModel.from_pretrained("funnel-transformer/medium-base") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import FunnelTokenizer, TFFunnelBaseModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/medium-base") model = TFFunnelBaseModel.from_pretrained("funnel-transformer/medium-base") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` ## Training data The BERT model was pretrained on: - [BookCorpus](https://yknzhu.wixsite.com/mbweb), a dataset consisting of 11,038 unpublished books, - [English Wikipedia](https://en.wikipedia.org/wiki/English_Wikipedia) (excluding lists, tables and headers), - [Clue Web](https://lemurproject.org/clueweb12/), a dataset of 733,019,372 English web pages, - [GigaWord](https://catalog.ldc.upenn.edu/LDC2011T07), an archive of newswire text data, - [Common Crawl](https://commoncrawl.org/), a dataset of raw web pages. ### BibTeX entry and citation info ```bibtex @misc{dai2020funneltransformer, title={Funnel-Transformer: Filtering out Sequential Redundancy for Efficient Language Processing}, author={Zihang Dai and Guokun Lai and Yiming Yang and Quoc V. Le}, year={2020}, eprint={2006.03236}, archivePrefix={arXiv}, primaryClass={cs.LG} } ```
{"language": "en", "license": "apache-2.0", "datasets": ["bookcorpus", "wikipedia", "gigaword"]}
feature-extraction
funnel-transformer/medium-base
[ "transformers", "pytorch", "tf", "funnel", "feature-extraction", "en", "dataset:bookcorpus", "dataset:wikipedia", "dataset:gigaword", "arxiv:2006.03236", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2006.03236" ]
[ "en" ]
TAGS #transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us
# Funnel Transformer medium model (B6-3x2-3x2 without decoder) Pretrained model on English language using a similar objective objective as ELECTRA. It was introduced in this paper and first released in this repository. This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. Note: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth of the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if you need one input per initial token. You should use the 'medium' model in that case. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: and in TensorFlow: ## Training data The BERT model was pretrained on: - BookCorpus, a dataset consisting of 11,038 unpublished books, - English Wikipedia (excluding lists, tables and headers), - Clue Web, a dataset of 733,019,372 English web pages, - GigaWord, an archive of newswire text data, - Common Crawl, a dataset of raw web pages. ### BibTeX entry and citation info
[ "# Funnel Transformer medium model (B6-3x2-3x2 without decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nNote: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth\nof the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if\nyou need one input per initial token. You should use the 'medium' model in that case.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ "TAGS\n#transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us \n", "# Funnel Transformer medium model (B6-3x2-3x2 without decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nNote: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth\nof the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if\nyou need one input per initial token. You should use the 'medium' model in that case.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ 72, 105, 288, 133, 32, 95, 11 ]
[ "passage: TAGS\n#transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us \n# Funnel Transformer medium model (B6-3x2-3x2 without decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nNote: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth\nof the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if\nyou need one input per initial token. You should use the 'medium' model in that case." ]
[ -0.0840676948428154, 0.14789658784866333, -0.004246928729116917, 0.05229368060827255, 0.10427729785442352, 0.00843088235706091, 0.04015902429819107, 0.06714960932731628, -0.09756524115800858, 0.09041661769151688, -0.013644535094499588, -0.056031692773103714, 0.09103187918663025, 0.05537758767604828, 0.08876620978116989, -0.21831941604614258, 0.05003061890602112, -0.036503881216049194, 0.14585475623607635, 0.09025008976459503, 0.061665404587984085, -0.088467538356781, 0.06494550406932831, -0.004417166579514742, -0.0498763807117939, -0.03428078442811966, -0.00905338954180479, -0.02503887750208378, 0.02829202264547348, 0.054349932819604874, 0.10507754236459732, -0.004619539249688387, 0.04065537825226784, -0.07600317895412445, 0.01616472564637661, 0.09047357738018036, 0.009475148282945156, 0.05340903252363205, 0.04043400660157204, -0.006763261277228594, 0.07524032890796661, -0.020779456943273544, 0.05487050861120224, 0.05900974199175835, -0.14217723906040192, -0.15852029621601105, -0.06533687561750412, 0.1285807490348816, 0.037151336669921875, 0.10615459829568863, -0.034960757941007614, 0.1042359322309494, -0.002673960290849209, 0.05896919220685959, 0.12558665871620178, -0.17165881395339966, -0.007496806792914867, -0.01675279438495636, -0.01607879437506199, 0.08655282109975815, -0.058622367680072784, -0.05923387408256531, -0.019601192325353622, 0.03569631278514862, 0.040814850479364395, -0.020016370341181755, -0.03064984828233719, -0.08162830770015717, -0.1199130043387413, -0.04797417297959328, 0.15242384374141693, -0.021237993612885475, -0.12894722819328308, -0.0636482834815979, -0.04801575094461441, -0.1266041100025177, 0.009008114226162434, -0.01206315029412508, -0.00035726933856494725, 0.015423368662595749, 0.041064176708459854, -0.03528692200779915, -0.10535848140716553, -0.030706096440553665, -0.05227011814713478, 0.12005369365215302, 0.02600054256618023, 0.05417818948626518, -0.07239240407943726, 0.11468391865491867, -0.09493086487054825, -0.05065149441361427, -0.0734926089644432, -0.04238941892981529, -0.10996552556753159, -0.04609685763716698, -0.031074093654751778, -0.13879667222499847, -0.0764489695429802, 0.043897416442632675, -0.10775714367628098, 0.04759720340371132, -0.0035277996212244034, 0.03380267694592476, 0.11258400976657867, 0.10791145265102386, -0.028247883543372154, 0.06793272495269775, -0.006154697388410568, -0.07739187777042389, 0.04143624007701874, -0.03669722005724907, -0.014320657588541508, -0.00033693009754642844, 0.0402841679751873, 0.047305043786764145, -0.000562049332074821, 0.05932506546378136, -0.0496772937476635, -0.050447165966033936, 0.14389805495738983, -0.10367431491613388, 0.0005318628973327577, 0.036462899297475815, -0.02647620625793934, 0.09365195780992508, 0.06389234215021133, -0.05826648324728012, -0.13232356309890747, 0.010147533379495144, -0.06603287905454636, -0.01082155667245388, -0.08151831477880478, -0.09925584495067596, 0.0034976876340806484, -0.04912246763706207, -0.09191107004880905, -0.09462819993495941, -0.19810324907302856, -0.07432730495929718, 0.020121686160564423, -0.015921102836728096, 0.015092261135578156, 0.014334037899971008, -0.0008546887547709048, -0.0176137313246727, -0.0003532489645294845, -0.12123909592628479, -0.011826439760625362, 0.03514830023050308, -0.04040171951055527, 0.050219710916280746, -0.0194418765604496, 0.03514159843325615, -0.1544201523065567, 0.008057334460318089, -0.20035548508167267, 0.117813840508461, 0.0012714221375063062, 0.004321638029068708, -0.05393730476498604, -0.012509208172559738, -0.07252122461795807, 0.010539153590798378, -0.009824779815971851, 0.1187915951013565, -0.12727224826812744, -0.032409701496362686, 0.205857515335083, -0.20160380005836487, 0.024706928059458733, 0.07670463621616364, -0.02455623261630535, 0.08963201940059662, 0.1643279641866684, 0.04661967605352402, 0.1696464717388153, -0.02714470960199833, -0.08547341823577881, 0.009919566102325916, -0.08306442946195602, 0.06947337090969086, 0.04096629470586777, -0.06680364906787872, -0.01857420802116394, 0.004503496456891298, -0.046311430633068085, -0.0320112481713295, -0.0111421849578619, -0.026597976684570312, -0.0004937331541441381, -0.018429506570100784, 0.005951513070613146, 0.021309345960617065, 0.009085829369723797, 0.014558599330484867, -0.10953792929649353, -0.024016916751861572, 0.09458843618631363, -0.07540394365787506, 0.037860527634620667, -0.07632379978895187, 0.02806725539267063, -0.06675327569246292, -0.012827499769628048, -0.19575951993465424, -0.02811627835035324, 0.04143597185611725, -0.08321873843669891, 0.07043308019638062, 0.042735181748867035, 0.02702745795249939, 0.09758498519659042, -0.01751742511987686, -0.011002583429217339, -0.010944634675979614, -0.009582668542861938, -0.05321592837572098, -0.08850396424531937, -0.06234823912382126, -0.04826046898961067, 0.03373648226261139, -0.059888795018196106, 0.01764284260571003, 0.04637368768453598, 0.0024303088430315256, 0.049031469970941544, -0.08712629228830338, 0.03751034662127495, 0.001525874831713736, 0.007034325506538153, -0.034922126680612564, 0.04043661803007126, 0.04906845837831497, -0.03705894574522972, 0.048245564103126526, -0.19057703018188477, -0.1416272222995758, 0.06113801524043083, 0.01994302310049534, -0.16527777910232544, 0.026949327439069748, 0.009384829550981522, -0.010871726088225842, -0.060634795576334, -0.07872480154037476, 0.22486452758312225, 0.010631504468619823, 0.08979826420545578, -0.05966269597411156, -0.018228424713015556, 0.030327841639518738, -0.026899918913841248, 0.000648049870505929, 0.05378033593297005, 0.00697457417845726, -0.15163716673851013, 0.059139981865882874, -0.0032252646051347256, 0.005694214254617691, 0.14818595349788666, 0.035947903990745544, -0.07586515694856644, -0.030225209891796112, 0.014633316546678543, 0.013696838170289993, 0.025613749399781227, -0.0441163070499897, 0.005876533687114716, 0.024727463722229004, 0.08943820744752884, 0.021092738956212997, -0.057469774037599564, 0.06942255049943924, 0.07203962653875351, -0.007169146090745926, -0.059268344193696976, -0.06515122205018997, -0.0436580590903759, 0.07544801384210587, 0.028491197153925896, 0.11389245092868805, 0.03750128671526909, -0.029297253116965294, -0.15194234251976013, 0.14931735396385193, -0.09789605438709259, -0.19583140313625336, -0.11651560664176941, -0.0023735288996249437, 0.022032661363482475, 0.04962383955717087, 0.04305538907647133, -0.05394485965371132, -0.059308089315891266, -0.10492333024740219, 0.08106416463851929, -0.05704681947827339, -0.05939580872654915, -0.03921554610133171, -0.06247914955019951, -0.009615670889616013, -0.08623574674129486, 0.010820284485816956, -0.002554505132138729, -0.10691455006599426, 0.015727534890174866, -0.039238423109054565, 0.046732258051633835, 0.16348892450332642, -0.017234578728675842, -0.020185112953186035, -0.002894449280574918, 0.187469944357872, -0.03319069743156433, 0.09937183558940887, 0.14537806808948517, -0.06751534342765808, 0.050407037138938904, 0.05774970352649689, 0.012915225699543953, -0.014257601462304592, 0.02637079916894436, 0.007162559777498245, -0.08437362313270569, -0.14118662476539612, -0.04356352612376213, -0.037810973823070526, 0.003908827435225248, 0.04870956763625145, 0.03708285838365555, 0.007046961225569248, 0.05832763761281967, -0.05317467078566551, -0.014185246080160141, 0.04402391240000725, 0.09818360954523087, -0.0002423243277007714, -0.027219258248806, 0.0481167733669281, -0.07397806644439697, 0.04066126048564911, 0.10034134238958359, -0.026215078309178352, 0.19306370615959167, -0.040169429033994675, 0.18727374076843262, 0.06428106874227524, -0.002030962146818638, 0.09998008608818054, 0.06381335109472275, -0.035053711384534836, 0.028020910918712616, -0.027432531118392944, -0.07179716974496841, -0.0651913583278656, -0.007016744930297136, -0.04524461552500725, 0.03162769228219986, -0.11779853701591492, -0.04243920370936394, 0.01928173378109932, 0.17402635514736176, 0.023628393188118935, -0.15613406896591187, -0.12453643232584, 0.021228516474366188, -0.0070253098383545876, -0.07903303951025009, 0.018457194790244102, 0.07429986447095871, -0.10549947619438171, -0.004255583044141531, -0.007583949714899063, 0.07727693766355515, -0.15621766448020935, 0.01304375659674406, -0.02022533491253853, 0.04588932916522026, -0.02669643610715866, 0.05934418737888336, -0.0623830184340477, 0.014930964447557926, 0.01159660704433918, 0.11568033695220947, -0.030747655779123306, 0.007196941412985325, -0.027596555650234222, 0.11424874514341354, 0.10362274944782257, 0.043399009853601456, -0.0417969711124897, -0.08185204118490219, -0.054243940860033035, 0.015871228650212288, 0.04700793698430061, -0.05954865366220474, 0.09521065652370453, -0.008612685836851597, 0.03277406468987465, -0.027813846245408058, -0.009830404072999954, -0.08603666722774506, -0.12372082471847534, 0.05159440636634827, -0.022879980504512787, 0.0885903537273407, -0.026637405157089233, -0.006928587798029184, 0.001972104189917445, 0.15054982900619507, -0.15183699131011963, -0.10785863548517227, -0.11869864165782928, -0.034632399678230286, 0.034056421369314194, -0.061221104115247726, -0.004454652313143015, -0.028818544000387192, 0.10856354236602783, 0.01643870398402214, -0.06591936945915222, 0.02444118820130825, -0.062361717224121094, -0.15514279901981354, -0.04808802902698517, 0.05693832039833069, 0.1391168087720871, 0.0386887826025486, -0.006708668544888496, 0.04161505028605461, -0.017316831275820732, -0.11468371003866196, -0.006473802030086517, 0.13608454167842865, 0.030711621046066284, 0.09518299251794815, -0.04103560000658035, -0.11302486807107925, -0.03079049475491047, -0.0019387506181374192, 0.11906901746988297, 0.12670394778251648, -0.07011788338422775, 0.13394908607006073, 0.2617189586162567, -0.11011116951704025, -0.20457452535629272, -0.009181585162878036, 0.02240496687591076, 0.014393751509487629, 0.008648985996842384, -0.1905054897069931, 0.09108259528875351, 0.0697777271270752, 0.0031889518722891808, -0.006852271500974894, -0.2754122018814087, -0.06783023476600647, 0.06575720012187958, 0.08581849932670593, 0.09718696027994156, -0.09726300835609436, -0.024269482120871544, -0.005453342571854591, -0.11599235981702805, 0.1527474820613861, -0.11460337787866592, 0.0786065086722374, 0.010015805251896381, -0.0273079015314579, 0.030613867565989494, -0.04357214644551277, 0.07150991261005402, 0.022238990291953087, 0.05147641524672508, -0.059535566717386246, 0.030238555744290352, 0.10617880523204803, -0.0331820473074913, 0.13413378596305847, 0.06393704563379288, 0.06295149773359299, -0.08418971300125122, -0.050849657505750656, -0.08044418692588806, 0.03021751157939434, -0.057793330401182175, -0.03415074571967125, -0.04889757186174393, 0.07088612020015717, 0.06496866047382355, -0.009458023123443127, -0.020245052874088287, -0.09498608857393265, 0.06352614611387253, 0.13378006219863892, 0.1345398873090744, 0.08527062833309174, -0.1764696091413498, -0.04886515066027641, -0.030603693798184395, 0.10987360775470734, -0.05845753848552704, 0.04045386239886284, 0.06531209498643875, 0.030567513778805733, 0.11163255572319031, 0.03766031190752983, -0.13835632801055908, 0.03360772132873535, 0.008495931513607502, -0.10496557503938675, -0.08429824560880661, 0.011025873012840748, 0.01866530068218708, -0.1165919154882431, 0.017441531643271446, 0.11487643420696259, -0.07370728999376297, -0.002192532876506448, -0.007499830797314644, 0.04719924554228783, 0.004566836170852184, 0.07609758526086807, 0.010077038779854774, 0.01786886900663376, -0.06295716017484665, 0.12122830003499985, 0.10430466383695602, -0.07410618662834167, 0.04989343881607056, 0.10496077686548233, -0.11103155463933945, -0.09360498934984207, -0.06490379571914673, 0.0848926305770874, -0.012719976715743542, -0.06630806624889374, 0.030761921778321266, -0.08627067506313324, 0.06246320158243179, 0.07869413495063782, -0.01749613508582115, 0.08368455618619919, -0.07051076740026474, 0.028246767818927765, -0.08718238025903702, 0.03948143124580383, -0.029184196144342422, 0.0027705607935786247, -0.05213705077767372, 0.164498969912529, 0.053319815546274185, 0.022079966962337494, -0.01230307761579752, -0.09225404262542725, -0.09523310512304306, -0.0014332580612972379, -0.07373075932264328, 0.008013692684471607, -0.03212868422269821, -0.01650342345237732, -0.013909736648201942, 0.0439898706972599, 0.017123784869909286, 0.020791569724678993, -0.03125149756669998, -0.00650188559666276, -0.023079413920640945, 0.026113364845514297, -0.08241740614175797, 0.045038945972919464, 0.03331002965569496, -0.03364028409123421, 0.0846598744392395, 0.024990014731884003, -0.04840071126818657, 0.02373589389026165, -0.0673242062330246, 0.06829666346311569, -0.06615488231182098, -0.04499223455786705, -0.019424796104431152, -0.08195801079273224, -0.013069509528577328, 0.0038741864264011383, -0.049999646842479706, 0.0064152018167078495, 0.10737565904855728, -0.07933109253644943, 0.11051489412784576, 0.036833666265010834, 0.029788201674818993, -0.11805067956447601, 0.05034004896879196, 0.013377745635807514, 0.03568979352712631, 0.09172341227531433, -0.03876544162631035, 0.08180808275938034, -0.12750467658042908, -0.012758000753819942, 0.05568899214267731, 0.027706913650035858, -0.03864073008298874, -0.01979333907365799, 0.04031975939869881, -0.023676101118326187, 0.04086528718471527, 0.0006122530903667212, 0.013870547525584698, 0.017351506277918816, -0.015282911248505116, -0.09632649272680283, 0.00819322932511568, 0.04375810921192169, -0.018200188875198364, -0.05571339279413223, 0.009255433455109596, 0.031021706759929657, -0.06920527666807175, -0.025195052847266197, 0.1667087972164154, 0.036226652562618256, 0.06166835501790047, 0.020627658814191818, -0.06394055485725403, -0.02540719136595726, -0.09833545982837677, -0.003008476924151182, -0.0013321817386895418, 0.02796473354101181, -0.03559188172221184, 0.07061314582824707, 0.1333603709936142, -0.02675575017929077, 0.12584646046161652, -0.00877049844712019, -0.06100327521562576, -0.07208859920501709, -0.23171503841876984, 0.02322513237595558, -0.004354001022875309, -0.04275961592793465, -0.11078424006700516, 0.0265495665371418, 0.04783911630511284, 0.005207979120314121, -0.03284264728426933, 0.13283978402614594, -0.09143826365470886, -0.09804244339466095, 0.011988786049187183, -0.013347713276743889, 0.05581655353307724, 0.017647458240389824, 0.06090483441948891, 0.07800054550170898, 0.07570341229438782, 0.0754956528544426, 0.10447493195533752, 0.07841552048921585, 0.007871824316680431, -0.03510317578911781, -0.08804149180650711, -0.003576706862077117, 0.011251142248511314, 0.025785449892282486, 0.16599375009536743, 0.025396935641765594, -0.04759487509727478, -0.00545918894931674, 0.1602485328912735, -0.07014798372983932, -0.03201058506965637, -0.13458967208862305, 0.23760810494422913, 0.008419714868068695, -0.03437967598438263, 0.024349644780158997, -0.11925642937421799, 0.04603574052453041, 0.17587660253047943, 0.11938347667455673, 0.014839159324765205, 0.030598819255828857, -0.0036370777525007725, -0.0009571753325872123, 0.026471203193068504, 0.11599133908748627, 0.005872735753655434, 0.3180105686187744, -0.06899794191122055, 0.20223292708396912, -0.021356692537665367, 0.0063367304392158985, -0.05853749066591263, 0.06916933506727219, -0.038663122802972794, 0.05259975790977478, -0.0799468532204628, 0.062170546501874924, -0.09897425025701523, -0.2311510443687439, 0.022538095712661743, 0.010717440396547318, -0.0266399085521698, -0.006526333279907703, -0.00999857485294342, 0.0284098070114851, 0.07454090565443039, 0.011464746668934822, 0.0076979100704193115, 0.18288864195346832, 0.021168943494558334, -0.07922391593456268, -0.07150411605834961, 0.07806313782930374, -0.061509981751441956, 0.15546968579292297, 0.013067742809653282, 0.08416788280010223, 0.08089554309844971, -0.009177505038678646, -0.11828842759132385, 0.07215593755245209, -0.06185741722583771, -0.023564042523503304, -0.0034973854199051857, 0.14397232234477997, -0.00034203275572508574, 0.11724836379289627, 0.04049514979124069, -0.08106833696365356, 0.010905992239713669, 0.01110114436596632, 0.01910356618463993, -0.07619532942771912, 0.05161571502685547, -0.033232200890779495, 0.1579117774963379, 0.1632130742073059, -0.011139961890876293, -0.03067188523709774, -0.045435577630996704, 0.01843862794339657, -0.025383146479725838, 0.048437170684337616, 0.002859553089365363, -0.10517445206642151, 0.008683192543685436, 0.04774145409464836, 0.0735861212015152, -0.23875661194324493, -0.03813493996858597, 0.02236051671206951, -0.022560717537999153, -0.008429528214037418, 0.049412332475185394, 0.01818668283522129, 0.05307677388191223, -0.049884650856256485, 0.08459098637104034, -0.0006267189746722579, 0.10806679725646973, -0.08894573897123337, -0.07919809222221375 ]
null
null
transformers
# Funnel Transformer medium model (B6-3x2-3x2 with decoder) Pretrained model on English language using a similar objective objective as [ELECTRA](https://huggingface.co/transformers/model_doc/electra.html). It was introduced in [this paper](https://arxiv.org/pdf/2006.03236.pdf) and first released in [this repository](https://github.com/laiguokun/Funnel-Transformer). This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=funnel-transformer) to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import FunnelTokenizer, FunnelModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/medium") model = FunneModel.from_pretrained("funnel-transformer/medium") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import FunnelTokenizer, TFFunnelModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/medium") model = TFFunnelModel.from_pretrained("funnel-transformer/medium") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` ## Training data The BERT model was pretrained on: - [BookCorpus](https://yknzhu.wixsite.com/mbweb), a dataset consisting of 11,038 unpublished books, - [English Wikipedia](https://en.wikipedia.org/wiki/English_Wikipedia) (excluding lists, tables and headers), - [Clue Web](https://lemurproject.org/clueweb12/), a dataset of 733,019,372 English web pages, - [GigaWord](https://catalog.ldc.upenn.edu/LDC2011T07), an archive of newswire text data, - [Common Crawl](https://commoncrawl.org/), a dataset of raw web pages. ### BibTeX entry and citation info ```bibtex @misc{dai2020funneltransformer, title={Funnel-Transformer: Filtering out Sequential Redundancy for Efficient Language Processing}, author={Zihang Dai and Guokun Lai and Yiming Yang and Quoc V. Le}, year={2020}, eprint={2006.03236}, archivePrefix={arXiv}, primaryClass={cs.LG} } ```
{"language": "en", "license": "apache-2.0", "datasets": ["bookcorpus", "wikipedia", "gigaword"]}
feature-extraction
funnel-transformer/medium
[ "transformers", "pytorch", "tf", "funnel", "feature-extraction", "en", "dataset:bookcorpus", "dataset:wikipedia", "dataset:gigaword", "arxiv:2006.03236", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2006.03236" ]
[ "en" ]
TAGS #transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us
# Funnel Transformer medium model (B6-3x2-3x2 with decoder) Pretrained model on English language using a similar objective objective as ELECTRA. It was introduced in this paper and first released in this repository. This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: and in TensorFlow: ## Training data The BERT model was pretrained on: - BookCorpus, a dataset consisting of 11,038 unpublished books, - English Wikipedia (excluding lists, tables and headers), - Clue Web, a dataset of 733,019,372 English web pages, - GigaWord, an archive of newswire text data, - Common Crawl, a dataset of raw web pages. ### BibTeX entry and citation info
[ "# Funnel Transformer medium model (B6-3x2-3x2 with decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ "TAGS\n#transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us \n", "# Funnel Transformer medium model (B6-3x2-3x2 with decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ 72, 105, 206, 133, 32, 95, 11 ]
[ "passage: TAGS\n#transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us \n# Funnel Transformer medium model (B6-3x2-3x2 with decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs." ]
[ -0.08007605373859406, 0.11824825406074524, -0.0030914631206542253, 0.04562826827168465, 0.11025729775428772, 0.007969631813466549, 0.04993690550327301, 0.06767290830612183, -0.0962742269039154, 0.06653006374835968, -0.013855229131877422, -0.024741994217038155, 0.09372496604919434, 0.07444097846746445, 0.06810694932937622, -0.21703413128852844, 0.061144132167100906, -0.038060277700424194, 0.10906282067298889, 0.09917908906936646, 0.06771780550479889, -0.09122337400913239, 0.062344085425138474, -0.010611457750201225, -0.06535890698432922, -0.01853288896381855, 0.0006128810928203166, -0.030594654381275177, 0.03805913031101227, 0.0731239840388298, 0.11549700051546097, 0.01724199950695038, 0.050124265253543854, -0.07774173468351364, 0.023502079769968987, 0.08053673058748245, 0.006292455364018679, 0.04977637156844139, 0.011103672906756401, -0.020554915070533752, 0.07727986574172974, 0.0046218689531087875, 0.0663856640458107, 0.06722798943519592, -0.13069286942481995, -0.17662645876407623, -0.043692875653505325, 0.11568833142518997, 0.02751031331717968, 0.11134262382984161, -0.038964204490184784, 0.1412789523601532, -0.015017843805253506, 0.06424329429864883, 0.12449225038290024, -0.1674732118844986, -0.006084526889026165, 0.012594914063811302, -0.03840311989188194, 0.11803684383630753, -0.04055614396929741, -0.054077453911304474, -0.010794080793857574, 0.0391952320933342, 0.05544491484761238, -0.011946589685976505, -0.01620681956410408, -0.0767807587981224, -0.12487254291772842, -0.05959662050008774, 0.16172757744789124, -0.027759399265050888, -0.13030357658863068, -0.09550046175718307, -0.05431757867336273, -0.09902564436197281, 0.005203994456678629, -0.025022706016898155, -0.012366761453449726, 0.015218336135149002, 0.014695916324853897, -0.03216874226927757, -0.10177246481180191, -0.041969578713178635, -0.03977314010262489, 0.11484484374523163, 0.04111533239483833, 0.05855102092027664, -0.09625474363565445, 0.1432104855775833, -0.07195024192333221, -0.06005962938070297, -0.06906142085790634, -0.07209397107362747, -0.08767552673816681, -0.030733294785022736, -0.032453738152980804, -0.14008720219135284, -0.07090993225574493, 0.056518275290727615, -0.07805731147527695, 0.05039922520518303, -0.013507695868611336, 0.03309660777449608, 0.12851381301879883, 0.10264064371585846, -0.02382628619670868, 0.11653728783130646, 0.009675205685198307, -0.07670219242572784, 0.03236940875649452, -0.0480428971350193, -0.02047104761004448, 0.004533817991614342, 0.025847645476460457, 0.06387615948915482, -0.0030167899094522, 0.06003280356526375, -0.05308632180094719, -0.06972334533929825, 0.12089242786169052, -0.11164853721857071, -0.02157970331609249, 0.034426722675561905, -0.033135347068309784, 0.12228534370660782, 0.06503438204526901, -0.03404080867767334, -0.12783057987689972, -0.006234507076442242, -0.061435963958501816, -0.019650515168905258, -0.08804582059383392, -0.10879505425691605, -0.0037798110861331224, -0.043918102979660034, -0.08504034578800201, -0.11016494035720825, -0.20133641362190247, -0.051881592720746994, 0.028348922729492188, -0.007883630692958832, 0.01284764800220728, 0.017519570887088776, 0.017126331105828285, -0.015924345701932907, -0.002028487389907241, -0.13742485642433167, -0.01077566109597683, 0.04630633443593979, -0.04193393513560295, 0.04981911554932594, -0.03319677338004112, 0.04123735800385475, -0.15741470456123352, -0.0062391567043960094, -0.20096132159233093, 0.11612021178007126, -0.012019393034279346, 0.006989687215536833, -0.05328638106584549, -0.020983412861824036, -0.07891855388879776, 0.0005252647097222507, -0.018207702785730362, 0.13417427241802216, -0.1352544128894806, -0.04482066258788109, 0.23137515783309937, -0.21464863419532776, 0.02208057790994644, 0.07201623171567917, -0.028760062530636787, 0.1190234124660492, 0.17542168498039246, 0.029465321451425552, 0.1595274806022644, -0.05192340165376663, -0.0684734508395195, 0.013867327943444252, -0.05722110718488693, 0.07415516674518585, 0.03611287474632263, -0.04510822519659996, -0.014992674812674522, 0.006939462386071682, -0.04029868170619011, -0.017341265454888344, -0.020895160734653473, -0.030852464959025383, -0.010207494720816612, -0.023054106160998344, 0.004618655890226364, 0.022652359679341316, 0.021018635481595993, 0.026350362226366997, -0.1074850931763649, -0.03575150668621063, 0.09535913169384003, -0.09850738197565079, 0.03369273990392685, -0.10245892405509949, 0.023362744599580765, -0.052194733172655106, -0.002318018116056919, -0.19961732625961304, -0.03603913262486458, 0.04189971461892128, -0.05735074728727341, 0.06927081942558289, 0.05290660262107849, 0.02592414617538452, 0.08163656294345856, -0.013642743229866028, -0.014179234392940998, -0.022300012409687042, -0.013631146401166916, -0.020060211420059204, -0.09340963512659073, -0.0569506473839283, -0.052321985363960266, 0.05507243424654007, -0.07480587810277939, 0.041845664381980896, 0.052354976534843445, -0.002213352359831333, 0.05617170408368111, -0.07487095892429352, 0.02154788374900818, 0.002642471343278885, -0.007658106740564108, -0.03512239456176758, 0.03923266381025314, 0.04102988541126251, -0.06616820394992828, 0.04749276861548424, -0.16936878859996796, -0.09683775901794434, 0.0758834108710289, 0.020553436130285263, -0.15638427436351776, 0.049957524985075, 0.0039979820139706135, -0.0036247950047254562, -0.04418660327792168, -0.07422574609518051, 0.21538837254047394, 0.00829989928752184, 0.09349582344293594, -0.06742431968450546, -0.02621917612850666, 0.03286997973918915, -0.029544496908783913, 0.0014013528125360608, 0.0461621955037117, 0.00862676091492176, -0.11842183023691177, 0.05371847748756409, 0.020803051069378853, 0.02471158467233181, 0.16008198261260986, 0.018226025626063347, -0.07446122169494629, -0.013792750425636768, 0.013025441206991673, 0.010471371002495289, 0.011056291870772839, 0.007026373874396086, 0.024013135582208633, 0.0315558984875679, 0.09935184568166733, 0.021412070840597153, -0.058461666107177734, 0.05515730753540993, 0.07856766879558563, -0.021918850019574165, -0.07172824442386627, -0.04318200796842575, -0.043475646525621414, 0.07709132879972458, 0.019931791350245476, 0.13382689654827118, 0.042031485587358475, -0.02553596720099449, -0.15314550697803497, 0.14638003706932068, -0.09683030843734741, -0.19555681943893433, -0.11720547080039978, -0.02420520968735218, 0.031363070011138916, 0.04375170171260834, 0.05596734210848808, -0.07382689416408539, -0.0713505744934082, -0.10761050879955292, 0.09771095216274261, -0.060449980199337006, -0.05810638144612312, -0.023730002343654633, -0.06551861763000488, -0.00979308970272541, -0.10612574964761734, 0.006964156404137611, 0.008847694844007492, -0.12408241629600525, 0.03306369110941887, -0.0635341927409172, 0.046473897993564606, 0.17027506232261658, -0.013263378292322159, -0.008756907656788826, -0.00830879993736744, 0.22509950399398804, -0.018420344218611717, 0.08530990034341812, 0.1733870506286621, -0.08979850262403488, 0.04477197676897049, 0.05734732747077942, 0.013315842486917973, -0.03018840029835701, 0.046159777790308, 0.0028361333534121513, -0.09480655193328857, -0.16395573318004608, -0.04752819985151291, -0.02841373346745968, -0.012122300453484058, 0.027541613206267357, 0.03926407918334007, 0.03271741420030594, 0.05821974575519562, -0.06709679961204529, -0.01022072322666645, 0.06373996287584305, 0.09903471171855927, 0.010126302018761635, -0.03305773064494133, 0.06611187011003494, -0.08664469420909882, 0.03819034621119499, 0.09915070980787277, -0.054382018744945526, 0.2326766699552536, -0.018068628385663033, 0.21066534519195557, 0.07259183377027512, 0.015974711626768112, 0.11461149901151657, 0.0749066025018692, -0.03533100336790085, 0.01575157605111599, -0.024417022243142128, -0.07187685370445251, -0.07729434221982956, -0.02001611888408661, -0.05283872410655022, 0.03187718614935875, -0.11956702917814255, -0.048385631293058395, 0.0022880234755575657, 0.18948015570640564, -0.002665150910615921, -0.1597628891468048, -0.12073151022195816, 0.029988180845975876, -0.0016444214852526784, -0.07203070819377899, 0.01830613985657692, 0.0617617666721344, -0.11962705850601196, -0.006933845113962889, -0.008890731260180473, 0.08866436779499054, -0.15877260267734528, 0.01035108882933855, -0.04163765534758568, 0.0052373674698174, -0.027657199651002884, 0.0639595165848732, -0.08162396401166916, 0.02578182891011238, 0.014547632075846195, 0.10872523486614227, -0.03522469103336334, 0.00539309810847044, -0.0395096093416214, 0.12866540253162384, 0.09549453854560852, 0.02951008453965187, -0.0309350173920393, -0.10411884635686874, -0.05631434917449951, 0.031338419765233994, 0.04267637059092522, -0.06432570517063141, 0.09556019306182861, -0.027214741334319115, 0.05169891566038132, -0.012206774204969406, -0.006124414503574371, -0.12542970478534698, -0.1300521194934845, 0.0471612848341465, -0.012874096632003784, 0.10912363976240158, -0.031044380739331245, -0.031366076320409775, 0.006776125635951757, 0.15476056933403015, -0.18276461958885193, -0.11437929421663284, -0.12965121865272522, -0.02945876307785511, 0.019848721101880074, -0.06137831136584282, 0.005630556959658861, -0.030048703774809837, 0.10905800759792328, 0.025886699557304382, -0.09223354607820511, 0.02142779529094696, -0.05973958596587181, -0.1452106237411499, -0.03460328280925751, 0.04378649219870567, 0.17353712022304535, 0.04405165836215019, -0.00919297058135271, 0.020747553557157516, -0.008398979902267456, -0.11958707869052887, -0.029458651319146156, 0.16224196553230286, 0.02411266416311264, 0.09865966439247131, -0.04696638509631157, -0.1372464895248413, -0.03833495453000069, -0.007639651652425528, 0.13124218583106995, 0.14589732885360718, -0.06093686446547508, 0.1411498785018921, 0.2558947205543518, -0.10575325787067413, -0.19095997512340546, -0.009015709161758423, 0.009770042262971401, 0.028746560215950012, 0.01801174506545067, -0.2091817408800125, 0.07029324769973755, 0.055731188505887985, 0.002325126202777028, -0.009590540081262589, -0.29427382349967957, -0.07531877607107162, 0.08949944376945496, 0.10380487143993378, 0.13570286333560944, -0.07994192838668823, -0.015023458749055862, -0.00735112139955163, -0.15772534906864166, 0.16155706346035004, -0.12282497435808182, 0.06994126737117767, 0.0063590132631361485, -0.009516960941255093, 0.020876245573163033, -0.03796464204788208, 0.060258910059928894, 0.004994404502213001, 0.06108342483639717, -0.07520658522844315, 0.04665514826774597, 0.11863354593515396, -0.022326625883579254, 0.12946978211402893, 0.05384919419884682, 0.060799311846494675, -0.08892806619405746, -0.06369402259588242, -0.07319072633981705, 0.03144088387489319, -0.055355191230773926, -0.051381297409534454, -0.055776555091142654, 0.058646075427532196, 0.0619099885225296, -0.0064127459190785885, -0.00882443506270647, -0.09773461520671844, 0.07443027943372726, 0.10856587439775467, 0.1387120634317398, 0.08743541687726974, -0.1421940177679062, -0.05369946360588074, -0.022571692243218422, 0.10604779422283173, -0.07004439830780029, 0.051968302577733994, 0.06580796092748642, 0.025851218029856682, 0.1091427356004715, 0.042666129767894745, -0.13728411495685577, 0.03934090957045555, 0.00017947777814697474, -0.12007088214159012, -0.07191623747348785, 0.017774665728211403, 0.01797187142074108, -0.10715688765048981, 0.015192808583378792, 0.12222953885793686, -0.06951563060283661, 0.0007512769079767168, -0.012715000659227371, 0.020027877762913704, 0.009421813301742077, 0.07047392427921295, 0.013777193613350391, 0.01006091758608818, -0.0683014765381813, 0.09899776428937912, 0.09745456278324127, -0.0730578750371933, 0.052227579057216644, 0.05916609242558479, -0.11662216484546661, -0.07822807133197784, -0.05390569940209389, 0.08352821320295334, -0.014944433234632015, -0.07311240583658218, 0.026548810303211212, -0.08572191745042801, 0.057394303381443024, 0.0845513865351677, -0.004279825370758772, 0.07975558936595917, -0.07920236140489578, 0.023847972974181175, -0.08039389550685883, 0.039582882076501846, -0.03183642774820328, -0.011563502252101898, -0.07787527143955231, 0.12650367617607117, 0.076842300593853, 0.01925748586654663, -0.0261075496673584, -0.11219262331724167, -0.09339259564876556, 0.010504373349249363, -0.08722999691963196, -0.0008371116127818823, -0.009582659229636192, -0.018519725650548935, -0.021522967144846916, 0.021608470007777214, 0.009534702636301517, 0.025272654369473457, -0.040102094411849976, 0.002003926318138838, -0.0274202823638916, 0.026574045419692993, -0.09344036877155304, 0.05286034941673279, 0.03147494047880173, -0.03532295674085617, 0.09024066478013992, 0.051124777644872665, -0.06614258885383606, 0.049907442182302475, -0.0744955986738205, 0.07979696989059448, -0.052365027368068695, -0.0351153165102005, -0.012541485019028187, -0.0717608630657196, -0.020055195316672325, 0.01241112221032381, -0.0570942685008049, 0.012288950383663177, 0.12040002644062042, -0.085968017578125, 0.10689886659383774, 0.044449687004089355, 0.01992192678153515, -0.11054368317127228, 0.05227016657590866, 0.011779040098190308, 0.05153582617640495, 0.11276337504386902, -0.03312797471880913, 0.08102943748235703, -0.1349753588438034, -0.022957278415560722, 0.06092856079339981, 0.041702475398778915, -0.03318334370851517, -0.03549431636929512, 0.022442739456892014, -0.027778610587120056, 0.06319262832403183, -0.026492493227124214, 0.016376536339521408, 0.01874999888241291, -0.030080731958150864, -0.07187081128358841, 0.00976321380585432, 0.0916493833065033, -0.0011153678642585874, -0.04093029722571373, 0.016277674585580826, 0.02565704472362995, -0.0691545307636261, -0.0480598621070385, 0.17367243766784668, 0.031199509277939796, 0.050052955746650696, 0.008643495850265026, -0.05972602590918541, -0.010290060192346573, -0.08122485131025314, -0.02588646113872528, -0.0014025475829839706, 0.018535364419221878, -0.017699074000120163, 0.10880273580551147, 0.1455806940793991, -0.022411301732063293, 0.10441272705793381, -0.003438528161495924, -0.05704620108008385, -0.09879575669765472, -0.24162918329238892, 0.01680455170571804, -0.03704238682985306, -0.03380449861288071, -0.10578817874193192, 0.018603011965751648, 0.05740091949701309, 0.023175569251179695, -0.0366867259144783, 0.14193157851696014, -0.08325596153736115, -0.10291366279125214, -0.009080471470952034, -0.014687461778521538, 0.030874105170369148, -0.0034453433472663164, 0.06478989869356155, 0.0826060101389885, 0.07888872921466827, 0.07112782448530197, 0.10855493694543839, 0.06081870198249817, 0.01024086493998766, -0.05116831883788109, -0.08417923748493195, -0.00796126201748848, 0.029399093240499496, 0.022485893219709396, 0.15700900554656982, 0.02088840678334236, -0.04705875366926193, -0.009057813324034214, 0.16451141238212585, -0.06507892161607742, -0.034818459302186966, -0.129121795296669, 0.2221502959728241, 0.00391518184915185, -0.018980955705046654, 0.0301752220839262, -0.1112767830491066, 0.027423720806837082, 0.17472834885120392, 0.13480916619300842, 0.01661301590502262, 0.028110094368457794, -0.01922634057700634, 0.0035335803404450417, 0.013978896662592888, 0.13576200604438782, 0.02712903544306755, 0.31183329224586487, -0.06832453608512878, 0.18292169272899628, -0.03677842393517494, -0.004551548510789871, -0.0610964372754097, 0.06939509510993958, -0.04102534428238869, 0.03148021921515465, -0.07887144386768341, 0.07673677057027817, -0.1044384241104126, -0.19865237176418304, 0.007342881988734007, 0.012742708437144756, -0.016034817323088646, -0.012083691544830799, -0.016690393909811974, 0.0046133557334542274, 0.056043870747089386, -0.009151811711490154, -0.0000127533967315685, 0.1760714054107666, 0.030868863686919212, -0.06906536966562271, -0.07298728078603745, 0.10086121410131454, -0.07651856541633606, 0.12831541895866394, 0.010313965380191803, 0.1203506588935852, 0.07336263358592987, -0.006718274671584368, -0.10011111944913864, 0.07180337607860565, -0.05794980376958847, -0.01114314142614603, -0.02687021717429161, 0.1393047720193863, -0.001418086001649499, 0.13446371257305145, 0.043076422065496445, -0.09498347342014313, 0.0167547557502985, -0.018328668549656868, 0.008249442093074322, -0.07063406705856323, 0.0535917691886425, -0.03610295057296753, 0.1399221271276474, 0.15347321331501007, -0.012676187790930271, -0.018049759790301323, -0.054596319794654846, 0.032575663179159164, -0.015629228204488754, 0.036328233778476715, -0.0006162820500321686, -0.13055512309074402, -0.006257145665585995, 0.07831955701112747, 0.05682025104761124, -0.25478842854499817, -0.04302917420864105, 0.0033094286918640137, -0.02209658920764923, -0.011059888638556004, 0.053602464497089386, 0.008288863115012646, 0.05933138355612755, -0.04151381552219391, 0.0766565352678299, -0.009510496631264687, 0.10195159912109375, -0.10083957761526108, -0.07437516003847122 ]
null
null
transformers
# Funnel Transformer small model (B4-4-4 without decoder) Pretrained model on English language using a similar objective objective as [ELECTRA](https://huggingface.co/transformers/model_doc/electra.html). It was introduced in [this paper](https://arxiv.org/pdf/2006.03236.pdf) and first released in [this repository](https://github.com/laiguokun/Funnel-Transformer). This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. **Note:** This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth of the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if you need one input per initial token. You should use the `small` model in that case. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=funnel-transformer) to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import FunnelTokenizer, FunnelBaseModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/small-base") model = FunnelBaseModel.from_pretrained("funnel-transformer/small-base") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import FunnelTokenizer, TFFunnelBaseModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/small-base") model = TFFunnelBaseModel.from_pretrained("funnel-transformer/small-base") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` ## Training data The BERT model was pretrained on: - [BookCorpus](https://yknzhu.wixsite.com/mbweb), a dataset consisting of 11,038 unpublished books, - [English Wikipedia](https://en.wikipedia.org/wiki/English_Wikipedia) (excluding lists, tables and headers), - [Clue Web](https://lemurproject.org/clueweb12/), a dataset of 733,019,372 English web pages, - [GigaWord](https://catalog.ldc.upenn.edu/LDC2011T07), an archive of newswire text data, - [Common Crawl](https://commoncrawl.org/), a dataset of raw web pages. ### BibTeX entry and citation info ```bibtex @misc{dai2020funneltransformer, title={Funnel-Transformer: Filtering out Sequential Redundancy for Efficient Language Processing}, author={Zihang Dai and Guokun Lai and Yiming Yang and Quoc V. Le}, year={2020}, eprint={2006.03236}, archivePrefix={arXiv}, primaryClass={cs.LG} } ```
{"language": "en", "license": "apache-2.0", "datasets": ["bookcorpus", "wikipedia", "gigaword"]}
feature-extraction
funnel-transformer/small-base
[ "transformers", "pytorch", "tf", "safetensors", "funnel", "feature-extraction", "en", "dataset:bookcorpus", "dataset:wikipedia", "dataset:gigaword", "arxiv:2006.03236", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2006.03236" ]
[ "en" ]
TAGS #transformers #pytorch #tf #safetensors #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us
# Funnel Transformer small model (B4-4-4 without decoder) Pretrained model on English language using a similar objective objective as ELECTRA. It was introduced in this paper and first released in this repository. This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. Note: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth of the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if you need one input per initial token. You should use the 'small' model in that case. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: and in TensorFlow: ## Training data The BERT model was pretrained on: - BookCorpus, a dataset consisting of 11,038 unpublished books, - English Wikipedia (excluding lists, tables and headers), - Clue Web, a dataset of 733,019,372 English web pages, - GigaWord, an archive of newswire text data, - Common Crawl, a dataset of raw web pages. ### BibTeX entry and citation info
[ "# Funnel Transformer small model (B4-4-4 without decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nNote: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth\nof the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if\nyou need one input per initial token. You should use the 'small' model in that case.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ "TAGS\n#transformers #pytorch #tf #safetensors #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us \n", "# Funnel Transformer small model (B4-4-4 without decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nNote: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth\nof the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if\nyou need one input per initial token. You should use the 'small' model in that case.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ 77, 102, 288, 133, 32, 95, 11 ]
[ "passage: TAGS\n#transformers #pytorch #tf #safetensors #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us \n# Funnel Transformer small model (B4-4-4 without decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nNote: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth\nof the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if\nyou need one input per initial token. You should use the 'small' model in that case." ]
[ -0.08551206439733505, 0.13384371995925903, -0.00473436526954174, 0.04879292845726013, 0.10591734945774078, -0.00027181170298717916, 0.04285525158047676, 0.06777814775705338, -0.11958020180463791, 0.10232622921466827, -0.001427307492122054, -0.0609910674393177, 0.09628809988498688, 0.05532669648528099, 0.09017116576433182, -0.22682403028011322, 0.07026385515928268, -0.04663846269249916, 0.1439637988805771, 0.07688664644956589, 0.07491472363471985, -0.09147585928440094, 0.07105935364961624, 0.007832441478967667, -0.04358801618218422, -0.03559332340955734, -0.015837935730814934, -0.023109152913093567, 0.029149992391467094, 0.05915702134370804, 0.09081650525331497, -0.01876574009656906, 0.043533992022275925, -0.07065719366073608, 0.016330469399690628, 0.08052852004766464, 0.02232922986149788, 0.05423597991466522, 0.04161892831325531, 0.01708151400089264, 0.06907356530427933, -0.03611398860812187, 0.04939710348844528, 0.05275627598166466, -0.1274990290403366, -0.18243171274662018, -0.0705944374203682, 0.12744583189487457, 0.03256305679678917, 0.09908254444599152, -0.037465766072273254, 0.11761586368083954, -0.009621690027415752, 0.03948492184281349, 0.11800938099622726, -0.17511959373950958, -0.0016183226834982634, -0.01726786606013775, -0.01259523257613182, 0.0823698416352272, -0.04995013773441315, -0.053208526223897934, -0.011210762895643711, 0.03465358167886734, 0.038689516484737396, -0.010648984462022781, -0.01233557891100645, -0.0749860405921936, -0.12198582291603088, -0.049865998327732086, 0.13198500871658325, -0.02223731391131878, -0.12293041497468948, -0.09442789852619171, -0.032863959670066833, -0.07746699452400208, 0.0006264847470447421, -0.027368782088160515, 0.003468177281320095, 0.016572540625929832, 0.06535540521144867, -0.032632164657115936, -0.1066175103187561, -0.017339374870061874, -0.06090952083468437, 0.09634731709957123, 0.02580517716705799, 0.05502285435795784, -0.06942234933376312, 0.08355890959501266, -0.12458368390798569, -0.054364755749702454, -0.07135171443223953, -0.04864448308944702, -0.07575883716344833, -0.038691822439432144, -0.04160473868250847, -0.10560856759548187, -0.052453551441431046, 0.03552650287747383, -0.11432243883609772, 0.0437614731490612, 0.001382689573802054, 0.027964377775788307, 0.11350199580192566, 0.11909349262714386, -0.03835512697696686, 0.07298392802476883, -0.002115135546773672, -0.04875344783067703, 0.05384876951575279, -0.033138033002614975, -0.033714957535266876, -0.0121932253241539, 0.04141216725111008, 0.04976430907845497, 0.006487926002591848, 0.052579548209905624, -0.03706581890583038, -0.04221906512975693, 0.16253584623336792, -0.10810321569442749, 0.016201283782720566, 0.04874083772301674, -0.03297530114650726, 0.10669215023517609, 0.04298495873808861, -0.04547450318932533, -0.12681280076503754, -0.002665653359144926, -0.0629180446267128, -0.012992830015718937, -0.08082999289035797, -0.09798289090394974, 0.0013010246912017465, -0.03682905063033104, -0.08132196217775345, -0.09942639619112015, -0.2021212875843048, -0.06903193891048431, 0.0021653235889971256, -0.016924427822232246, 0.009293051436543465, 0.012385237030684948, -0.008272909559309483, -0.01509244553744793, -0.0002824153343681246, -0.11848576366901398, -0.005845639854669571, 0.02597077004611492, -0.041697289794683456, 0.03582998737692833, -0.026859836652874947, 0.03174862638115883, -0.16228382289409637, 0.008367578499019146, -0.19794955849647522, 0.12694229185581207, 0.0048979828134179115, -0.0011652180692180991, -0.05411140248179436, -0.017810864374041557, -0.0801040306687355, 0.01748945564031601, -0.011781965382397175, 0.12835204601287842, -0.12846900522708893, -0.04066203907132149, 0.1794145256280899, -0.21088826656341553, 0.02243872359395027, 0.08844270557165146, -0.02760322205722332, 0.07211035490036011, 0.16827771067619324, 0.05722324550151825, 0.17264841496944427, -0.021069983020424843, -0.09513349831104279, 0.0201194416731596, -0.08259890973567963, 0.0748709887266159, 0.04677458107471466, -0.07930517941713333, -0.02370877005159855, 0.004798267502337694, -0.05852006748318672, -0.031772151589393616, -0.00739661231637001, -0.02753586322069168, -0.005782098975032568, -0.012842994183301926, 0.012071049772202969, 0.020871004089713097, 0.008283876813948154, 0.014312293380498886, -0.11187122017145157, -0.0020588096231222153, 0.10277987271547318, -0.06892900168895721, 0.030629659071564674, -0.08434948325157166, 0.05694666504859924, -0.06652899831533432, -0.01604369655251503, -0.1801365315914154, -0.021235307678580284, 0.043392643332481384, -0.10740315914154053, 0.06176648661494255, 0.028974564746022224, 0.032888349145650864, 0.09556923061609268, -0.001635445049032569, -0.008360299281775951, -0.004074373282492161, -0.01175902783870697, -0.05765464901924133, -0.08539660274982452, -0.06547516584396362, -0.045555584132671356, 0.04732901230454445, -0.0749434158205986, 0.01842794008553028, 0.04196777939796448, 0.002572577213868499, 0.02932463213801384, -0.07196129113435745, 0.02518484927713871, -0.004388747736811638, 0.01241760142147541, -0.03488383814692497, 0.03930027037858963, 0.0467536486685276, -0.03952004387974739, 0.045596420764923096, -0.21385031938552856, -0.14298956096172333, 0.07149961590766907, 0.04132837802171707, -0.1410321593284607, 0.006267934571951628, 0.01722687855362892, -0.001399492146447301, -0.05004701763391495, -0.08834969997406006, 0.2121758759021759, 0.005504744127392769, 0.08701612055301666, -0.07181529700756073, -0.007919320836663246, 0.022655755281448364, -0.034777674823999405, -0.00967190507799387, 0.049017250537872314, -0.016905492171645164, -0.13452847301959991, 0.06337400525808334, 0.016531504690647125, 0.01109184231609106, 0.13795164227485657, 0.01874777488410473, -0.07302200049161911, -0.02617385797202587, 0.010594741441309452, 0.007776640821248293, 0.004080712329596281, -0.015451407991349697, 0.008274085819721222, 0.03060181625187397, 0.08872057497501373, 0.03167297691106796, -0.04674350097775459, 0.07728847116231918, 0.0826437771320343, -0.004428465384989977, -0.035586319863796234, -0.06128091365098953, -0.02782653085887432, 0.06772857159376144, 0.03579515591263771, 0.10586768388748169, 0.04090886563062668, -0.026043063029646873, -0.15124188363552094, 0.15544509887695312, -0.097434401512146, -0.19050036370754242, -0.13628505170345306, 0.021408751606941223, 0.028760487213730812, 0.05494791269302368, 0.03550904616713524, -0.06590989232063293, -0.061662402004003525, -0.11227652430534363, 0.10296449810266495, -0.03898579254746437, -0.05711812898516655, -0.02291712909936905, -0.06491988897323608, -0.0006196317262947559, -0.09079994261264801, 0.009663296863436699, 0.0009083545883186162, -0.10853596031665802, -0.010021221823990345, -0.030690625309944153, 0.02347538433969021, 0.1587105095386505, -0.012430946342647076, -0.020716341212391853, 0.001388026517815888, 0.20249538123607635, -0.023568138480186462, 0.09076806157827377, 0.17770792543888092, -0.07586091011762619, 0.05402433127164841, 0.05617334321141243, 0.012332652695477009, -0.0015191048150882125, 0.016296204179525375, -0.006531333085149527, -0.06621618568897247, -0.12839959561824799, -0.058695290237665176, -0.019092993810772896, 0.006700297351926565, 0.03410691022872925, 0.029791364446282387, -0.0019971865694969893, 0.06978977471590042, -0.053498223423957825, -0.02076573297381401, 0.04098917171359062, 0.08785440772771835, 0.025669017806649208, -0.020110055804252625, 0.05070013925433159, -0.06851699203252792, 0.03320161998271942, 0.09732889384031296, -0.05995349586009979, 0.15799248218536377, -0.0408012680709362, 0.16982485353946686, 0.05400645360350609, 0.0051320018246769905, 0.10112819075584412, 0.0511774905025959, -0.039523426443338394, 0.01818203553557396, -0.02622360736131668, -0.0788973793387413, -0.06744249910116196, 0.000989436637610197, -0.038884811103343964, 0.05120716243982315, -0.13025836646556854, -0.020497899502515793, 0.022206531837582588, 0.14982926845550537, 0.023756740614771843, -0.18265904486179352, -0.13351742923259735, 0.034221772104501724, -0.005211784504354, -0.07859701663255692, 0.01193961687386036, 0.07622931897640228, -0.11636430025100708, -0.012662372551858425, 0.0013083098456263542, 0.06851507723331451, -0.14460189640522003, 0.02985820546746254, -0.04101620987057686, 0.06609801203012466, -0.01516348123550415, 0.064000204205513, -0.0514223612844944, -0.004724959377199411, 0.014243410900235176, 0.11668200045824051, -0.03535476699471474, 0.02301802672445774, -0.015956062823534012, 0.11353754252195358, 0.09863389283418655, 0.03352750837802887, -0.05785427242517471, -0.0839327946305275, -0.06484264135360718, 0.02516169846057892, 0.04922178387641907, -0.0550944022834301, 0.09585367888212204, -0.020633652806282043, 0.03634977340698242, -0.02466232143342495, 0.0011138425907120109, -0.0835542157292366, -0.1250411868095398, 0.04520350694656372, -0.027899030596017838, 0.09208843857049942, -0.0388958565890789, -0.007741475477814674, -0.03695281222462654, 0.12993179261684418, -0.1394185572862625, -0.11198320984840393, -0.10900937765836716, -0.03194275125861168, 0.0533573217689991, -0.06738274544477463, 0.00565937627106905, -0.029653139412403107, 0.12531869113445282, 0.01519543956965208, -0.06810934096574783, 0.014454232528805733, -0.06312660872936249, -0.15513107180595398, -0.030232468619942665, 0.0553244985640049, 0.14059697091579437, 0.03563108295202255, 0.00872777309268713, 0.048588648438453674, -0.01893213577568531, -0.11424580216407776, -0.015670744702219963, 0.14514486491680145, 0.026898493990302086, 0.0985681414604187, -0.03435128182172775, -0.12990258634090424, -0.02313954383134842, 0.011825370602309704, 0.12376586347818375, 0.1269606053829193, -0.07185603678226471, 0.1179586723446846, 0.27078777551651, -0.10620694607496262, -0.21773625910282135, -0.0069506471045315266, 0.024442071095108986, 0.022040924057364464, 0.019175097346305847, -0.18067499995231628, 0.10848499089479446, 0.07768122106790543, -0.004562240093946457, -0.040810029953718185, -0.28395965695381165, -0.06076665595173836, 0.08674335479736328, 0.06131477281451225, 0.07941608130931854, -0.10264299809932709, -0.03524424508213997, -0.012877421453595161, -0.08508052676916122, 0.14434772729873657, -0.1408061385154724, 0.05928083881735802, 0.015335115604102612, -0.025814395397901535, 0.040056001394987106, -0.03774505481123924, 0.08770584315061569, 0.003912587184458971, 0.053650856018066406, -0.06434037536382675, 0.023770974949002266, 0.08844000846147537, -0.032188501209020615, 0.14071862399578094, 0.051981713622808456, 0.06299246847629547, -0.08395858854055405, -0.048086706548929214, -0.0815555527806282, 0.0453299842774868, -0.05776750296354294, -0.024294676259160042, -0.05720421299338341, 0.0674605444073677, 0.05168257653713226, -0.0011158775305375457, -0.025507647544145584, -0.09083078801631927, 0.06553225964307785, 0.13806569576263428, 0.15070301294326782, 0.09380437433719635, -0.15457305312156677, -0.03191467002034187, -0.027833016589283943, 0.11388035863637924, -0.042489148676395416, 0.04662904888391495, 0.05728176236152649, 0.011583114042878151, 0.10431830585002899, 0.03218187019228935, -0.14269712567329407, 0.028227122500538826, 0.008751964196562767, -0.0927465558052063, -0.10780774801969528, 0.016164639964699745, 0.03432436287403107, -0.12425840646028519, 0.008652315475046635, 0.11992480605840683, -0.06899893283843994, -0.014100140891969204, -0.013991089537739754, 0.05654395744204521, 0.002637667115777731, 0.07425456494092941, 0.019801419228315353, 0.018891019746661186, -0.06104329973459244, 0.12045911699533463, 0.10015762597322464, -0.0802808552980423, 0.059513214975595474, 0.09983676671981812, -0.10094054043292999, -0.09238487482070923, -0.053123779594898224, 0.09641412645578384, -0.005765160545706749, -0.08534155040979385, 0.029429838061332703, -0.1115245595574379, 0.052928850054740906, 0.05718871206045151, -0.005409757141023874, 0.0761733278632164, -0.06650360673666, 0.030934041365981102, -0.08931407332420349, 0.04121801257133484, -0.01592211052775383, -0.005800995975732803, -0.05902789160609245, 0.16820481419563293, 0.042485933750867844, 0.019088126718997955, -0.0120230158790946, -0.07752848416566849, -0.09465444087982178, -0.012707071378827095, -0.08826562762260437, 0.020220283418893814, -0.03802727907896042, -0.022517314180731773, -0.010110978037118912, 0.04586980119347572, 0.021923460066318512, 0.02065597102046013, -0.026801124215126038, -0.00800407025963068, -0.03219132125377655, 0.033414438366889954, -0.0826275572180748, 0.05328688398003578, 0.0373803973197937, -0.030333418399095535, 0.08868767321109772, 0.028163669630885124, -0.040328625589609146, 0.019507423043251038, -0.0657845288515091, 0.046077173203229904, -0.06490885466337204, -0.04214648902416229, -0.020747171714901924, -0.09185413271188736, -0.0105973482131958, 0.01307512167841196, -0.05137414112687111, 0.0029902206733822823, 0.11806657910346985, -0.07170253247022629, 0.12083910405635834, 0.0442533902823925, 0.04346254840493202, -0.10782098025083542, 0.04203643649816513, 0.004884174093604088, 0.023192325606942177, 0.09666387736797333, -0.04414423182606697, 0.07506656646728516, -0.12610295414924622, -0.014720944687724113, 0.03406722843647003, 0.02302430011332035, -0.05007264390587807, -0.014681847766041756, 0.04157160222530365, -0.020180262625217438, 0.044822901487350464, 0.0021272983867675066, 0.0034282137639820576, 0.019369063898921013, 0.0012333485065028071, -0.05898826941847801, 0.021042702719569206, 0.050089433789253235, -0.009002349339425564, -0.05421609431505203, 0.016179051250219345, 0.024873679503798485, -0.0657971128821373, -0.022649001330137253, 0.15827438235282898, 0.05681834742426872, 0.06070144474506378, 0.03593877702951431, -0.04252585023641586, -0.022480394691228867, -0.08672331273555756, -0.0011487043229863048, -0.007655891124159098, 0.016106804832816124, -0.03895821422338486, 0.06876613199710846, 0.14428462088108063, -0.031503837555646896, 0.11809719353914261, -0.020132742822170258, -0.05866686627268791, -0.0906960591673851, -0.2262183576822281, 0.009334994480013847, -0.00955384224653244, -0.04211173951625824, -0.10736866295337677, 0.01996917463839054, 0.06890419125556946, 0.0006397416000254452, -0.03918059542775154, 0.1326552927494049, -0.09701529890298843, -0.09158779680728912, 0.005868968088179827, -0.024338483810424805, 0.05910838395357132, 0.03708062693476677, 0.04922432079911232, 0.08278504759073257, 0.0766957551240921, 0.08388422429561615, 0.10173901170492172, 0.08676999807357788, 0.01249769702553749, -0.047329869121313095, -0.0849759504199028, -0.004807774443179369, 0.0042410846799612045, 0.012487229891121387, 0.14952512085437775, 0.022604061290621758, -0.05366099998354912, -0.014299086295068264, 0.14840520918369293, -0.06858823448419571, -0.04966055974364281, -0.12695260345935822, 0.2307063192129135, -0.002865090500563383, -0.01307673379778862, 0.01415435504168272, -0.11765177547931671, 0.0394202396273613, 0.17959779500961304, 0.12893761694431305, 0.004839733708649874, 0.028810614719986916, -0.004035956226289272, -0.004192252177745104, 0.009537299163639545, 0.1181224137544632, 0.005696035921573639, 0.3120441436767578, -0.07554307579994202, 0.2126545011997223, -0.024552995339035988, 0.001046674558892846, -0.0798049047589302, 0.07447950541973114, -0.05642147362232208, 0.05730368196964264, -0.0643530786037445, 0.0504818856716156, -0.09488575160503387, -0.21902218461036682, -0.00004630927287507802, 0.014593624509871006, -0.026721064001321793, 0.001634353306144476, -0.0054382639937102795, 0.03757982328534126, 0.08200327306985855, 0.012692822143435478, 0.0036527086049318314, 0.16701556742191315, 0.009186241775751114, -0.07101243734359741, -0.0657976046204567, 0.0780322477221489, -0.06397334486246109, 0.16422295570373535, 0.02393365651369095, 0.09385524690151215, 0.08243376761674881, -0.014603003859519958, -0.1273500770330429, 0.058337267488241196, -0.0530538409948349, -0.014912167564034462, 0.006790586747229099, 0.11907236278057098, 0.01217968761920929, 0.11481857299804688, 0.04618017002940178, -0.0863606184720993, 0.01870407536625862, 0.01520212646573782, -0.003371539292857051, -0.06857802718877792, 0.03837542235851288, -0.03767000511288643, 0.15904876589775085, 0.15619894862174988, -0.013217656873166561, -0.039179131388664246, -0.03630213066935539, 0.027173815295100212, -0.025072406977415085, 0.06131665036082268, 0.009632362052798271, -0.11277373135089874, 0.011114515364170074, 0.04812365770339966, 0.08096975833177567, -0.22270642220973969, -0.03639973700046539, 0.015318638645112514, -0.016770442947745323, -0.006636693142354488, 0.042998261749744415, 0.0298205204308033, 0.045655257999897, -0.0506666861474514, 0.09136424958705902, -0.009408005513250828, 0.10745694488286972, -0.07526229321956635, -0.08023635298013687 ]
null
null
transformers
# Funnel Transformer small model (B4-4-4 with decoder) Pretrained model on English language using a similar objective objective as [ELECTRA](https://huggingface.co/transformers/model_doc/electra.html). It was introduced in [this paper](https://arxiv.org/pdf/2006.03236.pdf) and first released in [this repository](https://github.com/laiguokun/Funnel-Transformer). This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=funnel-transformer) to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import FunnelTokenizer, FunnelModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/small") model = FunneModel.from_pretrained("funnel-transformer/small") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import FunnelTokenizer, TFFunnelModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/small") model = TFFunnelModel.from_pretrained("funnel-transformer/small") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` ## Training data The BERT model was pretrained on: - [BookCorpus](https://yknzhu.wixsite.com/mbweb), a dataset consisting of 11,038 unpublished books, - [English Wikipedia](https://en.wikipedia.org/wiki/English_Wikipedia) (excluding lists, tables and headers), - [Clue Web](https://lemurproject.org/clueweb12/), a dataset of 733,019,372 English web pages, - [GigaWord](https://catalog.ldc.upenn.edu/LDC2011T07), an archive of newswire text data, - [Common Crawl](https://commoncrawl.org/), a dataset of raw web pages. ### BibTeX entry and citation info ```bibtex @misc{dai2020funneltransformer, title={Funnel-Transformer: Filtering out Sequential Redundancy for Efficient Language Processing}, author={Zihang Dai and Guokun Lai and Yiming Yang and Quoc V. Le}, year={2020}, eprint={2006.03236}, archivePrefix={arXiv}, primaryClass={cs.LG} } ```
{"language": "en", "license": "apache-2.0", "datasets": ["bookcorpus", "wikipedia", "gigaword"]}
feature-extraction
funnel-transformer/small
[ "transformers", "pytorch", "tf", "funnel", "feature-extraction", "en", "dataset:bookcorpus", "dataset:wikipedia", "dataset:gigaword", "arxiv:2006.03236", "license:apache-2.0", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2006.03236" ]
[ "en" ]
TAGS #transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #has_space #region-us
# Funnel Transformer small model (B4-4-4 with decoder) Pretrained model on English language using a similar objective objective as ELECTRA. It was introduced in this paper and first released in this repository. This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: and in TensorFlow: ## Training data The BERT model was pretrained on: - BookCorpus, a dataset consisting of 11,038 unpublished books, - English Wikipedia (excluding lists, tables and headers), - Clue Web, a dataset of 733,019,372 English web pages, - GigaWord, an archive of newswire text data, - Common Crawl, a dataset of raw web pages. ### BibTeX entry and citation info
[ "# Funnel Transformer small model (B4-4-4 with decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ "TAGS\n#transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #has_space #region-us \n", "# Funnel Transformer small model (B4-4-4 with decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ 76, 102, 206, 133, 32, 95, 11 ]
[ "passage: TAGS\n#transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #has_space #region-us \n# Funnel Transformer small model (B4-4-4 with decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs." ]
[ -0.08154674619436264, 0.11223237216472626, -0.0032308371737599373, 0.04878302663564682, 0.10668180137872696, 0.0070677390322089195, 0.05125908926129341, 0.07040929049253464, -0.10578405857086182, 0.07069940865039825, -0.012171502225100994, -0.03580605238676071, 0.09561049193143845, 0.07103224098682404, 0.07831726223230362, -0.23161830008029938, 0.0652780532836914, -0.04096424579620361, 0.10301417112350464, 0.09414330869913101, 0.07789295166730881, -0.09294440597295761, 0.06447935104370117, 0.0021801625844091177, -0.0622662678360939, -0.016377108171582222, -0.005441679619252682, -0.03141672536730766, 0.04303917661309242, 0.07732459902763367, 0.10430362075567245, 0.010654132813215256, 0.053819093853235245, -0.07589948922395706, 0.02332650125026703, 0.07998006790876389, 0.00864318199455738, 0.04920468106865883, 0.018374759703874588, -0.013607739470899105, 0.09565428644418716, -0.002577550010755658, 0.06480341404676437, 0.06148938834667206, -0.12389121949672699, -0.19605790078639984, -0.04385979473590851, 0.1203484982252121, 0.017586104571819305, 0.11090925335884094, -0.04131747782230377, 0.14139850437641144, -0.0193343423306942, 0.056569863110780716, 0.13061699271202087, -0.17431406676769257, -0.0012897795531898737, 0.01677810028195381, -0.026578182354569435, 0.10872132331132889, -0.035954706370830536, -0.04898499697446823, -0.007563307881355286, 0.03705732896924019, 0.052297964692115784, -0.004968510940670967, -0.0010701417922973633, -0.06392459571361542, -0.1286333054304123, -0.06592938303947449, 0.15529537200927734, -0.03219164162874222, -0.12115830928087234, -0.11018474400043488, -0.04560543969273567, -0.07891086488962173, -0.0017670788802206516, -0.033632081001996994, -0.0077720265835523605, 0.01758183166384697, 0.03094889596104622, -0.03218112513422966, -0.10235486179590225, -0.03304377943277359, -0.04849465936422348, 0.10010717809200287, 0.04451257362961769, 0.05643299221992493, -0.09643089026212692, 0.13561075925827026, -0.09016192704439163, -0.06361254304647446, -0.06754990667104721, -0.07101244479417801, -0.07771435379981995, -0.03120758756995201, -0.03832196444272995, -0.1256386786699295, -0.0612814836204052, 0.05035393685102463, -0.07924998551607132, 0.04641730338335037, -0.00950927846133709, 0.026568859815597534, 0.1315867155790329, 0.11359918117523193, -0.04179716855287552, 0.111074298620224, 0.009108350612223148, -0.06716383248567581, 0.02860306389629841, -0.04552493989467621, -0.032146066427230835, 0.0038681691512465477, 0.03507225215435028, 0.060241758823394775, 0.006045821588486433, 0.06172458082437515, -0.039035018533468246, -0.06344820559024811, 0.12700039148330688, -0.11641156673431396, -0.01574569195508957, 0.04237627238035202, -0.03988158702850342, 0.1266709417104721, 0.06103584170341492, -0.03386637568473816, -0.13451583683490753, -0.016748178750276566, -0.061842694878578186, -0.022647274658083916, -0.0811290442943573, -0.11244743317365646, -0.006990712136030197, -0.03151422739028931, -0.08039664477109909, -0.11589031666517258, -0.1964283138513565, -0.04938383400440216, 0.0251497570425272, -0.015751145780086517, 0.008019106462597847, 0.020760713145136833, 0.010555039159953594, -0.013025450520217419, -0.0005269937682896852, -0.12330704182386398, -0.011881678365170956, 0.039356306195259094, -0.04519624263048172, 0.04318727180361748, -0.023774875327944756, 0.04382455721497536, -0.15159514546394348, -0.006994709372520447, -0.1998857855796814, 0.12895575165748596, -0.011508987285196781, -0.009925353340804577, -0.060999833047389984, -0.030540822073817253, -0.08068440109491348, 0.003672819584608078, -0.01858660764992237, 0.1398070901632309, -0.1300959438085556, -0.04728333279490471, 0.21715223789215088, -0.21367913484573364, 0.018676448613405228, 0.07751264423131943, -0.03634115308523178, 0.11193360388278961, 0.1742401272058487, 0.03388009965419769, 0.169630229473114, -0.04856349900364876, -0.06744334846735, 0.025023916736245155, -0.049553461372852325, 0.0841866210103035, 0.03878745436668396, -0.04558984935283661, -0.020484665408730507, 0.005341008771210909, -0.0530240535736084, -0.018105167895555496, -0.019939562305808067, -0.029043978080153465, -0.014806663617491722, -0.017404958605766296, 0.014281390234827995, 0.028693631291389465, 0.023907648399472237, 0.029412081465125084, -0.11411219835281372, -0.012179218232631683, 0.0991644412279129, -0.09574687480926514, 0.03046921268105507, -0.10425379127264023, 0.04111042246222496, -0.05553701892495155, -0.004259963985532522, -0.2008158266544342, -0.03347747400403023, 0.04292900487780571, -0.07460912317037582, 0.07008102536201477, 0.06878969818353653, 0.03069956600666046, 0.07906096428632736, -0.009026714600622654, -0.006173059809952974, -0.021207710728049278, -0.008096421137452126, -0.023638378828763962, -0.09420966356992722, -0.06010206416249275, -0.05117630958557129, 0.0646233931183815, -0.08444435149431229, 0.03604499250650406, 0.05630674585700035, 0.0038457317277789116, 0.04378003627061844, -0.06383378058671951, 0.02030775137245655, -0.004008675459772348, -0.007389744743704796, -0.041585396975278854, 0.03390048071742058, 0.036733146756887436, -0.058936502784490585, 0.057793062180280685, -0.1821596920490265, -0.09999044239521027, 0.08454587310552597, 0.026340724900364876, -0.1476602405309677, 0.048205673694610596, 0.005576552357524633, -0.00099513353779912, -0.03888319060206413, -0.08737154304981232, 0.2242482304573059, 0.006506490055471659, 0.09369395673274994, -0.07475197315216064, -0.031237345188856125, 0.030853310599923134, -0.027858592569828033, 0.002153845503926277, 0.04893254488706589, -0.004734594374895096, -0.11283078044652939, 0.053149498999118805, 0.024080615490674973, 0.033373791724443436, 0.152401402592659, 0.011436570435762405, -0.07609713077545166, -0.010297768749296665, 0.009463481605052948, 0.006391777191311121, 0.004370823036879301, 0.0016927894903346896, 0.016970815137028694, 0.03721139207482338, 0.10011585056781769, 0.028239384293556213, -0.05442193150520325, 0.06246750056743622, 0.09005964547395706, -0.024559082463383675, -0.063806913793087, -0.04507536441087723, -0.03421007841825485, 0.08101155608892441, 0.028205540031194687, 0.12940643727779388, 0.037187837064266205, -0.023380231112241745, -0.15538758039474487, 0.15168815851211548, -0.09324292838573456, -0.1917342096567154, -0.12159183621406555, -0.009774924255907536, 0.03414591774344444, 0.04605361819267273, 0.05238204821944237, -0.07596683502197266, -0.07122372090816498, -0.11254926025867462, 0.11454535275697708, -0.056429922580718994, -0.0611187182366848, -0.031065119430422783, -0.06750026345252991, -0.00768709322437644, -0.11632276326417923, 0.008539263159036636, 0.01482471078634262, -0.1230802983045578, 0.01496812328696251, -0.06294281035661697, 0.03200312703847885, 0.16976462304592133, -0.016793420538306236, -0.007321314420551062, -0.009283149614930153, 0.23798784613609314, -0.020149029791355133, 0.08884551376104355, 0.1887294501066208, -0.0826597809791565, 0.041425738483667374, 0.05036213994026184, 0.01396977435797453, -0.022012725472450256, 0.03873845562338829, -0.0027141831815242767, -0.08869310468435287, -0.16725201904773712, -0.058745045214891434, -0.02520456351339817, -0.01799386367201805, 0.028384525328874588, 0.04172491282224655, 0.02310025691986084, 0.060444001108407974, -0.06512971967458725, -0.003775784745812416, 0.060784731060266495, 0.08955452591180801, 0.03104645386338234, -0.03058762475848198, 0.06645160168409348, -0.08683129400014877, 0.03561224415898323, 0.1018717959523201, -0.07190694659948349, 0.21390756964683533, -0.02865748107433319, 0.20287038385868073, 0.06972659379243851, 0.019782084971666336, 0.11142026633024216, 0.07485757768154144, -0.03749280422925949, 0.01774277165532112, -0.02425377629697323, -0.07404379546642303, -0.07693268358707428, -0.014485619030892849, -0.041688840836286545, 0.0354139618575573, -0.12376299500465393, -0.045931290835142136, -0.0045866696164011955, 0.17808544635772705, 0.0010852201376110315, -0.17043176293373108, -0.12307755649089813, 0.031816571950912476, -0.0031842272728681564, -0.07417833060026169, 0.018047776073217392, 0.06599495559930801, -0.12797172367572784, -0.009244666434824467, -0.001706017297692597, 0.08642062544822693, -0.15925487875938416, 0.013373227789998055, -0.04878619685769081, 0.011523543857038021, -0.022315792739391327, 0.07174697518348694, -0.07305708527565002, 0.022684210911393166, 0.01416821964085102, 0.10060883313417435, -0.04424956068396568, 0.014356706291437149, -0.03132672980427742, 0.13400517404079437, 0.09421643614768982, 0.029394282028079033, -0.03604396805167198, -0.09924326092004776, -0.056327883154153824, 0.03485793620347977, 0.03857249766588211, -0.06821291893720627, 0.09385310113430023, -0.0358082540333271, 0.05006118491292, -0.0123799629509449, 0.0004761549935210496, -0.1166892722249031, -0.13346774876117706, 0.0497208908200264, -0.02821207419037819, 0.09997554868459702, -0.03578733652830124, -0.03590761125087738, -0.023181358352303505, 0.14449267089366913, -0.16283032298088074, -0.11950439214706421, -0.13019075989723206, -0.02782929316163063, 0.03955506533384323, -0.06321362406015396, 0.011634094640612602, -0.026809662580490112, 0.11295173317193985, 0.027344390749931335, -0.09705402702093124, 0.017492279410362244, -0.06201328709721565, -0.14765742421150208, -0.02283027581870556, 0.042853593826293945, 0.17000724375247955, 0.04525583237409592, 0.001227637636475265, 0.02623705565929413, -0.01437359768897295, -0.1173085868358612, -0.03795832023024559, 0.15857195854187012, 0.016939735040068626, 0.09670943766832352, -0.038943566381931305, -0.13603048026561737, -0.030870957300066948, 0.005046611186116934, 0.12636059522628784, 0.13662801682949066, -0.06831731647253036, 0.13590127229690552, 0.2517032325267792, -0.1047087162733078, -0.1982477754354477, -0.009869898669421673, 0.012772059999406338, 0.0299700815230608, 0.0226057767868042, -0.2025090754032135, 0.08200681954622269, 0.05066714435815811, 0.00007871015259297565, -0.030204463750123978, -0.30713024735450745, -0.07429040223360062, 0.10235248506069183, 0.08986850827932358, 0.11466658115386963, -0.08790472894906998, -0.01601814106106758, -0.002657964825630188, -0.15100102126598358, 0.15137773752212524, -0.13974735140800476, 0.06141749024391174, 0.013424265198409557, -0.009664380922913551, 0.027719631791114807, -0.0337507389485836, 0.06731141358613968, 0.0031467683147639036, 0.05765066668391228, -0.08246906101703644, 0.04363030940294266, 0.11230410635471344, -0.023326991125941277, 0.14304915070533752, 0.05628731846809387, 0.06718692928552628, -0.09961216896772385, -0.06091051548719406, -0.0725732296705246, 0.038070812821388245, -0.055117227137088776, -0.04707646369934082, -0.06818805634975433, 0.057047560811042786, 0.06031883507966995, -0.007557165343314409, -0.018983354791998863, -0.09565889835357666, 0.07738231122493744, 0.11886093765497208, 0.1500626653432846, 0.0900828093290329, -0.148541659116745, -0.0445837676525116, -0.01725863479077816, 0.1038166806101799, -0.07648530602455139, 0.05486227944493294, 0.06436873227357864, 0.021943071857094765, 0.10534057766199112, 0.03522151708602905, -0.1440119445323944, 0.03385121747851372, -0.0025017503648996353, -0.11080818623304367, -0.10246201604604721, 0.011643961071968079, 0.017068225890398026, -0.10570995509624481, 0.002354075200855732, 0.125156432390213, -0.07520854473114014, -0.006992538459599018, -0.014660862274467945, 0.019349755719304085, 0.008965539745986462, 0.07767322659492493, 0.02265448495745659, 0.015474834479391575, -0.06744325906038284, 0.09319225698709488, 0.10436607897281647, -0.07587625831365585, 0.05950107425451279, 0.0560777522623539, -0.11189382523298264, -0.08020676672458649, -0.041353147476911545, 0.08528265357017517, -0.0058243353851139545, -0.07781713455915451, 0.029429124668240547, -0.09533508867025375, 0.05526241287589073, 0.0747372955083847, 0.0003881615411955863, 0.07954500615596771, -0.07825610041618347, 0.024308545514941216, -0.08158658444881439, 0.04497675970196724, -0.01769804209470749, -0.011058508418500423, -0.07418245822191238, 0.13317202031612396, 0.07386469095945358, 0.02827146090567112, -0.030906962230801582, -0.10355567187070847, -0.09149038791656494, 0.00273761129938066, -0.10163216292858124, 0.006695651914924383, -0.013111721724271774, -0.01714516431093216, -0.02126922644674778, 0.017098816111683846, 0.01321739237755537, 0.027933742851018906, -0.044054143130779266, 0.002096385695040226, -0.03243279457092285, 0.03545140475034714, -0.09401926398277283, 0.05581928417086601, 0.03603518754243851, -0.036506347358226776, 0.09433263540267944, 0.04679560661315918, -0.06916452199220657, 0.04634538292884827, -0.06381243467330933, 0.06684797257184982, -0.05933907628059387, -0.028378400951623917, -0.009244704619050026, -0.08848007023334503, -0.0118925292044878, 0.014714310877025127, -0.0576586052775383, 0.009618030861020088, 0.11796432733535767, -0.08533263206481934, 0.10804858058691025, 0.041158370673656464, 0.029929080978035927, -0.10137982666492462, 0.05535393953323364, 0.01011277548968792, 0.04027124494314194, 0.10874530673027039, -0.03215542808175087, 0.07619879394769669, -0.14084577560424805, -0.0287298783659935, 0.044961050152778625, 0.04095657169818878, -0.02972586080431938, -0.03657578304409981, 0.024198468774557114, -0.0262372475117445, 0.07555128633975983, -0.013833112083375454, 0.015832193195819855, 0.02088044211268425, -0.014122070744633675, -0.050193093717098236, 0.014000587165355682, 0.08929194509983063, -0.001185741857625544, -0.03859403729438782, 0.011314167641103268, 0.024127673357725143, -0.06849178671836853, -0.038535844534635544, 0.16729393601417542, 0.04245004430413246, 0.04383130371570587, 0.012257985770702362, -0.050697989761829376, -0.011497058905661106, -0.07798753678798676, -0.023299193009734154, -0.005983153358101845, 0.019151728600263596, -0.029255595058202744, 0.1006731390953064, 0.14913304150104523, -0.030280664563179016, 0.10936494171619415, -0.00920049101114273, -0.05431177094578743, -0.10371029376983643, -0.23453597724437714, 0.010016106069087982, -0.042524922639131546, -0.039289798587560654, -0.10646563023328781, 0.01821770891547203, 0.07074809819459915, 0.025202298536896706, -0.03944597393274307, 0.14319159090518951, -0.09141607582569122, -0.10563967376947403, -0.006647248286753893, -0.013700735755264759, 0.035009074956178665, 0.002364678308367729, 0.0628308355808258, 0.08550570160150528, 0.0794668197631836, 0.07090575248003006, 0.10746470838785172, 0.06537623703479767, 0.013741930015385151, -0.05588584393262863, -0.07971460372209549, -0.008047493174672127, 0.023357974365353584, 0.014086995273828506, 0.14997467398643494, 0.01990273967385292, -0.05089792609214783, -0.016252873465418816, 0.16588518023490906, -0.06352142989635468, -0.05149798095226288, -0.13200779259204865, 0.22700916230678558, -0.000940927944611758, -0.01343104150146246, 0.028118731454014778, -0.11324301362037659, 0.02472694031894207, 0.1851005256175995, 0.14732597768306732, 0.01489801611751318, 0.025792401283979416, -0.011808495037257671, 0.00245007430203259, 0.015158932656049728, 0.1389920711517334, 0.02261660248041153, 0.30954796075820923, -0.07176174223423004, 0.18457242846488953, -0.039597127586603165, -0.0076875146478414536, -0.06768596917390823, 0.0840938538312912, -0.04330591857433319, 0.027141690254211426, -0.07105221599340439, 0.07482253760099411, -0.10403519868850708, -0.20664158463478088, -0.005985601805150509, 0.0075834146700799465, -0.023494772613048553, -0.009598692879080772, -0.010953743942081928, 0.015441290102899075, 0.061239708214998245, -0.0069667985662817955, -0.00010135251068277285, 0.17345742881298065, 0.02875087969005108, -0.07247859984636307, -0.06668023765087128, 0.10418626666069031, -0.07802809029817581, 0.13415047526359558, 0.012433628551661968, 0.12179679423570633, 0.07880386710166931, -0.01003110222518444, -0.10978666692972183, 0.05774702876806259, -0.047928448766469955, -0.007762210443615913, -0.01810980588197708, 0.12889884412288666, 0.0057132956571877, 0.11797838658094406, 0.046141285449266434, -0.09692680835723877, 0.021761661395430565, -0.022898467257618904, 0.004542508628219366, -0.07342119514942169, 0.047209564596414566, -0.03705413267016411, 0.14143134653568268, 0.15437068045139313, -0.013322659768164158, -0.022339360788464546, -0.05567643418908119, 0.03239695355296135, -0.020866936072707176, 0.04822634533047676, 0.00393180875107646, -0.12808376550674438, -0.007458766922354698, 0.058648958802223206, 0.06178426742553711, -0.24276557564735413, -0.0410868339240551, 0.009427343495190144, -0.020155206322669983, -0.004441636614501476, 0.05871313437819481, 0.01709906756877899, 0.05773210898041725, -0.04317173734307289, 0.08393094688653946, -0.01807514578104019, 0.10205432772636414, -0.09764676541090012, -0.06986042112112045 ]
null
null
transformers
# Funnel Transformer xlarge model (B10-10-10 without decoder) Pretrained model on English language using a similar objective objective as [ELECTRA](https://huggingface.co/transformers/model_doc/electra.html). It was introduced in [this paper](https://arxiv.org/pdf/2006.03236.pdf) and first released in [this repository](https://github.com/laiguokun/Funnel-Transformer). This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. **Note:** This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth of the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if you need one input per initial token. You should use the `xlarge` model in that case. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=funnel-transformer) to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import FunnelTokenizer, FunnelBaseModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/xlarge-base") model = FunnelBaseModel.from_pretrained("funnel-transformer/xlarge-base") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import FunnelTokenizer, TFFunnelBaseModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/xlarge-base") model = TFFunnelBaseModel.from_pretrained("funnel-transformer/xlarge-base") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` ## Training data The BERT model was pretrained on: - [BookCorpus](https://yknzhu.wixsite.com/mbweb), a dataset consisting of 11,038 unpublished books, - [English Wikipedia](https://en.wikipedia.org/wiki/English_Wikipedia) (excluding lists, tables and headers), - [Clue Web](https://lemurproject.org/clueweb12/), a dataset of 733,019,372 English web pages, - [GigaWord](https://catalog.ldc.upenn.edu/LDC2011T07), an archive of newswire text data, - [Common Crawl](https://commoncrawl.org/), a dataset of raw web pages. ### BibTeX entry and citation info ```bibtex @misc{dai2020funneltransformer, title={Funnel-Transformer: Filtering out Sequential Redundancy for Efficient Language Processing}, author={Zihang Dai and Guokun Lai and Yiming Yang and Quoc V. Le}, year={2020}, eprint={2006.03236}, archivePrefix={arXiv}, primaryClass={cs.LG} } ```
{"language": "en", "license": "apache-2.0", "datasets": ["bookcorpus", "wikipedia", "gigaword"]}
feature-extraction
funnel-transformer/xlarge-base
[ "transformers", "pytorch", "tf", "safetensors", "funnel", "feature-extraction", "en", "dataset:bookcorpus", "dataset:wikipedia", "dataset:gigaword", "arxiv:2006.03236", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2006.03236" ]
[ "en" ]
TAGS #transformers #pytorch #tf #safetensors #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us
# Funnel Transformer xlarge model (B10-10-10 without decoder) Pretrained model on English language using a similar objective objective as ELECTRA. It was introduced in this paper and first released in this repository. This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. Note: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth of the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if you need one input per initial token. You should use the 'xlarge' model in that case. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: and in TensorFlow: ## Training data The BERT model was pretrained on: - BookCorpus, a dataset consisting of 11,038 unpublished books, - English Wikipedia (excluding lists, tables and headers), - Clue Web, a dataset of 733,019,372 English web pages, - GigaWord, an archive of newswire text data, - Common Crawl, a dataset of raw web pages. ### BibTeX entry and citation info
[ "# Funnel Transformer xlarge model (B10-10-10 without decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nNote: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth\nof the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if\nyou need one input per initial token. You should use the 'xlarge' model in that case.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ "TAGS\n#transformers #pytorch #tf #safetensors #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us \n", "# Funnel Transformer xlarge model (B10-10-10 without decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nNote: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth\nof the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if\nyou need one input per initial token. You should use the 'xlarge' model in that case.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ 77, 104, 289, 133, 32, 95, 11 ]
[ "passage: TAGS\n#transformers #pytorch #tf #safetensors #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #region-us \n# Funnel Transformer xlarge model (B10-10-10 without decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nNote: This model does not contain the decoder, so it ouputs hidden states that have a sequence length of one fourth\nof the inputs. It's good to use for tasks requiring a summary of the sentence (like sentence classification) but not if\nyou need one input per initial token. You should use the 'xlarge' model in that case." ]
[ -0.07125501334667206, 0.10804367810487747, -0.005233537871390581, 0.05262725055217743, 0.10012458264827728, -0.009352238848805428, 0.038332510739564896, 0.07851545512676239, -0.13262659311294556, 0.0990750715136528, -0.0056153773330152035, -0.046391427516937256, 0.09611152857542038, 0.052023209631443024, 0.09637338668107986, -0.23258303105831146, 0.06220749765634537, -0.04597070813179016, 0.12767541408538818, 0.07541222870349884, 0.08666475117206573, -0.08597681671380997, 0.07392984628677368, 0.01591400057077408, -0.045792222023010254, -0.029958628118038177, -0.00674477918073535, -0.033797916024923325, 0.038232024759054184, 0.07899630814790726, 0.08492270857095718, -0.03136978670954704, 0.037031129002571106, -0.05415002256631851, 0.015174128115177155, 0.09445606172084808, 0.006053289398550987, 0.04495975747704506, 0.03992820158600807, 0.028048226609826088, 0.06767289340496063, -0.04294081777334213, 0.05012387037277222, 0.04934616759419441, -0.12645594775676727, -0.20010514557361603, -0.06400974839925766, 0.13518548011779785, 0.035401009023189545, 0.09990793466567993, -0.03557814657688141, 0.12285760790109634, -0.015278635546565056, 0.05372312664985657, 0.13098376989364624, -0.17387425899505615, -0.007257802877575159, -0.04130301624536514, -0.015779651701450348, 0.08506640046834946, -0.03864435479044914, -0.05249566212296486, -0.017212800681591034, 0.035536233335733414, 0.033944468945264816, -0.01595510169863701, -0.02442733384668827, -0.06265685707330704, -0.12407621741294861, -0.04800654202699661, 0.13218927383422852, -0.037000883370637894, -0.11310257017612457, -0.09835732728242874, -0.03499535471200943, -0.053673408925533295, 0.010851159691810608, -0.034759972244501114, 0.0015632990980520844, 0.016569998115301132, 0.07462472468614578, -0.04495915398001671, -0.11277684569358826, -0.007742428220808506, -0.08559953421354294, 0.0910230427980423, 0.02540256269276142, 0.04897711053490639, -0.09104286879301071, 0.07265487313270569, -0.1092502698302269, -0.054365113377571106, -0.07256284356117249, -0.04043274745345116, -0.07837160676717758, -0.032343730330467224, -0.04520491883158684, -0.11392207443714142, -0.049646832048892975, 0.04282597824931145, -0.08660896867513657, 0.028327472507953644, -0.010171539150178432, 0.03096877411007881, 0.10882879048585892, 0.1325858235359192, -0.03937675803899765, 0.07260271906852722, 0.004492637701332569, -0.0553322434425354, 0.055685047060251236, -0.03978465124964714, -0.050204258412122726, -0.017701493576169014, 0.040651675313711166, 0.0487445704638958, 0.023973966017365456, 0.043322447687387466, -0.034832924604415894, -0.04445047304034233, 0.136178657412529, -0.11789115518331528, 0.0188473928719759, 0.040572091937065125, -0.027202831581234932, 0.1061873733997345, 0.042225487530231476, -0.04815668985247612, -0.13376158475875854, -0.007844304665923119, -0.06623276323080063, -0.014517231844365597, -0.07159993797540665, -0.1030396968126297, 0.0026835750322788954, -0.063683420419693, -0.07794912159442902, -0.11125130951404572, -0.1803903579711914, -0.0675995722413063, 0.00002551147736085113, -0.016639098525047302, 0.012920781038701534, 0.010280369780957699, -0.005797977559268475, -0.013268424198031425, 0.0052672796882689, -0.11562968045473099, 0.0036241873167455196, 0.03774585202336311, -0.052554864436388016, 0.030098458752036095, -0.032916110008955, 0.03685436025261879, -0.15982992947101593, 0.00014188408385962248, -0.1849505454301834, 0.1346074342727661, 0.01279470231384039, 0.0072164637967944145, -0.061583686619997025, -0.023183604702353477, -0.0752623900771141, 0.024134008213877678, -0.007077099289745092, 0.1141517236828804, -0.14785346388816833, -0.04950958117842674, 0.18769221007823944, -0.20721544325351715, 0.009967290796339512, 0.10052914172410965, -0.022504746913909912, 0.08218465000391006, 0.1648932546377182, 0.08004704862833023, 0.17286108434200287, -0.024806398898363113, -0.08150864392518997, 0.029845841228961945, -0.08102407306432724, 0.07734446972608566, 0.0551210418343544, -0.08714266121387482, -0.015037024393677711, -0.0013734634267166257, -0.04777159169316292, -0.023745344951748848, 0.0013899659970775247, -0.0255365539342165, -0.0017499338136985898, 0.0027405377477407455, 0.015051637776196003, 0.017927369102835655, 0.005216551944613457, 0.014848674647510052, -0.11190417408943176, 0.023061538115143776, 0.09119251370429993, -0.06686840206384659, 0.03746485710144043, -0.0873885303735733, 0.06036824360489845, -0.051979538053274155, -0.013146360404789448, -0.18435272574424744, -0.015727953985333443, 0.037487659603357315, -0.09289685636758804, 0.07121530175209045, 0.04375046491622925, 0.03015478141605854, 0.07779380679130554, 0.010465925559401512, -0.008445097133517265, -0.006459763273596764, -0.015028435736894608, -0.058163002133369446, -0.08723104745149612, -0.05398212745785713, -0.04671258479356766, 0.026122812181711197, -0.06250959634780884, 0.01670517399907112, 0.051645249128341675, -0.0030397262889891863, 0.039829887449741364, -0.06147421896457672, 0.0090390769764781, -0.000724600744433701, 0.00666681258007884, -0.03543519601225853, 0.031790055334568024, 0.043675944209098816, -0.04755685105919838, 0.057643696665763855, -0.2196640819311142, -0.13200660049915314, 0.06681853532791138, 0.04095923528075218, -0.1401967704296112, 0.0007584467530250549, 0.011232375167310238, 0.004822980612516403, -0.03592028096318245, -0.09647621214389801, 0.22843772172927856, -0.003742540953680873, 0.08291880041360855, -0.08715742081403732, -0.025816500186920166, 0.02694542147219181, -0.032507963478565216, -0.012811660766601562, 0.055841512978076935, -0.027656322345137596, -0.1487773358821869, 0.06146817281842232, 0.03241051360964775, 0.022517921403050423, 0.1461350917816162, 0.008771604858338833, -0.06808203458786011, -0.028175506740808487, 0.009182611480355263, 0.01816686987876892, -0.0035808885004371405, -0.008917747065424919, 0.003926908131688833, 0.0310718584805727, 0.07896723598241806, 0.02606438845396042, -0.05248266085982323, 0.07354740798473358, 0.09303232282400131, 0.002192805288359523, -0.027640951797366142, -0.04888143390417099, -0.024814698845148087, 0.07696619629859924, 0.045106250792741776, 0.11087583005428314, 0.04076281189918518, -0.021815022453665733, -0.1533508002758026, 0.15406477451324463, -0.11271212249994278, -0.20505189895629883, -0.13430148363113403, 0.03427928686141968, 0.035856980830430984, 0.048247430473566055, 0.04490148276090622, -0.06924448907375336, -0.0576767772436142, -0.11430097371339798, 0.10328398644924164, -0.041077565401792526, -0.04664035886526108, -0.025989817455410957, -0.056241314858198166, -0.015025329776108265, -0.09996938705444336, 0.008785717189311981, -0.0018777588848024607, -0.10584639012813568, -0.014345979318022728, -0.035614125430583954, 0.02071264572441578, 0.15155808627605438, -0.009432284161448479, -0.017010236158967018, -0.015685265883803368, 0.20342892408370972, -0.03466194495558739, 0.10366839170455933, 0.18966448307037354, -0.07236975431442261, 0.05034754052758217, 0.05074001103639603, 0.004338710568845272, 0.0038849404081702232, 0.012149520218372345, -0.00856021884828806, -0.06432711333036423, -0.13761387765407562, -0.05501914396882057, -0.025259636342525482, 0.0012045338517054915, 0.03586341813206673, 0.03819591552019119, 0.0098531823605299, 0.058438073843717575, -0.049372389912605286, 0.0004041742067784071, 0.03729227930307388, 0.08075996488332748, 0.04354404658079147, -0.018902316689491272, 0.05087847262620926, -0.06926462054252625, 0.027132263407111168, 0.10627579689025879, -0.06301189213991165, 0.15982988476753235, -0.0469810925424099, 0.1650308072566986, 0.046332888305187225, 0.004945347085595131, 0.09331440180540085, 0.05030927062034607, -0.05386458709836006, 0.01983865350484848, -0.028859207406640053, -0.08507576584815979, -0.06453845649957657, 0.010246559046208858, -0.02493979223072529, 0.042032480239868164, -0.12633979320526123, -0.02828524075448513, 0.03255277872085571, 0.14056292176246643, 0.012518931180238724, -0.15988537669181824, -0.1282450407743454, 0.023661267012357712, -0.005815410520881414, -0.07449235022068024, 0.011054436676204205, 0.0904596745967865, -0.11228947341442108, -0.012077857740223408, 0.003876602277159691, 0.06676682084798813, -0.11435823142528534, 0.026354799047112465, -0.03676748275756836, 0.06400064378976822, -0.01327422633767128, 0.0655621662735939, -0.07461947947740555, 0.0008262769551947713, 0.015792788937687874, 0.10852428525686264, -0.04617482051253319, 0.029837701469659805, -0.023656973615288734, 0.10881363600492477, 0.11247553676366806, 0.03833596408367157, -0.04547296464443207, -0.08160047978162766, -0.0567263700067997, 0.022098882123827934, 0.06439179182052612, -0.06339986622333527, 0.0852634385228157, -0.034596558660268784, 0.03193014860153198, -0.02208000235259533, -0.0016523413360118866, -0.06256171315908432, -0.13742774724960327, 0.04627291113138199, -0.034813396632671356, 0.07942105084657669, -0.042405419051647186, -0.012633081525564194, -0.029268812388181686, 0.11506509780883789, -0.15904268622398376, -0.11362555623054504, -0.11305614560842514, -0.04027581959962845, 0.058058056980371475, -0.07689637690782547, 0.01778324879705906, -0.02892470359802246, 0.11139006167650223, 0.00995478592813015, -0.08224102109670639, 0.02773674950003624, -0.05794049799442291, -0.14812453091144562, -0.029664870351552963, 0.06149628385901451, 0.13408346474170685, 0.04190659523010254, 0.012989330105483532, 0.04776706174015999, -0.02470659278333187, -0.10699761658906937, -0.020715763792395592, 0.1501534879207611, 0.01988002099096775, 0.09349210560321808, -0.03606542944908142, -0.1216278076171875, -0.03073415532708168, 0.008977538906037807, 0.13140122592449188, 0.10367552936077118, -0.07428552210330963, 0.13102854788303375, 0.25318828225135803, -0.10761044174432755, -0.2244618684053421, -0.007627875078469515, 0.03251876309514046, 0.019829731434583664, 0.02671217732131481, -0.18895743787288666, 0.09268593788146973, 0.06836100667715073, 0.0032152782659977674, -0.034073587507009506, -0.2916250228881836, -0.06740714609622955, 0.07362579554319382, 0.048331886529922485, 0.057959578931331635, -0.10327986627817154, -0.0322246253490448, 0.010265513323247433, -0.03866306319832802, 0.15823107957839966, -0.12265202403068542, 0.05352896451950073, 0.018359294161200523, -0.029519405215978622, 0.04282451793551445, -0.026474686339497566, 0.0689220130443573, -0.01359070185571909, 0.04795600473880768, -0.07271280139684677, -0.002091268077492714, 0.08890079706907272, -0.02867692895233631, 0.12545070052146912, 0.048816632479429245, 0.06754085421562195, -0.0813506469130516, -0.04459694027900696, -0.0906086191534996, 0.04918893054127693, -0.05870778113603592, -0.03771595284342766, -0.06550475209951401, 0.07484373450279236, 0.05016990751028061, -0.007771763950586319, -0.0027077957056462765, -0.08655638992786407, 0.07141058146953583, 0.10910798609256744, 0.15401165187358856, 0.09498460590839386, -0.15268254280090332, -0.03432371839880943, -0.024934763088822365, 0.11829279363155365, -0.04515505209565163, 0.04026013985276222, 0.05647008493542671, 0.0104221748188138, 0.10228946059942245, 0.0390787236392498, -0.1513548195362091, 0.02230544574558735, 0.010315661318600178, -0.08795183897018433, -0.11355230212211609, 0.001453928998671472, 0.03708801418542862, -0.11610209941864014, 0.002326984889805317, 0.12593962252140045, -0.060550469905138016, -0.01992461085319519, -0.012937637977302074, 0.05534830316901207, 0.004405815154314041, 0.06006457284092903, 0.03266081586480141, 0.014123800210654736, -0.050868142396211624, 0.10284598916769028, 0.09750716388225555, -0.09743920713663101, 0.0776885375380516, 0.11035455018281937, -0.08690066635608673, -0.09409920871257782, -0.03998030722141266, 0.09419023990631104, -0.013912547379732132, -0.07739270478487015, 0.03768683224916458, -0.11204305291175842, 0.05683198943734169, 0.08034714311361313, -0.011684819124639034, 0.07608489692211151, -0.06624823808670044, 0.027373958379030228, -0.07294183224439621, 0.0459868349134922, -0.03660891205072403, -0.014423064887523651, -0.05311904102563858, 0.1784413605928421, 0.037894561886787415, 0.032896384596824646, -0.01540167722851038, -0.07380271703004837, -0.09968394786119461, -0.012880811467766762, -0.0579427145421505, 0.023570284247398376, -0.025403907522559166, -0.018214266747236252, -0.008885367773473263, 0.042154308408498764, 0.031643159687519073, 0.020565088838338852, -0.025484442710876465, -0.002423142781481147, -0.032448407262563705, 0.03198747709393501, -0.08344858139753342, 0.0464775413274765, 0.02879537083208561, -0.022039776667952538, 0.10112417489290237, 0.017664939165115356, -0.04817599803209305, 0.02392149344086647, -0.0766308382153511, 0.04037310928106308, -0.07121733576059341, -0.028949536383152008, -0.011747782118618488, -0.07472002506256104, -0.00649255933240056, 0.008008996956050396, -0.06344026327133179, -0.00001188980877486756, 0.09724528342485428, -0.07473088055849075, 0.13449418544769287, 0.049157217144966125, 0.029796969145536423, -0.10944461077451706, 0.04105643182992935, -0.020583413541316986, 0.021223697811365128, 0.08524057269096375, -0.04584112763404846, 0.08406484872102737, -0.12118665874004364, -0.021209776401519775, 0.03268590196967125, 0.02827487327158451, -0.04425261169672012, -0.03219565004110336, 0.04000871255993843, -0.016000041738152504, 0.027841370552778244, 0.02536550909280777, 0.013827362097799778, 0.017136968672275543, -0.007415022701025009, -0.057660892605781555, 0.016705000773072243, 0.05412368103861809, -0.013959870673716068, -0.05266242474317551, -0.004224398639053106, 0.01073366403579712, -0.06859640777111053, -0.0019409620435908437, 0.1603514403104782, 0.05056465044617653, 0.07010046392679214, 0.03367166221141815, -0.03016234003007412, -0.03115292266011238, -0.0918867215514183, -0.024446668103337288, -0.0048040770925581455, 0.01948336884379387, -0.03607229143381119, 0.05125877261161804, 0.1478257179260254, -0.03296203911304474, 0.11641670018434525, -0.00961344689130783, -0.04922657087445259, -0.09469205886125565, -0.23731538653373718, 0.01051054336130619, 0.006751743145287037, -0.048613257706165314, -0.10702095180749893, 0.027094395831227303, 0.030840730294585228, 0.010469927452504635, -0.03772713616490364, 0.1313229650259018, -0.10059674829244614, -0.09782055765390396, 0.014789124950766563, -0.01460457406938076, 0.06899372488260269, 0.0396224744617939, 0.047003455460071564, 0.07888578623533249, 0.0706733912229538, 0.0794137567281723, 0.10373517125844955, 0.08103157579898834, 0.014383619651198387, -0.054885849356651306, -0.07867894321680069, -0.0004158085212111473, -0.0010813414119184017, -0.0005792116862721741, 0.14089561998844147, 0.03273090720176697, -0.05729737877845764, -0.013157111592590809, 0.1462380290031433, -0.06239759922027588, -0.05817436799407005, -0.11654123663902283, 0.21962350606918335, -0.001171969110146165, -0.007063268218189478, 0.006691321264952421, -0.11522532254457474, 0.030136357992887497, 0.18866731226444244, 0.13007180392742157, 0.01110483705997467, 0.03593512624502182, 0.003983958624303341, 0.0049425819888710976, 0.01128486916422844, 0.1340278834104538, 0.009447717107832432, 0.35051611065864563, -0.06881996244192123, 0.20543146133422852, -0.03031042031943798, 0.013047519139945507, -0.08567439019680023, 0.06957601755857468, -0.04554867371916771, 0.0474567711353302, -0.05167397856712341, 0.05374973639845848, -0.08796397596597672, -0.21279868483543396, 0.013307305984199047, -0.002252900041639805, -0.027903804555535316, -0.0015018777921795845, 0.023733042180538177, 0.034514497965574265, 0.08563987165689468, 0.01848151721060276, 0.00368850608356297, 0.15184639394283295, 0.015318230725824833, -0.07219769805669785, -0.056660253554582596, 0.08574075996875763, -0.0624593161046505, 0.14477324485778809, 0.019915899261832237, 0.08935628831386566, 0.08201995491981506, -0.0071210903115570545, -0.11805960536003113, 0.05860636755824089, -0.047281913459300995, -0.012144319713115692, 0.008065085858106613, 0.10818784683942795, 0.014582463540136814, 0.09400275349617004, 0.05222468450665474, -0.07894463837146759, 0.028383493423461914, 0.015967540442943573, -0.0008711499394848943, -0.07053305953741074, 0.05036640539765358, -0.04901118949055672, 0.16010373830795288, 0.1534317135810852, -0.016595957800745964, -0.04307654872536659, -0.0436549037694931, 0.015537888742983341, -0.01927301473915577, 0.060367245227098465, 0.0018957898719236255, -0.08885554224252701, 0.006511203944683075, 0.029311560094356537, 0.0804150253534317, -0.2064255177974701, -0.027704186737537384, 0.009460086934268475, -0.018613966181874275, -0.02824993245303631, 0.04489097371697426, 0.0356341190636158, 0.04115917906165123, -0.0415937565267086, 0.08638925850391388, -0.002553375670686364, 0.09793571382761002, -0.06913892179727554, -0.06336020678281784 ]
null
null
transformers
# Funnel Transformer xlarge model (B10-10-10 with decoder) Pretrained model on English language using a similar objective objective as [ELECTRA](https://huggingface.co/transformers/model_doc/electra.html). It was introduced in [this paper](https://arxiv.org/pdf/2006.03236.pdf) and first released in [this repository](https://github.com/laiguokun/Funnel-Transformer). This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=funnel-transformer) to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import FunnelTokenizer, FunnelModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/xlarge") model = FunneModel.from_pretrained("funnel-transformer/xlarge") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import FunnelTokenizer, TFFunnelModel tokenizer = FunnelTokenizer.from_pretrained("funnel-transformer/xlarge") model = TFFunnelModel.from_pretrained("funnel-transformer/xlarge") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` ## Training data The BERT model was pretrained on: - [BookCorpus](https://yknzhu.wixsite.com/mbweb), a dataset consisting of 11,038 unpublished books, - [English Wikipedia](https://en.wikipedia.org/wiki/English_Wikipedia) (excluding lists, tables and headers), - [Clue Web](https://lemurproject.org/clueweb12/), a dataset of 733,019,372 English web pages, - [GigaWord](https://catalog.ldc.upenn.edu/LDC2011T07), an archive of newswire text data, - [Common Crawl](https://commoncrawl.org/), a dataset of raw web pages. ### BibTeX entry and citation info ```bibtex @misc{dai2020funneltransformer, title={Funnel-Transformer: Filtering out Sequential Redundancy for Efficient Language Processing}, author={Zihang Dai and Guokun Lai and Yiming Yang and Quoc V. Le}, year={2020}, eprint={2006.03236}, archivePrefix={arXiv}, primaryClass={cs.LG} } ```
{"language": "en", "license": "apache-2.0", "datasets": ["bookcorpus", "wikipedia", "gigaword"]}
feature-extraction
funnel-transformer/xlarge
[ "transformers", "pytorch", "tf", "funnel", "feature-extraction", "en", "dataset:bookcorpus", "dataset:wikipedia", "dataset:gigaword", "arxiv:2006.03236", "license:apache-2.0", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2006.03236" ]
[ "en" ]
TAGS #transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #has_space #region-us
# Funnel Transformer xlarge model (B10-10-10 with decoder) Pretrained model on English language using a similar objective objective as ELECTRA. It was introduced in this paper and first released in this repository. This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description Funnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and the pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Intended uses & limitations You can use the raw model to extract a vector representation of a given text, but it's mostly intended to be fine-tuned on a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use Here is how to use this model to get the features of a given text in PyTorch: and in TensorFlow: ## Training data The BERT model was pretrained on: - BookCorpus, a dataset consisting of 11,038 unpublished books, - English Wikipedia (excluding lists, tables and headers), - Clue Web, a dataset of 733,019,372 English web pages, - GigaWord, an archive of newswire text data, - Common Crawl, a dataset of raw web pages. ### BibTeX entry and citation info
[ "# Funnel Transformer xlarge model (B10-10-10 with decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ "TAGS\n#transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #has_space #region-us \n", "# Funnel Transformer xlarge model (B10-10-10 with decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.", "## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.", "## Intended uses & limitations\n\nYou can use the raw model to extract a vector representation of a given text, but it's mostly intended to\nbe fine-tuned on a downstream task. See the model hub to look for\nfine-tuned versions on a task that interests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Training data\n\nThe BERT model was pretrained on:\n- BookCorpus, a dataset consisting of 11,038 unpublished books,\n- English Wikipedia (excluding lists, tables and headers),\n- Clue Web, a dataset of 733,019,372 English web pages,\n- GigaWord, an archive of newswire text data,\n- Common Crawl, a dataset of raw web pages.", "### BibTeX entry and citation info" ]
[ 76, 104, 206, 133, 32, 95, 11 ]
[ "passage: TAGS\n#transformers #pytorch #tf #funnel #feature-extraction #en #dataset-bookcorpus #dataset-wikipedia #dataset-gigaword #arxiv-2006.03236 #license-apache-2.0 #endpoints_compatible #has_space #region-us \n# Funnel Transformer xlarge model (B10-10-10 with decoder)\n\nPretrained model on English language using a similar objective objective as ELECTRA. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.\n\nDisclaimer: The team releasing Funnel Transformer did not write a model card for this model so this model card has been\nwritten by the Hugging Face team.## Model description\n\nFunnel Transformer is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, a small language model corrupts the input texts and serves as a generator of inputs for this model, and\nthe pretraining objective is to predict which token is an original and which one has been replaced, a bit like a GAN training.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs." ]
[ -0.07659371942281723, 0.08881308138370514, -0.0038287488278001547, 0.05368322879076004, 0.10168372094631195, -0.0009783729910850525, 0.047006286680698395, 0.08333979547023773, -0.1451214849948883, 0.07615883648395538, -0.012618103995919228, -0.023558054119348526, 0.09764357656240463, 0.05783650651574135, 0.08356794714927673, -0.24417486786842346, 0.07863081991672516, -0.03435901924967766, 0.10150054842233658, 0.08902756124734879, 0.0902387723326683, -0.09730060398578644, 0.07442338019609451, 0.02120395563542843, -0.06353429704904556, -0.021147452294826508, -0.0035362527705729008, -0.04538976028561592, 0.040783513337373734, 0.09432016313076019, 0.09544491767883301, 0.004059962462633848, 0.03841149061918259, -0.0545387901365757, 0.024307627230882645, 0.08122211694717407, 0.009624231606721878, 0.048010729253292084, 0.027744630351662636, 0.004286312963813543, 0.09121811389923096, -0.013023032806813717, 0.06544971466064453, 0.05743572115898132, -0.11128672957420349, -0.2150029093027115, -0.04458511248230934, 0.13187851011753082, 0.0005030141328461468, 0.11788633465766907, -0.041376303881406784, 0.15386317670345306, -0.034459132701158524, 0.05012018606066704, 0.15773582458496094, -0.16393372416496277, -0.001060393755324185, 0.0009607491083443165, -0.034912656992673874, 0.12262725830078125, -0.02994796819984913, -0.05862627550959587, -0.010331022553145885, 0.04184449836611748, 0.04022849351167679, 0.00539483642205596, -0.018991520628333092, -0.05095275118947029, -0.13385553658008575, -0.06421216577291489, 0.13390696048736572, -0.042522020637989044, -0.11306879669427872, -0.10920745879411697, -0.038727469742298126, -0.06922927498817444, -0.003910547122359276, -0.046027638018131256, -0.005895058624446392, 0.017103353515267372, 0.030241193249821663, -0.024670876562595367, -0.10370684415102005, -0.009858389385044575, -0.07994598150253296, 0.08836023509502411, 0.04801835119724274, 0.056861720979213715, -0.11554043740034103, 0.10339274257421494, -0.08631735295057297, -0.05748600512742996, -0.08129788190126419, -0.06140444055199623, -0.05383288487792015, -0.02745485119521618, -0.052184902131557465, -0.10658517479896545, -0.05098037049174309, 0.0369853638112545, -0.0718989223241806, 0.03804575651884079, -0.008850621059536934, 0.027112506330013275, 0.12128287553787231, 0.12924794852733612, -0.049214791506528854, 0.1322239488363266, 0.007274058181792498, -0.06045426055788994, 0.030357725918293, -0.0481090284883976, -0.050148289650678635, -0.02719019167125225, 0.03831363841891289, 0.05435878410935402, 0.010613239370286465, 0.05761237442493439, -0.037656813859939575, -0.06952687352895737, 0.12005615234375, -0.11954053491353989, -0.0012148356763646007, 0.04472757875919342, -0.04414932057261467, 0.14124657213687897, 0.05106714367866516, -0.032511625438928604, -0.14233902096748352, -0.022033296525478363, -0.06629976630210876, -0.017778264358639717, -0.0784849226474762, -0.10519450157880783, -0.004495130851864815, -0.02461443841457367, -0.07989570498466492, -0.11379310488700867, -0.18019244074821472, -0.04775787889957428, 0.008145153522491455, -0.00993774551898241, 0.0010945235844701529, 0.01526665035635233, 0.004560571629554033, -0.011477340944111347, 0.0062829419039189816, -0.10704828798770905, -0.006327503360807896, 0.045832812786102295, -0.05966034159064293, 0.03750554099678993, -0.028296370059251785, 0.05572732537984848, -0.15714025497436523, -0.017361296340823174, -0.18405559659004211, 0.13801267743110657, -0.003009866224601865, -0.01833639293909073, -0.05679108947515488, -0.042979829013347626, -0.09949320554733276, 0.004195105750113726, -0.021933211013674736, 0.12398774921894073, -0.1382128894329071, -0.07018564641475677, 0.1988213062286377, -0.2206704169511795, 0.012863249517977238, 0.08779893815517426, -0.04130129888653755, 0.10595237463712692, 0.17561517655849457, 0.027026599273085594, 0.17419938743114471, -0.03687063232064247, -0.08059078454971313, 0.032513994723558426, -0.05872940644621849, 0.07939358800649643, 0.04854736104607582, -0.06171237304806709, -0.013319229707121849, 0.002993683097884059, -0.05295553058385849, -0.02006925456225872, -0.01722007431089878, -0.028016695752739906, -0.018760371953248978, 0.00655899103730917, 0.014847325161099434, 0.021950777620077133, 0.023341936990618706, 0.034833960235118866, -0.11566685140132904, 0.01890191063284874, 0.1077500581741333, -0.08774037659168243, 0.04033017158508301, -0.11373519897460938, 0.049303337931632996, -0.04993521794676781, -0.012078252620995045, -0.19456414878368378, -0.011388864368200302, 0.03994893282651901, -0.07811260223388672, 0.0742780789732933, 0.05471497401595116, 0.031751058995723724, 0.057841211557388306, 0.004141072276979685, -0.0013812658144161105, -0.03256892040371895, -0.01671997457742691, -0.033373039215803146, -0.08726274222135544, -0.0625741109251976, -0.04151170328259468, 0.04849967733025551, -0.06815388798713684, 0.03606749325990677, 0.07192531228065491, -0.012569395825266838, 0.0338459350168705, -0.048368945717811584, -0.001616905676200986, -0.0036320812068879604, -0.014548580162227154, -0.03995722532272339, 0.030233606696128845, 0.03861537575721741, -0.06621503829956055, 0.05821084976196289, -0.19702774286270142, -0.09566093236207962, 0.09429055452346802, 0.02755352109670639, -0.1280379444360733, 0.02199503779411316, 0.0036610483657568693, 0.010877599939703941, -0.029124964028596878, -0.0811813622713089, 0.22488196194171906, -0.0003090497921220958, 0.08504495024681091, -0.09804122895002365, -0.0321430042386055, 0.02552192658185959, -0.029408777132630348, -0.012460672296583652, 0.043611638247966766, -0.004644579254090786, -0.0941118597984314, 0.06310625374317169, 0.04585481807589531, 0.028445499017834663, 0.1534888744354248, 0.00419472623616457, -0.07512149959802628, -0.005597453564405441, 0.0024005980230867863, 0.007470262702554464, -0.032998181879520416, 0.019546082243323326, 0.02511211298406124, 0.04226028919219971, 0.09010299295186996, 0.04244677722454071, -0.05660855025053024, 0.06407595425844193, 0.10662024468183517, -0.02086617611348629, -0.050754204392433167, -0.04294806346297264, -0.01725177653133869, 0.08407401293516159, 0.040623243898153305, 0.12548933923244476, 0.04691347852349281, -0.012651145458221436, -0.1511685848236084, 0.15189164876937866, -0.1007639616727829, -0.19614340364933014, -0.1285954713821411, 0.003857081290334463, 0.03724829852581024, 0.040671080350875854, 0.04738537594676018, -0.1020461767911911, -0.06729345768690109, -0.11099426448345184, 0.1414727121591568, -0.056134551763534546, -0.04515450447797775, -0.01682206429541111, -0.06423875689506531, -0.022895433008670807, -0.12410922348499298, 0.0056468709371984005, 0.010195753537118435, -0.12176311016082764, 0.0010182339465245605, -0.07186391949653625, 0.012281190603971481, 0.17567425966262817, -0.01408047042787075, -0.0038273499812930822, -0.014362448826432228, 0.24010047316551208, -0.01886330358684063, 0.08378443121910095, 0.21126188337802887, -0.08300112932920456, 0.046640101820230484, 0.040770791471004486, 0.006218076217919588, -0.007138202898204327, 0.031065670773386955, -0.007372973021119833, -0.0703103169798851, -0.166305273771286, -0.06062403321266174, -0.019953828305006027, -0.027403485029935837, 0.02496413327753544, 0.04808095470070839, 0.031043870374560356, 0.06774233281612396, -0.0635373517870903, -0.010859868489205837, 0.06863204389810562, 0.08122191578149796, 0.038239944726228714, -0.017172660678625107, 0.0678078904747963, -0.08656201511621475, 0.027867544442415237, 0.10612540692090988, -0.08181057125329971, 0.2001415342092514, -0.03292996436357498, 0.21163320541381836, 0.06342819333076477, 0.028665591031312943, 0.1264207810163498, 0.05637407302856445, -0.04331694915890694, 0.01084662415087223, -0.02532719448208809, -0.07618656009435654, -0.07421919703483582, -0.015203113667666912, -0.037030722945928574, 0.05028703808784485, -0.1332339346408844, -0.04146353527903557, 0.0001816389849409461, 0.1662442535161972, -0.02795771136879921, -0.19385074079036713, -0.13321641087532043, 0.02989235892891884, -0.0040850830264389515, -0.06510814279317856, 0.009188898839056492, 0.056980475783348083, -0.12764766812324524, -0.011622841469943523, 0.0062771509401500225, 0.07884597033262253, -0.12473432719707489, 0.023189354687929153, -0.053086213767528534, 0.03276097774505615, -0.020092125982046127, 0.07640467584133148, -0.10096348822116852, 0.0054137189872562885, 0.008589711971580982, 0.0907340869307518, -0.05625857040286064, 0.024925731122493744, -0.02917829528450966, 0.1356237381696701, 0.0884159728884697, 0.024006793275475502, -0.03611009940505028, -0.09867587685585022, -0.04240399971604347, 0.036504097282886505, 0.046131428331136703, -0.05043422058224678, 0.09187372773885727, -0.05460994690656662, 0.055218640714883804, -0.0022201649844646454, 0.025749102234840393, -0.08725904673337936, -0.14409194886684418, 0.049195803701877594, -0.034354522824287415, 0.10897587239742279, -0.03989796340465546, -0.051041971892118454, -0.04194112494587898, 0.13608801364898682, -0.14731982350349426, -0.12143479287624359, -0.12595728039741516, -0.025511544197797775, 0.05492538586258888, -0.07862292975187302, 0.020068297162652016, -0.03144831955432892, 0.1293952912092209, 0.023047566413879395, -0.10993698239326477, 0.011401101015508175, -0.04341818764805794, -0.1480354517698288, -0.009525774978101254, 0.03345024958252907, 0.16332681477069855, 0.052120842039585114, 0.003612598404288292, 0.02769077569246292, -0.03176579251885414, -0.1086043119430542, -0.048220086842775345, 0.17290377616882324, 0.004660591017454863, 0.08684832602739334, -0.03907899931073189, -0.13790690898895264, -0.03705379366874695, 0.010912886820733547, 0.1418720781803131, 0.1126355230808258, -0.06743839383125305, 0.12909585237503052, 0.23884807527065277, -0.10187646001577377, -0.20096933841705322, -0.0007499931962229311, 0.017990248277783394, 0.03538508713245392, 0.015327958390116692, -0.1989867389202118, 0.08163925260305405, 0.06911814957857132, 0.004156272858381271, -0.0675714984536171, -0.33006829023361206, -0.07249461114406586, 0.1086125299334526, 0.07283442467451096, 0.09315218776464462, -0.08616388589143753, -0.012957431375980377, -0.0048303971998393536, -0.09767582267522812, 0.16943852603435516, -0.1255148947238922, 0.05142571032047272, 0.01525439415127039, -0.010911058634519577, 0.032750897109508514, -0.020867016166448593, 0.0700693354010582, -0.0035253800451755524, 0.05926654487848282, -0.08676417917013168, 0.01904832012951374, 0.10354027152061462, -0.0135700274258852, 0.13158683478832245, 0.06375797837972641, 0.06440214067697525, -0.07964570075273514, -0.06403327733278275, -0.08893425017595291, 0.048825062811374664, -0.05606194585561752, -0.04365039989352226, -0.08137659728527069, 0.058887697756290436, 0.05629897117614746, -0.009302371181547642, 0.006648621056228876, -0.09215760231018066, 0.08516710996627808, 0.0920724868774414, 0.16283464431762695, 0.08532480895519257, -0.14936690032482147, -0.03476014733314514, -0.012410655617713928, 0.11575225740671158, -0.057258352637290955, 0.05766531080007553, 0.062126364558935165, 0.009098423644900322, 0.10279742628335953, 0.04819498211145401, -0.14621175825595856, 0.02335212752223015, -0.002796504180878401, -0.08709679543972015, -0.11059542745351791, 0.011405854485929012, 0.025705881416797638, -0.11825098842382431, -0.01310007181018591, 0.11587020754814148, -0.06771193444728851, -0.023256050422787666, -0.012033592909574509, 0.02755492366850376, 0.010715468786656857, 0.07266868650913239, 0.033771272748708725, 0.007542106322944164, -0.06348715722560883, 0.09056761860847473, 0.1104118749499321, -0.10135353356599808, 0.07127244025468826, 0.059749189764261246, -0.09808462858200073, -0.07987045496702194, -0.012445875443518162, 0.08667928725481033, -0.02436685375869274, -0.0794762447476387, 0.025845034047961235, -0.11340560019016266, 0.06791435927152634, 0.07284059375524521, 0.0020811648573726416, 0.08025703579187393, -0.08145774900913239, 0.023273315280675888, -0.07418754696846008, 0.040081050246953964, -0.023527037352323532, -0.015829382464289665, -0.07352949678897858, 0.16833551228046417, 0.06667229533195496, 0.032774679362773895, -0.028679830953478813, -0.090913325548172, -0.08996307104825974, -0.004648077767342329, -0.09810492396354675, 0.020909370854496956, -0.009484095498919487, -0.019502870738506317, -0.02189898118376732, 0.018253063783049583, 0.021303411573171616, 0.026824306696653366, -0.04974329099059105, 0.002795365871861577, -0.03431253880262375, 0.03565559908747673, -0.08689773827791214, 0.07080545276403427, 0.04118974506855011, -0.020073071122169495, 0.10178881883621216, 0.047414444386959076, -0.07187218219041824, 0.038546353578567505, -0.07218597829341888, 0.041733499616384506, -0.06047039479017258, -0.022036584094166756, -0.011179182678461075, -0.08294448256492615, -0.003101260168477893, 0.02447371557354927, -0.06707840412855148, 0.003635588102042675, 0.11751673370599747, -0.08263415843248367, 0.1190386414527893, 0.0423685759305954, 0.03395041450858116, -0.09554488956928253, 0.05057467892765999, -0.005151537247002125, 0.03161302208900452, 0.10650515556335449, -0.031181495636701584, 0.07261504232883453, -0.14100781083106995, -0.02923722006380558, 0.02793974056839943, 0.041328463703393936, -0.046218715608119965, -0.04876700043678284, 0.02860429510474205, -0.017707929015159607, 0.05374697595834732, -0.0005898591480217874, -0.0005867011495865881, 0.017387572675943375, 0.008330140262842178, -0.04615127667784691, 0.021465685218572617, 0.10859151184558868, 0.0018467819318175316, -0.03700830042362213, 0.03903307393193245, 0.009659458883106709, -0.06952118128538132, -0.02242695353925228, 0.1746709644794464, 0.06351949274539948, 0.061734236776828766, 0.014907834120094776, -0.030775343999266624, -0.017838630825281143, -0.06589915603399277, -0.03331223502755165, -0.01081306952983141, -0.0006820445996709168, -0.02662476897239685, 0.09153910726308823, 0.15907472372055054, -0.02980782650411129, 0.1168203204870224, -0.003877694718539715, -0.04754484072327614, -0.12168433517217636, -0.24496294558048248, -0.0006324414280243218, -0.040301162749528885, -0.04292634502053261, -0.10956863313913345, 0.012950731441378593, 0.07237468659877777, 0.031218720600008965, -0.04698975384235382, 0.13477490842342377, -0.10157442092895508, -0.12643027305603027, -0.00886924285441637, -0.02264413796365261, 0.044688425958156586, -0.0035177036188542843, 0.05580420792102814, 0.09615319222211838, 0.08532850444316864, 0.07649853825569153, 0.10773372650146484, 0.08109830319881439, 0.013732852414250374, -0.06594521552324295, -0.07367418706417084, -0.005199206527322531, 0.009775451384484768, 0.0064469631761312485, 0.1413601189851761, 0.01841532438993454, -0.05113193765282631, -0.017902016639709473, 0.14752501249313354, -0.05504065752029419, -0.0680307000875473, -0.11479467153549194, 0.2249658703804016, -0.004020546562969685, -0.007477656472474337, 0.025914885103702545, -0.1003727838397026, 0.004648125264793634, 0.18756361305713654, 0.14010614156723022, 0.006679137237370014, 0.03250968083739281, -0.0020060017704963684, 0.0048756953328847885, 0.006293085869401693, 0.15144392848014832, 0.03387101739645004, 0.3169015347957611, -0.07766686379909515, 0.19506828486919403, -0.039985328912734985, 0.008490681648254395, -0.07321702688932419, 0.0851716548204422, -0.05561104044318199, 0.027701135724782944, -0.06016488000750542, 0.05530521646142006, -0.10446227341890335, -0.18850180506706238, -0.022370491176843643, -0.003938579000532627, -0.02905557118356228, -0.008084647357463837, -0.01034385897219181, 0.020543411374092102, 0.06299727410078049, 0.0050477986223995686, -0.0006739050731994212, 0.15776926279067993, 0.02819718047976494, -0.08026767522096634, -0.05969151854515076, 0.12140784412622452, -0.08362528681755066, 0.11484148353338242, 0.020220238715410233, 0.11706314980983734, 0.07908539474010468, -0.0025752633810043335, -0.11021609604358673, 0.05694892257452011, -0.04813021421432495, 0.011755079962313175, -0.008935505524277687, 0.09265609085559845, 0.020527755841612816, 0.10653550177812576, 0.05548001080751419, -0.09933487325906754, 0.0293668694794178, -0.021038608625531197, -0.00215912819840014, -0.06732221692800522, 0.04182063415646553, -0.044197726994752884, 0.1371874213218689, 0.15981732308864594, -0.016095168888568878, -0.03705858066678047, -0.05846882984042168, 0.030967356637120247, -0.021704383194446564, 0.05684104934334755, 0.0004576572682708502, -0.11524982005357742, -0.005449744872748852, 0.036341592669487, 0.0642600953578949, -0.21549426019191742, -0.036491695791482925, 0.01182575710117817, -0.03048909455537796, -0.01671595312654972, 0.05183523893356323, 0.01689983159303665, 0.05556797236204147, -0.035200681537389755, 0.1059122383594513, -0.015557747334241867, 0.10111113637685776, -0.09082566201686859, -0.056995220482349396 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-base-finetuned-bbc-headline This model is a fine-tuned version of [t5-base](https://huggingface.co/t5-base) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 12 - eval_batch_size: 12 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:| | No log | 1.0 | 167 | 2.2978 | 31.8313 | 10.3824 | 29.6182 | 29.4336 | 10.3153 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.9.1 - Datasets 1.12.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "t5-base-finetuned-bbc-headline", "results": []}]}
text2text-generation
furyhawk/t5-base-finetuned-bbc-headline
[ "transformers", "pytorch", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
t5-base-finetuned-bbc-headline ============================== This model is a fine-tuned version of t5-base on the None dataset. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 12 * eval\_batch\_size: 12 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 1 ### Training results ### Framework versions * Transformers 4.11.3 * Pytorch 1.9.1 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 12\n* eval\\_batch\\_size: 12\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.1\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 12\n* eval\\_batch\\_size: 12\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.1\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 63, 98, 4, 31 ]
[ "passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 12\n* eval\\_batch\\_size: 12\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.1\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.07917294651269913, 0.04864596948027611, -0.002182974945753813, 0.11597494035959244, 0.16170091927051544, 0.023381227627396584, 0.1268475502729416, 0.1353602409362793, -0.12793263792991638, 0.004200286231935024, 0.13233067095279694, 0.1732514351606369, 0.015047364868223667, 0.1225942000746727, -0.04851919785141945, -0.2595739960670471, 0.003491318551823497, 0.030370812863111496, -0.053154151886701584, 0.15092484652996063, 0.10372843593358994, -0.10602402687072754, 0.09179793298244476, -0.009956271387636662, -0.1954755038022995, 0.021206684410572052, 0.009479016065597534, -0.0678381398320198, 0.1597500741481781, 0.031234456226229668, 0.09475749731063843, 0.005969724152237177, 0.0739540159702301, -0.19461055099964142, 0.006758211180567741, 0.05588430166244507, -0.0017394390888512135, 0.07880964875221252, 0.05605163052678108, -0.006745294667780399, 0.18102601170539856, -0.07245244830846786, 0.03939914330840111, 0.03011893481016159, -0.12203200906515121, -0.20767086744308472, -0.07834647595882416, 0.027755821123719215, 0.07545594125986099, 0.1257125437259674, -0.008379369974136353, 0.13614404201507568, -0.08870742470026016, 0.10964812338352203, 0.2475784420967102, -0.29367268085479736, -0.05683545768260956, 0.00858201552182436, 0.021947896108031273, 0.08609814196825027, -0.0776451900601387, -0.024788009002804756, 0.05525949224829674, 0.05692436546087265, 0.1330861896276474, -0.03949585184454918, -0.14409998059272766, 0.01625540480017662, -0.1458403319120407, -0.04747922345995903, 0.1848125010728836, 0.02783810906112194, -0.028332818299531937, -0.02866641990840435, -0.08043739199638367, -0.14394859969615936, -0.01735721528530121, -0.011590752750635147, 0.046005792915821075, -0.015027809888124466, -0.05867008864879608, -0.020282797515392303, -0.10439889878034592, -0.06459491699934006, -0.0782221183180809, 0.1274353563785553, 0.04248498007655144, 0.014073709957301617, -0.04170067980885506, 0.1098250150680542, 0.010781916789710522, -0.12665338814258575, 0.01267610490322113, 0.03233743831515312, 0.025454988703131676, -0.0360545888543129, -0.0774466022849083, -0.05749311298131943, 0.021632956340909004, 0.13089124858379364, -0.06047814339399338, 0.0375778004527092, 0.028540702536702156, 0.036148134618997574, -0.09879960119724274, 0.1717275232076645, -0.032432232052087784, -0.02568834088742733, 0.020183155313134193, 0.03595779091119766, 0.02902360074222088, -0.014657155610620975, -0.12229155749082565, -0.002003754023462534, 0.1166999414563179, 0.032489508390426636, -0.06424079090356827, 0.07983476668596268, -0.0410219207406044, -0.022466609254479408, -0.02365809679031372, -0.10729651898145676, 0.008829837664961815, -0.013412442058324814, -0.08581885695457458, -0.009087814018130302, 0.035785481333732605, 0.008341792970895767, -0.05586150288581848, 0.08606988936662674, -0.08130613714456558, 0.031661443412303925, -0.0859742984175682, -0.10590288043022156, 0.016301007941365242, -0.054724402725696564, 0.019059399142861366, -0.10911612957715988, -0.20408034324645996, -0.0060902670957148075, 0.054139990359544754, -0.02135828696191311, -0.06969144940376282, -0.06282404065132141, -0.08462017774581909, 0.008311533369123936, -0.03187226876616478, 0.13361595571041107, -0.07317100465297699, 0.10968617349863052, 0.04083363711833954, 0.06288418173789978, -0.06025770306587219, 0.060586750507354736, -0.11563419550657272, -0.006405573803931475, -0.16084645688533783, 0.054075658321380615, -0.0237534549087286, 0.0735791027545929, -0.09211394935846329, -0.11124932020902634, 0.02001090906560421, 0.0028156719636172056, 0.06675279140472412, 0.10179691761732101, -0.15602850914001465, -0.08324094861745834, 0.18071676790714264, -0.07088713347911835, -0.145315483212471, 0.1176769956946373, -0.050866276025772095, 0.06691048294305801, 0.08196251839399338, 0.16587403416633606, 0.04812927544116974, -0.06522224843502045, 0.029671616852283478, 0.012560950592160225, 0.04294712841510773, -0.05770638957619667, 0.07109889388084412, 0.0019204005366191268, -0.007813677191734314, 0.029585057869553566, -0.023979030549526215, 0.06941793113946915, -0.0950561985373497, -0.08989889919757843, -0.054036784917116165, -0.0900176391005516, 0.02659023366868496, 0.0514310821890831, 0.07773882150650024, -0.11525621265172958, -0.08128651976585388, 0.055760256946086884, 0.08447016030550003, -0.06897364556789398, 0.04200078919529915, -0.050458598881959915, 0.0586102269589901, -0.011208506301045418, -0.0018682981608435512, -0.18445372581481934, -0.015266409143805504, 0.00802867766469717, 0.016746001318097115, 0.033844079822301865, 0.010173344053328037, 0.06033734604716301, 0.054895106703042984, -0.05385657399892807, -0.01835251785814762, -0.031085694208741188, -0.010518829338252544, -0.11746839433908463, -0.18655246496200562, -0.015998881310224533, -0.007434413768351078, 0.13998503983020782, -0.19114567339420319, 0.04195865988731384, -0.02720390446484089, 0.06069082021713257, -0.0026713472325354815, 0.007444408256560564, -0.0494484081864357, 0.07772965729236603, -0.05592822656035423, -0.05114695057272911, 0.07545025646686554, 0.008129552938044071, -0.09110482782125473, -0.029204517602920532, -0.1155395582318306, 0.16304399073123932, 0.14563970267772675, -0.14782515168190002, -0.07683531194925308, 0.014928432181477547, -0.055428024381399155, -0.03157001733779907, -0.03562331572175026, 0.01791822537779808, 0.17356795072555542, -0.01878063939511776, 0.16189827024936676, -0.07987437397241592, -0.04437435418367386, 0.014093045145273209, -0.03821251913905144, 0.050295889377593994, 0.11940842121839523, 0.11228225380182266, -0.08414044976234436, 0.14385724067687988, 0.1655689924955368, -0.09182858467102051, 0.12820276618003845, -0.04848407581448555, -0.07948124408721924, 0.00002448004852340091, -0.010429966263473034, -0.013915013521909714, 0.04936806857585907, -0.15002308785915375, -0.0027403489220887423, 0.024861622601747513, 0.03305019438266754, 0.027690794318914413, -0.22518807649612427, -0.03511664643883705, 0.03221513330936432, -0.06230024993419647, -0.013023708947002888, -0.018810927867889404, 0.004570200107991695, 0.11352880299091339, 0.002986007835716009, -0.09462493658065796, 0.04585839807987213, 0.008652381598949432, -0.09134341776371002, 0.21737702190876007, -0.09602175652980804, -0.17031769454479218, -0.12248946726322174, -0.08362303674221039, -0.06060134992003441, 0.011906039901077747, 0.08844396471977234, -0.0836438238620758, -0.03157703951001167, -0.08117186278104782, 0.050867706537246704, -0.009387906640768051, 0.028569290414452553, 0.009459033608436584, 0.0009586209780536592, 0.06381252408027649, -0.12261784821748734, -0.016405237838625908, -0.03798028081655502, -0.08762212842702866, 0.054675132036209106, 0.011160111054778099, 0.10526394098997116, 0.15908639132976532, -0.01700807735323906, 0.011576665565371513, -0.03816065192222595, 0.23812063038349152, -0.0559934601187706, -0.024514706805348396, 0.15771928429603577, 0.011463457718491554, 0.05219537392258644, 0.0998229905962944, 0.05060592293739319, -0.09869121015071869, 0.0273350290954113, 0.023173099383711815, -0.034020423889160156, -0.24133281409740448, -0.04669547826051712, -0.056445538997650146, 0.006195428781211376, 0.08657041192054749, 0.03312470391392708, 0.0636700689792633, 0.055714283138513565, 0.0283619724214077, 0.08778470009565353, -0.01567726954817772, 0.06136526167392731, 0.14176006615161896, 0.04307707026600838, 0.1284196674823761, -0.05214658007025719, -0.05822296440601349, 0.05673961341381073, -0.019337236881256104, 0.2169407606124878, 0.001874482142738998, 0.13623656332492828, 0.05239889770746231, 0.16065138578414917, -0.019314777106046677, 0.09314564615488052, 0.000461625779280439, -0.026003846898674965, -0.02267128974199295, -0.04848017916083336, -0.0454828180372715, 0.028179306536912918, -0.10243590921163559, 0.05213579535484314, -0.1221335306763649, 0.0019694704096764326, 0.06714151054620743, 0.2684391736984253, 0.03608883544802666, -0.32653719186782837, -0.09239652752876282, 0.0007718034903518856, -0.05826488509774208, -0.014133013784885406, 0.04479465261101723, 0.07495958358049393, -0.09203127771615982, 0.043900541961193085, -0.05179411545395851, 0.10738614946603775, -0.02407153695821762, 0.0624307245016098, 0.06781759858131409, 0.09177780896425247, 0.015578394755721092, 0.09185373783111572, -0.33205702900886536, 0.26835739612579346, -0.003577787196263671, 0.06276372820138931, -0.08174394071102142, 0.008718178607523441, 0.03530208393931389, 0.08003135025501251, 0.05364309251308441, -0.008142431266605854, -0.024887317791581154, -0.15706895291805267, -0.04272302985191345, 0.04112551361322403, 0.09826569259166718, -0.024683617055416107, 0.09390317648649216, -0.0496649444103241, 0.014979330822825432, 0.07631254941225052, 0.0017103089485317469, -0.056198496371507645, -0.08583644777536392, -0.0023330370895564556, 0.028971241787075996, -0.0062789772637188435, -0.06180766597390175, -0.10324937850236893, -0.1055026426911354, 0.16354644298553467, 0.00019141972006764263, -0.04078663885593414, -0.10495822876691818, 0.07012168318033218, 0.06829486042261124, -0.09223593771457672, 0.04198463633656502, 0.01008344255387783, 0.03946889936923981, 0.042267944663763046, -0.09203319251537323, 0.1196160838007927, -0.06014034152030945, -0.16210290789604187, -0.040775008499622345, 0.10710137337446213, 0.008463850244879723, 0.06051621213555336, -0.01287116575986147, 0.004631017334759235, -0.057121675461530685, -0.09244589507579803, -0.0053913528099656105, -0.016798846423625946, 0.06365916877985, 0.02939794771373272, -0.060023993253707886, 0.010561967268586159, -0.07215899974107742, -0.05443762615323067, 0.2077518105506897, 0.23712407052516937, -0.08173257857561111, 0.026812033727765083, 0.041956912726163864, -0.06933805346488953, -0.19465497136116028, 0.010941562242805958, 0.04259354621171951, -0.008906024508178234, 0.029733000323176384, -0.19201257824897766, 0.1136719286441803, 0.11429408937692642, -0.004513062071055174, 0.10425625741481781, -0.3607324957847595, -0.13044631481170654, 0.11236605048179626, 0.1345929205417633, 0.14448387920856476, -0.1559794396162033, -0.017831839621067047, -0.04936124384403229, -0.1304577887058258, 0.11261636763811111, -0.11683966219425201, 0.12968771159648895, -0.027483250945806503, 0.10147949308156967, -0.0022210560273379087, -0.042627133429050446, 0.11152711510658264, 0.018609317019581795, 0.09249046444892883, -0.07352491468191147, 0.016301482915878296, 0.04118962585926056, -0.033165786415338516, 0.04096019268035889, -0.11841290444135666, 0.042949266731739044, -0.08685306459665298, -0.021075839176774025, -0.07476785033941269, 0.04265621304512024, -0.029700979590415955, -0.07056175172328949, -0.031369514763355255, -0.008610855787992477, 0.06331949681043625, -0.018467111513018608, 0.13270927965641022, 0.015573037788271904, 0.14292199909687042, 0.11068668961524963, 0.07161762565374374, -0.09083005785942078, -0.03497166931629181, -0.02081504464149475, -0.015218832530081272, 0.056337278336286545, -0.16488102078437805, 0.030580317601561546, 0.13437685370445251, 0.010747263208031654, 0.14757898449897766, 0.08591947704553604, -0.015998221933841705, 0.0022985879331827164, 0.06368091702461243, -0.17762933671474457, -0.08380116522312164, -0.020524989813566208, -0.05800590664148331, -0.09254223108291626, 0.059287525713443756, 0.10684067010879517, -0.07528511434793472, -0.012088022194802761, -0.0155298737809062, 0.01303514838218689, -0.07470716536045074, 0.18417030572891235, 0.04380703344941139, 0.0416082888841629, -0.10496323555707932, 0.08876176923513412, 0.04748396947979927, -0.06782644987106323, 0.015543593093752861, 0.11840908229351044, -0.08720020949840546, -0.05243012309074402, 0.09453917294740677, 0.16343289613723755, -0.091640904545784, -0.05113108083605766, -0.13414767384529114, -0.12483752518892288, 0.0872647613286972, 0.16963893175125122, 0.10865944623947144, 0.021409064531326294, -0.053335223346948624, 0.006011987570673227, -0.1117209941148758, 0.0749647468328476, 0.03931387513875961, 0.05792732536792755, -0.12505660951137543, 0.16819581389427185, 0.013944197446107864, 0.051054637879133224, -0.028418276458978653, 0.019502654671669006, -0.09257013350725174, 0.01912163756787777, -0.14538252353668213, -0.03226806968450546, -0.023024072870612144, 0.005268795415759087, 0.001100825727917254, -0.04833751171827316, -0.05720861628651619, 0.020183727145195007, -0.12049810588359833, -0.03097335435450077, 0.014501587487757206, 0.06270521134138107, -0.12307331711053848, -0.03522821515798569, 0.02362128160893917, -0.06732599437236786, 0.08398976922035217, 0.06531856954097748, -0.009664547629654408, 0.08084560930728912, -0.1569766104221344, 0.002632519928738475, 0.07538976520299911, 0.024373812600970268, 0.0566331148147583, -0.07422428578138351, -0.008858864195644855, 0.016629446297883987, 0.06310401111841202, 0.02029404230415821, 0.06627143174409866, -0.1366523653268814, -0.0063933213241398335, -0.027860848233103752, -0.0918816477060318, -0.06330176442861557, 0.02487439103424549, 0.0753958448767662, 0.007582054473459721, 0.19915933907032013, -0.08219407498836517, 0.036102019250392914, -0.20592761039733887, 0.006831109989434481, -0.011449619196355343, -0.11950411647558212, -0.14060477912425995, -0.07696956396102905, 0.06025188788771629, -0.05471418425440788, 0.1399163156747818, 0.023025546222925186, 0.05182741954922676, 0.02828139066696167, -0.007445800118148327, 0.01381876040250063, 0.012731031514704227, 0.23461133241653442, 0.04216904565691948, -0.031675584614276886, 0.04054085910320282, 0.04571684077382088, 0.11197195202112198, 0.09748857468366623, 0.20480012893676758, 0.14066432416439056, -0.0033939466811716557, 0.10284251719713211, 0.0219118669629097, -0.05517038702964783, -0.15106071531772614, 0.003117592539638281, -0.01956818625330925, 0.11502469331026077, -0.023438332602381706, 0.21759678423404694, 0.08355186879634857, -0.14840272068977356, 0.03885175660252571, -0.05681624636054039, -0.07637591660022736, -0.12010952085256577, -0.03224341198801994, -0.08112258464097977, -0.16587525606155396, -0.010297865606844425, -0.11785360425710678, 0.035987552255392075, 0.11200784146785736, 0.015206746757030487, -0.03296986594796181, 0.14225643873214722, 0.021206123754382133, -0.010289841331541538, 0.05122530087828636, -0.01026699785143137, -0.021738942712545395, -0.10647067427635193, -0.07321388274431229, -0.013525839895009995, -0.009562790393829346, 0.03500599414110184, -0.04077242314815521, -0.061778102070093155, 0.038428861647844315, -0.05837540701031685, -0.09244310110807419, 0.01757095381617546, 0.02658015489578247, 0.05065922439098358, 0.05860096216201782, 0.011353785172104836, -0.0033786254934966564, 0.005374784581363201, 0.24470336735248566, -0.0919509083032608, -0.11948703974485397, -0.09069705009460449, 0.28744497895240784, 0.04647303745150566, -0.00890024658292532, 0.029164299368858337, -0.05938807502388954, -0.022610485553741455, 0.2589184045791626, 0.20510615408420563, -0.08972618728876114, -0.014466590248048306, -0.006199606694281101, -0.0030199240427464247, -0.015341260470449924, 0.1219402402639389, 0.1583179086446762, 0.04660766199231148, -0.08424834161996841, -0.02008887007832527, -0.04903056472539902, -0.0010604590643197298, -0.05726759508252144, 0.06355644762516022, 0.029035842046141624, -0.009726980701088905, -0.023800551891326904, 0.05728472024202347, -0.07442960143089294, -0.07502104341983795, 0.019478987902402878, -0.18666775524616241, -0.1480935513973236, -0.010445859283208847, 0.11946103721857071, 0.012139055877923965, 0.06353101879358292, -0.020150642842054367, 0.0038059880025684834, 0.07451078295707703, -0.01605049893260002, -0.10726287961006165, -0.07784072309732437, 0.10310747474431992, -0.14511466026306152, 0.17739343643188477, -0.039333708584308624, 0.0655146911740303, 0.12102089822292328, 0.0674641877412796, -0.07357370108366013, 0.09280652552843094, 0.04161441698670387, -0.04643237590789795, 0.034564316272735596, 0.06928757578134537, -0.03630826994776726, 0.04778393357992172, 0.04295654967427254, -0.12098854780197144, 0.008995948359370232, -0.023153331130743027, -0.052637819200754166, -0.03233223780989647, -0.06409633159637451, -0.05685875192284584, 0.1183834969997406, 0.21719400584697723, -0.04028778895735741, 0.019457794725894928, -0.09246940165758133, 0.00046589059638790786, 0.051135577261447906, 0.0009051461238414049, -0.06633755564689636, -0.22266824543476105, -0.006728063337504864, 0.07690277695655823, -0.012270068749785423, -0.23744411766529083, -0.07731765508651733, -0.00855524092912674, -0.06254848092794418, -0.11943590641021729, 0.09409618377685547, 0.08532225340604782, 0.04100821539759636, -0.04578138142824173, -0.07077021896839142, -0.0743459016084671, 0.16456227004528046, -0.14998897910118103, -0.10071311146020889 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-base-finetuned-bbc This model is a fine-tuned version of [t5-base](https://huggingface.co/t5-base) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 6 - eval_batch_size: 6 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:| | No log | 1.0 | 334 | 0.1500 | 24.5024 | 21.4979 | 24.0227 | 24.0303 | 19.0 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.9.1 - Datasets 1.12.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "t5-base-finetuned-bbc", "results": []}]}
text2text-generation
furyhawk/t5-base-finetuned-bbc
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
t5-base-finetuned-bbc ===================== This model is a fine-tuned version of t5-base on the None dataset. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.0003 * train\_batch\_size: 6 * eval\_batch\_size: 6 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 1 ### Training results ### Framework versions * Transformers 4.11.3 * Pytorch 1.9.1 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 6\n* eval\\_batch\\_size: 6\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.1\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 6\n* eval\\_batch\\_size: 6\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.1\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 67, 97, 4, 31 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 6\n* eval\\_batch\\_size: 6\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.1\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.07713507860898972, 0.049754854291677475, -0.0030080201104283333, 0.11260558664798737, 0.14560267329216003, 0.020395636558532715, 0.1372489184141159, 0.13907669484615326, -0.11508522927761078, 0.022086776793003082, 0.1324935108423233, 0.16786542534828186, 0.016723619773983955, 0.09847183525562286, -0.048044826835393906, -0.2620338797569275, -0.006808123551309109, 0.04087843745946884, -0.05744755268096924, 0.13657048344612122, 0.09460321813821793, -0.11522477865219116, 0.09173021465539932, 0.0005339315976016223, -0.18749447166919708, 0.026118962094187737, -0.0022447865922003984, -0.056839071214199066, 0.1517561972141266, 0.030200257897377014, 0.10742940753698349, 0.010026181116700172, 0.06981122493743896, -0.19725528359413147, 0.010659812018275261, 0.055406808853149414, -0.005625320132821798, 0.07922391593456268, 0.05183795094490051, 0.00020239075820427388, 0.19532735645771027, -0.07274607568979263, 0.04829353094100952, 0.027325766161084175, -0.1168907880783081, -0.20087945461273193, -0.07850524038076401, 0.03502911329269409, 0.08055902272462845, 0.11775945872068405, -0.007969626225531101, 0.13160103559494019, -0.07626853138208389, 0.10985471308231354, 0.24442832171916962, -0.2943490445613861, -0.06282661855220795, 0.014660965651273727, 0.027277208864688873, 0.07739263027906418, -0.08556096255779266, -0.0178869366645813, 0.053387660533189774, 0.05304048955440521, 0.12779641151428223, -0.03606516867876053, -0.11169180274009705, 0.01859155297279358, -0.14552876353263855, -0.05387245491147041, 0.18202875554561615, 0.031385306268930435, -0.027659088373184204, -0.051871445029973984, -0.07734899967908859, -0.14942140877246857, -0.0202828086912632, -0.017381569370627403, 0.04154243320226669, -0.0205521322786808, -0.05717579275369644, -0.03971968963742256, -0.11911787837743759, -0.06638310849666595, -0.07300979644060135, 0.12308730185031891, 0.04534994810819626, 0.010289501398801804, -0.04759971797466278, 0.11235026270151138, 0.009603935293853283, -0.1282004714012146, 0.010115782730281353, 0.03591937571763992, 0.022659916430711746, -0.030820900574326515, -0.06200052797794342, -0.09383013844490051, 0.020547591149806976, 0.1202741488814354, -0.05921401455998421, 0.04978149011731148, 0.023681309074163437, 0.042762137949466705, -0.09840181469917297, 0.16986748576164246, -0.01683359406888485, -0.005742150358855724, 0.006609649397432804, 0.038569290190935135, 0.039146289229393005, -0.018242061138153076, -0.1311252862215042, 0.009403237141668797, 0.09520440548658371, 0.022548969835042953, -0.06046606972813606, 0.07579242438077927, -0.03915951028466225, -0.019327275454998016, -0.015912434086203575, -0.09640973806381226, 0.018542421981692314, -0.0008537705289199948, -0.07203211635351181, -0.0077690607868134975, 0.04108855873346329, 0.008045957423746586, -0.04820864647626877, 0.08988532423973083, -0.07539994269609451, 0.024755991995334625, -0.09447618573904037, -0.11087137460708618, 0.020999284461140633, -0.08844985067844391, 0.02351437322795391, -0.10404589772224426, -0.19832804799079895, -0.0026598137337714434, 0.059838369488716125, -0.022360602393746376, -0.05962758883833885, -0.06726624816656113, -0.08315905183553696, 0.017755214124917984, -0.020562022924423218, 0.13914866745471954, -0.06534399837255478, 0.09593387693166733, 0.03404654189944267, 0.0585094578564167, -0.05038684234023094, 0.05266447737812996, -0.09455019235610962, 0.00600836519151926, -0.1468605250120163, 0.045085739344358444, -0.0342864952981472, 0.06226322799921036, -0.08799800276756287, -0.09752943366765976, -0.005964096635580063, 0.0055550942197442055, 0.06657296419143677, 0.10304346680641174, -0.16618603467941284, -0.08663906902074814, 0.16610687971115112, -0.07554381340742111, -0.1376349925994873, 0.1314002424478531, -0.0559370331466198, 0.06031375750899315, 0.06289778649806976, 0.17639820277690887, 0.06302104145288467, -0.09485325217247009, 0.01699359342455864, 0.021305996924638748, 0.041395410895347595, -0.05442441254854202, 0.061993446201086044, -0.006229146849364042, 0.031230779364705086, 0.024649109691381454, -0.014201107434928417, 0.05399313569068909, -0.08524884283542633, -0.09005074948072433, -0.051978353410959244, -0.0791478306055069, 0.02677459642291069, 0.062093302607536316, 0.07969227433204651, -0.11218912899494171, -0.09463990479707718, 0.04248705506324768, 0.08246616274118423, -0.07651782780885696, 0.04821651428937912, -0.05573870241641998, 0.07265469431877136, -0.02878059260547161, -0.004206904210150242, -0.17289145290851593, -0.02063392847776413, 0.010751347057521343, 0.004215562250465155, 0.029870329424738884, 0.022205889225006104, 0.07009249180555344, 0.05797772482037544, -0.051525115966796875, -0.019971365109086037, -0.040041644126176834, -0.011101558804512024, -0.12470060586929321, -0.19516533613204956, -0.026978112757205963, -0.023220814764499664, 0.12494757771492004, -0.20170453190803528, 0.04986932873725891, 0.0052205640822649, 0.06314686685800552, 0.007693496532738209, -0.0019170864252373576, -0.04751376807689667, 0.06849879026412964, -0.06453507393598557, -0.04963333159685135, 0.07451796531677246, 0.013888600282371044, -0.09976961463689804, -0.030905188992619514, -0.12577076256275177, 0.1405624896287918, 0.14026933908462524, -0.13782840967178345, -0.06579769402742386, -0.0005857564974576235, -0.05993625894188881, -0.03520798683166504, -0.03376509249210358, 0.00918208435177803, 0.1945532262325287, -0.013630641624331474, 0.15716314315795898, -0.08098335564136505, -0.055231351405382156, 0.02363089844584465, -0.037586312741041183, 0.02554210275411606, 0.13466036319732666, 0.1112281084060669, -0.0684739425778389, 0.1442151963710785, 0.15765024721622467, -0.09426993131637573, 0.14896360039710999, -0.048563193529844284, -0.08108117431402206, -0.00560530973598361, -0.0005177365383133292, -0.0035881183575838804, 0.07249671965837479, -0.1830713003873825, -0.005765347275882959, 0.02246391773223877, 0.02513614110648632, 0.03396409749984741, -0.2282266616821289, -0.027923405170440674, 0.044509805738925934, -0.057381320744752884, -0.003760763444006443, -0.008661233820021152, -0.002573535544797778, 0.10613756626844406, -0.006960798054933548, -0.0813874676823616, 0.044345367699861526, 0.000192971812793985, -0.09093751758337021, 0.21843163669109344, -0.07942254096269608, -0.15538859367370605, -0.1264282763004303, -0.08294931054115295, -0.05878828093409538, 0.009965997189283371, 0.08695048838853836, -0.09153324365615845, -0.029038619250059128, -0.09280538558959961, 0.0398138053715229, -0.014559194445610046, 0.025614945217967033, 0.009665173478424549, 0.001915191882289946, 0.05206732451915741, -0.118735671043396, -0.015645325183868408, -0.05067894980311394, -0.0704190731048584, 0.04466584324836731, 0.017171315848827362, 0.11505577713251114, 0.1562863290309906, -0.014595494605600834, 0.020275479182600975, -0.03568369522690773, 0.2098781019449234, -0.06291306763887405, -0.021962573751807213, 0.15790988504886627, -0.0016813230467960238, 0.05637887865304947, 0.09229826182126999, 0.05623749643564224, -0.0844372883439064, 0.016824640333652496, 0.03386394679546356, -0.04520172253251076, -0.2420046031475067, -0.036681681871414185, -0.0637628510594368, 0.012327317148447037, 0.08725253492593765, 0.03608521819114685, 0.05420762300491333, 0.061802588403224945, 0.030863633379340172, 0.07565761357545853, -0.019724411889910698, 0.06889336556196213, 0.13488446176052094, 0.037348803132772446, 0.12192712724208832, -0.06036188453435898, -0.05849883705377579, 0.047525376081466675, -0.0018129858653992414, 0.22966933250427246, 0.006936235353350639, 0.16372495889663696, 0.06476939469575882, 0.15026339888572693, -0.01953667216002941, 0.08435999602079391, -0.015994945541024208, -0.037066828459501266, -0.01870005391538143, -0.04784387722611427, -0.028774524107575417, 0.03263524919748306, -0.09080608189105988, 0.06208430230617523, -0.12285779416561127, 0.023049918934702873, 0.05639117583632469, 0.2673361897468567, 0.03448214754462242, -0.3187207281589508, -0.08541902154684067, 0.012885665521025658, -0.04723382741212845, -0.012140240520238876, 0.040817782282829285, 0.10165537148714066, -0.08153710514307022, 0.03987188637256622, -0.06678277254104614, 0.099214106798172, -0.032214656472206116, 0.055079057812690735, 0.060575149953365326, 0.08128807693719864, 0.009768405929207802, 0.0943370908498764, -0.3239816427230835, 0.27744656801223755, -0.0009335125796496868, 0.07116647064685822, -0.08484194427728653, 0.016164032742381096, 0.029099630191922188, 0.06119225174188614, 0.08122079074382782, -0.01895437017083168, -0.041682660579681396, -0.14854955673217773, -0.04981555417180061, 0.03310496360063553, 0.093827985227108, -0.02677409164607525, 0.10047659277915955, -0.036961451172828674, 0.013112979009747505, 0.07589239627122879, 0.022215373814105988, -0.049654342234134674, -0.11018075793981552, -0.006433801259845495, 0.021784363314509392, -0.06096949800848961, -0.06094510108232498, -0.10440818965435028, -0.11845270544290543, 0.15854312479496002, -0.03165675699710846, -0.0425594188272953, -0.10629453510046005, 0.06412474066019058, 0.05181621387600899, -0.092099629342556, 0.03628517687320709, 0.011309683322906494, 0.062018249183893204, 0.024579612538218498, -0.08354775607585907, 0.10789451003074646, -0.0705985352396965, -0.1624729484319687, -0.05092731490731239, 0.11357646435499191, 0.018728062510490417, 0.06003253906965256, -0.011821296066045761, 0.0031903237104415894, -0.054372113198041916, -0.09048809111118317, 0.0196099691092968, -0.01603790372610092, 0.07506633549928665, 0.0074232215993106365, -0.06000794842839241, 0.025375302881002426, -0.06411997228860855, -0.048116158694028854, 0.20740091800689697, 0.2407546490430832, -0.08378823101520538, 0.03041762486100197, 0.026528863236308098, -0.08415648341178894, -0.19700351357460022, 0.009936423040926456, 0.04403161257505417, -0.0005888697924092412, 0.031816691160202026, -0.18683049082756042, 0.1056305319070816, 0.09884404391050339, -0.006762670818716288, 0.10198741406202316, -0.36205920577049255, -0.13041405379772186, 0.12114126235246658, 0.14197081327438354, 0.11659522354602814, -0.1538597196340561, -0.022325700148940086, -0.030261877924203873, -0.11756020039319992, 0.10736743360757828, -0.12056554853916168, 0.12434714287519455, -0.030328726395964622, 0.10350486636161804, 0.003568046959117055, -0.053040727972984314, 0.1075945496559143, 0.009599040262401104, 0.09479928761720657, -0.067377470433712, 0.00011462299153208733, 0.04849671572446823, -0.0445026196539402, 0.035975538194179535, -0.12376980483531952, 0.03012862615287304, -0.09786737710237503, -0.022598179057240486, -0.06881946325302124, 0.04830536991357803, -0.03889552503824234, -0.06849157065153122, -0.036132000386714935, -0.015353402122855186, 0.05921820178627968, -0.008670805022120476, 0.16319407522678375, 0.013287266716361046, 0.14923392236232758, 0.1300676316022873, 0.08909765630960464, -0.07309751957654953, -0.05384968966245651, -0.021476007997989655, -0.012129117734730244, 0.053646959364414215, -0.1658409833908081, 0.026707546785473824, 0.13597652316093445, 0.02101585827767849, 0.1437358856201172, 0.08974984288215637, -0.031366556882858276, 0.017436286434531212, 0.055838532745838165, -0.1711520105600357, -0.1036645770072937, -0.022123057395219803, -0.051051393151283264, -0.09831754118204117, 0.06598365306854248, 0.1048685684800148, -0.07273546606302261, -0.001698114676401019, -0.013230055570602417, 0.011101905256509781, -0.0610693022608757, 0.1779823899269104, 0.04122376814484596, 0.04370623826980591, -0.09409204125404358, 0.08727701753377914, 0.03641573712229729, -0.08878932148218155, 0.02519521676003933, 0.10332654416561127, -0.07340171933174133, -0.05555886775255203, 0.07681480795145035, 0.18245254456996918, -0.05905018001794815, -0.05137176439166069, -0.14534366130828857, -0.12898141145706177, 0.08834388852119446, 0.16799502074718475, 0.1042545959353447, 0.014543558470904827, -0.07001232355833054, 0.01795324869453907, -0.11936396360397339, 0.09707512706518173, 0.03702014684677124, 0.06314510852098465, -0.13660210371017456, 0.14894181489944458, 0.010678711347281933, 0.04788295924663544, -0.02323881722986698, 0.01979818008840084, -0.09239622205495834, 0.01725032366812229, -0.1259957104921341, -0.02311224490404129, -0.017781034111976624, -0.0003392654180061072, -0.0019160077208653092, -0.03969546779990196, -0.06162641569972038, 0.024419022724032402, -0.11459995806217194, -0.023042287677526474, 0.027960389852523804, 0.06406988203525543, -0.11106552183628082, -0.03249773383140564, 0.019969871267676353, -0.07172377407550812, 0.07915950566530228, 0.055171702057123184, 0.0013219027314335108, 0.06739291548728943, -0.14790014922618866, 0.022061625495553017, 0.07552481442689896, 0.02949145808815956, 0.05102304369211197, -0.07425442337989807, -0.014221052639186382, 0.005644613411277533, 0.04876900836825371, 0.012977655977010727, 0.0635298565030098, -0.1348281055688858, 0.00884560588747263, -0.02493700012564659, -0.09797756373882294, -0.0672747790813446, 0.040356479585170746, 0.06725040078163147, 0.013064899481832981, 0.19882015883922577, -0.0833733007311821, 0.03942596912384033, -0.20926812291145325, 0.011468823067843914, 0.01013835147023201, -0.1109241396188736, -0.1235235333442688, -0.07336015999317169, 0.05721574276685715, -0.06777480244636536, 0.12926441431045532, 0.021914003416895866, 0.0284879133105278, 0.029356317594647408, -0.020043911412358284, 0.02340114675462246, 0.00899664219468832, 0.2287716418504715, 0.025396335870027542, -0.03511469066143036, 0.0434630811214447, 0.03584134578704834, 0.11300873011350632, 0.1177181825041771, 0.20790205895900726, 0.14357976615428925, -0.018912527710199356, 0.11538314819335938, 0.03325822576880455, -0.0505029559135437, -0.16106146574020386, 0.030803661793470383, -0.020140642300248146, 0.12171057611703873, -0.023347336798906326, 0.21378013491630554, 0.10481731593608856, -0.15287868678569794, 0.03724066540598869, -0.04442082718014717, -0.07681300491094589, -0.1269378811120987, -0.0720774456858635, -0.08047596365213394, -0.14765013754367828, 0.002674597082659602, -0.1196514368057251, 0.03338788077235222, 0.10291013866662979, 0.01884029246866703, -0.02668249048292637, 0.15535689890384674, 0.025453288108110428, 0.006181187927722931, 0.053733933717012405, -0.004024814814329147, -0.02627541683614254, -0.1069190576672554, -0.06355924904346466, -0.00824856385588646, -0.024463968351483345, 0.028086097911000252, -0.044150080531835556, -0.042039528489112854, 0.03498825430870056, -0.038637980818748474, -0.09184905886650085, 0.011397046968340874, 0.02227451279759407, 0.06227099150419235, 0.05451233312487602, 0.010045337490737438, -0.0009563766070641577, -0.004763301461935043, 0.2185421586036682, -0.0879824087023735, -0.08251716196537018, -0.08151248097419739, 0.2561185956001282, 0.04303160682320595, -0.008059856481850147, 0.02427646517753601, -0.062010277062654495, -0.013420701958239079, 0.2585541903972626, 0.20372888445854187, -0.06886766105890274, -0.01272591669112444, 0.005312043707817793, -0.0010838485322892666, -0.019910646602511406, 0.11275728791952133, 0.14508824050426483, 0.06051143258810043, -0.08464232832193375, -0.03496673330664635, -0.04373485594987869, 0.0016658891690894961, -0.05736193433403969, 0.08592547476291656, 0.04020396247506142, -0.00451523344963789, -0.022215120494365692, 0.049028705805540085, -0.07215847074985504, -0.07416723668575287, 0.0280220378190279, -0.199037104845047, -0.14286743104457855, -0.0015634113224223256, 0.12728051841259003, -0.00047468641423620284, 0.06859506666660309, -0.0264794509857893, -0.001770825358107686, 0.07856906950473785, -0.016733018681406975, -0.11070074141025543, -0.05530725046992302, 0.09393337368965149, -0.15346799790859222, 0.20310702919960022, -0.04465622827410698, 0.057790838181972504, 0.12810175120830536, 0.05788804218173027, -0.07013047486543655, 0.07793056219816208, 0.042054783552885056, -0.05124843493103981, 0.030165575444698334, 0.07802391052246094, -0.030490299686789513, 0.0550764724612236, 0.04664875566959381, -0.12412387132644653, 0.015107493847608566, -0.0377899631857872, -0.057688821107149124, -0.03234582766890526, -0.042274296283721924, -0.06429233402013779, 0.12082181870937347, 0.2073042243719101, -0.02961086854338646, 0.004870428703725338, -0.08299192786216736, 0.005438422784209251, 0.059302978217601776, 0.0261481124907732, -0.05204824358224869, -0.21670256555080414, 0.001604248653165996, 0.0628993883728981, -0.02172735705971718, -0.26343631744384766, -0.09143883734941483, 0.0017780911875888705, -0.06408309191465378, -0.1163816973567009, 0.07034844905138016, 0.11135251075029373, 0.04832874611020088, -0.05408567935228348, -0.05387600511312485, -0.07241332530975342, 0.1590440422296524, -0.14159518480300903, -0.0910242572426796 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-small-finetuned-bbc-headline This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 12 - eval_batch_size: 12 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:-------:|:------:|:------:|:---------:|:-------:| | No log | 1.0 | 167 | 3.6454 | 22.4311 | 5.9878 | 20.118 | 20.482 | 18.9009 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.9.1 - Datasets 1.12.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "t5-small-finetuned-bbc-headline", "results": []}]}
text2text-generation
furyhawk/t5-small-finetuned-bbc-headline
[ "transformers", "pytorch", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
t5-small-finetuned-bbc-headline =============================== This model is a fine-tuned version of t5-small on the None dataset. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 12 * eval\_batch\_size: 12 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 1 ### Training results ### Framework versions * Transformers 4.11.3 * Pytorch 1.9.1 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 12\n* eval\\_batch\\_size: 12\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.1\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 12\n* eval\\_batch\\_size: 12\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.1\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 63, 98, 4, 31 ]
[ "passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 12\n* eval\\_batch\\_size: 12\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.1\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.07917294651269913, 0.04864596948027611, -0.002182974945753813, 0.11597494035959244, 0.16170091927051544, 0.023381227627396584, 0.1268475502729416, 0.1353602409362793, -0.12793263792991638, 0.004200286231935024, 0.13233067095279694, 0.1732514351606369, 0.015047364868223667, 0.1225942000746727, -0.04851919785141945, -0.2595739960670471, 0.003491318551823497, 0.030370812863111496, -0.053154151886701584, 0.15092484652996063, 0.10372843593358994, -0.10602402687072754, 0.09179793298244476, -0.009956271387636662, -0.1954755038022995, 0.021206684410572052, 0.009479016065597534, -0.0678381398320198, 0.1597500741481781, 0.031234456226229668, 0.09475749731063843, 0.005969724152237177, 0.0739540159702301, -0.19461055099964142, 0.006758211180567741, 0.05588430166244507, -0.0017394390888512135, 0.07880964875221252, 0.05605163052678108, -0.006745294667780399, 0.18102601170539856, -0.07245244830846786, 0.03939914330840111, 0.03011893481016159, -0.12203200906515121, -0.20767086744308472, -0.07834647595882416, 0.027755821123719215, 0.07545594125986099, 0.1257125437259674, -0.008379369974136353, 0.13614404201507568, -0.08870742470026016, 0.10964812338352203, 0.2475784420967102, -0.29367268085479736, -0.05683545768260956, 0.00858201552182436, 0.021947896108031273, 0.08609814196825027, -0.0776451900601387, -0.024788009002804756, 0.05525949224829674, 0.05692436546087265, 0.1330861896276474, -0.03949585184454918, -0.14409998059272766, 0.01625540480017662, -0.1458403319120407, -0.04747922345995903, 0.1848125010728836, 0.02783810906112194, -0.028332818299531937, -0.02866641990840435, -0.08043739199638367, -0.14394859969615936, -0.01735721528530121, -0.011590752750635147, 0.046005792915821075, -0.015027809888124466, -0.05867008864879608, -0.020282797515392303, -0.10439889878034592, -0.06459491699934006, -0.0782221183180809, 0.1274353563785553, 0.04248498007655144, 0.014073709957301617, -0.04170067980885506, 0.1098250150680542, 0.010781916789710522, -0.12665338814258575, 0.01267610490322113, 0.03233743831515312, 0.025454988703131676, -0.0360545888543129, -0.0774466022849083, -0.05749311298131943, 0.021632956340909004, 0.13089124858379364, -0.06047814339399338, 0.0375778004527092, 0.028540702536702156, 0.036148134618997574, -0.09879960119724274, 0.1717275232076645, -0.032432232052087784, -0.02568834088742733, 0.020183155313134193, 0.03595779091119766, 0.02902360074222088, -0.014657155610620975, -0.12229155749082565, -0.002003754023462534, 0.1166999414563179, 0.032489508390426636, -0.06424079090356827, 0.07983476668596268, -0.0410219207406044, -0.022466609254479408, -0.02365809679031372, -0.10729651898145676, 0.008829837664961815, -0.013412442058324814, -0.08581885695457458, -0.009087814018130302, 0.035785481333732605, 0.008341792970895767, -0.05586150288581848, 0.08606988936662674, -0.08130613714456558, 0.031661443412303925, -0.0859742984175682, -0.10590288043022156, 0.016301007941365242, -0.054724402725696564, 0.019059399142861366, -0.10911612957715988, -0.20408034324645996, -0.0060902670957148075, 0.054139990359544754, -0.02135828696191311, -0.06969144940376282, -0.06282404065132141, -0.08462017774581909, 0.008311533369123936, -0.03187226876616478, 0.13361595571041107, -0.07317100465297699, 0.10968617349863052, 0.04083363711833954, 0.06288418173789978, -0.06025770306587219, 0.060586750507354736, -0.11563419550657272, -0.006405573803931475, -0.16084645688533783, 0.054075658321380615, -0.0237534549087286, 0.0735791027545929, -0.09211394935846329, -0.11124932020902634, 0.02001090906560421, 0.0028156719636172056, 0.06675279140472412, 0.10179691761732101, -0.15602850914001465, -0.08324094861745834, 0.18071676790714264, -0.07088713347911835, -0.145315483212471, 0.1176769956946373, -0.050866276025772095, 0.06691048294305801, 0.08196251839399338, 0.16587403416633606, 0.04812927544116974, -0.06522224843502045, 0.029671616852283478, 0.012560950592160225, 0.04294712841510773, -0.05770638957619667, 0.07109889388084412, 0.0019204005366191268, -0.007813677191734314, 0.029585057869553566, -0.023979030549526215, 0.06941793113946915, -0.0950561985373497, -0.08989889919757843, -0.054036784917116165, -0.0900176391005516, 0.02659023366868496, 0.0514310821890831, 0.07773882150650024, -0.11525621265172958, -0.08128651976585388, 0.055760256946086884, 0.08447016030550003, -0.06897364556789398, 0.04200078919529915, -0.050458598881959915, 0.0586102269589901, -0.011208506301045418, -0.0018682981608435512, -0.18445372581481934, -0.015266409143805504, 0.00802867766469717, 0.016746001318097115, 0.033844079822301865, 0.010173344053328037, 0.06033734604716301, 0.054895106703042984, -0.05385657399892807, -0.01835251785814762, -0.031085694208741188, -0.010518829338252544, -0.11746839433908463, -0.18655246496200562, -0.015998881310224533, -0.007434413768351078, 0.13998503983020782, -0.19114567339420319, 0.04195865988731384, -0.02720390446484089, 0.06069082021713257, -0.0026713472325354815, 0.007444408256560564, -0.0494484081864357, 0.07772965729236603, -0.05592822656035423, -0.05114695057272911, 0.07545025646686554, 0.008129552938044071, -0.09110482782125473, -0.029204517602920532, -0.1155395582318306, 0.16304399073123932, 0.14563970267772675, -0.14782515168190002, -0.07683531194925308, 0.014928432181477547, -0.055428024381399155, -0.03157001733779907, -0.03562331572175026, 0.01791822537779808, 0.17356795072555542, -0.01878063939511776, 0.16189827024936676, -0.07987437397241592, -0.04437435418367386, 0.014093045145273209, -0.03821251913905144, 0.050295889377593994, 0.11940842121839523, 0.11228225380182266, -0.08414044976234436, 0.14385724067687988, 0.1655689924955368, -0.09182858467102051, 0.12820276618003845, -0.04848407581448555, -0.07948124408721924, 0.00002448004852340091, -0.010429966263473034, -0.013915013521909714, 0.04936806857585907, -0.15002308785915375, -0.0027403489220887423, 0.024861622601747513, 0.03305019438266754, 0.027690794318914413, -0.22518807649612427, -0.03511664643883705, 0.03221513330936432, -0.06230024993419647, -0.013023708947002888, -0.018810927867889404, 0.004570200107991695, 0.11352880299091339, 0.002986007835716009, -0.09462493658065796, 0.04585839807987213, 0.008652381598949432, -0.09134341776371002, 0.21737702190876007, -0.09602175652980804, -0.17031769454479218, -0.12248946726322174, -0.08362303674221039, -0.06060134992003441, 0.011906039901077747, 0.08844396471977234, -0.0836438238620758, -0.03157703951001167, -0.08117186278104782, 0.050867706537246704, -0.009387906640768051, 0.028569290414452553, 0.009459033608436584, 0.0009586209780536592, 0.06381252408027649, -0.12261784821748734, -0.016405237838625908, -0.03798028081655502, -0.08762212842702866, 0.054675132036209106, 0.011160111054778099, 0.10526394098997116, 0.15908639132976532, -0.01700807735323906, 0.011576665565371513, -0.03816065192222595, 0.23812063038349152, -0.0559934601187706, -0.024514706805348396, 0.15771928429603577, 0.011463457718491554, 0.05219537392258644, 0.0998229905962944, 0.05060592293739319, -0.09869121015071869, 0.0273350290954113, 0.023173099383711815, -0.034020423889160156, -0.24133281409740448, -0.04669547826051712, -0.056445538997650146, 0.006195428781211376, 0.08657041192054749, 0.03312470391392708, 0.0636700689792633, 0.055714283138513565, 0.0283619724214077, 0.08778470009565353, -0.01567726954817772, 0.06136526167392731, 0.14176006615161896, 0.04307707026600838, 0.1284196674823761, -0.05214658007025719, -0.05822296440601349, 0.05673961341381073, -0.019337236881256104, 0.2169407606124878, 0.001874482142738998, 0.13623656332492828, 0.05239889770746231, 0.16065138578414917, -0.019314777106046677, 0.09314564615488052, 0.000461625779280439, -0.026003846898674965, -0.02267128974199295, -0.04848017916083336, -0.0454828180372715, 0.028179306536912918, -0.10243590921163559, 0.05213579535484314, -0.1221335306763649, 0.0019694704096764326, 0.06714151054620743, 0.2684391736984253, 0.03608883544802666, -0.32653719186782837, -0.09239652752876282, 0.0007718034903518856, -0.05826488509774208, -0.014133013784885406, 0.04479465261101723, 0.07495958358049393, -0.09203127771615982, 0.043900541961193085, -0.05179411545395851, 0.10738614946603775, -0.02407153695821762, 0.0624307245016098, 0.06781759858131409, 0.09177780896425247, 0.015578394755721092, 0.09185373783111572, -0.33205702900886536, 0.26835739612579346, -0.003577787196263671, 0.06276372820138931, -0.08174394071102142, 0.008718178607523441, 0.03530208393931389, 0.08003135025501251, 0.05364309251308441, -0.008142431266605854, -0.024887317791581154, -0.15706895291805267, -0.04272302985191345, 0.04112551361322403, 0.09826569259166718, -0.024683617055416107, 0.09390317648649216, -0.0496649444103241, 0.014979330822825432, 0.07631254941225052, 0.0017103089485317469, -0.056198496371507645, -0.08583644777536392, -0.0023330370895564556, 0.028971241787075996, -0.0062789772637188435, -0.06180766597390175, -0.10324937850236893, -0.1055026426911354, 0.16354644298553467, 0.00019141972006764263, -0.04078663885593414, -0.10495822876691818, 0.07012168318033218, 0.06829486042261124, -0.09223593771457672, 0.04198463633656502, 0.01008344255387783, 0.03946889936923981, 0.042267944663763046, -0.09203319251537323, 0.1196160838007927, -0.06014034152030945, -0.16210290789604187, -0.040775008499622345, 0.10710137337446213, 0.008463850244879723, 0.06051621213555336, -0.01287116575986147, 0.004631017334759235, -0.057121675461530685, -0.09244589507579803, -0.0053913528099656105, -0.016798846423625946, 0.06365916877985, 0.02939794771373272, -0.060023993253707886, 0.010561967268586159, -0.07215899974107742, -0.05443762615323067, 0.2077518105506897, 0.23712407052516937, -0.08173257857561111, 0.026812033727765083, 0.041956912726163864, -0.06933805346488953, -0.19465497136116028, 0.010941562242805958, 0.04259354621171951, -0.008906024508178234, 0.029733000323176384, -0.19201257824897766, 0.1136719286441803, 0.11429408937692642, -0.004513062071055174, 0.10425625741481781, -0.3607324957847595, -0.13044631481170654, 0.11236605048179626, 0.1345929205417633, 0.14448387920856476, -0.1559794396162033, -0.017831839621067047, -0.04936124384403229, -0.1304577887058258, 0.11261636763811111, -0.11683966219425201, 0.12968771159648895, -0.027483250945806503, 0.10147949308156967, -0.0022210560273379087, -0.042627133429050446, 0.11152711510658264, 0.018609317019581795, 0.09249046444892883, -0.07352491468191147, 0.016301482915878296, 0.04118962585926056, -0.033165786415338516, 0.04096019268035889, -0.11841290444135666, 0.042949266731739044, -0.08685306459665298, -0.021075839176774025, -0.07476785033941269, 0.04265621304512024, -0.029700979590415955, -0.07056175172328949, -0.031369514763355255, -0.008610855787992477, 0.06331949681043625, -0.018467111513018608, 0.13270927965641022, 0.015573037788271904, 0.14292199909687042, 0.11068668961524963, 0.07161762565374374, -0.09083005785942078, -0.03497166931629181, -0.02081504464149475, -0.015218832530081272, 0.056337278336286545, -0.16488102078437805, 0.030580317601561546, 0.13437685370445251, 0.010747263208031654, 0.14757898449897766, 0.08591947704553604, -0.015998221933841705, 0.0022985879331827164, 0.06368091702461243, -0.17762933671474457, -0.08380116522312164, -0.020524989813566208, -0.05800590664148331, -0.09254223108291626, 0.059287525713443756, 0.10684067010879517, -0.07528511434793472, -0.012088022194802761, -0.0155298737809062, 0.01303514838218689, -0.07470716536045074, 0.18417030572891235, 0.04380703344941139, 0.0416082888841629, -0.10496323555707932, 0.08876176923513412, 0.04748396947979927, -0.06782644987106323, 0.015543593093752861, 0.11840908229351044, -0.08720020949840546, -0.05243012309074402, 0.09453917294740677, 0.16343289613723755, -0.091640904545784, -0.05113108083605766, -0.13414767384529114, -0.12483752518892288, 0.0872647613286972, 0.16963893175125122, 0.10865944623947144, 0.021409064531326294, -0.053335223346948624, 0.006011987570673227, -0.1117209941148758, 0.0749647468328476, 0.03931387513875961, 0.05792732536792755, -0.12505660951137543, 0.16819581389427185, 0.013944197446107864, 0.051054637879133224, -0.028418276458978653, 0.019502654671669006, -0.09257013350725174, 0.01912163756787777, -0.14538252353668213, -0.03226806968450546, -0.023024072870612144, 0.005268795415759087, 0.001100825727917254, -0.04833751171827316, -0.05720861628651619, 0.020183727145195007, -0.12049810588359833, -0.03097335435450077, 0.014501587487757206, 0.06270521134138107, -0.12307331711053848, -0.03522821515798569, 0.02362128160893917, -0.06732599437236786, 0.08398976922035217, 0.06531856954097748, -0.009664547629654408, 0.08084560930728912, -0.1569766104221344, 0.002632519928738475, 0.07538976520299911, 0.024373812600970268, 0.0566331148147583, -0.07422428578138351, -0.008858864195644855, 0.016629446297883987, 0.06310401111841202, 0.02029404230415821, 0.06627143174409866, -0.1366523653268814, -0.0063933213241398335, -0.027860848233103752, -0.0918816477060318, -0.06330176442861557, 0.02487439103424549, 0.0753958448767662, 0.007582054473459721, 0.19915933907032013, -0.08219407498836517, 0.036102019250392914, -0.20592761039733887, 0.006831109989434481, -0.011449619196355343, -0.11950411647558212, -0.14060477912425995, -0.07696956396102905, 0.06025188788771629, -0.05471418425440788, 0.1399163156747818, 0.023025546222925186, 0.05182741954922676, 0.02828139066696167, -0.007445800118148327, 0.01381876040250063, 0.012731031514704227, 0.23461133241653442, 0.04216904565691948, -0.031675584614276886, 0.04054085910320282, 0.04571684077382088, 0.11197195202112198, 0.09748857468366623, 0.20480012893676758, 0.14066432416439056, -0.0033939466811716557, 0.10284251719713211, 0.0219118669629097, -0.05517038702964783, -0.15106071531772614, 0.003117592539638281, -0.01956818625330925, 0.11502469331026077, -0.023438332602381706, 0.21759678423404694, 0.08355186879634857, -0.14840272068977356, 0.03885175660252571, -0.05681624636054039, -0.07637591660022736, -0.12010952085256577, -0.03224341198801994, -0.08112258464097977, -0.16587525606155396, -0.010297865606844425, -0.11785360425710678, 0.035987552255392075, 0.11200784146785736, 0.015206746757030487, -0.03296986594796181, 0.14225643873214722, 0.021206123754382133, -0.010289841331541538, 0.05122530087828636, -0.01026699785143137, -0.021738942712545395, -0.10647067427635193, -0.07321388274431229, -0.013525839895009995, -0.009562790393829346, 0.03500599414110184, -0.04077242314815521, -0.061778102070093155, 0.038428861647844315, -0.05837540701031685, -0.09244310110807419, 0.01757095381617546, 0.02658015489578247, 0.05065922439098358, 0.05860096216201782, 0.011353785172104836, -0.0033786254934966564, 0.005374784581363201, 0.24470336735248566, -0.0919509083032608, -0.11948703974485397, -0.09069705009460449, 0.28744497895240784, 0.04647303745150566, -0.00890024658292532, 0.029164299368858337, -0.05938807502388954, -0.022610485553741455, 0.2589184045791626, 0.20510615408420563, -0.08972618728876114, -0.014466590248048306, -0.006199606694281101, -0.0030199240427464247, -0.015341260470449924, 0.1219402402639389, 0.1583179086446762, 0.04660766199231148, -0.08424834161996841, -0.02008887007832527, -0.04903056472539902, -0.0010604590643197298, -0.05726759508252144, 0.06355644762516022, 0.029035842046141624, -0.009726980701088905, -0.023800551891326904, 0.05728472024202347, -0.07442960143089294, -0.07502104341983795, 0.019478987902402878, -0.18666775524616241, -0.1480935513973236, -0.010445859283208847, 0.11946103721857071, 0.012139055877923965, 0.06353101879358292, -0.020150642842054367, 0.0038059880025684834, 0.07451078295707703, -0.01605049893260002, -0.10726287961006165, -0.07784072309732437, 0.10310747474431992, -0.14511466026306152, 0.17739343643188477, -0.039333708584308624, 0.0655146911740303, 0.12102089822292328, 0.0674641877412796, -0.07357370108366013, 0.09280652552843094, 0.04161441698670387, -0.04643237590789795, 0.034564316272735596, 0.06928757578134537, -0.03630826994776726, 0.04778393357992172, 0.04295654967427254, -0.12098854780197144, 0.008995948359370232, -0.023153331130743027, -0.052637819200754166, -0.03233223780989647, -0.06409633159637451, -0.05685875192284584, 0.1183834969997406, 0.21719400584697723, -0.04028778895735741, 0.019457794725894928, -0.09246940165758133, 0.00046589059638790786, 0.051135577261447906, 0.0009051461238414049, -0.06633755564689636, -0.22266824543476105, -0.006728063337504864, 0.07690277695655823, -0.012270068749785423, -0.23744411766529083, -0.07731765508651733, -0.00855524092912674, -0.06254848092794418, -0.11943590641021729, 0.09409618377685547, 0.08532225340604782, 0.04100821539759636, -0.04578138142824173, -0.07077021896839142, -0.0743459016084671, 0.16456227004528046, -0.14998897910118103, -0.10071311146020889 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-small-finetuned-bbc This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.3238 - Rouge1: 21.2266 - Rouge2: 16.0927 - Rougel: 19.6785 - Rougelsum: 19.8849 - Gen Len: 19.0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:| | 0.4882 | 1.0 | 1001 | 0.3238 | 21.2266 | 16.0927 | 19.6785 | 19.8849 | 19.0 | ### Framework versions - Transformers 4.12.0 - Pytorch 1.10.0 - Datasets 1.14.0 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["rouge"], "model-index": [{"name": "t5-small-finetuned-bbc", "results": []}]}
text2text-generation
furyhawk/t5-small-finetuned-bbc
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
t5-small-finetuned-bbc ====================== This model is a fine-tuned version of t5-small on the None dataset. It achieves the following results on the evaluation set: * Loss: 0.3238 * Rouge1: 21.2266 * Rouge2: 16.0927 * Rougel: 19.6785 * Rougelsum: 19.8849 * Gen Len: 19.0 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 2 * eval\_batch\_size: 2 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 1 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.12.0 * Pytorch 1.10.0 * Datasets 1.14.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 2\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.0\n* Pytorch 1.10.0\n* Datasets 1.14.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 2\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.0\n* Pytorch 1.10.0\n* Datasets 1.14.0\n* Tokenizers 0.10.3" ]
[ 67, 113, 4, 30 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 2\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.0\n* Pytorch 1.10.0\n* Datasets 1.14.0\n* Tokenizers 0.10.3" ]
[ -0.06782788038253784, 0.04433724656701088, -0.0035998334642499685, 0.09652519971132278, 0.14275497198104858, 0.015429266728460789, 0.1330752968788147, 0.12975476682186127, -0.1051301583647728, 0.026657886803150177, 0.11661048978567123, 0.14387692511081696, 0.031686823815107346, 0.105348601937294, -0.051702648401260376, -0.28164613246917725, -0.000021436148017528467, 0.04579269886016846, -0.06063133850693703, 0.1376054733991623, 0.09252828359603882, -0.11194471269845963, 0.07953189313411713, 0.012627150863409042, -0.16403554379940033, 0.02137131802737713, 0.002087414963170886, -0.056814055889844894, 0.14766059815883636, 0.04134934023022652, 0.11505349725484848, 0.010109550319612026, 0.07201773673295975, -0.20944343507289886, 0.010537943802773952, 0.06467604637145996, 0.0040346053428947926, 0.08706402033567429, 0.07452190667390823, 0.006003061309456825, 0.16334740817546844, -0.05794606730341911, 0.05370039865374565, 0.028131656348705292, -0.11398454755544662, -0.22781585156917572, -0.07693566381931305, 0.04308208450675011, 0.0813128873705864, 0.11028821021318436, -0.013428544625639915, 0.10616841167211533, -0.06701444089412689, 0.10445152223110199, 0.24147748947143555, -0.28660890460014343, -0.06231407821178436, -0.0027231755666434765, 0.04494253918528557, 0.07830056548118591, -0.08254013955593109, -0.02683071233332157, 0.03391525521874428, 0.055605459958314896, 0.13027556240558624, -0.02524024434387684, -0.09210877120494843, 0.005101039074361324, -0.14727051556110382, -0.048706766217947006, 0.14578410983085632, 0.03818109259009361, -0.02564958482980728, -0.05905092507600784, -0.07782602310180664, -0.1718481034040451, -0.03385491296648979, -0.019560419023036957, 0.042065054178237915, -0.01544724591076374, -0.05269087851047516, -0.03544958308339119, -0.11443790793418884, -0.06579707562923431, -0.07079295068979263, 0.12356285005807877, 0.04753381386399269, 0.00012766027066390961, -0.04364969581365585, 0.11161283403635025, -0.008199862204492092, -0.1254054754972458, 0.018108325079083443, 0.028564492240548134, 0.0053420900367200375, -0.025781933218240738, -0.06013994663953781, -0.09443621337413788, 0.005965488497167826, 0.13299871981143951, -0.07495728880167007, 0.0532926544547081, -0.0038984084967523813, 0.05108484998345375, -0.1021685004234314, 0.1687321960926056, -0.035651225596666336, -0.009970627725124359, -0.0016499374760314822, 0.04703643172979355, 0.03056427463889122, -0.01738944835960865, -0.11615189909934998, 0.005445741582661867, 0.1007419228553772, 0.019203118979930878, -0.05051674321293831, 0.06858287006616592, -0.045036956667900085, -0.02390499971807003, -0.020186861976981163, -0.09543820470571518, 0.023588230833411217, 0.001586501020938158, -0.06760028749704361, 0.0033244772348552942, 0.04605631157755852, 0.015313544310629368, -0.047257453203201294, 0.10498610138893127, -0.07329992204904556, 0.02897300384938717, -0.09538904577493668, -0.12136410921812057, 0.028001312166452408, -0.06227850541472435, 0.009625358507037163, -0.10077358782291412, -0.1790349781513214, -0.009091285988688469, 0.061474937945604324, -0.03600281849503517, -0.056246835738420486, -0.047558512538671494, -0.07250577211380005, 0.027577312663197517, -0.027235515415668488, 0.1474483609199524, -0.06102651730179787, 0.09904813021421432, 0.028310393914580345, 0.05316472426056862, -0.03857502341270447, 0.06141665205359459, -0.09223145246505737, 0.020594166591763496, -0.1566881388425827, 0.049100589007139206, -0.03550577163696289, 0.052464134991168976, -0.10089586675167084, -0.10165266692638397, -0.01787670888006687, 0.0028115366585552692, 0.08343285322189331, 0.08830803632736206, -0.154879629611969, -0.08590175211429596, 0.17103959619998932, -0.07867550849914551, -0.11992273479700089, 0.13132688403129578, -0.05130386725068092, 0.0422637015581131, 0.05065455287694931, 0.17037270963191986, 0.06139059737324715, -0.08944401144981384, 0.022792501375079155, -0.000908329791855067, 0.045353252440690994, -0.03368982672691345, 0.074342280626297, -0.00708856200799346, 0.019181907176971436, 0.019995950162410736, -0.00992581993341446, 0.06628614664077759, -0.0861194059252739, -0.08347377926111221, -0.048306722193956375, -0.0699964165687561, 0.029123414307832718, 0.056319598108530045, 0.07359984517097473, -0.09987644851207733, -0.10616382211446762, 0.061124272644519806, 0.07610936462879181, -0.08183436095714569, 0.051134560257196426, -0.058165550231933594, 0.060025401413440704, -0.03257395699620247, -0.0018190336413681507, -0.18381784856319427, -0.018535656854510307, 0.009918633848428726, -0.022135058417916298, 0.031517669558525085, 0.0135117769241333, 0.069485604763031, 0.060007091611623764, -0.05561831220984459, -0.02867540903389454, -0.04680934548377991, -0.006918735336512327, -0.12213654071092606, -0.1972578912973404, -0.030484922230243683, -0.012241514399647713, 0.12219353765249252, -0.21091949939727783, 0.043242983520030975, -0.006392347160726786, 0.08306142687797546, 0.014123060740530491, -0.0006961054168641567, -0.04939413443207741, 0.07595488429069519, -0.05848483368754387, -0.03989946097135544, 0.07565265893936157, 0.01367438305169344, -0.10390675812959671, -0.00695889862254262, -0.14047613739967346, 0.13661620020866394, 0.13013988733291626, -0.13621343672275543, -0.06768522411584854, -0.009981220588088036, -0.06540030986070633, -0.047182608395814896, -0.03572717681527138, 0.0012024646857753396, 0.18902093172073364, 0.0012805620208382607, 0.1623096466064453, -0.07561662793159485, -0.04972439631819725, 0.02255043014883995, -0.03338634595274925, 0.017333175987005234, 0.13210684061050415, 0.1098245158791542, -0.07674012333154678, 0.1413242667913437, 0.15904557704925537, -0.08908900618553162, 0.14896896481513977, -0.03629208356142044, -0.08987785130739212, -0.017193863168358803, -0.006527523044496775, 0.0001317002606811002, 0.0674353539943695, -0.16973525285720825, 0.009769772179424763, 0.028077440336346626, 0.027976306155323982, 0.03525811433792114, -0.2209646850824356, -0.015932954847812653, 0.04296252131462097, -0.05502373352646828, -0.005261301063001156, -0.009557788260281086, 0.01595301553606987, 0.11048020422458649, 0.0013257486280053854, -0.07359479367733002, 0.03619956225156784, -0.004742119926959276, -0.08891066163778305, 0.20220457017421722, -0.08130957186222076, -0.18631812930107117, -0.13219432532787323, -0.07919999212026596, -0.045776575803756714, 0.004416164476424456, 0.07504019141197205, -0.07572651654481888, -0.03056749328970909, -0.08638820797204971, 0.0542326383292675, -0.02526708133518696, 0.023552464321255684, 0.0026545210275799036, 0.005429068114608526, 0.07235877960920334, -0.10571255534887314, -0.012938103638589382, -0.04263373836874962, -0.05644642561674118, 0.043252378702163696, 0.03509129211306572, 0.10376886278390884, 0.16163957118988037, -0.012684289366006851, 0.014108268544077873, -0.03339694067835808, 0.19491468369960785, -0.06574923545122147, -0.018825577571988106, 0.16035422682762146, -0.014659375883638859, 0.05777197703719139, 0.11627084761857986, 0.059506241232156754, -0.07241041958332062, 0.01793687231838703, 0.04258417710661888, -0.03670738637447357, -0.2411254644393921, -0.04287393391132355, -0.06099496781826019, 0.021754002198576927, 0.09574591368436813, 0.026170386001467705, 0.04640398174524307, 0.054071396589279175, 0.023063933476805687, 0.06767836213111877, -0.008479711599647999, 0.07457855343818665, 0.1588844358921051, 0.031557243317365646, 0.13433611392974854, -0.04440002143383026, -0.06056072935461998, 0.04810607060790062, -0.004757370334118605, 0.21462886035442352, 0.007350333966314793, 0.15533018112182617, 0.0604400709271431, 0.14641764760017395, -0.0010903246002271771, 0.07187599688768387, -0.002061997540295124, -0.03102552890777588, -0.012867079116404057, -0.047187209129333496, -0.0290084145963192, 0.03438453748822212, -0.06704654544591904, 0.04497829079627991, -0.12404672056436539, -0.006146666593849659, 0.044157061725854874, 0.2704029083251953, 0.03533807396888733, -0.3126929700374603, -0.09411459416151047, 0.011351626366376877, -0.0651094987988472, -0.023988202214241028, 0.03420967608690262, 0.09722533822059631, -0.08823759108781815, 0.044016871601343155, -0.08351439237594604, 0.10565251111984253, -0.037209220230579376, 0.04778248444199562, 0.06682052463293076, 0.08947866410017014, 0.012446824461221695, 0.08703675866127014, -0.31043413281440735, 0.2716985046863556, -0.003472156822681427, 0.05457907170057297, -0.07274822890758514, 0.01784026250243187, 0.026269137859344482, 0.03929935395717621, 0.06368310004472733, -0.024405889213085175, -0.05732078477740288, -0.16301462054252625, -0.06130891293287277, 0.017685336992144585, 0.09744621813297272, -0.030025692656636238, 0.11098957061767578, -0.04561527073383331, 0.0094941770657897, 0.07494211941957474, 0.003192265285179019, -0.07271026819944382, -0.10502028465270996, 0.010933009907603264, 0.029795924201607704, -0.02603544481098652, -0.07096627354621887, -0.1100049614906311, -0.09830038994550705, 0.16939325630664825, -0.03848700970411301, -0.037215057760477066, -0.1058887392282486, 0.08862464874982834, 0.07448191940784454, -0.0879577100276947, 0.03733263909816742, 0.005270204972475767, 0.08567387610673904, 0.024934934452176094, -0.08112850040197372, 0.1172589659690857, -0.07157692313194275, -0.173759326338768, -0.05194651708006859, 0.1284428983926773, 0.020696815103292465, 0.06342977285385132, -0.020360836759209633, 0.007517158053815365, -0.04789453372359276, -0.0816565677523613, 0.016606448218226433, -0.005045270547270775, 0.06617199629545212, 0.022706134244799614, -0.06583625078201294, 0.017087340354919434, -0.0562903955578804, -0.05144750326871872, 0.20406830310821533, 0.2272561937570572, -0.08523885160684586, 0.03562590479850769, 0.03400197625160217, -0.07929209619760513, -0.18780772387981415, 0.009321293793618679, 0.06541794538497925, -0.001064034760929644, 0.04427838698029518, -0.19181248545646667, 0.0907185971736908, 0.10248927772045135, -0.011204205453395844, 0.09844070672988892, -0.3535374402999878, -0.13651707768440247, 0.11917940527200699, 0.13439476490020752, 0.08732422441244125, -0.15252214670181274, -0.025699082762002945, -0.030122386291623116, -0.11855534464120865, 0.12990324199199677, -0.09653998911380768, 0.1263178288936615, -0.027052273973822594, 0.1017284095287323, 0.011597820557653904, -0.05660867691040039, 0.1083865538239479, -0.017207637429237366, 0.07680467516183853, -0.0685083270072937, 0.014888228848576546, 0.039680711925029755, -0.043469373136758804, 0.027088427916169167, -0.09784199297428131, 0.019310524687170982, -0.09197678416967392, -0.03318614140152931, -0.06989653408527374, 0.03402944654226303, -0.0369064025580883, -0.06216585636138916, -0.0340457446873188, 0.004841904621571302, 0.05903713405132294, -0.007343661040067673, 0.159297376871109, -0.0008433745824731886, 0.1517000049352646, 0.1346580684185028, 0.08820624649524689, -0.051863137632608414, -0.06950891762971878, -0.021303437650203705, -0.017218485474586487, 0.05146022140979767, -0.138535276055336, 0.032383911311626434, 0.14723645150661469, 0.00759586738422513, 0.15235212445259094, 0.08449626713991165, -0.04190165176987648, 0.01325757522135973, 0.0594119094312191, -0.16008125245571136, -0.12325162440538406, -0.023386411368846893, -0.025552209466695786, -0.10943756997585297, 0.04295787215232849, 0.12464161217212677, -0.07224418222904205, -0.0016978512285277247, 0.0036801521200686693, 0.019625667482614517, -0.055217478424310684, 0.18272171914577484, 0.026783250272274017, 0.044397588819265366, -0.08806183934211731, 0.0871124118566513, 0.048711854964494705, -0.10883699357509613, 0.018754037097096443, 0.11065531522035599, -0.06439591944217682, -0.04763481393456459, 0.04939225688576698, 0.15708573162555695, -0.06708485633134842, -0.058406926691532135, -0.139895498752594, -0.13521715998649597, 0.0970127210021019, 0.1483108252286911, 0.08172658830881119, 0.012950403615832329, -0.06330493092536926, 0.020034942775964737, -0.11185197532176971, 0.1044667586684227, 0.04856828972697258, 0.060219570994377136, -0.1360442042350769, 0.1572965532541275, 0.015142573043704033, 0.040383171290159225, -0.020403554663062096, 0.019273169338703156, -0.09009118378162384, 0.014664721675217152, -0.14593584835529327, -0.024966999888420105, -0.020036034286022186, -0.003656235756352544, -0.003292836481705308, -0.041110992431640625, -0.06540784984827042, 0.01760144717991352, -0.11208710819482803, -0.030276624485850334, 0.01177323516458273, 0.05850179120898247, -0.11056350916624069, -0.02932637184858322, 0.026814982295036316, -0.07052388787269592, 0.08332556486129761, 0.053068146109580994, 0.008242658339440823, 0.052083250135183334, -0.13420480489730835, 0.024329829961061478, 0.06261385232210159, 0.024745382368564606, 0.039725374430418015, -0.09713469445705414, -0.010691859759390354, 0.0016797082498669624, 0.03914453461766243, 0.01041419617831707, 0.060009025037288666, -0.13983100652694702, -0.005829171277582645, -0.01847526803612709, -0.08468226343393326, -0.0696980282664299, 0.033437296748161316, 0.043832533061504364, 0.033573251217603683, 0.18562637269496918, -0.08800291270017624, 0.05130266770720482, -0.21998275816440582, 0.015899289399385452, 0.004399178083986044, -0.1054014340043068, -0.07845035195350647, -0.07302767783403397, 0.057581450790166855, -0.05912204459309578, 0.13096047937870026, 0.02100873924791813, 0.042978547513484955, 0.03779754042625427, -0.037644512951374054, 0.02153756283223629, 0.017344634979963303, 0.2160089612007141, 0.029126187786459923, -0.04123394563794136, 0.03959515318274498, 0.03190670534968376, 0.10950565338134766, 0.12942340970039368, 0.20989695191383362, 0.15021678805351257, -0.005242424085736275, 0.10026582330465317, 0.03722582384943962, -0.061271265149116516, -0.16943159699440002, 0.047235775738954544, -0.03190228343009949, 0.1321544498205185, -0.03011135756969452, 0.24007612466812134, 0.09705600887537003, -0.15639717876911163, 0.04778449982404709, -0.04502851516008377, -0.07592654228210449, -0.12004204839468002, -0.07353256642818451, -0.08225590735673904, -0.1540689468383789, -0.007024738006293774, -0.1183137595653534, 0.04465652257204056, 0.09252751618623734, 0.02208302728831768, -0.02425597421824932, 0.13976499438285828, 0.0276450514793396, -0.0041780052706599236, 0.0632861778140068, 0.005037613678723574, -0.013289486058056355, -0.10195589810609818, -0.07397058606147766, 0.0038068678695708513, -0.010872447863221169, 0.03479336202144623, -0.03575409948825836, -0.05036678537726402, 0.033362455666065216, -0.026288021355867386, -0.09478498250246048, 0.010138877667486668, 0.020014122128486633, 0.06966263800859451, 0.06459751725196838, 0.005065075121819973, 0.002537456341087818, -0.01341671496629715, 0.23253825306892395, -0.08814657479524612, -0.0757504478096962, -0.09162289649248123, 0.24705106019973755, 0.031794480979442596, -0.013326828368008137, 0.026852257549762726, -0.06005813553929329, -0.017049314454197884, 0.24420884251594543, 0.18938791751861572, -0.0674339160323143, -0.012819290161132812, 0.015370165929198265, -0.005572113674134016, -0.0169990211725235, 0.10618248581886292, 0.14500907063484192, 0.07651906460523605, -0.08806303143501282, -0.02986014261841774, -0.04577194154262543, -0.0012935699196532369, -0.049234818667173386, 0.08968830853700638, 0.03686191514134407, 0.0014610213693231344, -0.024351324886083603, 0.05653424561023712, -0.06637740880250931, -0.0845448300242424, 0.009547140449285507, -0.20177145302295685, -0.15383431315422058, -0.013123972341418266, 0.12216848880052567, -0.007151500321924686, 0.055742595344781876, -0.02547372505068779, -0.002277882769703865, 0.08620477467775345, -0.016901638358831406, -0.08455804735422134, -0.06504016369581223, 0.09019066393375397, -0.12091828137636185, 0.19755621254444122, -0.041123129427433014, 0.043049056082963943, 0.12850162386894226, 0.061680927872657776, -0.06995117664337158, 0.07302333414554596, 0.043950825929641724, -0.06434544920921326, 0.03708767890930176, 0.10836237668991089, -0.031869929283857346, 0.06717094033956528, 0.05043937265872955, -0.14209182560443878, 0.015362357720732689, -0.06898941099643707, -0.06725975126028061, -0.025171125307679176, -0.04000160098075867, -0.06134680286049843, 0.12406925857067108, 0.2169332355260849, -0.03776654973626137, -0.001417885534465313, -0.08086749166250229, 0.0010421638144180179, 0.05000630021095276, 0.05861825495958328, -0.03542450815439224, -0.2283639907836914, 0.008655952289700508, 0.06491170823574066, -0.009428288787603378, -0.26759713888168335, -0.09679044038057327, 0.014385036192834377, -0.05962013825774193, -0.11849819123744965, 0.08195292204618454, 0.09952130913734436, 0.04477189481258392, -0.05061391368508339, -0.07911429554224014, -0.06806057691574097, 0.16341359913349152, -0.14012354612350464, -0.07576154172420502 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-small-finetuned-xsum This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the xsum dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:-------:|:------:|:-------:|:---------:|:-------:| | No log | 1.0 | 128 | 2.9003 | 19.4784 | 2.8529 | 14.7786 | 15.0614 | 18.9825 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.9.1 - Datasets 1.12.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["xsum"], "model-index": [{"name": "t5-small-finetuned-xsum", "results": []}]}
text2text-generation
furyhawk/t5-small-finetuned-xsum
[ "transformers", "pytorch", "t5", "text2text-generation", "generated_from_trainer", "dataset:xsum", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #t5 #text2text-generation #generated_from_trainer #dataset-xsum #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
t5-small-finetuned-xsum ======================= This model is a fine-tuned version of t5-small on the xsum dataset. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 1 ### Training results ### Framework versions * Transformers 4.11.3 * Pytorch 1.9.1 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.1\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #generated_from_trainer #dataset-xsum #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.1\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 69, 98, 4, 31 ]
[ "passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #generated_from_trainer #dataset-xsum #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.1\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.07997394353151321, 0.058514554053545, -0.002649132162332535, 0.1261199563741684, 0.15193350613117218, 0.026369627565145493, 0.12548786401748657, 0.13108357787132263, -0.10050759464502335, 0.01585061475634575, 0.13210514187812805, 0.16160060465335846, 0.019460707902908325, 0.12679901719093323, -0.05587637051939964, -0.2556529641151428, -0.004739534575492144, 0.030445706099271774, -0.028010740876197815, 0.1455075442790985, 0.1033482626080513, -0.11575791239738464, 0.09492339193820953, -0.00708753103390336, -0.17941832542419434, 0.0022996540647000074, -0.002411878202110529, -0.0519631989300251, 0.1526709496974945, 0.01896682009100914, 0.09814922511577606, -0.002404453931376338, 0.0744895339012146, -0.19247294962406158, 0.0043382360599935055, 0.044070035219192505, 0.005497777368873358, 0.08225111663341522, 0.04088501259684563, 0.0070083001628518105, 0.15700232982635498, -0.06322188675403595, 0.04657133296132088, 0.025218019261956215, -0.12660053372383118, -0.20619162917137146, -0.08135835826396942, 0.030648157000541687, 0.08282074332237244, 0.11040718108415604, -0.01215808279812336, 0.12471233308315277, -0.08302633464336395, 0.10254371911287308, 0.24338991940021515, -0.2939198315143585, -0.0584167018532753, 0.024668581783771515, 0.025240210816264153, 0.08683480322360992, -0.082910917699337, -0.03183119371533394, 0.04338839650154114, 0.06151605769991875, 0.14326442778110504, -0.03112361766397953, -0.11123025417327881, 0.015365704894065857, -0.14773279428482056, -0.05950355529785156, 0.19830799102783203, 0.03930151090025902, -0.03143500164151192, -0.04501767084002495, -0.07598430663347244, -0.15809956192970276, -0.01793818548321724, -0.014891414903104305, 0.04426715895533562, -0.01871540956199169, -0.05096578970551491, -0.018237009644508362, -0.10350244492292404, -0.06214640289545059, -0.0710250586271286, 0.12496379762887955, 0.03955591842532158, 0.014814547263085842, -0.03660585731267929, 0.10958905518054962, -0.00420890050008893, -0.12558503448963165, 0.010729532688856125, 0.024662889540195465, 0.02945450320839882, -0.030458861961960793, -0.0729139894247055, -0.06610063463449478, 0.010917002335190773, 0.13359329104423523, -0.05373327434062958, 0.03317446634173393, 0.03324054554104805, 0.041498880833387375, -0.08493804186582565, 0.158800408244133, -0.01566447503864765, -0.0356268472969532, 0.011996269226074219, 0.042100537568330765, 0.033878084272146225, -0.01057179644703865, -0.1248745247721672, 0.006840249989181757, 0.10240121930837631, 0.023197894915938377, -0.06507381051778793, 0.0724816843867302, -0.04800545796751976, -0.027688872069120407, -0.012442097999155521, -0.09303636848926544, 0.011547458358108997, -0.015028631314635277, -0.08063871413469315, -0.024527736008167267, 0.031195184215903282, 0.014867114834487438, -0.042482081800699234, 0.09139516949653625, -0.08111586421728134, 0.0209968201816082, -0.08340218663215637, -0.10469868034124374, 0.016334623098373413, -0.08378718048334122, 0.019259708002209663, -0.10399243235588074, -0.22115971148014069, 0.001275678165256977, 0.05889567732810974, -0.026650138199329376, -0.07209663093090057, -0.06526123732328415, -0.07452844828367233, 0.011429236270487309, -0.024729253724217415, 0.13111473619937897, -0.06958390772342682, 0.1052197813987732, 0.03387148305773735, 0.06581956893205643, -0.05215982347726822, 0.05822012573480606, -0.11168491095304489, 0.004903921391814947, -0.14817555248737335, 0.05375230684876442, -0.02267894335091114, 0.055175911635160446, -0.08773524314165115, -0.10473816096782684, 0.005263062659651041, -0.010287889279425144, 0.066351018846035, 0.10478734970092773, -0.16125445067882538, -0.07249816507101059, 0.17709650099277496, -0.07067068666219711, -0.1494017094373703, 0.13412804901599884, -0.058103255927562714, 0.06323982775211334, 0.0728367269039154, 0.17793597280979156, 0.04058841988444328, -0.08404674381017685, 0.016882671043276787, 0.004961005412042141, 0.04579762741923332, -0.055986884981393814, 0.08184624463319778, 0.00695486506447196, 0.017832381650805473, 0.03024030476808548, -0.020409315824508667, 0.06016942486166954, -0.0939459502696991, -0.09394029527902603, -0.05388641729950905, -0.09053239971399307, 0.03518497943878174, 0.05069806054234505, 0.0735338106751442, -0.11152952164411545, -0.07704455405473709, 0.03584175929427147, 0.07771380990743637, -0.06554672122001648, 0.03767182677984238, -0.05129692703485489, 0.0751439705491066, -0.029956219717860222, -0.006965094245970249, -0.17402005195617676, -0.015488014556467533, 0.017827019095420837, 0.025516094639897346, 0.023351887241005898, 0.008440791629254818, 0.06109919399023056, 0.0659000501036644, -0.058036111295223236, -0.015383083373308182, -0.04434098303318024, -0.006147237494587898, -0.12796184420585632, -0.17502492666244507, -0.03255235031247139, -0.010602534748613834, 0.13638831675052643, -0.19926461577415466, 0.04192586615681648, -0.02939549833536148, 0.06687702238559723, 0.0020050909370183945, -0.009885949082672596, -0.049434490501880646, 0.06635420769453049, -0.05717303231358528, -0.04638989269733429, 0.07703658938407898, 0.012666009366512299, -0.09481421858072281, -0.03681453317403793, -0.10541471838951111, 0.1496015340089798, 0.1322111338376999, -0.13231100142002106, -0.06696859002113342, -0.00455106608569622, -0.06401146948337555, -0.033218178898096085, -0.04965675622224808, 0.0041285851038992405, 0.17800696194171906, -0.010582235641777515, 0.15511099994182587, -0.08445284515619278, -0.0486748069524765, 0.02149941213428974, -0.03898110240697861, 0.04156067222356796, 0.12376950681209564, 0.11296025663614273, -0.09155967086553574, 0.14615203440189362, 0.16125927865505219, -0.07732709497213364, 0.12919828295707703, -0.04543595761060715, -0.07451257109642029, -0.020021704956889153, -0.013784199021756649, -0.01518173236399889, 0.06829994171857834, -0.15428569912910461, 0.007282684091478586, 0.033931609243154526, 0.036071863025426865, 0.01931975781917572, -0.21790891885757446, -0.03774366155266762, 0.04277941584587097, -0.05578772351145744, -0.02396317385137081, -0.007546956185251474, 0.0031150993891060352, 0.10903681069612503, -0.00038029701681807637, -0.0825638696551323, 0.04728563129901886, 0.0050191194750368595, -0.08818954229354858, 0.2059849202632904, -0.08240596204996109, -0.16105183959007263, -0.11788029968738556, -0.08750247955322266, -0.05937453731894493, 0.011594978161156178, 0.08599645644426346, -0.07498139142990112, -0.026227787137031555, -0.08787465840578079, 0.04376724362373352, -0.01400178112089634, 0.02145102433860302, 0.012477410025894642, -0.003229339374229312, 0.05078411102294922, -0.11126895993947983, -0.02319890633225441, -0.046904049813747406, -0.06511526554822922, 0.05098444223403931, 0.011630653403699398, 0.10975563526153564, 0.14182519912719727, -0.014767087064683437, 0.01693933829665184, -0.03464080020785332, 0.2551378309726715, -0.05598245561122894, -0.018406137824058533, 0.15981997549533844, 0.005050293169915676, 0.05915122851729393, 0.10024429857730865, 0.0534806028008461, -0.093075230717659, 0.014800571836531162, 0.02303101308643818, -0.04497664049267769, -0.23265735805034637, -0.04876033589243889, -0.06404580920934677, 0.010174375027418137, 0.09363153576850891, 0.023519128561019897, 0.04279599338769913, 0.06398368626832962, 0.02904442884027958, 0.08502773940563202, -0.02243868261575699, 0.06951139122247696, 0.1421663463115692, 0.04281369224190712, 0.12626267969608307, -0.04015299677848816, -0.053521573543548584, 0.05445672571659088, -0.003400311805307865, 0.23117060959339142, 0.00403992086648941, 0.1413174569606781, 0.05910303443670273, 0.15940187871456146, -0.024757560342550278, 0.08927419036626816, -0.008729781024158001, -0.023835474625229836, -0.031669434159994125, -0.045303210616111755, -0.050093866884708405, 0.029769735410809517, -0.08303549140691757, 0.06033848598599434, -0.11797545105218887, 0.009352345950901508, 0.06100612133741379, 0.2679465711116791, 0.0345560684800148, -0.32827380299568176, -0.10064329206943512, 0.008748961612582207, -0.05120231956243515, -0.011571792885661125, 0.04322980344295502, 0.09254497289657593, -0.09684320539236069, 0.03802366554737091, -0.06117615848779678, 0.1040673702955246, -0.032205160707235336, 0.055158767849206924, 0.05844539776444435, 0.07644907385110855, 0.014066076837480068, 0.08924854546785355, -0.3133293688297272, 0.27166375517845154, -0.00529985036700964, 0.061156149953603745, -0.08480563014745712, 0.006723362952470779, 0.022355325520038605, 0.05982886627316475, 0.07312562316656113, -0.0009203396621160209, -0.013162265531718731, -0.16251851618289948, -0.04690636694431305, 0.032621774822473526, 0.08376285433769226, -0.04150301218032837, 0.099564328789711, -0.0398155115544796, 0.019869856536388397, 0.06857410818338394, 0.03692419081926346, -0.0494358129799366, -0.09151892364025116, -0.006282983813434839, 0.037052396684885025, -0.026015590876340866, -0.051273494958877563, -0.10026004910469055, -0.10645751655101776, 0.14177651703357697, -0.00937336403876543, -0.04987914860248566, -0.10378793627023697, 0.07152146846055984, 0.06711222976446152, -0.08635975420475006, 0.03958100080490112, 0.009345748461782932, 0.056557547301054, 0.025046154856681824, -0.07315059006214142, 0.1076507568359375, -0.07148158550262451, -0.16593849658966064, -0.05691763013601303, 0.11878407001495361, 0.02191450260579586, 0.06526581197977066, -0.010794282890856266, 0.013529124669730663, -0.061708517372608185, -0.08847595006227493, 0.02000255137681961, -0.0175459161400795, 0.08243745565414429, 0.02546462044119835, -0.05581970885396004, 0.029721811413764954, -0.06610950827598572, -0.056342680007219315, 0.20697271823883057, 0.24301345646381378, -0.08580175787210464, 0.032470010221004486, 0.027023056522011757, -0.068837471306324, -0.17428964376449585, -0.007821486331522465, 0.0537816546857357, -0.004708756692707539, 0.04488014802336693, -0.18280944228172302, 0.10610432922840118, 0.11062043905258179, -0.010413560084998608, 0.10061949491500854, -0.3606530725955963, -0.12448471784591675, 0.10286383330821991, 0.13927340507507324, 0.14426590502262115, -0.14637362957000732, -0.018606694415211678, -0.034801315516233444, -0.14949147403240204, 0.12675565481185913, -0.08541656285524368, 0.13169606029987335, -0.03354617953300476, 0.11654913425445557, 0.0013742247829213738, -0.04756172001361847, 0.10883041471242905, 0.02790527418255806, 0.08624210208654404, -0.06913415342569351, -0.0065808966755867004, 0.041602373123168945, -0.04126899689435959, 0.04147631675004959, -0.11524982750415802, 0.038154225796461105, -0.11252336204051971, -0.02098742127418518, -0.07133897393941879, 0.03417963534593582, -0.03256843239068985, -0.07596223801374435, -0.033484600484371185, 0.0017185518518090248, 0.07622073590755463, -0.010539303533732891, 0.13395829498767853, 0.02475128509104252, 0.14030829071998596, 0.11898857355117798, 0.06495541334152222, -0.07947299629449844, -0.06267054378986359, -0.03268807381391525, -0.014466951601207256, 0.046810995787382126, -0.1527949869632721, 0.025146836414933205, 0.13799728453159332, 0.017676809802651405, 0.15261851251125336, 0.0881342887878418, -0.01519074197858572, 0.010927753522992134, 0.05028318241238594, -0.17568272352218628, -0.09246961027383804, -0.01917967014014721, -0.051782358437776566, -0.10323657095432281, 0.05293538048863411, 0.1065724715590477, -0.07940518856048584, -0.006910948548465967, -0.02019619755446911, 0.02552017569541931, -0.05987841635942459, 0.17718112468719482, 0.03738729655742645, 0.04587649926543236, -0.09944941848516464, 0.08577782660722733, 0.04877568408846855, -0.05867059901356697, 0.017827987670898438, 0.09298241138458252, -0.08834537863731384, -0.04618917778134346, 0.0634305402636528, 0.1587466299533844, -0.07454291731119156, -0.041274286806583405, -0.13226769864559174, -0.12231069803237915, 0.0921829417347908, 0.14302995800971985, 0.10440896451473236, 0.018397869542241096, -0.0667494684457779, 0.004321410786360502, -0.10902354121208191, 0.09569721668958664, 0.03720599040389061, 0.05731917917728424, -0.12802813947200775, 0.14361049234867096, 0.006070116069167852, 0.043912265449762344, -0.02302122861146927, 0.020692996680736542, -0.08685092628002167, 0.017053775489330292, -0.11970220506191254, -0.02654484659433365, -0.018644243478775024, 0.001955503597855568, -0.013528920710086823, -0.052109573036432266, -0.05630413815379143, 0.02157507836818695, -0.11287444829940796, -0.025555286556482315, 0.024866845458745956, 0.06478238850831985, -0.11686132103204727, -0.04054051265120506, 0.02523687481880188, -0.06482114642858505, 0.08537165075540543, 0.06606142967939377, -0.004604043439030647, 0.056021224707365036, -0.1400790512561798, 0.01738286204636097, 0.06659705191850662, 0.028357578441500664, 0.056625593453645706, -0.08970612287521362, -0.015325277112424374, 0.00828730221837759, 0.04178221896290779, 0.022264990955591202, 0.07083622366189957, -0.13826629519462585, 0.00006531712278956547, -0.013711351901292801, -0.08758881688117981, -0.06387151777744293, 0.019696474075317383, 0.07149218767881393, 0.0156180988997221, 0.19836798310279846, -0.06972478330135345, 0.046642303466796875, -0.215156689286232, 0.003210781142115593, 0.005168197210878134, -0.11641579866409302, -0.13991892337799072, -0.07530112564563751, 0.05261170119047165, -0.055927574634552, 0.14479126036167145, 0.030283741652965546, 0.03275442123413086, 0.024011513218283653, 0.0007646132144145668, 0.010213825851678848, -0.0025430850218981504, 0.2231471985578537, 0.03413620591163635, -0.044169388711452484, 0.05171043798327446, 0.0395808108150959, 0.10996422916650772, 0.10972891747951508, 0.20767509937286377, 0.13877546787261963, 0.009574095718562603, 0.09246347099542618, 0.02337447553873062, -0.04925518110394478, -0.17311343550682068, 0.011703728698194027, -0.008440353907644749, 0.11860626935958862, -0.019692301750183105, 0.2287842035293579, 0.0688234269618988, -0.16370011866092682, 0.04337853193283081, -0.05432121828198433, -0.07331223785877228, -0.11230150610208511, -0.060768116265535355, -0.07695990055799484, -0.15051180124282837, -0.004323579370975494, -0.12489455938339233, 0.03795618191361427, 0.11710694432258606, 0.013127937912940979, -0.030370311811566353, 0.13228961825370789, 0.010669762268662453, 0.005048188380897045, 0.05020754039287567, -0.009179891087114811, -0.02401065267622471, -0.10865964740514755, -0.07153553515672684, -0.010136660188436508, -0.010626839473843575, 0.039866022765636444, -0.04398851469159126, -0.04849483072757721, 0.024330291897058487, -0.039402443915605545, -0.09445745497941971, 0.007615449372678995, 0.02206769399344921, 0.058850303292274475, 0.05333884432911873, 0.006736386101692915, 0.011601516976952553, 0.0004319900763221085, 0.23339533805847168, -0.08048217743635178, -0.09123951941728592, -0.08645922690629959, 0.24532751739025116, 0.04519444331526756, -0.016790809109807014, 0.03463512659072876, -0.059778597205877304, -0.014220603741705418, 0.24719657003879547, 0.20010711252689362, -0.08854307234287262, -0.018826495856046677, 0.0022698012180626392, -0.005705161020159721, -0.006043159868568182, 0.10799553990364075, 0.1433725357055664, 0.052917785942554474, -0.0872928649187088, -0.0075827608816325665, -0.05905787646770477, -0.0024047077167779207, -0.04484754055738449, 0.07206238061189651, 0.033971380442380905, -0.0029045299161225557, -0.027076639235019684, 0.055137984454631805, -0.07651779800653458, -0.08100073039531708, 0.017619969323277473, -0.20880794525146484, -0.14927539229393005, -0.008770769461989403, 0.10609032213687897, 0.010972407646477222, 0.06605372577905655, -0.025835761800408363, 0.007030332926660776, 0.08267879486083984, -0.01873067580163479, -0.09580681473016739, -0.071219302713871, 0.10465652495622635, -0.12693268060684204, 0.19960369169712067, -0.041105713695287704, 0.06797536462545395, 0.11908379942178726, 0.05847109109163284, -0.07419919222593307, 0.08181881159543991, 0.04547201842069626, -0.026650013402104378, 0.02455892227590084, 0.07806308567523956, -0.031666651368141174, 0.06686554104089737, 0.0477120541036129, -0.12801969051361084, 0.004760604351758957, -0.003558507189154625, -0.05070265755057335, -0.031327664852142334, -0.05067463964223862, -0.05734685808420181, 0.12692934274673462, 0.2120850682258606, -0.04573134705424309, 0.0010185821447521448, -0.08538109064102173, 0.00878242775797844, 0.060667600482702255, 0.015562833286821842, -0.05436599999666214, -0.2207716405391693, -0.005116496235132217, 0.09205278009176254, -0.01434277929365635, -0.2479126900434494, -0.09090379625558853, -0.006911981385201216, -0.06366978585720062, -0.1099991425871849, 0.07806404680013657, 0.08916788548231125, 0.040905892848968506, -0.05276968330144882, -0.05579075217247009, -0.07620201259851456, 0.1575630009174347, -0.13635453581809998, -0.09170141071081161 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-cased-wikitext2 This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 6.8575 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 7.0964 | 1.0 | 2346 | 7.0532 | | 6.9055 | 2.0 | 4692 | 6.8710 | | 6.8574 | 3.0 | 7038 | 6.8917 | ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.0
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "bert-base-cased-wikitext2", "results": []}]}
fill-mask
fznmhmmd/bert-base-cased-wikitext2
[ "transformers", "pytorch", "tensorboard", "bert", "fill-mask", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #bert #fill-mask #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
bert-base-cased-wikitext2 ========================= This model is a fine-tuned version of bert-base-cased on the None dataset. It achieves the following results on the evaluation set: * Loss: 6.8575 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.16.2 * Pytorch 1.10.0+cu111 * Datasets 1.18.3 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #fill-mask #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ 55, 98, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #fill-mask #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ -0.11804281920194626, 0.04380122572183609, -0.0019370403606444597, 0.12602244317531586, 0.1647389829158783, 0.028444208204746246, 0.11101580411195755, 0.1190485879778862, -0.0952233076095581, 0.024750210344791412, 0.14073480665683746, 0.17312008142471313, 0.010017718188464642, 0.13018493354320526, -0.034064289182424545, -0.23650521039962769, -0.0074819233268499374, 0.04061378538608551, -0.08609285950660706, 0.13523592054843903, 0.08519299328327179, -0.13556796312332153, 0.0778646320104599, 0.015889624133706093, -0.2121465653181076, 0.015155384317040443, 0.024733226746320724, -0.062431223690509796, 0.14757393300533295, 0.00335792638361454, 0.13839960098266602, -0.0034233410842716694, 0.0813794881105423, -0.15107986330986023, 0.014457213692367077, 0.05730382725596428, 0.004481148906052113, 0.08171287178993225, 0.042253416031599045, 0.0016652600606903434, 0.10222579538822174, -0.09761053323745728, 0.05542087182402611, 0.01701461337506771, -0.1240006759762764, -0.2341778725385666, -0.08821330219507217, 0.004965659696608782, 0.06559538096189499, 0.10977016389369965, 0.007444808259606361, 0.15547573566436768, -0.09247920662164688, 0.08825135231018066, 0.2591892182826996, -0.305569052696228, -0.06883496046066284, 0.012846847996115685, 0.015679296106100082, 0.04391634464263916, -0.09923990815877914, -0.015577412210404873, 0.04645175114274025, 0.054055750370025635, 0.1474062204360962, -0.040628183633089066, -0.0998924970626831, 0.014100217260420322, -0.1366090625524521, -0.0343216210603714, 0.09488831460475922, 0.025102952495217323, -0.039751701056957245, -0.03305305540561676, -0.06800784915685654, -0.16057835519313812, -0.044221751391887665, -0.006320864427834749, 0.04308323189616203, -0.043970927596092224, -0.08010801672935486, 0.0007154530612751842, -0.10511204600334167, -0.07497060298919678, -0.080658920109272, 0.16750821471214294, 0.039139170199632645, 0.029620232060551643, -0.03474011272192001, 0.10190372169017792, -0.007340844254940748, -0.14511996507644653, 0.025757571682333946, 0.038713037967681885, -0.008619643747806549, -0.030617808923125267, -0.0773618072271347, -0.08853942900896072, 0.015680156648159027, 0.11308540403842926, -0.04697342589497566, 0.04710708558559418, 0.04681159555912018, 0.0488470196723938, -0.11703190207481384, 0.18604300916194916, -0.04334332421422005, -0.008586817421019077, 0.016930824145674706, 0.03946799039840698, 0.026341520249843597, -0.008955956436693668, -0.10834125429391861, 0.0019009934039786458, 0.08616573363542557, 0.012726632878184319, -0.06344529241323471, 0.06253841519355774, -0.05766166001558304, -0.010655421763658524, 0.018061529844999313, -0.1011914536356926, 0.033509593456983566, -0.01454149093478918, -0.07106493413448334, -0.033047154545784, 0.0390450619161129, 0.011514749377965927, -0.014845089986920357, 0.1278228610754013, -0.08489412814378738, 0.03719111531972885, -0.11077708005905151, -0.1078823059797287, 0.009475366212427616, -0.09270947426557541, 0.021988440304994583, -0.09683303534984589, -0.1663757860660553, -0.002327755792066455, 0.07262835651636124, -0.02430587448179722, -0.04477836564183235, -0.020375577732920647, -0.06984473019838333, 0.010257979854941368, -0.011880878359079361, 0.1779651790857315, -0.058230768889188766, 0.10888061672449112, 0.04771232604980469, 0.08574418723583221, -0.05826292186975479, 0.050699230283498764, -0.09026525169610977, 0.008492072112858295, -0.2014256864786148, 0.01604226790368557, -0.03955468162894249, 0.06804287433624268, -0.08638668805360794, -0.10679927468299866, -0.003103539114817977, -0.007560164667665958, 0.0893423855304718, 0.09201755374670029, -0.17207470536231995, -0.07306616753339767, 0.1679992526769638, -0.06376945227384567, -0.10960184037685394, 0.11921048164367676, -0.05405937135219574, 0.0326697863638401, 0.057351112365722656, 0.1262178122997284, 0.061294637620449066, -0.11125387251377106, 0.03707810118794441, -0.0023384084925055504, 0.04175690561532974, -0.08625709265470505, 0.06625509262084961, -0.0063677942380309105, -0.005535550881177187, 0.03430755436420441, -0.021161379292607307, 0.0584605447947979, -0.09160876274108887, -0.10170692950487137, -0.04694350063800812, -0.10778844356536865, 0.059680476784706116, 0.06975018233060837, 0.08225454390048981, -0.10446668416261673, -0.08672603964805603, 0.02631298080086708, 0.0665205717086792, -0.043557167053222656, 0.0355561338365078, -0.062209006398916245, 0.06539351493120193, -0.06189906597137451, -0.02852027118206024, -0.18891644477844238, -0.012556396424770355, 0.0028196140192449093, -0.01227441243827343, 0.010359530337154865, 0.006348817143589258, 0.0871533676981926, 0.0605459026992321, -0.05477818101644516, -0.008355767466127872, -0.03250564634799957, -0.009513906203210354, -0.13642925024032593, -0.19620776176452637, -0.03996844217181206, -0.02358425036072731, 0.10635720193386078, -0.162708580493927, 0.024271681904792786, -0.051287177950143814, 0.06569698452949524, 0.0016680000117048621, -0.01215280219912529, -0.05214974656701088, 0.08933845907449722, -0.016516102477908134, -0.05054641142487526, 0.07193320989608765, 0.004557866603136063, -0.08591938018798828, -0.04080183058977127, -0.08731227368116379, 0.18515987694263458, 0.13819725811481476, -0.1157020777463913, -0.08300478756427765, 0.03269633650779724, -0.0642930343747139, -0.03280108422040939, -0.04752422124147415, 0.04930582642555237, 0.1819184422492981, -0.0034518877509981394, 0.14169350266456604, -0.06293953955173492, -0.04209220036864281, 0.03977322205901146, -0.03436525911092758, 0.03195391595363617, 0.09114322811365128, 0.13092485070228577, -0.05141114071011543, 0.12811914086341858, 0.16612187027931213, -0.11540187895298004, 0.11540087312459946, -0.03231939300894737, -0.07831571996212006, -0.01557270810008049, -0.019823187962174416, 0.012511294335126877, 0.11796671897172928, -0.1368808001279831, -0.007232255768030882, 0.023025233298540115, 0.007109243422746658, 0.02124127745628357, -0.23482745885849, -0.04562273249030113, 0.03253849595785141, -0.03506677597761154, -0.0308355912566185, -0.009288652800023556, 0.004121100530028343, 0.09963531792163849, 0.005722259636968374, -0.08233589679002762, 0.045467376708984375, 0.00776287168264389, -0.07047373801469803, 0.2145027220249176, -0.08141524344682693, -0.15361367166042328, -0.11945625394582748, -0.0890214592218399, -0.036605097353458405, 0.00962420180439949, 0.06247447058558464, -0.09549237042665482, -0.03673836961388588, -0.04533272981643677, -0.0039050867781043053, 0.011339972727000713, 0.059349264949560165, 0.012408195063471794, -0.01641847752034664, 0.0882515162229538, -0.10582064837217331, -0.01509937085211277, -0.050021667033433914, -0.06542141735553741, 0.05520399659872055, 0.055143680423498154, 0.12207236886024475, 0.14256060123443604, -0.020432937890291214, 0.008855195716023445, -0.019681021571159363, 0.22348535060882568, -0.0640089362859726, -0.035934846848249435, 0.13654841482639313, -0.0052816555835306644, 0.06255381554365158, 0.0997561514377594, 0.07310500741004944, -0.08551324903964996, 0.005988655611872673, 0.027352115139365196, -0.050158288329839706, -0.2082744836807251, -0.03369887173175812, -0.06809115409851074, -0.057437099516391754, 0.09784254431724548, 0.03157193586230278, 0.04112813249230385, 0.07434644550085068, 0.050612013787031174, 0.08679576963186264, -0.06646223366260529, 0.046455904841423035, 0.0679221898317337, 0.05484931543469429, 0.12578162550926208, -0.044067587703466415, -0.07012410461902618, 0.02745419181883335, -0.01776811107993126, 0.22902807593345642, 0.0005660378374159336, 0.11708169430494308, 0.0678674578666687, 0.20849977433681488, -0.005598689429461956, 0.10808858275413513, -0.007765171583741903, -0.05905207246541977, -0.008160042576491833, -0.04880543798208237, -0.0343928299844265, 0.0135470200330019, -0.05212332680821419, 0.06555550545454025, -0.11096290498971939, -0.0061170319095253944, 0.03900687396526337, 0.2810724079608917, 0.03953344374895096, -0.3332221806049347, -0.07627875357866287, -0.011782193556427956, -0.017931954935193062, -0.020283188670873642, 0.005273426417261362, 0.08731384575366974, -0.08936777710914612, 0.0406225211918354, -0.07628025114536285, 0.08893841505050659, 0.00800961721688509, 0.042640719562768936, 0.07640742510557175, 0.1029590368270874, 0.015578928403556347, 0.06712612509727478, -0.31981998682022095, 0.2913988530635834, 0.004939526319503784, 0.08948660641908646, -0.09562091529369354, 0.008863426744937897, 0.041147198528051376, 0.03126874566078186, 0.07330508530139923, -0.015202529728412628, -0.01870608516037464, -0.1828000694513321, -0.05549265071749687, 0.03685615956783295, 0.08337903767824173, -0.00923593994230032, 0.08993825316429138, -0.017643718048930168, -0.006193134002387524, 0.07317669689655304, 0.030204059556126595, -0.05849152430891991, -0.09049216657876968, -0.0076278806664049625, 0.029768861830234528, -0.08801275491714478, -0.06909743696451187, -0.116856150329113, -0.12983949482440948, 0.1643747091293335, 0.02071698009967804, -0.029401935636997223, -0.11805640906095505, 0.07572171837091446, 0.08734524995088577, -0.08806941658258438, 0.059352997690439224, 0.002371825510635972, 0.05785829946398735, 0.02116193249821663, -0.07942129671573639, 0.1130242794752121, -0.07029989361763, -0.13835594058036804, -0.06968587636947632, 0.09459025412797928, 0.028792887926101685, 0.06611110270023346, -0.013058412820100784, 0.023887798190116882, -0.040874578058719635, -0.083754763007164, 0.04066748544573784, -0.04222496598958969, 0.08363120257854462, 0.014347530901432037, -0.04581888020038605, 0.012429156340658665, -0.05032961815595627, -0.024159932509064674, 0.17651064693927765, 0.22368252277374268, -0.10330293327569962, 0.02648788131773472, 0.029956072568893433, -0.05171600729227066, -0.20776796340942383, 0.035785093903541565, 0.0635417252779007, 0.017508529126644135, 0.051163073629140854, -0.16679999232292175, 0.13864223659038544, 0.09279060363769531, -0.014938496053218842, 0.12802423536777496, -0.31889188289642334, -0.13062649965286255, 0.1302788108587265, 0.1595543771982193, 0.16295458376407623, -0.14142939448356628, -0.016994241625070572, -0.028927326202392578, -0.13126100599765778, 0.06145376339554787, -0.09851256757974625, 0.12173397839069366, -0.03613658994436264, 0.09119261056184769, -0.0032316073775291443, -0.07603484392166138, 0.1200607419013977, 0.0026890027802437544, 0.09091728925704956, -0.060595400631427765, -0.019805310294032097, 0.05042818933725357, -0.03583792969584465, 0.006433306261897087, -0.0863368809223175, 0.024987168610095978, -0.05785362049937248, -0.014613762497901917, -0.0850311815738678, 0.053535666316747665, -0.028109349310398102, -0.05022181198000908, -0.021885525435209274, 0.014819792471826077, 0.04452717676758766, -0.021614842116832733, 0.11506634950637817, 0.03247493878006935, 0.15955336391925812, 0.09705106168985367, 0.04334186762571335, -0.07853426039218903, -0.09041804075241089, -0.010989396832883358, -0.020766466856002808, 0.06552372127771378, -0.12402325868606567, 0.023533865809440613, 0.1251315027475357, 0.028368091210722923, 0.11196838319301605, 0.08565036207437515, -0.0286878552287817, 0.01722603850066662, 0.07643895596265793, -0.16210436820983887, -0.05568522587418556, 0.011629904620349407, -0.06269457936286926, -0.113002710044384, 0.044576551765203476, 0.07234475761651993, -0.06154422089457512, -0.009677132591605186, -0.015435127541422844, 0.0007478023180738091, -0.08194349706172943, 0.21394851803779602, 0.05317794531583786, 0.050266481935977936, -0.10756353288888931, 0.04959588870406151, 0.03915925323963165, -0.07637780159711838, -0.0021641452331095934, 0.06535358726978302, -0.07612822204828262, -0.03620601445436478, 0.1295357495546341, 0.17615853250026703, -0.031484924256801605, -0.04254510626196861, -0.1443806141614914, -0.11752092838287354, 0.07547447085380554, 0.16355840861797333, 0.11046659201383591, 0.0028174228500574827, -0.05023225396871567, 0.01306987926363945, -0.11479470878839493, 0.06534779816865921, 0.04548530653119087, 0.0716017335653305, -0.12297023832798004, 0.16153503954410553, 0.010421526618301868, 0.06235371530056, -0.023094667121767998, 0.03876030445098877, -0.0914972722530365, 0.01864350400865078, -0.12651784718036652, -0.02886820398271084, -0.023397646844387054, -0.011946108192205429, -0.005664207972586155, -0.057543542236089706, -0.06526517122983932, 0.024624457582831383, -0.12388962507247925, -0.03429226577281952, 0.0382060669362545, 0.028826672583818436, -0.11728166043758392, -0.0423617959022522, 0.03032313846051693, -0.06075683608651161, 0.047325145453214645, 0.06299196928739548, 0.012214184738695621, 0.06433548778295517, -0.1431974321603775, -0.018033470958471298, 0.07079382240772247, 0.008472702465951443, 0.07716771960258484, -0.07865816354751587, -0.010133996605873108, 0.003768827999010682, 0.07078733295202255, 0.011619948782026768, 0.08592028170824051, -0.14898467063903809, 0.0007522878004238009, -0.024211201816797256, -0.08870268613100052, -0.060016121715307236, 0.015431396663188934, 0.09363534301519394, 0.009739382192492485, 0.1969122737646103, -0.08996067196130753, 0.04933410510420799, -0.2061324417591095, 0.0003733888443093747, -0.02574039064347744, -0.09418626129627228, -0.11384168267250061, -0.049209024757146835, 0.07158297300338745, -0.05119439586997032, 0.13430117070674896, 0.02141626365482807, 0.049019139260053635, 0.022209780290722847, -0.019015496596693993, 0.018226729705929756, 0.015237372368574142, 0.2129012495279312, 0.030207302421331406, -0.03816622495651245, 0.07378116995096207, 0.06436997652053833, 0.09579547494649887, 0.12657169997692108, 0.21515780687332153, 0.15264743566513062, 0.03723854199051857, 0.106281578540802, 0.01692812144756317, -0.049311041831970215, -0.15166856348514557, 0.014028948731720448, -0.056521013379096985, 0.09805663675069809, -0.013886705972254276, 0.19419294595718384, 0.06898408383131027, -0.1684456616640091, 0.05461963638663292, -0.048928286880254745, -0.0844298005104065, -0.11320801079273224, -0.04675600677728653, -0.08085701614618301, -0.123316191136837, 0.007435597013682127, -0.08439688384532928, 0.020802989602088928, 0.11734781414270401, -0.0003254633629694581, -0.015211940743029118, 0.192257359623909, 0.02587692253291607, 0.03999999538064003, 0.0424739271402359, 0.009135142900049686, -0.022089680656790733, -0.07023867219686508, -0.05801369994878769, -0.027151048183441162, -0.018985049799084663, 0.04132065922021866, -0.06869017332792282, -0.07762522995471954, 0.054636355489492416, -0.011190407909452915, -0.10214953124523163, 0.012651550583541393, 0.011565843597054482, 0.07234829664230347, 0.05156566575169563, 0.013867085799574852, 0.030351907014846802, -0.024509122595191002, 0.18758414685726166, -0.0858890563249588, -0.0925723984837532, -0.09777165204286575, 0.25701993703842163, 0.04113214835524559, -0.021381262689828873, 0.024440934881567955, -0.058231059461832047, -0.002263482194393873, 0.26586535573005676, 0.2162809669971466, -0.08462007343769073, -0.00022664129210170358, 0.007978697307407856, -0.013263524509966373, -0.0356750525534153, 0.1258949339389801, 0.13940764963626862, 0.06719311326742172, -0.10296238213777542, -0.05220576748251915, -0.06273677945137024, -0.012972511351108551, -0.07131515443325043, 0.045457687228918076, 0.03874102607369423, 0.0026129225734621286, -0.03687943518161774, 0.04821845144033432, -0.04179731011390686, -0.10496954619884491, 0.09342876076698303, -0.1972687542438507, -0.16262021660804749, -0.008009281940758228, 0.11053233593702316, 0.005408870987594128, 0.07233873009681702, -0.027352599427103996, 0.007825185544788837, 0.06908247619867325, -0.01912372000515461, -0.08321236073970795, -0.09353167563676834, 0.10173162817955017, -0.111968494951725, 0.21642418205738068, -0.03633418679237366, 0.06745132058858871, 0.12475226074457169, 0.08215309679508209, -0.07225337624549866, 0.06257951259613037, 0.03351308032870293, -0.09485068917274475, 0.02828112617135048, 0.09288430958986282, -0.037087131291627884, 0.012329824268817902, 0.027008742094039917, -0.10421109944581985, 0.027711138129234314, -0.08152216672897339, -0.032381001859903336, -0.03579189255833626, -0.03806816413998604, -0.06333442777395248, 0.11839384585618973, 0.21559050679206848, -0.01668320596218109, 0.013468259014189243, -0.07872966676950455, 0.014305299147963524, 0.06043969839811325, 0.0177000779658556, -0.1041441336274147, -0.21887429058551788, 0.020203853026032448, 0.045899685472249985, -0.03186710923910141, -0.24231094121932983, -0.10092182457447052, 0.004197039641439915, -0.0861874520778656, -0.08961018919944763, 0.06149911507964134, 0.07163489609956741, 0.058766067028045654, -0.049371760338544846, -0.084843210875988, -0.07383713871240616, 0.1494603157043457, -0.16338983178138733, -0.09480001777410507 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-cola This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.8273 - Matthews Correlation: 0.5544 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.5256 | 1.0 | 535 | 0.5419 | 0.4248 | | 0.3486 | 2.0 | 1070 | 0.5187 | 0.4999 | | 0.2406 | 3.0 | 1605 | 0.6580 | 0.5054 | | 0.1692 | 4.0 | 2140 | 0.7455 | 0.5403 | | 0.1343 | 5.0 | 2675 | 0.8273 | 0.5544 | ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.0
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["glue"], "metrics": ["matthews_correlation"], "model-index": [{"name": "distilbert-base-uncased-finetuned-cola", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "glue", "type": "glue", "args": "cola"}, "metrics": [{"type": "matthews_correlation", "value": 0.5543972545286807, "name": "Matthews Correlation"}]}]}]}
text-classification
fznmhmmd/distilbert-base-uncased-finetuned-cola
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "dataset:glue", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-cola ====================================== This model is a fine-tuned version of distilbert-base-uncased on the glue dataset. It achieves the following results on the evaluation set: * Loss: 0.8273 * Matthews Correlation: 0.5544 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 5 ### Training results ### Framework versions * Transformers 4.16.2 * Pytorch 1.10.0+cu111 * Datasets 1.18.3 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ 67, 98, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ -0.10348068922758102, 0.09148605167865753, -0.0022954836022108793, 0.12302655726671219, 0.16473764181137085, 0.029876047745347023, 0.11664267629384995, 0.1302110105752945, -0.0878436341881752, 0.02543756179511547, 0.12662576138973236, 0.16262155771255493, 0.020631218329072, 0.12087254971265793, -0.05344507843255997, -0.26481878757476807, -0.00954040139913559, 0.05130685120820999, -0.04174657538533211, 0.13426357507705688, 0.09329058229923248, -0.1231277585029602, 0.09115666896104813, 0.013420876115560532, -0.19516946375370026, 0.00201223767362535, -0.0007724217721261084, -0.05339557304978371, 0.14357911050319672, 0.02452739141881466, 0.11937303096055984, 0.0003319440584164113, 0.08145023137331009, -0.1945306956768036, 0.009916625916957855, 0.046701375395059586, 0.0018440545536577702, 0.09148869663476944, 0.04440341889858246, -0.0011416783090680838, 0.12547406554222107, -0.0883527398109436, 0.05178067460656166, 0.021715596318244934, -0.11551861464977264, -0.2127065807580948, -0.08081891387701035, 0.0405232310295105, 0.08268959075212479, 0.10872946679592133, -0.007196844555437565, 0.12247898429632187, -0.07745738327503204, 0.09489422291517258, 0.23106767237186432, -0.2972404956817627, -0.0659942477941513, 0.03629834204912186, 0.013412061147391796, 0.03745477646589279, -0.10173027217388153, -0.0377054437994957, 0.05005799978971481, 0.05379036068916321, 0.12692078948020935, -0.03019030950963497, -0.11055943369865417, 0.002316690282896161, -0.1385059505701065, -0.03447825089097023, 0.16383424401283264, 0.0412207655608654, -0.0308991689234972, -0.06073944643139839, -0.05710354447364807, -0.1502271145582199, -0.03864529728889465, -0.010285157710313797, 0.046585649251937866, -0.02382180094718933, -0.04229508712887764, -0.004785689990967512, -0.11021022498607635, -0.05849221721291542, -0.07634159922599792, 0.11212288588285446, 0.03714621812105179, 0.00965169258415699, -0.03245272859930992, 0.10942050069570541, -0.006381617393344641, -0.12516553699970245, 0.014764266088604927, 0.021977951750159264, 0.016854027286171913, -0.04099957272410393, -0.05396966263651848, -0.06450658291578293, 0.007764738518744707, 0.1275479942560196, -0.05921218544244766, 0.045907724648714066, 0.046832021325826645, 0.047279953956604004, -0.0910499319434166, 0.19001732766628265, -0.03149630129337311, -0.02014079876244068, 0.009288782253861427, 0.040842894464731216, 0.019353672862052917, -0.008877582848072052, -0.12351246923208237, 0.004562444984912872, 0.08937069028615952, 0.007729888428002596, -0.06502804160118103, 0.077152319252491, -0.054169218987226486, -0.021537424996495247, 0.004974089562892914, -0.0925840437412262, 0.026125459000468254, -0.0023696627467870712, -0.07028466463088989, -0.022876888513565063, 0.03493975102901459, 0.014959706924855709, -0.02497624047100544, 0.11131783574819565, -0.08554515242576599, 0.029842987656593323, -0.09330154955387115, -0.10611137002706528, 0.02050638571381569, -0.10477912425994873, 0.026218537241220474, -0.09508968144655228, -0.18332253396511078, -0.017160901799798012, 0.062236882746219635, -0.026304490864276886, -0.059443339705467224, -0.055562593042850494, -0.06969424337148666, 0.014923201873898506, -0.010747545398771763, 0.1182631254196167, -0.0635029524564743, 0.08913702517747879, 0.026541225612163544, 0.06336144357919693, -0.04123564437031746, 0.056380730122327805, -0.1029263436794281, 0.016929831355810165, -0.15925705432891846, 0.03990291431546211, -0.04568052291870117, 0.07734546065330505, -0.08471763879060745, -0.10755743831396103, 0.0045159063301980495, -0.004773437511175871, 0.06634347140789032, 0.09433982521295547, -0.17825455963611603, -0.07311838120222092, 0.1635434776544571, -0.07277669757604599, -0.12628166377544403, 0.11865793913602829, -0.05997232720255852, 0.05510609224438667, 0.05736212432384491, 0.17833933234214783, 0.0720689445734024, -0.08016322553157806, -0.003538931719958782, 0.022151730954647064, 0.046557385474443436, -0.07095295190811157, 0.06665494292974472, 0.009641564451158047, 0.02324853278696537, 0.03641578182578087, -0.02319316193461418, 0.06043113395571709, -0.08411320298910141, -0.094823457300663, -0.043874356895685196, -0.07993423193693161, 0.03350040689110756, 0.0764167308807373, 0.07151050120592117, -0.09737885743379593, -0.0809127688407898, 0.04480326920747757, 0.0776895210146904, -0.05769297108054161, 0.029527835547924042, -0.05515516549348831, 0.07536843419075012, -0.029270611703395844, -0.02229178138077259, -0.18096449971199036, -0.029233552515506744, 0.005905236583203077, 0.00461921002715826, 0.011473044753074646, 0.02054179273545742, 0.062115442007780075, 0.053953222930431366, -0.04901926591992378, -0.016520880162715912, -0.021534211933612823, -0.0016557901399210095, -0.12969975173473358, -0.1899591088294983, -0.03136550262570381, -0.024618536233901978, 0.1403314471244812, -0.206315815448761, 0.04675181582570076, -0.010046504437923431, 0.071916364133358, 0.00943793635815382, -0.005788105074316263, -0.037703968584537506, 0.06980443000793457, -0.045524198561906815, -0.05132804438471794, 0.08034015446901321, 0.021152136847376823, -0.08919230103492737, -0.042948246002197266, -0.0944969430565834, 0.15587735176086426, 0.12980902194976807, -0.10666801035404205, -0.07886876165866852, -0.01752552203834057, -0.06929495185613632, -0.03253386914730072, -0.046629101037979126, 0.03089221939444542, 0.19334089756011963, -0.0057763331569731236, 0.15134920179843903, -0.07007114589214325, -0.049630433320999146, 0.022067880257964134, -0.03617172688245773, 0.013486909680068493, 0.12855865061283112, 0.13056056201457977, -0.07010336965322495, 0.15201197564601898, 0.14409585297107697, -0.08407939225435257, 0.13985194265842438, -0.040423717349767685, -0.06506288051605225, -0.015965327620506287, -0.02710880897939205, -0.010020623914897442, 0.09792278707027435, -0.15398730337619781, -0.002388850785791874, 0.03264133632183075, 0.019953329116106033, 0.02561068721115589, -0.22168494760990143, -0.03851085156202316, 0.03620898723602295, -0.03936419263482094, -0.010801783762872219, -0.006293339654803276, 0.005123737268149853, 0.099776990711689, 0.004696644842624664, -0.08376874029636383, 0.04186678305268288, 0.004644679371267557, -0.08690396696329117, 0.2162937968969345, -0.07906312495470047, -0.17329132556915283, -0.12459393590688705, -0.07870804518461227, -0.043099645525217056, -0.0006197145557962358, 0.07219891995191574, -0.09086521714925766, -0.03347395732998848, -0.07394793629646301, 0.016268594190478325, 0.004127702210098505, 0.02735956944525242, 0.01253552921116352, 0.004940738435834646, 0.06401343643665314, -0.1063295230269432, -0.019773203879594803, -0.05651792883872986, -0.03786171227693558, 0.0446893535554409, 0.027988433837890625, 0.10908740013837814, 0.1475294679403305, -0.01488849613815546, 0.016142575070261955, -0.03245636075735092, 0.236946702003479, -0.05981217697262764, -0.020328663289546967, 0.1417233794927597, -0.009224441833794117, 0.05207998678088188, 0.12015313655138016, 0.06853050738573074, -0.08067213743925095, 0.005674567073583603, 0.03243220970034599, -0.038105472922325134, -0.23070302605628967, -0.05341356247663498, -0.05923246592283249, 0.007216841448098421, 0.0935586467385292, 0.026574047282338142, 0.027094939723610878, 0.07268910109996796, 0.04202091693878174, 0.07887616008520126, -0.039956413209438324, 0.06022903323173523, 0.1276325136423111, 0.03587205708026886, 0.12549570202827454, -0.0485922247171402, -0.06260202825069427, 0.04125089943408966, -0.012665070593357086, 0.22441771626472473, 0.0036260392516851425, 0.12858140468597412, 0.060763128101825714, 0.16073182225227356, -0.006428951397538185, 0.08137209713459015, -0.013494585640728474, -0.04077665135264397, -0.015004271641373634, -0.036497749388217926, -0.03784942254424095, 0.027807995676994324, -0.07401497662067413, 0.060699883848428726, -0.12397698312997818, 0.01996294967830181, 0.060566410422325134, 0.2576342821121216, 0.03713158145546913, -0.323432058095932, -0.10025621950626373, 0.00282878614962101, -0.03632260859012604, -0.025572864338755608, 0.029513590037822723, 0.08909446001052856, -0.09776820987462997, 0.03476474806666374, -0.07485097646713257, 0.09763772040605545, -0.04730924963951111, 0.04995490610599518, 0.08224654942750931, 0.08347439765930176, 0.012666335329413414, 0.09324385225772858, -0.2918814420700073, 0.2761648893356323, -0.00007573797483928502, 0.0620449036359787, -0.08132880181074142, 0.010391105897724628, 0.03727884590625763, 0.06091126799583435, 0.09131623059511185, -0.012225004844367504, -0.04733143001794815, -0.17605245113372803, -0.06611157208681107, 0.02757733128964901, 0.06398440152406693, -0.030406340956687927, 0.08697547018527985, -0.03065989352762699, 0.007524167653173208, 0.06976216286420822, 0.005401842761784792, -0.045676689594984055, -0.10834278911352158, -0.00778288533911109, 0.023803863674402237, -0.06611955165863037, -0.059602025896310806, -0.11702249944210052, -0.12276234477758408, 0.16779576241970062, -0.028637932613492012, -0.04264632984995842, -0.10973475873470306, 0.08625014871358871, 0.055805761367082596, -0.09044497460126877, 0.04301720857620239, 0.002081508282572031, 0.08180484920740128, 0.020375272259116173, -0.07547292858362198, 0.10499080270528793, -0.0771421492099762, -0.15239882469177246, -0.06422466784715652, 0.10794258862733841, 0.02980002574622631, 0.06435586512088776, -0.008842872455716133, 0.00901352521032095, -0.0486224964261055, -0.08938624709844589, 0.019282618537545204, -0.00013774285616818815, 0.08739374577999115, 0.00877909641712904, -0.07083219289779663, 0.012592953629791737, -0.0597652792930603, -0.032124925404787064, 0.2084427773952484, 0.2118198424577713, -0.10192637890577316, 0.026571858674287796, 0.01737801730632782, -0.07349719107151031, -0.19828394055366516, 0.031650908291339874, 0.05889834091067314, 0.009060108102858067, 0.035099007189273834, -0.1763526350259781, 0.13956797122955322, 0.102257639169693, -0.01299660187214613, 0.10153043270111084, -0.31127604842185974, -0.12243517488241196, 0.13350917398929596, 0.13117966055870056, 0.1107979342341423, -0.1280149519443512, -0.019227130338549614, -0.02045135386288166, -0.14110441505908966, 0.11007409542798996, -0.08445243537425995, 0.11649177223443985, -0.03637200966477394, 0.0832626223564148, 0.002716791583225131, -0.06038431450724602, 0.11320134252309799, 0.03001723624765873, 0.09173953533172607, -0.06272721290588379, -0.03565295785665512, 0.02873510681092739, -0.04839441180229187, 0.04129466414451599, -0.09495443850755692, 0.029764125123620033, -0.11069756001234055, -0.027970068156719208, -0.06641346961259842, 0.04448976367712021, -0.042088717222213745, -0.06358592212200165, -0.03718508407473564, 0.023409102112054825, 0.05508594214916229, -0.007334549445658922, 0.13400959968566895, 0.026072608307003975, 0.1427386850118637, 0.0996645838022232, 0.07817801833152771, -0.07685045152902603, -0.07546738535165787, -0.024036915972828865, -0.011588722467422485, 0.04755742475390434, -0.14280779659748077, 0.02313205972313881, 0.15387830138206482, 0.01909896917641163, 0.14599280059337616, 0.08460244536399841, -0.019058987498283386, 0.001374579151161015, 0.059949882328510284, -0.16713014245033264, -0.0878666341304779, -0.013119746930897236, -0.06208950653672218, -0.1194811686873436, 0.043869271874427795, 0.09051049500703812, -0.06727153062820435, -0.009456361643970013, -0.006581311114132404, 0.01648966409265995, -0.05074736848473549, 0.17930923402309418, 0.056837521493434906, 0.046653520315885544, -0.09937667101621628, 0.06777936965227127, 0.04454825818538666, -0.07685873657464981, 0.011012088507413864, 0.07821232080459595, -0.08650293946266174, -0.0552859865128994, 0.06975916773080826, 0.18555298447608948, -0.04588611051440239, -0.04852474480867386, -0.1402791440486908, -0.12592436373233795, 0.08435630053281784, 0.14124564826488495, 0.11801300942897797, 0.009990457445383072, -0.06496996432542801, 0.0005833751056343317, -0.11324189603328705, 0.10048721730709076, 0.04907653108239174, 0.06291599571704865, -0.14508862793445587, 0.13627180457115173, 0.014503736048936844, 0.05412086844444275, -0.018928060308098793, 0.02754272148013115, -0.0992475226521492, 0.008631311357021332, -0.09962839633226395, -0.01056043989956379, -0.03920818120241165, 0.00988402683287859, -0.005131166893988848, -0.04514012113213539, -0.058122653514146805, 0.014701725915074348, -0.10730111598968506, -0.022076278924942017, 0.02723095752298832, 0.06627681106328964, -0.10553426295518875, -0.036801356822252274, 0.02543506771326065, -0.0630459114909172, 0.0752045214176178, 0.04854367673397064, 0.014264537021517754, 0.04823144152760506, -0.13089507818222046, 0.01804414764046669, 0.07707779109477997, 0.02609756588935852, 0.06487300246953964, -0.1012471541762352, -0.005991428624838591, 0.0008756459574215114, 0.0364258699119091, 0.02079150080680847, 0.07286219298839569, -0.14232607185840607, 0.0032555432990193367, -0.019487107172608376, -0.08063076436519623, -0.06793943792581558, 0.02705864980816841, 0.08977404981851578, 0.017028121277689934, 0.19965432584285736, -0.07695881277322769, 0.048752423375844955, -0.21314099431037903, 0.006563444156199694, -0.007979915477335453, -0.10863091051578522, -0.09884380549192429, -0.06806447356939316, 0.056490324437618256, -0.05831785127520561, 0.1521150767803192, 0.04621584340929985, 0.02097199112176895, 0.027592875063419342, -0.014663904905319214, 0.013873129151761532, 0.012499130330979824, 0.18968160450458527, 0.02315196767449379, -0.03892238438129425, 0.060869134962558746, 0.04344213008880615, 0.10514671355485916, 0.11975841224193573, 0.20499777793884277, 0.14335976541042328, 0.0007634255453012884, 0.09935532510280609, 0.03613463044166565, -0.053582388907670975, -0.16469132900238037, 0.04401427507400513, -0.039850927889347076, 0.11092086881399155, -0.018751390278339386, 0.21082064509391785, 0.0682586207985878, -0.17147260904312134, 0.04892030730843544, -0.057178497314453125, -0.08601470291614532, -0.1134747788310051, -0.05985361710190773, -0.0800875797867775, -0.1297968029975891, -0.000034089443943230435, -0.11826374381780624, -0.0005235447897575796, 0.1174224391579628, 0.005633013788610697, -0.02731461264193058, 0.16010835766792297, 0.011170967482030392, 0.02594037726521492, 0.06165660545229912, 0.008989768102765083, -0.03613077849149704, -0.12293252348899841, -0.052642516791820526, -0.016276879236102104, -0.015594038181006908, 0.033774614334106445, -0.05931494012475014, -0.037554118782281876, 0.031184900552034378, -0.02027048170566559, -0.09173652529716492, 0.0036402372643351555, 0.016913199797272682, 0.057751670479774475, 0.048560429364442825, 0.009997212328016758, 0.019075101241469383, -0.0059278663247823715, 0.19885186851024628, -0.07337722182273865, -0.06241562217473984, -0.10892566293478012, 0.23533464968204498, 0.041218534111976624, -0.022455492988228798, 0.031782735139131546, -0.06612449139356613, 0.004328879527747631, 0.24956606328487396, 0.21047116816043854, -0.07244722545146942, -0.0065431902185082436, 0.012562385760247707, -0.007097634486854076, -0.017912808805704117, 0.0986594632267952, 0.14931195974349976, 0.05848231539130211, -0.09368746727705002, -0.05104219168424606, -0.059819914400577545, -0.017858268693089485, -0.039675503969192505, 0.07364355772733688, 0.0478375218808651, 0.007639316841959953, -0.032984185963869095, 0.05411234498023987, -0.07018603384494781, -0.08971085399389267, 0.05477668344974518, -0.21682670712471008, -0.16399972140789032, -0.009443135000765324, 0.09615889936685562, 0.00434968201443553, 0.06322154402732849, -0.027622783556580544, -0.006259012967348099, 0.09283244609832764, -0.019432447850704193, -0.09740570932626724, -0.06426984071731567, 0.08471661806106567, -0.11873972415924072, 0.22185829281806946, -0.045842207968235016, 0.0528714694082737, 0.12712854146957397, 0.0699344202876091, -0.07376046478748322, 0.06245974823832512, 0.03862829506397247, -0.038276441395282745, 0.02886485867202282, 0.07490130513906479, -0.035239145159721375, 0.05644586682319641, 0.04707338288426399, -0.1422886699438095, 0.01882581040263176, -0.051350630819797516, -0.06144268065690994, -0.04812309890985489, -0.021385405212640762, -0.060594189912080765, 0.13178621232509613, 0.218593567609787, -0.027916565537452698, -0.010838805697858334, -0.07056831568479538, 0.011763928458094597, 0.050375860184431076, 0.022016096860170364, -0.05682146176695824, -0.2126062661409378, 0.019520791247487068, 0.04937693104147911, -0.01826678402721882, -0.24639299511909485, -0.10040940344333649, 0.0044407895766198635, -0.0742439329624176, -0.09766552597284317, 0.07167235016822815, 0.08758994191884995, 0.05211486667394638, -0.05936238914728165, -0.04031859710812569, -0.07389482855796814, 0.14774809777736664, -0.1392611712217331, -0.09294098615646362 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # gpt2-wikitext2 This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on the None dataset. It achieves the following results on the evaluation set: - Loss: 6.1112 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 6.5571 | 1.0 | 2249 | 6.4684 | | 6.1921 | 2.0 | 4498 | 6.1984 | | 6.0016 | 3.0 | 6747 | 6.1112 | ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.0
{"license": "mit", "tags": ["generated_from_trainer"], "model-index": [{"name": "gpt2-wikitext2", "results": []}]}
text-generation
fznmhmmd/gpt2-wikitext2
[ "transformers", "pytorch", "tensorboard", "gpt2", "text-generation", "generated_from_trainer", "license:mit", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
gpt2-wikitext2 ============== This model is a fine-tuned version of gpt2 on the None dataset. It achieves the following results on the evaluation set: * Loss: 6.1112 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.16.2 * Pytorch 1.10.0+cu111 * Datasets 1.18.3 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ 63, 98, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ -0.09154090285301208, 0.04790894687175751, -0.002184151206165552, 0.11117149889469147, 0.16774266958236694, 0.03017689846456051, 0.13103847205638885, 0.13129031658172607, -0.11401329189538956, 0.036095719784498215, 0.1389191448688507, 0.1702653169631958, 0.014559632167220116, 0.1101122498512268, -0.04969068989157677, -0.2709689140319824, -0.015016303397715092, 0.050961051136255264, -0.05571942776441574, 0.13819219172000885, 0.09048023819923401, -0.12517641484737396, 0.08742985129356384, -0.0038411198183894157, -0.22832395136356354, 0.014837736263871193, 0.014296517707407475, -0.05370286479592323, 0.15024344623088837, 0.03847575932741165, 0.1145910695195198, 0.01663755066692829, 0.07478562742471695, -0.17500877380371094, 0.013498245738446712, 0.05366280674934387, -0.003114348975941539, 0.09061431884765625, 0.059214431792497635, -0.00558617664501071, 0.18062038719654083, -0.0740666314959526, 0.039723802357912064, 0.017404830083251, -0.13080930709838867, -0.19098035991191864, -0.07488781213760376, 0.023442968726158142, 0.059575099498033524, 0.1109687015414238, -0.018535133451223373, 0.13270270824432373, -0.07684239000082016, 0.10215362906455994, 0.24175173044204712, -0.2988654673099518, -0.06552857905626297, 0.0436723455786705, 0.03397605940699577, 0.08428090065717697, -0.10248701274394989, -0.007487060967832804, 0.06268195062875748, 0.04670692980289459, 0.11779480427503586, -0.03867389261722565, -0.08324722945690155, 0.02222873643040657, -0.14630725979804993, -0.03131650760769844, 0.1440073698759079, 0.019422130659222603, -0.027908289805054665, -0.048285484313964844, -0.06324312835931778, -0.16276739537715912, -0.0255263801664114, -0.015109935775399208, 0.0339057520031929, -0.029044510796666145, -0.09711529314517975, -0.03294030576944351, -0.1168546974658966, -0.07095075398683548, -0.0732005313038826, 0.14518849551677704, 0.034968893975019455, 0.009237493388354778, -0.034638963639736176, 0.11207779496908188, -0.01635655388236046, -0.12074065208435059, 0.01513468474149704, 0.039468470960855484, 0.01789081282913685, -0.051539089530706406, -0.0685238391160965, -0.09255342185497284, 0.009481295011937618, 0.10241615772247314, -0.07119669020175934, 0.04612160846590996, 0.03238658234477043, 0.048917971551418304, -0.08200409263372421, 0.17944076657295227, -0.038824185729026794, 0.006819301750510931, 0.0010240797419101, 0.041862066835165024, 0.015114696696400642, -0.025720130652189255, -0.13514064252376556, 0.0055731358006596565, 0.10797539353370667, 0.013417274691164494, -0.07508037239313126, 0.07962367683649063, -0.04357932507991791, -0.0204959437251091, -0.015236489474773407, -0.09157538414001465, 0.031108368188142776, -0.005381562281399965, -0.08511962741613388, -0.01427430845797062, 0.015953296795487404, 0.005789796821773052, -0.036435846239328384, 0.11993834376335144, -0.09618431329727173, 0.03975940868258476, -0.09586972743272781, -0.12300056964159012, 0.012343509122729301, -0.0837741494178772, 0.016628554090857506, -0.09020568430423737, -0.1729010045528412, -0.019898565486073494, 0.04723161458969116, -0.033654890954494476, -0.046043761074543, -0.07077806442975998, -0.07548027485609055, 0.012515530921518803, -0.020145010203123093, 0.13305892050266266, -0.06216910853981972, 0.11445122212171555, 0.044230058789253235, 0.06574102491140366, -0.06125045567750931, 0.054452694952487946, -0.09305593371391296, 0.0005818717181682587, -0.1635381281375885, 0.05132431164383888, -0.03240293636918068, 0.05487579479813576, -0.07702431827783585, -0.11089999228715897, 0.0013434237334877253, 0.019256291911005974, 0.07539182901382446, 0.09179570525884628, -0.16162334382534027, -0.10028079897165298, 0.18144936859607697, -0.0688043162226677, -0.11640964448451996, 0.12525390088558197, -0.06351755559444427, 0.062307510524988174, 0.07654168456792831, 0.18675817549228668, 0.04239583760499954, -0.08253256976604462, 0.010934832505881786, 0.004951097536832094, 0.042934220284223557, -0.057707127183675766, 0.055139172822237015, -0.0015863884473219514, 0.03225238248705864, 0.023985927924513817, -0.008778751827776432, 0.049245018512010574, -0.10635518282651901, -0.08048561960458755, -0.036274660378694534, -0.08494063466787338, 0.03540873900055885, 0.06762945652008057, 0.08905865252017975, -0.1162978857755661, -0.09223170578479767, 0.07310468703508377, 0.07138915359973907, -0.0798463523387909, 0.03495784476399422, -0.060681115835905075, 0.07502477616071701, -0.035629451274871826, -0.016626732423901558, -0.1734222024679184, -0.024074802175164223, 0.006135197356343269, 0.027318937703967094, 0.040848392993211746, 0.03859094902873039, 0.07171530276536942, 0.0674448013305664, -0.054635316133499146, -0.00585352024063468, -0.018305521458387375, -0.01821196638047695, -0.1364089548587799, -0.1907189041376114, -0.017547817900776863, -0.016391750425100327, 0.1295745074748993, -0.22565092146396637, 0.043938230723142624, 0.014436748810112476, 0.059468258172273636, 0.005165283568203449, -0.016510430723428726, -0.047136664390563965, 0.08210225403308868, -0.05232483893632889, -0.04764464870095253, 0.07584637403488159, 0.00610940158367157, -0.09755333513021469, -0.045637618750333786, -0.14725539088249207, 0.164504274725914, 0.14566995203495026, -0.14257663488388062, -0.0912485420703888, -0.011670369654893875, -0.05547601357102394, -0.023535236716270447, -0.04589021950960159, 0.017533576115965843, 0.18956686556339264, -0.014782675541937351, 0.16159431636333466, -0.07225091755390167, -0.048522576689720154, 0.024529000744223595, -0.04061124846339226, 0.030337361618876457, 0.12413414567708969, 0.11849547922611237, -0.0766391009092331, 0.14823590219020844, 0.12058093398809433, -0.08998166769742966, 0.1599632352590561, -0.02773555926978588, -0.07613857835531235, -0.013474292121827602, -0.004752865992486477, -0.00026723751216195524, 0.08003447949886322, -0.15477991104125977, -0.02079842984676361, 0.010766426101326942, 0.02213730849325657, 0.04015021398663521, -0.23782092332839966, -0.04297001659870148, 0.033444419503211975, -0.05020155757665634, -0.0023494898341596127, -0.016890864819288254, -0.004998047836124897, 0.11300087720155716, 0.008105495944619179, -0.07091471552848816, 0.039591990411281586, 0.007141388487070799, -0.08682527393102646, 0.21803393959999084, -0.07016240805387497, -0.15930421650409698, -0.12653207778930664, -0.0810994803905487, -0.05214521288871765, 0.01865457370877266, 0.06716002523899078, -0.09542781859636307, -0.021622413769364357, -0.07565654069185257, 0.04743213579058647, -0.023878660053014755, 0.027900641784071922, -0.0012413286603987217, -0.0056598735973238945, 0.04161794111132622, -0.1153336688876152, -0.012898201122879982, -0.06382767856121063, -0.08974321186542511, 0.05935664847493172, 0.024129273369908333, 0.11165563762187958, 0.1697143316268921, -0.02155318669974804, 0.01872202754020691, -0.043657537549734116, 0.2248871922492981, -0.07478951662778854, -0.03436831384897232, 0.12070286273956299, -0.00023143450380302966, 0.05238955467939377, 0.09198525547981262, 0.0663284957408905, -0.09835929423570633, 0.013079585507512093, 0.03485707938671112, -0.0475134514272213, -0.22200295329093933, -0.036950837820768356, -0.05650486797094345, 0.004685133695602417, 0.0904216542840004, 0.03741206228733063, 0.055356215685606, 0.07155811041593552, 0.03987801820039749, 0.08369835466146469, -0.02772979997098446, 0.05985988304018974, 0.11807525902986526, 0.04045854136347771, 0.13311928510665894, -0.04680924862623215, -0.07103113830089569, 0.04681263864040375, -0.019989067688584328, 0.22321604192256927, 0.00025650180759839714, 0.15681356191635132, 0.05108335241675377, 0.1450379639863968, 0.0004529485304374248, 0.07553030550479889, -0.013010270893573761, -0.0496259443461895, -0.012975066900253296, -0.042996812611818314, -0.0329095795750618, 0.023700371384620667, -0.067669577896595, 0.036275964230298996, -0.12394816428422928, -0.0029683236498385668, 0.05970390886068344, 0.2182581126689911, 0.04356873780488968, -0.3291844427585602, -0.09277503192424774, 0.0015261439839378, -0.028401004150509834, -0.016770705580711365, 0.026843098923563957, 0.11081304401159286, -0.08086773008108139, 0.033318862318992615, -0.0689031183719635, 0.09495903551578522, -0.0556633323431015, 0.06040404736995697, 0.057139452546834946, 0.10534163564443588, -0.00788985937833786, 0.08592315018177032, -0.3220860958099365, 0.27272671461105347, 0.010078266263008118, 0.07507113367319107, -0.08086104691028595, -0.002042392734438181, 0.01698460802435875, 0.050316352397203445, 0.058848705142736435, -0.017167238518595695, -0.0301820058375597, -0.18649150431156158, -0.03751383349299431, 0.03317003324627876, 0.11979252099990845, -0.0241579320281744, 0.10174338519573212, -0.027875088155269623, 0.02004375122487545, 0.07519017904996872, -0.023011846467852592, -0.03515256568789482, -0.10312490165233612, 0.0023722522892057896, 0.01044845674186945, -0.042911745607852936, -0.05406617000699043, -0.1096707358956337, -0.12660086154937744, 0.1866525262594223, -0.012655875645577908, -0.0350058488547802, -0.10568028688430786, 0.09974654018878937, 0.056002356112003326, -0.08879096060991287, 0.029206762090325356, 0.016990039497613907, 0.05756596848368645, 0.02310051955282688, -0.06709477305412292, 0.12245291471481323, -0.04706496000289917, -0.15953341126441956, -0.04993072897195816, 0.11123108863830566, 0.030761703848838806, 0.06066351756453514, -0.013277517631649971, 0.013936015777289867, -0.04412936046719551, -0.09660474956035614, 0.0326232872903347, -0.032835524529218674, 0.05634452402591705, 0.01710325852036476, -0.04453078657388687, 0.03426390513777733, -0.059848327189683914, -0.04480806365609169, 0.20472773909568787, 0.2506873309612274, -0.0821322649717331, 0.014953149482607841, 0.03573226556181908, -0.0774463340640068, -0.20423917472362518, 0.048619646579027176, 0.042975254356861115, 0.005744313821196556, 0.02591707371175289, -0.19422277808189392, 0.09951306134462357, 0.11930137872695923, -0.0024077296257019043, 0.13587796688079834, -0.34987401962280273, -0.13601331412792206, 0.10504686087369919, 0.14712201058864594, 0.13058118522167206, -0.16082815825939178, -0.024699924513697624, -0.02536102756857872, -0.1093539372086525, 0.11684046685695648, -0.10077027231454849, 0.14016874134540558, -0.024085361510515213, 0.10793174803256989, 0.008021670393645763, -0.05888748541474342, 0.11063754558563232, 0.014005998149514198, 0.09095386415719986, -0.07122839242219925, -0.02792314440011978, 0.04805930331349373, -0.026308270171284676, 0.018072867766022682, -0.10109025239944458, 0.017324823886156082, -0.08431388437747955, -0.03238210082054138, -0.0697421208024025, 0.0450812429189682, -0.033819567412137985, -0.07101496309041977, -0.04594295471906662, 0.002644988941028714, 0.03119884431362152, -0.0119785126298666, 0.13828054070472717, -0.002143767662346363, 0.1634291708469391, 0.10911955684423447, 0.08047578483819962, -0.08439774811267853, -0.04220839589834213, -0.004841073881834745, -0.010871374048292637, 0.053971607238054276, -0.148051917552948, 0.01805998757481575, 0.1482211947441101, 0.023678189143538475, 0.14894337952136993, 0.09266475588083267, -0.03593441843986511, 0.028939301148056984, 0.06735216081142426, -0.17313742637634277, -0.10105223953723907, -0.01440330222249031, -0.08304141461849213, -0.09071098268032074, 0.06334608793258667, 0.10437328368425369, -0.06583280116319656, -0.005749199539422989, -0.009935025125741959, 0.0022947376128286123, -0.061088334769010544, 0.19549089670181274, 0.051168084144592285, 0.042740222066640854, -0.09930147230625153, 0.06030723825097084, 0.04070674255490303, -0.07585228979587555, 0.023495038971304893, 0.10471191257238388, -0.06933919340372086, -0.045418910682201385, 0.0664716362953186, 0.18332023918628693, -0.08397449553012848, -0.03427346050739288, -0.1412811130285263, -0.11850735545158386, 0.07746182382106781, 0.16571715474128723, 0.10304321348667145, 0.01799110695719719, -0.059322845190763474, 0.027668209746479988, -0.14229020476341248, 0.07141375541687012, 0.03831510618329048, 0.06594479084014893, -0.12171030789613724, 0.19278888404369354, 0.011119788512587547, 0.040664300322532654, -0.029830198734998703, 0.018930474296212196, -0.11600326001644135, 0.015583639033138752, -0.1172579899430275, -0.03199515491724014, -0.03222142159938812, -0.005573422182351351, -0.002334869233891368, -0.04409262537956238, -0.05663740634918213, 0.005006554536521435, -0.12165345996618271, -0.02114926278591156, 0.027485378086566925, 0.040046609938144684, -0.11752359569072723, -0.023545444011688232, 0.011530530638992786, -0.05748510733246803, 0.0790523886680603, 0.050424523651599884, 0.01030760258436203, 0.07404220104217529, -0.16823682188987732, 0.021833905950188637, 0.07212761789560318, 0.007535671815276146, 0.05725701525807381, -0.049993474036455154, -0.004372365772724152, 0.002520917449146509, 0.0770244300365448, 0.02960074320435524, 0.062169868499040604, -0.1328713297843933, 0.012860669754445553, -0.03635582700371742, -0.08316214382648468, -0.07094733417034149, 0.05476260557770729, 0.06240183860063553, 0.016008060425519943, 0.18278248608112335, -0.0929349884390831, 0.03670749068260193, -0.20967181026935577, 0.00640035979449749, 0.0017599391285330057, -0.11635550111532211, -0.1027316004037857, -0.07409337908029556, 0.0711076408624649, -0.05817238241434097, 0.14708782732486725, 0.03511781990528107, 0.023101946339011192, 0.027400095015764236, -0.014749379828572273, 0.02008034661412239, 0.012644943781197071, 0.22630324959754944, 0.04021883010864258, -0.043688636273145676, 0.03443204611539841, 0.06186903268098831, 0.11395983397960663, 0.10914915800094604, 0.21618333458900452, 0.11927526444196701, -0.02905486337840557, 0.09996999800205231, 0.033717069774866104, -0.060363829135894775, -0.14930124580860138, 0.04406102001667023, -0.038323573768138885, 0.10138288140296936, -0.031061161309480667, 0.20221248269081116, 0.09068119525909424, -0.15017008781433105, 0.035696495324373245, -0.051961444318294525, -0.0942218154668808, -0.11942566186189651, -0.057162776589393616, -0.085393525660038, -0.14703741669654846, 0.007185769733041525, -0.12116332352161407, 0.03353458270430565, 0.10444089770317078, 0.020532622933387756, -0.02924007922410965, 0.1633773148059845, 0.030277660116553307, 0.01075546070933342, 0.05729609355330467, -0.0036821342073380947, -0.0178205668926239, -0.11147824674844742, -0.06225251033902168, -0.013823808170855045, -0.015464290976524353, 0.046145033091306686, -0.04499934986233711, -0.04862174019217491, 0.03196078911423683, -0.04182630777359009, -0.09612369537353516, 0.008089038543403149, 0.0373714379966259, 0.06798700243234634, 0.03997175768017769, 0.0016033934662118554, -0.006821238901466131, -0.014671352691948414, 0.222270205616951, -0.07574975490570068, -0.07885903865098953, -0.07425349950790405, 0.27256056666374207, 0.04048144444823265, 0.0027663386426866055, 0.020481429994106293, -0.0705634281039238, 0.005923576653003693, 0.2550033926963806, 0.22300535440444946, -0.09006433933973312, -0.0054147569462656975, 0.00497685931622982, 0.0002551785728428513, -0.008909730240702629, 0.1174522265791893, 0.12613384425640106, 0.07293765246868134, -0.09188392013311386, -0.0416770838201046, -0.04544425755739212, -0.003677608445286751, -0.04051583260297775, 0.06380369514226913, 0.05816878005862236, 0.015301080420613289, -0.0418560728430748, 0.05463914945721626, -0.08380051702260971, -0.08702948689460754, 0.03645637631416321, -0.20973928272724152, -0.14964690804481506, 0.0016248500905930996, 0.10823892056941986, -0.0035807976964861155, 0.07889149338006973, -0.02822762355208397, -0.0022887871600687504, 0.04286777228116989, -0.01525236014276743, -0.11194998770952225, -0.05760727450251579, 0.08612610399723053, -0.1337067186832428, 0.17946220934391022, -0.05095742642879486, 0.06767892092466354, 0.12914401292800903, 0.0567118376493454, -0.056170325726270676, 0.08150850236415863, 0.034504249691963196, -0.06736637651920319, 0.030149884521961212, 0.09147025644779205, -0.03410087898373604, 0.04303671419620514, 0.05057705193758011, -0.13637639582157135, 0.023661542683839798, -0.06979118287563324, -0.05279651656746864, -0.03606823831796646, -0.05947493389248848, -0.060641318559646606, 0.12458920478820801, 0.2188384234905243, -0.01997010037302971, 0.013635996729135513, -0.07608854025602341, 0.005272256210446358, 0.058167118579149246, 0.044576648622751236, -0.06452834606170654, -0.2454196661710739, -0.004605409689247608, 0.08185486495494843, -0.026521064341068268, -0.27561089396476746, -0.07278566807508469, -0.0028352790977805853, -0.05790458992123604, -0.1036747470498085, 0.08536790311336517, 0.09567233175039291, 0.054930638521909714, -0.05790930613875389, -0.0649113580584526, -0.07362701743841171, 0.16253699362277985, -0.14485016465187073, -0.09287360310554504 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-common_voice-es-demo This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the COMMON_VOICE - ES dataset. It achieves the following results on the evaluation set: - Loss: 0.1788 - Wer: 1.0239 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 15.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:-----:|:---------------:|:------:| | No log | 0.02 | 100 | 6.6465 | 1.0 | | No log | 0.04 | 200 | 3.0150 | 1.0 | | No log | 0.05 | 300 | 2.8622 | 1.0003 | | No log | 0.07 | 400 | 0.9506 | 0.9771 | | 5.1598 | 0.09 | 500 | 0.4883 | 1.0009 | | 5.1598 | 0.11 | 600 | 0.3893 | 1.0203 | | 5.1598 | 0.13 | 700 | 0.3417 | 1.0283 | | 5.1598 | 0.14 | 800 | 0.3352 | 1.0335 | | 5.1598 | 0.16 | 900 | 0.2987 | 1.0168 | | 0.3671 | 0.18 | 1000 | 0.2921 | 1.0159 | | 0.3671 | 0.2 | 1100 | 0.2770 | 1.0096 | | 0.3671 | 0.22 | 1200 | 0.2790 | 1.0398 | | 0.3671 | 0.24 | 1300 | 0.2659 | 1.0190 | | 0.3671 | 0.25 | 1400 | 0.2657 | 1.0528 | | 0.289 | 0.27 | 1500 | 0.2556 | 1.0301 | | 0.289 | 0.29 | 1600 | 0.2514 | 1.0193 | | 0.289 | 0.31 | 1700 | 0.2708 | 1.0699 | | 0.289 | 0.33 | 1800 | 0.2455 | 1.0723 | | 0.289 | 0.34 | 1900 | 0.2456 | 1.0100 | | 0.271 | 0.36 | 2000 | 0.2338 | 1.0533 | | 0.271 | 0.38 | 2100 | 0.2479 | 1.0128 | | 0.271 | 0.4 | 2200 | 0.2483 | 1.0386 | | 0.271 | 0.42 | 2300 | 0.2436 | 1.0528 | | 0.271 | 0.43 | 2400 | 0.2382 | 1.0476 | | 0.2634 | 0.45 | 2500 | 0.2329 | 1.0680 | | 0.2634 | 0.47 | 2600 | 0.2433 | 1.0581 | | 0.2634 | 0.49 | 2700 | 0.2354 | 1.0641 | | 0.2634 | 0.51 | 2800 | 0.2318 | 1.0504 | | 0.2634 | 0.52 | 2900 | 0.2325 | 1.0500 | | 0.2522 | 0.54 | 3000 | 0.2344 | 1.0380 | | 0.2522 | 0.56 | 3100 | 0.2244 | 1.0663 | | 0.2522 | 0.58 | 3200 | 0.2340 | 1.0647 | | 0.2522 | 0.6 | 3300 | 0.2288 | 1.0538 | | 0.2522 | 0.61 | 3400 | 0.2212 | 1.0614 | | 0.2468 | 0.63 | 3500 | 0.2487 | 1.0557 | | 0.2468 | 0.65 | 3600 | 0.2330 | 1.0510 | | 0.2468 | 0.67 | 3700 | 0.2308 | 1.0506 | | 0.2468 | 0.69 | 3800 | 0.2320 | 1.0451 | | 0.2468 | 0.71 | 3900 | 0.2261 | 1.0701 | | 0.2505 | 0.72 | 4000 | 0.2281 | 1.0713 | | 0.2505 | 0.74 | 4100 | 0.2277 | 1.0741 | | 0.2505 | 0.76 | 4200 | 0.2253 | 1.0814 | | 0.2505 | 0.78 | 4300 | 0.2215 | 1.0437 | | 0.2505 | 0.8 | 4400 | 0.2220 | 1.0557 | | 0.2434 | 0.81 | 4500 | 0.2184 | 1.0533 | | 0.2434 | 0.83 | 4600 | 0.2222 | 1.0819 | | 0.2434 | 0.85 | 4700 | 0.2162 | 1.0238 | | 0.2434 | 0.87 | 4800 | 0.2132 | 1.0457 | | 0.2434 | 0.89 | 4900 | 0.2068 | 1.0611 | | 0.2347 | 0.9 | 5000 | 0.2166 | 1.0332 | | 0.2347 | 0.92 | 5100 | 0.2087 | 1.0433 | | 0.2347 | 0.94 | 5200 | 0.2100 | 1.0292 | | 0.2347 | 0.96 | 5300 | 0.2067 | 1.0734 | | 0.2347 | 0.98 | 5400 | 0.2148 | 1.0279 | | 0.2333 | 0.99 | 5500 | 0.2125 | 1.0277 | | 0.2333 | 1.01 | 5600 | 0.2054 | 1.0453 | | 0.2333 | 1.03 | 5700 | 0.2091 | 1.0557 | | 0.2333 | 1.05 | 5800 | 0.2086 | 1.0239 | | 0.2333 | 1.07 | 5900 | 0.2051 | 1.0645 | | 0.2087 | 1.09 | 6000 | 0.2103 | 1.0240 | | 0.2087 | 1.1 | 6100 | 0.2145 | 1.0197 | | 0.2087 | 1.12 | 6200 | 0.2136 | 1.0248 | | 0.2087 | 1.14 | 6300 | 0.2045 | 1.0443 | | 0.2087 | 1.16 | 6400 | 0.2089 | 1.0397 | | 0.2013 | 1.18 | 6500 | 0.2012 | 1.0654 | | 0.2013 | 1.19 | 6600 | 0.2054 | 1.0414 | | 0.2013 | 1.21 | 6700 | 0.2081 | 1.0632 | | 0.2013 | 1.23 | 6800 | 0.2104 | 1.0190 | | 0.2013 | 1.25 | 6900 | 0.2045 | 1.0813 | | 0.2092 | 1.27 | 7000 | 0.2096 | 1.0751 | | 0.2092 | 1.28 | 7100 | 0.2103 | 1.0328 | | 0.2092 | 1.3 | 7200 | 0.2044 | 1.0011 | | 0.2092 | 1.32 | 7300 | 0.2089 | 1.0260 | | 0.2092 | 1.34 | 7400 | 0.2063 | 1.0551 | | 0.2076 | 1.36 | 7500 | 0.2029 | 1.0075 | | 0.2076 | 1.37 | 7600 | 0.2040 | 1.0528 | | 0.2076 | 1.39 | 7700 | 0.2075 | 1.0398 | | 0.2076 | 1.41 | 7800 | 0.2023 | 1.0231 | | 0.2076 | 1.43 | 7900 | 0.2049 | 1.0318 | | 0.2028 | 1.45 | 8000 | 0.2072 | 1.0763 | | 0.2028 | 1.47 | 8100 | 0.2075 | 1.0762 | | 0.2028 | 1.48 | 8200 | 0.2052 | 1.0838 | | 0.2028 | 1.5 | 8300 | 0.2053 | 1.0407 | | 0.2028 | 1.52 | 8400 | 0.2066 | 1.0266 | | 0.2025 | 1.54 | 8500 | 0.2037 | 1.0628 | | 0.2025 | 1.56 | 8600 | 0.2010 | 1.0351 | | 0.2025 | 1.57 | 8700 | 0.1961 | 1.0812 | | 0.2025 | 1.59 | 8800 | 0.1963 | 1.0868 | | 0.2025 | 1.61 | 8900 | 0.2022 | 1.0710 | | 0.1997 | 1.63 | 9000 | 0.2051 | 1.0764 | | 0.1997 | 1.65 | 9100 | 0.1987 | 1.0581 | | 0.1997 | 1.66 | 9200 | 0.2051 | 1.0611 | | 0.1997 | 1.68 | 9300 | 0.1999 | 1.0808 | | 0.1997 | 1.7 | 9400 | 0.1972 | 1.0703 | | 0.1983 | 1.72 | 9500 | 0.1961 | 1.0584 | | 0.1983 | 1.74 | 9600 | 0.2031 | 1.0938 | | 0.1983 | 1.75 | 9700 | 0.2019 | 1.0891 | | 0.1983 | 1.77 | 9800 | 0.2006 | 1.0542 | | 0.1983 | 1.79 | 9900 | 0.1925 | 1.0627 | | 0.1961 | 1.81 | 10000 | 0.1976 | 1.0751 | | 0.1961 | 1.83 | 10100 | 0.2051 | 1.0611 | | 0.1961 | 1.85 | 10200 | 0.2037 | 1.0656 | | 0.1961 | 1.86 | 10300 | 0.2025 | 1.0291 | | 0.1961 | 1.88 | 10400 | 0.1977 | 1.0525 | | 0.2025 | 1.9 | 10500 | 0.2030 | 1.0670 | | 0.2025 | 1.92 | 10600 | 0.1980 | 1.0765 | | 0.2025 | 1.94 | 10700 | 0.1975 | 1.0254 | | 0.2025 | 1.95 | 10800 | 0.1986 | 1.0636 | | 0.2025 | 1.97 | 10900 | 0.1956 | 1.0352 | | 0.2025 | 1.99 | 11000 | 0.1954 | 1.0265 | | 0.2025 | 2.01 | 11100 | 0.1957 | 1.0752 | | 0.2025 | 2.03 | 11200 | 0.1943 | 1.0784 | | 0.2025 | 2.04 | 11300 | 0.1898 | 1.0341 | | 0.2025 | 2.06 | 11400 | 0.1921 | 1.0301 | | 0.1805 | 2.08 | 11500 | 0.1910 | 1.0230 | | 0.1805 | 2.1 | 11600 | 0.1961 | 1.0203 | | 0.1805 | 2.12 | 11700 | 0.1973 | 1.0776 | | 0.1805 | 2.13 | 11800 | 0.1876 | 1.0788 | | 0.1805 | 2.15 | 11900 | 0.1934 | 1.0251 | | 0.177 | 2.17 | 12000 | 0.1967 | 1.0340 | | 0.177 | 2.19 | 12100 | 0.1932 | 1.0131 | | 0.177 | 2.21 | 12200 | 0.1926 | 1.0078 | | 0.177 | 2.23 | 12300 | 0.1947 | 0.9991 | | 0.177 | 2.24 | 12400 | 0.1914 | 1.0213 | | 0.1782 | 2.26 | 12500 | 0.1962 | 0.9882 | | 0.1782 | 2.28 | 12600 | 0.1960 | 1.0562 | | 0.1782 | 2.3 | 12700 | 0.2006 | 1.0401 | | 0.1782 | 2.32 | 12800 | 0.1950 | 1.0688 | | 0.1782 | 2.33 | 12900 | 0.1920 | 1.0435 | | 0.1796 | 2.35 | 13000 | 0.1926 | 1.0667 | | 0.1796 | 2.37 | 13100 | 0.1949 | 1.0859 | | 0.1796 | 2.39 | 13200 | 0.1932 | 1.0670 | | 0.1796 | 2.41 | 13300 | 0.1882 | 1.0663 | | 0.1796 | 2.42 | 13400 | 0.1877 | 1.0760 | | 0.1775 | 2.44 | 13500 | 0.1893 | 1.0859 | | 0.1775 | 2.46 | 13600 | 0.1936 | 1.0702 | | 0.1775 | 2.48 | 13700 | 0.1871 | 1.0414 | | 0.1775 | 2.5 | 13800 | 0.1917 | 1.0430 | | 0.1775 | 2.51 | 13900 | 0.1922 | 1.0422 | | 0.1778 | 2.53 | 14000 | 0.1875 | 1.0585 | | 0.1778 | 2.55 | 14100 | 0.1876 | 1.0603 | | 0.1778 | 2.57 | 14200 | 0.1888 | 1.0628 | | 0.1778 | 2.59 | 14300 | 0.1948 | 1.0782 | | 0.1778 | 2.6 | 14400 | 0.1942 | 1.0695 | | 0.1784 | 2.62 | 14500 | 0.1842 | 1.0863 | | 0.1784 | 2.64 | 14600 | 0.1850 | 1.0543 | | 0.1784 | 2.66 | 14700 | 0.1824 | 1.0683 | | 0.1784 | 2.68 | 14800 | 0.1888 | 1.0693 | | 0.1784 | 2.7 | 14900 | 0.1871 | 1.0175 | | 0.1753 | 2.71 | 15000 | 0.1889 | 1.0549 | | 0.1753 | 2.73 | 15100 | 0.1865 | 1.0544 | | 0.1753 | 2.75 | 15200 | 0.1918 | 1.0726 | | 0.1753 | 2.77 | 15300 | 0.1964 | 1.0915 | | 0.1753 | 2.79 | 15400 | 0.1900 | 1.0610 | | 0.1768 | 2.8 | 15500 | 0.1894 | 1.0763 | | 0.1768 | 2.82 | 15600 | 0.1882 | 1.0548 | | 0.1768 | 2.84 | 15700 | 0.1861 | 1.0902 | | 0.1768 | 2.86 | 15800 | 0.1860 | 1.0551 | | 0.1768 | 2.88 | 15900 | 0.1879 | 1.0581 | | 0.1761 | 2.89 | 16000 | 0.1899 | 1.0544 | | 0.1761 | 2.91 | 16100 | 0.1860 | 1.0530 | | 0.1761 | 2.93 | 16200 | 0.1894 | 1.0596 | | 0.1761 | 2.95 | 16300 | 0.1835 | 1.0394 | | 0.1761 | 2.97 | 16400 | 0.1852 | 1.0445 | | 0.1754 | 2.98 | 16500 | 0.1847 | 1.0390 | | 0.1754 | 3.0 | 16600 | 0.1828 | 1.0440 | | 0.1754 | 3.02 | 16700 | 0.1869 | 1.0560 | | 0.1754 | 3.04 | 16800 | 0.1882 | 1.0573 | | 0.1754 | 3.06 | 16900 | 0.1912 | 1.0600 | | 0.1592 | 3.08 | 17000 | 0.1921 | 1.0529 | | 0.1592 | 3.09 | 17100 | 0.1881 | 1.0175 | | 0.1592 | 3.11 | 17200 | 0.1891 | 1.0654 | | 0.1592 | 3.13 | 17300 | 0.1889 | 1.0687 | | 0.1592 | 3.15 | 17400 | 0.1916 | 1.0642 | | 0.1556 | 3.17 | 17500 | 0.1850 | 1.0295 | | 0.1556 | 3.18 | 17600 | 0.1875 | 1.0273 | | 0.1556 | 3.2 | 17700 | 0.1894 | 1.0051 | | 0.1556 | 3.22 | 17800 | 0.1870 | 1.0462 | | 0.1556 | 3.24 | 17900 | 0.1831 | 1.0308 | | 0.1557 | 3.26 | 18000 | 0.1878 | 1.0603 | | 0.1557 | 3.27 | 18100 | 0.1850 | 1.0566 | | 0.1557 | 3.29 | 18200 | 0.1843 | 1.0629 | | 0.1557 | 3.31 | 18300 | 0.1886 | 1.0378 | | 0.1557 | 3.33 | 18400 | 0.1892 | 1.0381 | | 0.159 | 3.35 | 18500 | 0.1942 | 1.0519 | | 0.159 | 3.36 | 18600 | 0.1829 | 1.0622 | | 0.159 | 3.38 | 18700 | 0.1894 | 1.0557 | | 0.159 | 3.4 | 18800 | 0.1895 | 1.0627 | | 0.159 | 3.42 | 18900 | 0.1863 | 1.0362 | | 0.1582 | 3.44 | 19000 | 0.1888 | 1.0491 | | 0.1582 | 3.46 | 19100 | 0.1854 | 1.0483 | | 0.1582 | 3.47 | 19200 | 0.1797 | 0.9787 | | 0.1582 | 3.49 | 19300 | 0.1785 | 1.0086 | | 0.1582 | 3.51 | 19400 | 0.1797 | 0.9915 | | 0.1507 | 3.53 | 19500 | 0.1873 | 1.0266 | | 0.1507 | 3.55 | 19600 | 0.1838 | 1.0299 | | 0.1507 | 3.56 | 19700 | 0.1817 | 1.0355 | | 0.1507 | 3.58 | 19800 | 0.1819 | 1.0271 | | 0.1507 | 3.6 | 19900 | 0.1883 | 1.0248 | | 0.1601 | 3.62 | 20000 | 0.1823 | 1.0406 | | 0.1601 | 3.64 | 20100 | 0.1801 | 1.0261 | | 0.1601 | 3.65 | 20200 | 0.1783 | 1.0329 | | 0.1601 | 3.67 | 20300 | 0.1857 | 1.0162 | | 0.1601 | 3.69 | 20400 | 0.1814 | 1.0212 | | 0.1552 | 3.71 | 20500 | 0.1837 | 1.0232 | | 0.1552 | 3.73 | 20600 | 0.1843 | 1.0314 | | 0.1552 | 3.74 | 20700 | 0.1842 | 1.0258 | | 0.1552 | 3.76 | 20800 | 0.1821 | 1.0479 | | 0.1552 | 3.78 | 20900 | 0.1864 | 1.0459 | | 0.1576 | 3.8 | 21000 | 0.1831 | 1.0364 | | 0.1576 | 3.82 | 21100 | 0.1852 | 1.0271 | | 0.1576 | 3.83 | 21200 | 0.1865 | 1.0204 | | 0.1576 | 3.85 | 21300 | 0.1794 | 1.0324 | | 0.1576 | 3.87 | 21400 | 0.1826 | 1.0315 | | 0.1585 | 3.89 | 21500 | 0.1824 | 1.0327 | | 0.1585 | 3.91 | 21600 | 0.1838 | 1.0208 | | 0.1585 | 3.93 | 21700 | 0.1850 | 1.0199 | | 0.1585 | 3.94 | 21800 | 0.1841 | 1.0050 | | 0.1585 | 3.96 | 21900 | 0.1783 | 1.0003 | | 0.1572 | 3.98 | 22000 | 0.1787 | 1.0115 | | 0.1572 | 4.0 | 22100 | 0.1810 | 1.0235 | | 0.1572 | 4.02 | 22200 | 0.1763 | 1.0191 | | 0.1572 | 4.03 | 22300 | 0.1764 | 1.0332 | | 0.1572 | 4.05 | 22400 | 0.1794 | 1.0429 | | 0.1406 | 4.07 | 22500 | 0.1905 | 1.0288 | | 0.1406 | 4.09 | 22600 | 0.1776 | 1.0244 | | 0.1406 | 4.11 | 22700 | 0.1782 | 1.0451 | | 0.1406 | 4.12 | 22800 | 0.1771 | 1.0387 | | 0.1406 | 4.14 | 22900 | 0.1788 | 1.0435 | | 0.14 | 4.16 | 23000 | 0.1792 | 1.0421 | | 0.14 | 4.18 | 23100 | 0.1841 | 1.0241 | | 0.14 | 4.2 | 23200 | 0.1769 | 1.0546 | | 0.14 | 4.21 | 23300 | 0.1815 | 1.0602 | | 0.14 | 4.23 | 23400 | 0.1784 | 1.0369 | | 0.1394 | 4.25 | 23500 | 0.1809 | 1.0406 | | 0.1394 | 4.27 | 23600 | 0.1744 | 1.0133 | | 0.1394 | 4.29 | 23700 | 0.1771 | 1.0214 | | 0.1394 | 4.31 | 23800 | 0.1765 | 1.0064 | | 0.1394 | 4.32 | 23900 | 0.1793 | 1.0200 | | 0.14 | 4.34 | 24000 | 0.1776 | 1.0352 | | 0.14 | 4.36 | 24100 | 0.1775 | 1.0294 | | 0.14 | 4.38 | 24200 | 0.1763 | 1.0213 | | 0.14 | 4.4 | 24300 | 0.1697 | 1.0302 | | 0.14 | 4.41 | 24400 | 0.1771 | 1.0259 | | 0.1408 | 4.43 | 24500 | 0.1747 | 1.0409 | | 0.1408 | 4.45 | 24600 | 0.1769 | 1.0278 | | 0.1408 | 4.47 | 24700 | 0.1767 | 1.0190 | | 0.1408 | 4.49 | 24800 | 0.1745 | 1.0281 | | 0.1408 | 4.5 | 24900 | 0.1738 | 1.0356 | | 0.1391 | 4.52 | 25000 | 0.1781 | 1.0429 | | 0.1391 | 4.54 | 25100 | 0.1784 | 1.0076 | | 0.1391 | 4.56 | 25200 | 0.1771 | 1.0157 | | 0.1391 | 4.58 | 25300 | 0.1758 | 1.0337 | | 0.1391 | 4.59 | 25400 | 0.1758 | 1.0466 | | 0.1398 | 4.61 | 25500 | 0.1724 | 1.0403 | | 0.1398 | 4.63 | 25600 | 0.1765 | 1.0481 | | 0.1398 | 4.65 | 25700 | 0.1757 | 1.0320 | | 0.1398 | 4.67 | 25800 | 0.1814 | 1.0479 | | 0.1398 | 4.69 | 25900 | 0.1713 | 1.0251 | | 0.1427 | 4.7 | 26000 | 0.1735 | 1.0340 | | 0.1427 | 4.72 | 26100 | 0.1765 | 1.0358 | | 0.1427 | 4.74 | 26200 | 0.1731 | 1.0220 | | 0.1427 | 4.76 | 26300 | 0.1769 | 1.0261 | | 0.1427 | 4.78 | 26400 | 0.1747 | 1.0139 | | 0.1424 | 4.79 | 26500 | 0.1791 | 1.0406 | | 0.1424 | 4.81 | 26600 | 0.1735 | 1.0497 | | 0.1424 | 4.83 | 26700 | 0.1710 | 1.0433 | | 0.1424 | 4.85 | 26800 | 0.1771 | 1.0002 | | 0.1424 | 4.87 | 26900 | 0.1748 | 1.0046 | | 0.1419 | 4.88 | 27000 | 0.1794 | 1.0332 | | 0.1419 | 4.9 | 27100 | 0.1772 | 1.0558 | | 0.1419 | 4.92 | 27200 | 0.1757 | 1.0477 | | 0.1419 | 4.94 | 27300 | 0.1735 | 1.0324 | | 0.1419 | 4.96 | 27400 | 0.1758 | 1.0260 | | 0.1433 | 4.97 | 27500 | 0.1767 | 1.0422 | | 0.1433 | 4.99 | 27600 | 0.1695 | 1.0386 | | 0.1433 | 5.01 | 27700 | 0.1763 | 1.0571 | | 0.1433 | 5.03 | 27800 | 0.1743 | 1.0367 | | 0.1433 | 5.05 | 27900 | 0.1804 | 1.0255 | | 0.1306 | 5.07 | 28000 | 0.1803 | 1.0377 | | 0.1306 | 5.08 | 28100 | 0.1750 | 1.0552 | | 0.1306 | 5.1 | 28200 | 0.1743 | 1.0512 | | 0.1306 | 5.12 | 28300 | 0.1777 | 1.0584 | | 0.1306 | 5.14 | 28400 | 0.1726 | 1.0374 | | 0.123 | 5.16 | 28500 | 0.1776 | 1.0439 | | 0.123 | 5.17 | 28600 | 0.1759 | 1.0682 | | 0.123 | 5.19 | 28700 | 0.1724 | 1.0511 | | 0.123 | 5.21 | 28800 | 0.1677 | 1.0560 | | 0.123 | 5.23 | 28900 | 0.1699 | 1.0421 | | 0.1217 | 5.25 | 29000 | 0.1803 | 1.0370 | | 0.1217 | 5.26 | 29100 | 0.1770 | 1.0474 | | 0.1217 | 5.28 | 29200 | 0.1733 | 1.0332 | | 0.1217 | 5.3 | 29300 | 0.1746 | 1.0158 | | 0.1217 | 5.32 | 29400 | 0.1763 | 1.0341 | | 0.1246 | 5.34 | 29500 | 0.1775 | 1.0348 | | 0.1246 | 5.35 | 29600 | 0.1730 | 1.0492 | | 0.1246 | 5.37 | 29700 | 0.1730 | 1.0503 | | 0.1246 | 5.39 | 29800 | 0.1727 | 1.0437 | | 0.1246 | 5.41 | 29900 | 0.1744 | 1.0539 | | 0.127 | 5.43 | 30000 | 0.1748 | 1.0463 | | 0.127 | 5.44 | 30100 | 0.1746 | 1.0555 | | 0.127 | 5.46 | 30200 | 0.1810 | 1.0558 | | 0.127 | 5.48 | 30300 | 0.1773 | 1.0407 | | 0.127 | 5.5 | 30400 | 0.1722 | 1.0489 | | 0.1276 | 5.52 | 30500 | 0.1720 | 1.0520 | | 0.1276 | 5.54 | 30600 | 0.1777 | 1.0347 | | 0.1276 | 5.55 | 30700 | 0.1685 | 1.0347 | | 0.1276 | 5.57 | 30800 | 0.1659 | 1.0338 | | 0.1276 | 5.59 | 30900 | 0.1756 | 1.0228 | | 0.1246 | 5.61 | 31000 | 0.1717 | 1.0409 | | 0.1246 | 5.63 | 31100 | 0.1764 | 1.0202 | | 0.1246 | 5.64 | 31200 | 0.1693 | 1.0314 | | 0.1246 | 5.66 | 31300 | 0.1731 | 1.0319 | | 0.1246 | 5.68 | 31400 | 0.1688 | 1.0380 | | 0.1271 | 5.7 | 31500 | 0.1671 | 1.0350 | | 0.1271 | 5.72 | 31600 | 0.1676 | 1.0430 | | 0.1271 | 5.73 | 31700 | 0.1656 | 1.0441 | | 0.1271 | 5.75 | 31800 | 0.1664 | 1.0403 | | 0.1271 | 5.77 | 31900 | 0.1691 | 1.0152 | | 0.1259 | 5.79 | 32000 | 0.1702 | 1.0018 | | 0.1259 | 5.81 | 32100 | 0.1664 | 1.0246 | | 0.1259 | 5.82 | 32200 | 0.1737 | 1.0340 | | 0.1259 | 5.84 | 32300 | 0.1742 | 1.0449 | | 0.1259 | 5.86 | 32400 | 0.1707 | 1.0279 | | 0.1273 | 5.88 | 32500 | 0.1697 | 1.0471 | | 0.1273 | 5.9 | 32600 | 0.1668 | 1.0322 | | 0.1273 | 5.92 | 32700 | 0.1706 | 1.0378 | | 0.1273 | 5.93 | 32800 | 0.1704 | 1.0350 | | 0.1273 | 5.95 | 32900 | 0.1725 | 1.0244 | | 0.123 | 5.97 | 33000 | 0.1678 | 1.0447 | | 0.123 | 5.99 | 33100 | 0.1681 | 1.0438 | | 0.123 | 6.01 | 33200 | 0.1689 | 1.0297 | | 0.123 | 6.02 | 33300 | 0.1690 | 1.0333 | | 0.123 | 6.04 | 33400 | 0.1734 | 1.0296 | | 0.1163 | 6.06 | 33500 | 0.1748 | 1.0307 | | 0.1163 | 6.08 | 33600 | 0.1715 | 1.0123 | | 0.1163 | 6.1 | 33700 | 0.1668 | 1.0117 | | 0.1163 | 6.11 | 33800 | 0.1690 | 1.0230 | | 0.1163 | 6.13 | 33900 | 0.1693 | 1.0166 | | 0.1101 | 6.15 | 34000 | 0.1728 | 1.0162 | | 0.1101 | 6.17 | 34100 | 0.1683 | 1.0107 | | 0.1101 | 6.19 | 34200 | 0.1703 | 0.9814 | | 0.1101 | 6.2 | 34300 | 0.1692 | 1.0007 | | 0.1101 | 6.22 | 34400 | 0.1690 | 1.0000 | | 0.1118 | 6.24 | 34500 | 0.1734 | 0.9972 | | 0.1118 | 6.26 | 34600 | 0.1739 | 1.0096 | | 0.1118 | 6.28 | 34700 | 0.1749 | 1.0047 | | 0.1118 | 6.3 | 34800 | 0.1709 | 1.0111 | | 0.1118 | 6.31 | 34900 | 0.1717 | 1.0179 | | 0.1153 | 6.33 | 35000 | 0.1690 | 1.0155 | | 0.1153 | 6.35 | 35100 | 0.1710 | 1.0144 | | 0.1153 | 6.37 | 35200 | 0.1719 | 1.0030 | | 0.1153 | 6.39 | 35300 | 0.1690 | 1.0272 | | 0.1153 | 6.4 | 35400 | 0.1673 | 1.0103 | | 0.1106 | 6.42 | 35500 | 0.1710 | 1.0222 | | 0.1106 | 6.44 | 35600 | 0.1747 | 1.0173 | | 0.1106 | 6.46 | 35700 | 0.1721 | 0.9933 | | 0.1106 | 6.48 | 35800 | 0.1670 | 1.0184 | | 0.1106 | 6.49 | 35900 | 0.1714 | 1.0122 | | 0.1116 | 6.51 | 36000 | 0.1717 | 1.0035 | | 0.1116 | 6.53 | 36100 | 0.1685 | 1.0099 | | 0.1116 | 6.55 | 36200 | 0.1687 | 1.0288 | | 0.1116 | 6.57 | 36300 | 0.1664 | 1.0314 | | 0.1116 | 6.58 | 36400 | 0.1665 | 1.0264 | | 0.1128 | 6.6 | 36500 | 0.1681 | 1.0420 | | 0.1128 | 6.62 | 36600 | 0.1682 | 1.0409 | | 0.1128 | 6.64 | 36700 | 0.1717 | 1.0271 | | 0.1128 | 6.66 | 36800 | 0.1717 | 1.0166 | | 0.1128 | 6.68 | 36900 | 0.1755 | 1.0175 | | 0.1134 | 6.69 | 37000 | 0.1623 | 1.0185 | | 0.1134 | 6.71 | 37100 | 0.1674 | 1.0302 | | 0.1134 | 6.73 | 37200 | 0.1633 | 1.0325 | | 0.1134 | 6.75 | 37300 | 0.1628 | 1.0228 | | 0.1134 | 6.77 | 37400 | 0.1636 | 1.0243 | | 0.1102 | 6.78 | 37500 | 0.1667 | 1.0282 | | 0.1102 | 6.8 | 37600 | 0.1623 | 1.0212 | | 0.1102 | 6.82 | 37700 | 0.1639 | 1.0140 | | 0.1102 | 6.84 | 37800 | 0.1587 | 1.0258 | | 0.1102 | 6.86 | 37900 | 0.1610 | 1.0087 | | 0.1113 | 6.87 | 38000 | 0.1647 | 1.0199 | | 0.1113 | 6.89 | 38100 | 0.1609 | 1.0054 | | 0.1113 | 6.91 | 38200 | 0.1602 | 1.0145 | | 0.1113 | 6.93 | 38300 | 0.1602 | 1.0144 | | 0.1113 | 6.95 | 38400 | 0.1602 | 1.0375 | | 0.1071 | 6.96 | 38500 | 0.1592 | 1.0259 | | 0.1071 | 6.98 | 38600 | 0.1612 | 1.0236 | | 0.1071 | 7.0 | 38700 | 0.1621 | 1.0277 | | 0.1071 | 7.02 | 38800 | 0.1669 | 1.0367 | | 0.1071 | 7.04 | 38900 | 0.1742 | 1.0484 | | 0.1062 | 7.05 | 39000 | 0.1752 | 1.0302 | | 0.1062 | 7.07 | 39100 | 0.1676 | 1.0244 | | 0.1062 | 7.09 | 39200 | 0.1723 | 1.0300 | | 0.1062 | 7.11 | 39300 | 0.1727 | 1.0294 | | 0.1062 | 7.13 | 39400 | 0.1711 | 1.0255 | | 0.1021 | 7.15 | 39500 | 0.1699 | 1.0471 | | 0.1021 | 7.16 | 39600 | 0.1682 | 1.0426 | | 0.1021 | 7.18 | 39700 | 0.1713 | 1.0233 | | 0.1021 | 7.2 | 39800 | 0.1682 | 1.0259 | | 0.1021 | 7.22 | 39900 | 0.1710 | 1.0162 | | 0.103 | 7.24 | 40000 | 0.1725 | 1.0283 | | 0.103 | 7.25 | 40100 | 0.1729 | 1.0264 | | 0.103 | 7.27 | 40200 | 0.1665 | 1.0451 | | 0.103 | 7.29 | 40300 | 0.1671 | 1.0386 | | 0.103 | 7.31 | 40400 | 0.1671 | 1.0316 | | 0.0981 | 7.33 | 40500 | 0.1708 | 1.0257 | | 0.0981 | 7.34 | 40600 | 0.1642 | 1.0152 | | 0.0981 | 7.36 | 40700 | 0.1707 | 1.0110 | | 0.0981 | 7.38 | 40800 | 0.1675 | 1.0186 | | 0.0981 | 7.4 | 40900 | 0.1702 | 1.0123 | | 0.1005 | 7.42 | 41000 | 0.1699 | 1.0159 | | 0.1005 | 7.43 | 41100 | 0.1703 | 1.0219 | | 0.1005 | 7.45 | 41200 | 0.1707 | 1.0194 | | 0.1005 | 7.47 | 41300 | 0.1644 | 1.0016 | | 0.1005 | 7.49 | 41400 | 0.1716 | 0.9941 | | 0.1021 | 7.51 | 41500 | 0.1670 | 1.0159 | | 0.1021 | 7.53 | 41600 | 0.1667 | 1.0033 | | 0.1021 | 7.54 | 41700 | 0.1667 | 1.0176 | | 0.1021 | 7.56 | 41800 | 0.1679 | 1.0194 | | 0.1021 | 7.58 | 41900 | 0.1632 | 1.0418 | | 0.0963 | 7.6 | 42000 | 0.1712 | 1.0152 | | 0.0963 | 7.62 | 42100 | 0.1632 | 1.0364 | | 0.0963 | 7.63 | 42200 | 0.1702 | 1.0229 | | 0.0963 | 7.65 | 42300 | 0.1655 | 1.0179 | | 0.0963 | 7.67 | 42400 | 0.1698 | 1.0329 | | 0.1014 | 7.69 | 42500 | 0.1691 | 1.0398 | | 0.1014 | 7.71 | 42600 | 0.1638 | 1.0487 | | 0.1014 | 7.72 | 42700 | 0.1617 | 1.0210 | | 0.1014 | 7.74 | 42800 | 0.1648 | 1.0124 | | 0.1014 | 7.76 | 42900 | 0.1608 | 1.0202 | | 0.1008 | 7.78 | 43000 | 0.1611 | 1.0353 | | 0.1008 | 7.8 | 43100 | 0.1633 | 1.0319 | | 0.1008 | 7.81 | 43200 | 0.1640 | 1.0032 | | 0.1008 | 7.83 | 43300 | 0.1589 | 0.9985 | | 0.1008 | 7.85 | 43400 | 0.1630 | 0.9975 | | 0.0988 | 7.87 | 43500 | 0.1604 | 1.0053 | | 0.0988 | 7.89 | 43600 | 0.1687 | 1.0063 | | 0.0988 | 7.91 | 43700 | 0.1619 | 1.0096 | | 0.0988 | 7.92 | 43800 | 0.1565 | 0.9901 | | 0.0988 | 7.94 | 43900 | 0.1619 | 0.9742 | | 0.102 | 7.96 | 44000 | 0.1598 | 0.9593 | | 0.102 | 7.98 | 44100 | 0.1635 | 0.9718 | | 0.102 | 8.0 | 44200 | 0.1624 | 0.9903 | | 0.102 | 8.01 | 44300 | 0.1605 | 0.9882 | | 0.102 | 8.03 | 44400 | 0.1657 | 1.0128 | | 0.0961 | 8.05 | 44500 | 0.1651 | 1.0155 | | 0.0961 | 8.07 | 44600 | 0.1680 | 1.0194 | | 0.0961 | 8.09 | 44700 | 0.1694 | 1.0112 | | 0.0961 | 8.1 | 44800 | 0.1665 | 1.0073 | | 0.0961 | 8.12 | 44900 | 0.1612 | 1.0200 | | 0.0894 | 8.14 | 45000 | 0.1652 | 1.0337 | | 0.0894 | 8.16 | 45100 | 0.1626 | 1.0086 | | 0.0894 | 8.18 | 45200 | 0.1639 | 1.0083 | | 0.0894 | 8.19 | 45300 | 0.1634 | 1.0223 | | 0.0894 | 8.21 | 45400 | 0.1631 | 1.0339 | | 0.0887 | 8.23 | 45500 | 0.1640 | 1.0311 | | 0.0887 | 8.25 | 45600 | 0.1661 | 1.0264 | | 0.0887 | 8.27 | 45700 | 0.1650 | 1.0315 | | 0.0887 | 8.29 | 45800 | 0.1624 | 1.0390 | | 0.0887 | 8.3 | 45900 | 0.1624 | 1.0350 | | 0.0884 | 8.32 | 46000 | 0.1615 | 1.0318 | | 0.0884 | 8.34 | 46100 | 0.1628 | 1.0410 | | 0.0884 | 8.36 | 46200 | 0.1627 | 1.0429 | | 0.0884 | 8.38 | 46300 | 0.1644 | 1.0320 | | 0.0884 | 8.39 | 46400 | 0.1633 | 1.0177 | | 0.0893 | 8.41 | 46500 | 0.1654 | 1.0189 | | 0.0893 | 8.43 | 46600 | 0.1598 | 1.0154 | | 0.0893 | 8.45 | 46700 | 0.1618 | 1.0250 | | 0.0893 | 8.47 | 46800 | 0.1639 | 1.0402 | | 0.0893 | 8.48 | 46900 | 0.1616 | 1.0336 | | 0.0869 | 8.5 | 47000 | 0.1613 | 1.0296 | | 0.0869 | 8.52 | 47100 | 0.1648 | 1.0568 | | 0.0869 | 8.54 | 47200 | 0.1625 | 1.0256 | | 0.0869 | 8.56 | 47300 | 0.1609 | 1.0390 | | 0.0869 | 8.57 | 47400 | 0.1606 | 1.0450 | | 0.0894 | 8.59 | 47500 | 0.1605 | 1.0445 | | 0.0894 | 8.61 | 47600 | 0.1660 | 1.0402 | | 0.0894 | 8.63 | 47700 | 0.1618 | 1.0444 | | 0.0894 | 8.65 | 47800 | 0.1669 | 1.0333 | | 0.0894 | 8.66 | 47900 | 0.1627 | 1.0364 | | 0.0885 | 8.68 | 48000 | 0.1616 | 1.0334 | | 0.0885 | 8.7 | 48100 | 0.1626 | 1.0564 | | 0.0885 | 8.72 | 48200 | 0.1624 | 1.0396 | | 0.0885 | 8.74 | 48300 | 0.1623 | 1.0396 | | 0.0885 | 8.76 | 48400 | 0.1612 | 1.0112 | | 0.0888 | 8.77 | 48500 | 0.1638 | 1.0292 | | 0.0888 | 8.79 | 48600 | 0.1639 | 0.9988 | | 0.0888 | 8.81 | 48700 | 0.1618 | 1.0127 | | 0.0888 | 8.83 | 48800 | 0.1584 | 1.0042 | | 0.0888 | 8.85 | 48900 | 0.1615 | 1.0041 | | 0.0887 | 8.86 | 49000 | 0.1637 | 1.0269 | | 0.0887 | 8.88 | 49100 | 0.1627 | 0.9989 | | 0.0887 | 8.9 | 49200 | 0.1583 | 1.0104 | | 0.0887 | 8.92 | 49300 | 0.1600 | 1.0214 | | 0.0887 | 8.94 | 49400 | 0.1599 | 1.0126 | | 0.0893 | 8.95 | 49500 | 0.1595 | 1.0516 | | 0.0893 | 8.97 | 49600 | 0.1625 | 1.0464 | | 0.0893 | 8.99 | 49700 | 0.1595 | 1.0361 | | 0.0893 | 9.01 | 49800 | 0.1614 | 1.0469 | | 0.0893 | 9.03 | 49900 | 0.1612 | 1.0304 | | 0.0834 | 9.04 | 50000 | 0.1643 | 1.0335 | | 0.0834 | 9.06 | 50100 | 0.1640 | 1.0175 | | 0.0834 | 9.08 | 50200 | 0.1655 | 1.0264 | | 0.0834 | 9.1 | 50300 | 0.1678 | 1.0243 | | 0.0834 | 9.12 | 50400 | 0.1659 | 1.0145 | | 0.079 | 9.14 | 50500 | 0.1644 | 1.0316 | | 0.079 | 9.15 | 50600 | 0.1630 | 1.0326 | | 0.079 | 9.17 | 50700 | 0.1634 | 1.0154 | | 0.079 | 9.19 | 50800 | 0.1697 | 1.0095 | | 0.079 | 9.21 | 50900 | 0.1678 | 1.0050 | | 0.078 | 9.23 | 51000 | 0.1626 | 1.0159 | | 0.078 | 9.24 | 51100 | 0.1666 | 1.0238 | | 0.078 | 9.26 | 51200 | 0.1644 | 1.0244 | | 0.078 | 9.28 | 51300 | 0.1655 | 1.0345 | | 0.078 | 9.3 | 51400 | 0.1615 | 1.0237 | | 0.0776 | 9.32 | 51500 | 0.1664 | 1.0180 | | 0.0776 | 9.33 | 51600 | 0.1603 | 1.0208 | | 0.0776 | 9.35 | 51700 | 0.1594 | 1.0230 | | 0.0776 | 9.37 | 51800 | 0.1622 | 1.0201 | | 0.0776 | 9.39 | 51900 | 0.1596 | 1.0039 | | 0.0782 | 9.41 | 52000 | 0.1645 | 1.0204 | | 0.0782 | 9.42 | 52100 | 0.1640 | 1.0318 | | 0.0782 | 9.44 | 52200 | 0.1621 | 1.0290 | | 0.0782 | 9.46 | 52300 | 0.1638 | 1.0318 | | 0.0782 | 9.48 | 52400 | 0.1613 | 1.0217 | | 0.0782 | 9.5 | 52500 | 0.1609 | 1.0261 | | 0.0782 | 9.52 | 52600 | 0.1625 | 1.0101 | | 0.0782 | 9.53 | 52700 | 0.1613 | 1.0058 | | 0.0782 | 9.55 | 52800 | 0.1599 | 1.0068 | | 0.0782 | 9.57 | 52900 | 0.1600 | 1.0110 | | 0.0797 | 9.59 | 53000 | 0.1594 | 1.0171 | | 0.0797 | 9.61 | 53100 | 0.1583 | 1.0124 | | 0.0797 | 9.62 | 53200 | 0.1646 | 1.0093 | | 0.0797 | 9.64 | 53300 | 0.1580 | 1.0201 | | 0.0797 | 9.66 | 53400 | 0.1599 | 1.0207 | | 0.0783 | 9.68 | 53500 | 0.1577 | 1.0226 | | 0.0783 | 9.7 | 53600 | 0.1593 | 1.0160 | | 0.0783 | 9.71 | 53700 | 0.1570 | 1.0173 | | 0.0783 | 9.73 | 53800 | 0.1614 | 1.0299 | | 0.0783 | 9.75 | 53900 | 0.1610 | 1.0184 | | 0.0779 | 9.77 | 54000 | 0.1606 | 1.0173 | | 0.0779 | 9.79 | 54100 | 0.1577 | 1.0032 | | 0.0779 | 9.8 | 54200 | 0.1590 | 1.0070 | | 0.0779 | 9.82 | 54300 | 0.1580 | 1.0257 | | 0.0779 | 9.84 | 54400 | 0.1592 | 1.0108 | | 0.0778 | 9.86 | 54500 | 0.1617 | 0.9907 | | 0.0778 | 9.88 | 54600 | 0.1605 | 1.0189 | | 0.0778 | 9.89 | 54700 | 0.1605 | 1.0177 | | 0.0778 | 9.91 | 54800 | 0.1536 | 1.0275 | | 0.0778 | 9.93 | 54900 | 0.1658 | 1.0282 | | 0.0777 | 9.95 | 55000 | 0.1543 | 1.0385 | | 0.0777 | 9.97 | 55100 | 0.1559 | 1.0375 | | 0.0777 | 9.99 | 55200 | 0.1590 | 1.0215 | | 0.0777 | 10.0 | 55300 | 0.1624 | 1.0242 | | 0.0777 | 10.02 | 55400 | 0.1635 | 1.0244 | | 0.0712 | 10.04 | 55500 | 0.1629 | 1.0298 | | 0.0712 | 10.06 | 55600 | 0.1601 | 1.0299 | | 0.0712 | 10.08 | 55700 | 0.1625 | 1.0117 | | 0.0712 | 10.09 | 55800 | 0.1650 | 1.0233 | | 0.0712 | 10.11 | 55900 | 0.1631 | 1.0061 | | 0.0667 | 10.13 | 56000 | 0.1637 | 1.0226 | | 0.0667 | 10.15 | 56100 | 0.1607 | 1.0042 | | 0.0667 | 10.17 | 56200 | 0.1599 | 1.0117 | | 0.0667 | 10.18 | 56300 | 0.1623 | 1.0246 | | 0.0667 | 10.2 | 56400 | 0.1639 | 1.0294 | | 0.0695 | 10.22 | 56500 | 0.1650 | 1.0232 | | 0.0695 | 10.24 | 56600 | 0.1620 | 1.0289 | | 0.0695 | 10.26 | 56700 | 0.1667 | 1.0209 | | 0.0695 | 10.27 | 56800 | 0.1580 | 1.0163 | | 0.0695 | 10.29 | 56900 | 0.1646 | 1.0293 | | 0.0686 | 10.31 | 57000 | 0.1636 | 1.0106 | | 0.0686 | 10.33 | 57100 | 0.1586 | 1.0044 | | 0.0686 | 10.35 | 57200 | 0.1582 | 1.0213 | | 0.0686 | 10.37 | 57300 | 0.1627 | 1.0151 | | 0.0686 | 10.38 | 57400 | 0.1619 | 1.0248 | | 0.0686 | 10.4 | 57500 | 0.1596 | 1.0098 | | 0.0686 | 10.42 | 57600 | 0.1606 | 1.0031 | | 0.0686 | 10.44 | 57700 | 0.1620 | 1.0046 | | 0.0686 | 10.46 | 57800 | 0.1592 | 1.0018 | | 0.0686 | 10.47 | 57900 | 0.1592 | 1.0058 | | 0.0669 | 10.49 | 58000 | 0.1605 | 0.9961 | | 0.0669 | 10.51 | 58100 | 0.1632 | 1.0102 | | 0.0669 | 10.53 | 58200 | 0.1593 | 1.0061 | | 0.0669 | 10.55 | 58300 | 0.1586 | 1.0091 | | 0.0669 | 10.56 | 58400 | 0.1603 | 1.0085 | | 0.068 | 10.58 | 58500 | 0.1579 | 1.0031 | | 0.068 | 10.6 | 58600 | 0.1591 | 1.0021 | | 0.068 | 10.62 | 58700 | 0.1590 | 1.0163 | | 0.068 | 10.64 | 58800 | 0.1584 | 1.0045 | | 0.068 | 10.65 | 58900 | 0.1594 | 1.0158 | | 0.0693 | 10.67 | 59000 | 0.1568 | 1.0052 | | 0.0693 | 10.69 | 59100 | 0.1581 | 0.9955 | | 0.0693 | 10.71 | 59200 | 0.1622 | 0.9917 | | 0.0693 | 10.73 | 59300 | 0.1580 | 1.0018 | | 0.0693 | 10.75 | 59400 | 0.1601 | 1.0077 | | 0.0699 | 10.76 | 59500 | 0.1605 | 0.9997 | | 0.0699 | 10.78 | 59600 | 0.1585 | 1.0009 | | 0.0699 | 10.8 | 59700 | 0.1541 | 1.0058 | | 0.0699 | 10.82 | 59800 | 0.1583 | 1.0026 | | 0.0699 | 10.84 | 59900 | 0.1592 | 0.9992 | | 0.0671 | 10.85 | 60000 | 0.1590 | 1.0004 | | 0.0671 | 10.87 | 60100 | 0.1585 | 1.0060 | | 0.0671 | 10.89 | 60200 | 0.1579 | 1.0063 | | 0.0671 | 10.91 | 60300 | 0.1582 | 0.9949 | | 0.0671 | 10.93 | 60400 | 0.1562 | 1.0004 | | 0.0661 | 10.94 | 60500 | 0.1560 | 0.9950 | | 0.0661 | 10.96 | 60600 | 0.1564 | 0.9990 | | 0.0661 | 10.98 | 60700 | 0.1552 | 0.9982 | | 0.0661 | 11.0 | 60800 | 0.1596 | 1.0018 | | 0.0661 | 11.02 | 60900 | 0.1618 | 0.9905 | | 0.0634 | 11.03 | 61000 | 0.1652 | 0.9890 | | 0.0634 | 11.05 | 61100 | 0.1649 | 0.9886 | | 0.0634 | 11.07 | 61200 | 0.1668 | 0.9870 | | 0.0634 | 11.09 | 61300 | 0.1663 | 0.9921 | | 0.0634 | 11.11 | 61400 | 0.1650 | 0.9919 | | 0.0587 | 11.13 | 61500 | 0.1674 | 0.9831 | | 0.0587 | 11.14 | 61600 | 0.1633 | 0.9793 | | 0.0587 | 11.16 | 61700 | 0.1665 | 0.9781 | | 0.0587 | 11.18 | 61800 | 0.1642 | 0.9821 | | 0.0587 | 11.2 | 61900 | 0.1638 | 0.9797 | | 0.0581 | 11.22 | 62000 | 0.1628 | 0.9727 | | 0.0581 | 11.23 | 62100 | 0.1661 | 0.9796 | | 0.0581 | 11.25 | 62200 | 0.1641 | 0.9830 | | 0.0581 | 11.27 | 62300 | 0.1601 | 0.9867 | | 0.0581 | 11.29 | 62400 | 0.1626 | 0.9757 | | 0.0584 | 11.31 | 62500 | 0.1632 | 1.0014 | | 0.0584 | 11.32 | 62600 | 0.1626 | 1.0052 | | 0.0584 | 11.34 | 62700 | 0.1586 | 1.0098 | | 0.0584 | 11.36 | 62800 | 0.1597 | 1.0151 | | 0.0584 | 11.38 | 62900 | 0.1624 | 1.0054 | | 0.0589 | 11.4 | 63000 | 0.1618 | 1.0018 | | 0.0589 | 11.41 | 63100 | 0.1635 | 1.0032 | | 0.0589 | 11.43 | 63200 | 0.1654 | 1.0142 | | 0.0589 | 11.45 | 63300 | 0.1646 | 1.0031 | | 0.0589 | 11.47 | 63400 | 0.1618 | 1.0118 | | 0.0579 | 11.49 | 63500 | 0.1634 | 1.0218 | | 0.0579 | 11.51 | 63600 | 0.1616 | 1.0179 | | 0.0579 | 11.52 | 63700 | 0.1603 | 1.0036 | | 0.0579 | 11.54 | 63800 | 0.1610 | 1.0150 | | 0.0579 | 11.56 | 63900 | 0.1605 | 1.0285 | | 0.0572 | 11.58 | 64000 | 0.1621 | 1.0261 | | 0.0572 | 11.6 | 64100 | 0.1625 | 1.0252 | | 0.0572 | 11.61 | 64200 | 0.1677 | 1.0257 | | 0.0572 | 11.63 | 64300 | 0.1656 | 1.0243 | | 0.0572 | 11.65 | 64400 | 0.1669 | 1.0270 | | 0.0592 | 11.67 | 64500 | 0.1605 | 1.0305 | | 0.0592 | 11.69 | 64600 | 0.1633 | 1.0277 | | 0.0592 | 11.7 | 64700 | 0.1606 | 1.0176 | | 0.0592 | 11.72 | 64800 | 0.1618 | 1.0249 | | 0.0592 | 11.74 | 64900 | 0.1609 | 1.0113 | | 0.0595 | 11.76 | 65000 | 0.1609 | 1.0254 | | 0.0595 | 11.78 | 65100 | 0.1662 | 1.0275 | | 0.0595 | 11.79 | 65200 | 0.1652 | 1.0164 | | 0.0595 | 11.81 | 65300 | 0.1638 | 1.0266 | | 0.0595 | 11.83 | 65400 | 0.1589 | 1.0274 | | 0.0588 | 11.85 | 65500 | 0.1607 | 1.0136 | | 0.0588 | 11.87 | 65600 | 0.1592 | 1.0136 | | 0.0588 | 11.88 | 65700 | 0.1581 | 1.0183 | | 0.0588 | 11.9 | 65800 | 0.1587 | 1.0133 | | 0.0588 | 11.92 | 65900 | 0.1596 | 1.0170 | | 0.0558 | 11.94 | 66000 | 0.1590 | 1.0161 | | 0.0558 | 11.96 | 66100 | 0.1597 | 1.0193 | | 0.0558 | 11.98 | 66200 | 0.1590 | 1.0193 | | 0.0558 | 11.99 | 66300 | 0.1608 | 1.0242 | | 0.0558 | 12.01 | 66400 | 0.1642 | 1.0231 | | 0.0555 | 12.03 | 66500 | 0.1679 | 1.0168 | | 0.0555 | 12.05 | 66600 | 0.1674 | 1.0083 | | 0.0555 | 12.07 | 66700 | 0.1658 | 1.0069 | | 0.0555 | 12.08 | 66800 | 0.1661 | 1.0134 | | 0.0555 | 12.1 | 66900 | 0.1682 | 1.0274 | | 0.0508 | 12.12 | 67000 | 0.1702 | 1.0219 | | 0.0508 | 12.14 | 67100 | 0.1694 | 1.0219 | | 0.0508 | 12.16 | 67200 | 0.1667 | 1.0236 | | 0.0508 | 12.17 | 67300 | 0.1672 | 1.0253 | | 0.0508 | 12.19 | 67400 | 0.1640 | 1.0215 | | 0.0513 | 12.21 | 67500 | 0.1649 | 1.0242 | | 0.0513 | 12.23 | 67600 | 0.1687 | 1.0262 | | 0.0513 | 12.25 | 67700 | 0.1655 | 1.0231 | | 0.0513 | 12.26 | 67800 | 0.1692 | 1.0176 | | 0.0513 | 12.28 | 67900 | 0.1675 | 1.0202 | | 0.0519 | 12.3 | 68000 | 0.1644 | 1.0241 | | 0.0519 | 12.32 | 68100 | 0.1651 | 1.0297 | | 0.0519 | 12.34 | 68200 | 0.1661 | 1.0287 | | 0.0519 | 12.36 | 68300 | 0.1665 | 1.0257 | | 0.0519 | 12.37 | 68400 | 0.1685 | 1.0233 | | 0.0522 | 12.39 | 68500 | 0.1636 | 1.0177 | | 0.0522 | 12.41 | 68600 | 0.1709 | 1.0200 | | 0.0522 | 12.43 | 68700 | 0.1684 | 1.0164 | | 0.0522 | 12.45 | 68800 | 0.1666 | 1.0119 | | 0.0522 | 12.46 | 68900 | 0.1683 | 1.0136 | | 0.05 | 12.48 | 69000 | 0.1696 | 1.0127 | | 0.05 | 12.5 | 69100 | 0.1708 | 1.0184 | | 0.05 | 12.52 | 69200 | 0.1654 | 1.0282 | | 0.05 | 12.54 | 69300 | 0.1700 | 1.0235 | | 0.05 | 12.55 | 69400 | 0.1688 | 1.0257 | | 0.0513 | 12.57 | 69500 | 0.1646 | 1.0274 | | 0.0513 | 12.59 | 69600 | 0.1660 | 1.0247 | | 0.0513 | 12.61 | 69700 | 0.1657 | 1.0188 | | 0.0513 | 12.63 | 69800 | 0.1654 | 1.0087 | | 0.0513 | 12.64 | 69900 | 0.1681 | 1.0146 | | 0.0512 | 12.66 | 70000 | 0.1660 | 1.0185 | | 0.0512 | 12.68 | 70100 | 0.1690 | 1.0214 | | 0.0512 | 12.7 | 70200 | 0.1683 | 1.0160 | | 0.0512 | 12.72 | 70300 | 0.1695 | 1.0198 | | 0.0512 | 12.74 | 70400 | 0.1666 | 1.0193 | | 0.0484 | 12.75 | 70500 | 0.1654 | 1.0142 | | 0.0484 | 12.77 | 70600 | 0.1598 | 1.0154 | | 0.0484 | 12.79 | 70700 | 0.1623 | 1.0139 | | 0.0484 | 12.81 | 70800 | 0.1662 | 1.0180 | | 0.0484 | 12.83 | 70900 | 0.1659 | 1.0232 | | 0.0501 | 12.84 | 71000 | 0.1662 | 1.0202 | | 0.0501 | 12.86 | 71100 | 0.1639 | 1.0161 | | 0.0501 | 12.88 | 71200 | 0.1666 | 1.0151 | | 0.0501 | 12.9 | 71300 | 0.1644 | 1.0129 | | 0.0501 | 12.92 | 71400 | 0.1642 | 1.0171 | | 0.0482 | 12.93 | 71500 | 0.1635 | 1.0162 | | 0.0482 | 12.95 | 71600 | 0.1637 | 1.0186 | | 0.0482 | 12.97 | 71700 | 0.1639 | 1.0142 | | 0.0482 | 12.99 | 71800 | 0.1643 | 1.0122 | | 0.0482 | 13.01 | 71900 | 0.1679 | 1.0156 | | 0.0483 | 13.02 | 72000 | 0.1717 | 1.0224 | | 0.0483 | 13.04 | 72100 | 0.1742 | 1.0229 | | 0.0483 | 13.06 | 72200 | 0.1718 | 1.0237 | | 0.0483 | 13.08 | 72300 | 0.1742 | 1.0266 | | 0.0483 | 13.1 | 72400 | 0.1736 | 1.0257 | | 0.0443 | 13.12 | 72500 | 0.1741 | 1.0275 | | 0.0443 | 13.13 | 72600 | 0.1745 | 1.0325 | | 0.0443 | 13.15 | 72700 | 0.1737 | 1.0296 | | 0.0443 | 13.17 | 72800 | 0.1722 | 1.0303 | | 0.0443 | 13.19 | 72900 | 0.1702 | 1.0305 | | 0.0424 | 13.21 | 73000 | 0.1733 | 1.0241 | | 0.0424 | 13.22 | 73100 | 0.1748 | 1.0243 | | 0.0424 | 13.24 | 73200 | 0.1760 | 1.0231 | | 0.0424 | 13.26 | 73300 | 0.1745 | 1.0241 | | 0.0424 | 13.28 | 73400 | 0.1772 | 1.0217 | | 0.0424 | 13.3 | 73500 | 0.1755 | 1.0206 | | 0.0424 | 13.31 | 73600 | 0.1743 | 1.0242 | | 0.0424 | 13.33 | 73700 | 0.1738 | 1.0208 | | 0.0424 | 13.35 | 73800 | 0.1736 | 1.0249 | | 0.0424 | 13.37 | 73900 | 0.1747 | 1.0271 | | 0.0437 | 13.39 | 74000 | 0.1707 | 1.0241 | | 0.0437 | 13.4 | 74100 | 0.1731 | 1.0269 | | 0.0437 | 13.42 | 74200 | 0.1743 | 1.0290 | | 0.0437 | 13.44 | 74300 | 0.1739 | 1.0266 | | 0.0437 | 13.46 | 74400 | 0.1763 | 1.0246 | | 0.0443 | 13.48 | 74500 | 0.1724 | 1.0209 | | 0.0443 | 13.49 | 74600 | 0.1744 | 1.0244 | | 0.0443 | 13.51 | 74700 | 0.1717 | 1.0232 | | 0.0443 | 13.53 | 74800 | 0.1754 | 1.0217 | | 0.0443 | 13.55 | 74900 | 0.1721 | 1.0234 | | 0.0435 | 13.57 | 75000 | 0.1751 | 1.0197 | | 0.0435 | 13.59 | 75100 | 0.1727 | 1.0285 | | 0.0435 | 13.6 | 75200 | 0.1715 | 1.0221 | | 0.0435 | 13.62 | 75300 | 0.1746 | 1.0247 | | 0.0435 | 13.64 | 75400 | 0.1712 | 1.0231 | | 0.0436 | 13.66 | 75500 | 0.1719 | 1.0228 | | 0.0436 | 13.68 | 75600 | 0.1727 | 1.0197 | | 0.0436 | 13.69 | 75700 | 0.1750 | 1.0252 | | 0.0436 | 13.71 | 75800 | 0.1702 | 1.0241 | | 0.0436 | 13.73 | 75900 | 0.1720 | 1.0250 | | 0.0433 | 13.75 | 76000 | 0.1744 | 1.0210 | | 0.0433 | 13.77 | 76100 | 0.1735 | 1.0211 | | 0.0433 | 13.78 | 76200 | 0.1727 | 1.0205 | | 0.0433 | 13.8 | 76300 | 0.1706 | 1.0218 | | 0.0433 | 13.82 | 76400 | 0.1709 | 1.0238 | | 0.0431 | 13.84 | 76500 | 0.1705 | 1.0197 | | 0.0431 | 13.86 | 76600 | 0.1734 | 1.0223 | | 0.0431 | 13.87 | 76700 | 0.1695 | 1.0250 | | 0.0431 | 13.89 | 76800 | 0.1734 | 1.0232 | | 0.0431 | 13.91 | 76900 | 0.1724 | 1.0219 | | 0.041 | 13.93 | 77000 | 0.1706 | 1.0236 | | 0.041 | 13.95 | 77100 | 0.1689 | 1.0220 | | 0.041 | 13.97 | 77200 | 0.1738 | 1.0230 | | 0.041 | 13.98 | 77300 | 0.1727 | 1.0254 | | 0.041 | 14.0 | 77400 | 0.1721 | 1.0261 | | 0.041 | 14.02 | 77500 | 0.1760 | 1.0261 | | 0.041 | 14.04 | 77600 | 0.1772 | 1.0202 | | 0.041 | 14.06 | 77700 | 0.1782 | 1.0202 | | 0.041 | 14.07 | 77800 | 0.1777 | 1.0222 | | 0.041 | 14.09 | 77900 | 0.1787 | 1.0203 | | 0.0383 | 14.11 | 78000 | 0.1790 | 1.0236 | | 0.0383 | 14.13 | 78100 | 0.1812 | 1.0245 | | 0.0383 | 14.15 | 78200 | 0.1778 | 1.0224 | | 0.0383 | 14.16 | 78300 | 0.1771 | 1.0231 | | 0.0383 | 14.18 | 78400 | 0.1782 | 1.0242 | | 0.0391 | 14.2 | 78500 | 0.1785 | 1.0262 | | 0.0391 | 14.22 | 78600 | 0.1791 | 1.0261 | | 0.0391 | 14.24 | 78700 | 0.1770 | 1.0254 | | 0.0391 | 14.25 | 78800 | 0.1810 | 1.0257 | | 0.0391 | 14.27 | 78900 | 0.1794 | 1.0241 | | 0.0387 | 14.29 | 79000 | 0.1774 | 1.0256 | | 0.0387 | 14.31 | 79100 | 0.1774 | 1.0236 | | 0.0387 | 14.33 | 79200 | 0.1759 | 1.0222 | | 0.0387 | 14.35 | 79300 | 0.1787 | 1.0237 | | 0.0387 | 14.36 | 79400 | 0.1788 | 1.0227 | | 0.0372 | 14.38 | 79500 | 0.1789 | 1.0232 | | 0.0372 | 14.4 | 79600 | 0.1771 | 1.0254 | | 0.0372 | 14.42 | 79700 | 0.1777 | 1.0244 | | 0.0372 | 14.44 | 79800 | 0.1791 | 1.0225 | | 0.0372 | 14.45 | 79900 | 0.1786 | 1.0237 | | 0.0385 | 14.47 | 80000 | 0.1782 | 1.0243 | | 0.0385 | 14.49 | 80100 | 0.1770 | 1.0236 | | 0.0385 | 14.51 | 80200 | 0.1782 | 1.0240 | | 0.0385 | 14.53 | 80300 | 0.1764 | 1.0243 | | 0.0385 | 14.54 | 80400 | 0.1748 | 1.0248 | | 0.039 | 14.56 | 80500 | 0.1758 | 1.0232 | | 0.039 | 14.58 | 80600 | 0.1763 | 1.0246 | | 0.039 | 14.6 | 80700 | 0.1770 | 1.0220 | | 0.039 | 14.62 | 80800 | 0.1788 | 1.0225 | | 0.039 | 14.63 | 80900 | 0.1781 | 1.0230 | | 0.039 | 14.65 | 81000 | 0.1779 | 1.0230 | | 0.039 | 14.67 | 81100 | 0.1755 | 1.0212 | | 0.039 | 14.69 | 81200 | 0.1765 | 1.0226 | | 0.039 | 14.71 | 81300 | 0.1787 | 1.0241 | | 0.039 | 14.72 | 81400 | 0.1782 | 1.0250 | | 0.0368 | 14.74 | 81500 | 0.1780 | 1.0248 | | 0.0368 | 14.76 | 81600 | 0.1782 | 1.0242 | | 0.0368 | 14.78 | 81700 | 0.1782 | 1.0242 | | 0.0368 | 14.8 | 81800 | 0.1792 | 1.0241 | | 0.0368 | 14.82 | 81900 | 0.1796 | 1.0238 | | 0.0378 | 14.83 | 82000 | 0.1795 | 1.0236 | | 0.0378 | 14.85 | 82100 | 0.1796 | 1.0239 | | 0.0378 | 14.87 | 82200 | 0.1792 | 1.0236 | | 0.0378 | 14.89 | 82300 | 0.1789 | 1.0239 | | 0.0378 | 14.91 | 82400 | 0.1788 | 1.0238 | | 0.0386 | 14.92 | 82500 | 0.1787 | 1.0239 | | 0.0386 | 14.94 | 82600 | 0.1786 | 1.0236 | | 0.0386 | 14.96 | 82700 | 0.1786 | 1.0237 | | 0.0386 | 14.98 | 82800 | 0.1787 | 1.0239 | | 0.0386 | 15.0 | 82900 | 0.1788 | 1.0238 | ### Framework versions - Transformers 4.16.0.dev0 - Pytorch 1.10.1 - Datasets 1.17.0 - Tokenizers 0.10.3
{"language": ["es"], "license": "apache-2.0", "tags": ["automatic-speech-recognition", "common_voice", "generated_from_trainer"], "datasets": ["common_voice"], "model-index": [{"name": "wav2vec2-common_voice-es-demo", "results": []}]}
automatic-speech-recognition
gabrieljg/wav2vec2-common_voice-es-demo
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "common_voice", "generated_from_trainer", "es", "dataset:common_voice", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "es" ]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #common_voice #generated_from_trainer #es #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us
wav2vec2-common\_voice-es-demo ============================== This model is a fine-tuned version of facebook/wav2vec2-large-xlsr-53 on the COMMON\_VOICE - ES dataset. It achieves the following results on the evaluation set: * Loss: 0.1788 * Wer: 1.0239 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.0003 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * gradient\_accumulation\_steps: 2 * total\_train\_batch\_size: 32 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 500 * num\_epochs: 15.0 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.16.0.dev0 * Pytorch 1.10.1 * Datasets 1.17.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 15.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.0.dev0\n* Pytorch 1.10.1\n* Datasets 1.17.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #common_voice #generated_from_trainer #es #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 15.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.0.dev0\n* Pytorch 1.10.1\n* Datasets 1.17.0\n* Tokenizers 0.10.3" ]
[ 69, 159, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #common_voice #generated_from_trainer #es #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 15.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.16.0.dev0\n* Pytorch 1.10.1\n* Datasets 1.17.0\n* Tokenizers 0.10.3" ]
[ -0.10831723362207413, 0.06830528378486633, -0.003474173601716757, 0.037817761301994324, 0.11993183940649033, 0.01966193877160549, 0.0930694192647934, 0.1449575424194336, -0.09907979518175125, 0.07138242572546005, 0.08626111596822739, 0.06868524104356766, 0.06436414271593094, 0.09930981695652008, -0.0158182755112648, -0.3102611303329468, 0.012758049182593822, 0.001773053198121488, -0.09621258825063705, 0.10495293885469437, 0.10441071540117264, -0.10490093380212784, 0.007436338346451521, 0.018276624381542206, -0.1076497882604599, 0.004678310360759497, -0.022968964651226997, -0.054814886301755905, 0.131913423538208, 0.0599164217710495, 0.08263958245515823, 0.02398690953850746, 0.08886335045099258, -0.28757351636886597, 0.016790928319096565, 0.05135900527238846, 0.042523544281721115, 0.06704649329185486, 0.09855291247367859, -0.010859531350433826, 0.13870806992053986, -0.06371452659368515, 0.0697844997048378, 0.05277116596698761, -0.09592412412166595, -0.3314160108566284, -0.07925018668174744, 0.027668653056025505, 0.13551756739616394, 0.09852837771177292, -0.03988182172179222, 0.058063969016075134, -0.08764515817165375, 0.09884899854660034, 0.22598247230052948, -0.23406986892223358, -0.06994196027517319, -0.028179632499814034, 0.05215078964829445, 0.035188473761081696, -0.10538091510534286, -0.02477641962468624, 0.025311648845672607, 0.04645022749900818, 0.08877574652433395, 0.008861773647367954, -0.04531579837203026, 0.01258801482617855, -0.13557247817516327, -0.04447941854596138, 0.12202378362417221, 0.08024346083402634, -0.024667873978614807, -0.10459505021572113, -0.012278122827410698, -0.20187950134277344, -0.04811130091547966, 0.002905498258769512, 0.024515293538570404, -0.026451444253325462, -0.09491954743862152, 0.020099444314837456, -0.08390095829963684, -0.08914653211832047, 0.010090618394315243, 0.13410231471061707, 0.04416476562619209, -0.044467683881521225, 0.009347688406705856, 0.09918497502803802, 0.03955623507499695, -0.13077165186405182, 0.010761630721390247, 0.05315578728914261, -0.1038317158818245, -0.018513156101107597, -0.038908932358026505, -0.06149539351463318, 0.016324568539857864, 0.10374373197555542, -0.0234654750674963, 0.08343681693077087, 0.0009019221761263907, 0.025413725525140762, -0.0632857158780098, 0.15582840144634247, -0.04447588697075844, -0.05947999283671379, -0.04594336822628975, 0.08710236847400665, -0.004553183447569609, -0.01664572022855282, -0.07720998674631119, 0.014777365140616894, 0.10630109161138535, 0.05148376524448395, -0.011971051804721355, 0.008435518480837345, -0.07708794623613358, -0.01659688539803028, -0.012715736404061317, -0.1019042506814003, 0.040633633732795715, 0.03988751396536827, -0.04593510925769806, 0.0012052390957251191, 0.006522228941321373, 0.03801446780562401, -0.019661566242575645, 0.12318692356348038, -0.04321301728487015, 0.005582091864198446, -0.06223510578274727, -0.11129751056432724, 0.03477193042635918, -0.007921443320810795, 0.0026713211555033922, -0.06816652417182922, -0.08297428488731384, -0.053992386907339096, 0.05280296877026558, -0.04545700177550316, -0.0841078832745552, -0.07678450644016266, -0.05590914562344551, 0.052245210856199265, -0.027726516127586365, 0.1915205419063568, -0.06348779797554016, 0.11393450945615768, 0.019558802247047424, 0.030686378479003906, 0.044964827597141266, 0.08022448420524597, -0.03517642617225647, 0.031828343868255615, -0.13135206699371338, 0.08814236521720886, -0.08095067739486694, 0.030818797647953033, -0.1445338875055313, -0.1211397647857666, -0.004442253150045872, 0.0022154701873660088, 0.09853255748748779, 0.09714207053184509, -0.2067231386899948, -0.10269375145435333, 0.15956634283065796, -0.07452615350484848, -0.07767918705940247, 0.16360726952552795, -0.03107190690934658, -0.01857486553490162, 0.04613308236002922, 0.17396758496761322, 0.10465142875909805, -0.09365088492631912, 0.03137556463479996, -0.049192845821380615, 0.13515609502792358, 0.05176440626382828, 0.09539193660020828, -0.05548439919948578, 0.009541671723127365, -0.002068739617243409, -0.009470123797655106, 0.08230984956026077, -0.09076303243637085, -0.0834953561425209, -0.013305644504725933, -0.07365052402019501, 0.027919024229049683, 0.053029775619506836, 0.01589266024529934, -0.10185357183218002, -0.12359708547592163, 0.02335042878985405, 0.11463901400566101, -0.10368002206087112, 0.04362734034657478, -0.06612618267536163, 0.03729669004678726, -0.006248949095606804, -0.011733888648450375, -0.16197346150875092, 0.023226333782076836, 0.030246146023273468, -0.04803089424967766, 0.03827216476202011, 0.0013054298469796777, 0.06908828020095825, 0.04838966205716133, -0.06818457692861557, -0.06381192058324814, -0.055688485503196716, 0.007658648304641247, -0.07726546376943588, -0.25305303931236267, -0.07110034674406052, -0.032030947506427765, 0.16850735247135162, -0.2114313244819641, 0.0016228208551183343, 0.014926745556294918, 0.11735499650239944, 0.03010910004377365, -0.045939620584249496, -0.005329027771949768, 0.09745728224515915, -0.016911525279283524, -0.05122638866305351, 0.03304455056786537, 0.007460631430149078, -0.14651061594486237, 0.013410378247499466, -0.12545688450336456, 0.06873268634080887, 0.09670870006084442, -0.025133566930890083, -0.09625978767871857, -0.06966759264469147, -0.05798708274960518, -0.06327987462282181, -0.033764224499464035, 0.0032118838280439377, 0.21545132994651794, 0.03007577918469906, 0.11055127531290054, -0.06145593523979187, -0.037076640874147415, 0.032584358006715775, 0.008867933414876461, -0.011847802437841892, 0.1371697038412094, 0.06138702481985092, -0.056667327880859375, 0.08856123685836792, 0.07555323094129562, -0.07159633189439774, 0.15739616751670837, -0.0707598403096199, -0.10150160640478134, -0.03446941077709198, 0.015546586364507675, 0.017583588138222694, 0.10048343986272812, -0.17529939115047455, -0.006438846699893475, 0.01984899677336216, 0.025706062093377113, 0.023376060649752617, -0.20395146310329437, -0.006023449823260307, 0.05145290121436119, -0.07002629339694977, -0.03821225464344025, -0.01560446247458458, -0.003932872787117958, 0.08460165560245514, 0.0062621524557471275, -0.060212116688489914, -0.012116771191358566, -0.030398819595575333, -0.09135410934686661, 0.17608146369457245, -0.1164572536945343, -0.13701999187469482, -0.12126914411783218, -0.046278756111860275, 0.00213300297036767, -0.013592981733381748, 0.07016966491937637, -0.12045154720544815, -0.028301585465669632, -0.05753050372004509, 0.05907348170876503, -0.08462769538164139, 0.029425831511616707, -0.024134427309036255, 0.005845377687364817, 0.06774943321943283, -0.1066851019859314, 0.020593659952282906, -0.00958868209272623, -0.01919555477797985, 0.02050091326236725, 0.03855849429965019, 0.07901467382907867, 0.1764761060476303, 0.048394881188869476, 0.006800372153520584, -0.056989431381225586, 0.1636185497045517, -0.1158023253083229, -0.030640484765172005, 0.10535204410552979, -0.002588635543361306, 0.028726449236273766, 0.15118125081062317, 0.05658270791172981, -0.07461432367563248, 0.017665384337306023, 0.03124394454061985, -0.009905984625220299, -0.2507857084274292, -0.047920096665620804, -0.07332775741815567, -0.016197973862290382, 0.08965589851140976, 0.02721291035413742, -0.024317419156432152, 0.01021570060402155, -0.03051355481147766, 0.0013623705599457026, 0.018870985135436058, 0.05844006687402725, 0.10391008853912354, 0.024335216730833054, 0.11510367691516876, -0.012725857086479664, -0.028250640258193016, 0.03531026840209961, -0.009337464347481728, 0.22521409392356873, 0.014013861306011677, 0.16718868911266327, 0.0517808273434639, 0.15973544120788574, 0.01001548022031784, 0.036570265889167786, 0.025860287249088287, -0.015304247848689556, 0.011940802447497845, -0.05201875790953636, -0.03835636377334595, 0.033818040043115616, 0.1316852569580078, 0.023939764127135277, -0.1272224336862564, -0.03193547576665878, 0.008775580674409866, 0.35453590750694275, 0.06813634186983109, -0.25529956817626953, -0.09797059744596481, 0.014611547812819481, -0.09006814658641815, -0.03559667617082596, 0.02668871358036995, 0.1060703694820404, -0.09904175996780396, 0.0673568993806839, -0.051959291100502014, 0.09268766641616821, -0.05742403119802475, 0.011669819243252277, 0.05433938652276993, 0.069215789437294, -0.0013723537558689713, 0.07155872136354446, -0.27110475301742554, 0.30417323112487793, -0.012394336983561516, 0.07048985362052917, -0.03634205088019371, 0.03899301216006279, 0.023809615522623062, -0.04179012030363083, 0.05563245713710785, -0.012102687731385231, -0.11857149004936218, -0.17940106987953186, -0.061286017298698425, 0.021484531462192535, 0.11614590138196945, -0.04833691567182541, 0.10964841395616531, -0.030623266473412514, -0.014272195287048817, 0.061229806393384933, -0.04192117229104042, -0.1058315634727478, -0.11506275087594986, 0.00971479807049036, 0.04477864131331444, 0.0932001918554306, -0.09720370173454285, -0.10814916342496872, -0.08103442937135696, 0.1720251888036728, -0.0819949060678482, -0.005564010236412287, -0.1157938614487648, 0.09520141780376434, 0.16175860166549683, -0.06841330230236053, 0.04858282208442688, 0.026870405301451683, 0.1112864539027214, 0.009206279180943966, -0.000019882469132426195, 0.11635927110910416, -0.0811271145939827, -0.19682981073856354, -0.06834494322538376, 0.1861848533153534, 0.04502066224813461, 0.09196116775274277, -0.030157823115587234, 0.02788425050675869, -0.020168302580714226, -0.05855527147650719, 0.064134381711483, 0.04324949160218239, 0.0023237282875925303, 0.07569321244955063, -0.036649297922849655, -0.040821872651576996, -0.07994929701089859, -0.10307922959327698, 0.17434492707252502, 0.30480414628982544, -0.08883696794509888, 0.06712683290243149, 0.051246047019958496, -0.05267278477549553, -0.12391749024391174, 0.0019549974240362644, 0.13891898095607758, 0.045499902218580246, 0.029110783711075783, -0.22536398470401764, 0.025080474093556404, 0.07752583920955658, -0.018609024584293365, 0.04656969755887985, -0.3332669138908386, -0.1354578137397766, 0.12103323638439178, 0.07901634275913239, -0.016339998692274094, -0.14046256244182587, -0.053632210940122604, -0.02570459619164467, -0.10088080167770386, 0.03724873065948486, -0.01957506686449051, 0.13339458405971527, 0.010505287908017635, 0.059681519865989685, 0.029939789324998856, -0.04034380987286568, 0.14256663620471954, -0.02721017599105835, 0.04390588402748108, -0.00823727436363697, 0.0414082296192646, -0.04472815617918968, -0.03751303628087044, -0.007486490532755852, -0.09345754235982895, 0.012069394811987877, -0.11565303802490234, -0.038222700357437134, -0.07961530238389969, 0.01598532125353813, -0.04094555228948593, -0.05008118972182274, -0.02443070523440838, 0.03287340700626373, 0.07118701934814453, -0.0033476834651082754, 0.11196926981210709, -0.06656856834888458, 0.1731892079114914, 0.08553014695644379, 0.10673024505376816, 0.0048268320970237255, -0.09770970791578293, -0.01796826906502247, -0.028957802802324295, 0.04789678752422333, -0.09813667833805084, 0.03157075121998787, 0.13401491940021515, 0.04645487293601036, 0.16077116131782532, 0.04824754223227501, -0.08389776945114136, 0.02226494625210762, 0.054662153124809265, -0.06968826800584793, -0.15052033960819244, -0.016705846413969994, 0.059961430728435516, -0.13983935117721558, -0.018306046724319458, 0.11525038629770279, -0.045658260583877563, -0.020915722474455833, 0.017311403527855873, 0.03008919022977352, -0.06133439019322395, 0.2305586189031601, 0.00010442999337101355, 0.07750128954648972, -0.09482063353061676, 0.062322914600372314, 0.07279013842344284, -0.17547759413719177, 0.03973294794559479, 0.07552798837423325, -0.03329210355877876, -0.02722143568098545, 0.041215475648641586, 0.08671603351831436, 0.02554658241569996, -0.05260971561074257, -0.0979870930314064, -0.16400203108787537, 0.08481553196907043, 0.07439867407083511, 0.02972145937383175, 0.024951016530394554, -0.04631530120968819, 0.044321827590465546, -0.09641663730144501, 0.09686674177646637, 0.10047457367181778, 0.06021999567747116, -0.12255950272083282, 0.1685749590396881, 0.016713101416826248, -0.0028610078152269125, 0.009997155517339706, -0.01195724867284298, -0.08377803862094879, 0.037159234285354614, -0.12986432015895844, -0.030253790318965912, -0.04593480005860329, 0.008046645671129227, 0.013397793285548687, -0.053396835923194885, -0.03826945647597313, 0.019396327435970306, -0.11724293231964111, -0.04231038689613342, -0.02031642012298107, 0.07996401190757751, -0.09450695663690567, -0.02714126743376255, 0.04086510092020035, -0.10014671832323074, 0.09176506847143173, 0.06321773678064346, 0.016718588769435883, 0.037725869566202164, -0.12777255475521088, -0.002954186173155904, 0.0451122410595417, 0.0025302995927631855, 0.009342875331640244, -0.19321414828300476, -0.025229094550013542, -0.012308198027312756, 0.024485507979989052, -0.0025290451012551785, 0.03509903699159622, -0.12913557887077332, -0.04532093554735184, -0.03248606622219086, -0.07771778106689453, -0.0596628300845623, 0.040157150477170944, 0.056927651166915894, 0.045269690454006195, 0.16336530447006226, -0.10206259787082672, 0.07187876105308533, -0.21654650568962097, 0.00950478482991457, -0.050214603543281555, -0.059923723340034485, -0.09759427607059479, -0.03746842220425606, 0.08593204617500305, -0.05721491947770119, 0.08063247799873352, -0.05405398830771446, 0.06286115199327469, 0.027623100206255913, -0.11862469464540482, 0.026168711483478546, 0.03544669225811958, 0.2246401011943817, 0.04929327964782715, -0.0350758396089077, 0.057136572897434235, 0.0014597551198676229, 0.05750352144241333, 0.20276698470115662, 0.12170226871967316, 0.15684601664543152, 0.06694678962230682, 0.07514619082212448, 0.06495290249586105, -0.12847831845283508, -0.11632629483938217, 0.1120481863617897, -0.02301599271595478, 0.13317571580410004, -0.018415383994579315, 0.26252657175064087, 0.09659671038389206, -0.19880624115467072, 0.0724945217370987, -0.04280557110905647, -0.08190613240003586, -0.0955020859837532, -0.05018918216228485, -0.07523199915885925, -0.18381617963314056, 0.0029726356733590364, -0.10270554572343826, 0.058236874639987946, 0.02960720844566822, 0.03768931329250336, 0.02607111819088459, 0.10259909182786942, 0.04235660284757614, -0.01461736112833023, 0.10175365954637527, 0.014980798587203026, -0.012627406045794487, -0.0704638808965683, -0.0893634706735611, 0.06656462699174881, -0.0420454666018486, 0.042839087545871735, -0.03463199734687805, -0.10963023453950882, 0.06030009686946869, 0.008679918013513088, -0.11244747787714005, 0.02803177572786808, -0.018691357225179672, 0.07836131006479263, 0.13117176294326782, 0.041142143309116364, -0.005994702689349651, -0.011646319180727005, 0.24043546617031097, -0.10355240851640701, -0.0794016569852829, -0.12450195848941803, 0.2590564489364624, 0.008530175313353539, -0.018590502440929413, 0.02160031720995903, -0.06803333759307861, -0.005844017956405878, 0.13463474810123444, 0.17057302594184875, -0.002110502915456891, -0.012373129837214947, 0.023284457623958588, -0.014714868739247322, -0.04664544016122818, 0.06413646042346954, 0.1357046663761139, 0.06028023734688759, -0.06022518500685692, -0.006549651734530926, -0.05766531080007553, -0.04492848739027977, -0.01665627583861351, 0.08217541873455048, 0.029824700206518173, -0.022566623985767365, -0.009603582322597504, 0.12221545726060867, -0.05271247401833534, -0.15185678005218506, 0.003592139109969139, -0.18534575402736664, -0.17406143248081207, -0.033252645283937454, 0.08179497718811035, 0.057107362896203995, 0.04738996922969818, -0.014416498132050037, -0.011743891052901745, 0.121448814868927, -0.0011210566153749824, -0.04240282252430916, -0.12108755111694336, 0.09296219050884247, -0.07835060358047485, 0.15563137829303741, -0.034082524478435516, 0.051601577550172806, 0.11323007941246033, 0.09364881366491318, -0.050705842673778534, 0.06162717938423157, 0.07838407903909683, -0.12672068178653717, 0.062070224434137344, 0.19527339935302734, -0.04662754386663437, 0.1617979109287262, 0.05218140780925751, -0.10975640267133713, 0.04032281041145325, -0.10900266468524933, -0.06755891442298889, -0.04542579874396324, 0.009913410060107708, -0.044816650450229645, 0.12858302891254425, 0.2024172991514206, -0.06679065525531769, -0.02939848229289055, -0.059870023280382156, -0.0037589739076793194, 0.04911191761493683, 0.14018204808235168, -0.04852232709527016, -0.2708273231983185, 0.021421223878860474, 0.015467624180018902, 0.025058574974536896, -0.24139036238193512, -0.10877297073602676, 0.036809492856264114, -0.05829871445894241, -0.06551746279001236, 0.11118371039628983, 0.06108971685171127, 0.033686865121126175, -0.05680760368704796, -0.11567647755146027, -0.01999126374721527, 0.1811380237340927, -0.16528168320655823, -0.055715322494506836 ]
null
null
transformers
# Tagalog DialoGPT This is an extension of the base Tagalog DialoGPT model (https://huggingface.co/gabtan99/dialogpt-tagalog-medium). This model is trained on 52K original conversations and 52K synthetic conversations, where 10% of tokens in each utterance in the synthetic conversation are machine-generated tokens.
{"language": ["tl"], "tags": ["conversational", "tagalog", "filipino"]}
text-generation
gabtan99/dialogpt-tagalog-medium-10
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "tagalog", "filipino", "tl", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "tl" ]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #tagalog #filipino #tl #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Tagalog DialoGPT This is an extension of the base Tagalog DialoGPT model (URL This model is trained on 52K original conversations and 52K synthetic conversations, where 10% of tokens in each utterance in the synthetic conversation are machine-generated tokens.
[ "# Tagalog DialoGPT\n\nThis is an extension of the base Tagalog DialoGPT model (URL \n\nThis model is trained on 52K original conversations and 52K synthetic conversations, where 10% of tokens in each utterance in the synthetic conversation are machine-generated tokens." ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #tagalog #filipino #tl #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Tagalog DialoGPT\n\nThis is an extension of the base Tagalog DialoGPT model (URL \n\nThis model is trained on 52K original conversations and 52K synthetic conversations, where 10% of tokens in each utterance in the synthetic conversation are machine-generated tokens." ]
[ 61, 65 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #tagalog #filipino #tl #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Tagalog DialoGPT\n\nThis is an extension of the base Tagalog DialoGPT model (URL \n\nThis model is trained on 52K original conversations and 52K synthetic conversations, where 10% of tokens in each utterance in the synthetic conversation are machine-generated tokens." ]
[ 0.008769840002059937, -0.08318886160850525, -0.004494436085224152, 0.024684054777026176, 0.10473398864269257, -0.003882759250700474, 0.10993286222219467, 0.16117209196090698, 0.0040648262947797775, -0.016684407368302345, 0.03729601576924324, 0.13874590396881104, 0.06578053534030914, 0.11826031655073166, 0.002807477954775095, -0.26690173149108887, 0.06583648175001144, 0.04609055444598198, 0.13871945440769196, 0.11746340245008469, 0.06321883946657181, -0.0173651073127985, 0.06862630695104599, 0.003958004526793957, -0.1435043215751648, 0.06355112046003342, 0.03293651342391968, -0.11888081580400467, 0.07795554399490356, 0.03469875082373619, -0.012373858131468296, -0.011456099338829517, -0.05290709435939789, -0.12039102613925934, 0.019880875945091248, -0.021136639639735222, -0.01144452579319477, 0.006236798129975796, -0.1381322294473648, -0.024373676627874374, 0.11275068670511246, 0.11300541460514069, 0.04381557181477547, 0.06084455922245979, -0.16044701635837555, -0.12296898663043976, 0.014042653143405914, 0.11455769091844559, 0.038631338626146317, 0.028337009251117706, -0.08068545907735825, 0.1686526983976364, -0.052251528948545456, 0.0726194903254509, 0.15005110204219818, -0.3367586135864258, -0.02062801830470562, 0.1629936695098877, 0.08480296283960342, 0.11371847242116928, -0.02188975363969803, 0.06365594267845154, 0.04619845375418663, 0.01862367056310177, -0.04031306877732277, -0.10812778770923615, -0.043621502816677094, -0.04681924358010292, -0.08003728091716766, 0.02132645808160305, 0.17633657157421112, -0.022997833788394928, -0.007253524847328663, -0.1250266134738922, -0.04317231848835945, -0.12238063663244247, -0.09685083478689194, -0.06878478825092316, -0.052168238908052444, 0.0903363898396492, -0.0031972937285900116, -0.07451420277357101, -0.0956588014960289, -0.05820200592279434, -0.13757002353668213, 0.16833236813545227, 0.05792253464460373, 0.029823889955878258, -0.17584246397018433, 0.06911862641572952, -0.10009455680847168, -0.11377128958702087, -0.010985618457198143, -0.08642833679914474, 0.033933598548173904, 0.06751670688390732, -0.0005344106466509402, -0.040041789412498474, 0.09374455362558365, 0.08264687657356262, 0.005713813938200474, 0.07948412001132965, -0.06920474022626877, 0.04083609953522682, 0.013823178596794605, 0.144405797123909, -0.009957706555724144, -0.0299356821924448, 0.024717101827263832, -0.1013948917388916, 0.01287281047552824, -0.055926743894815445, -0.17269104719161987, 0.0009128273813985288, 0.011661997064948082, 0.05103583633899689, -0.02596917934715748, 0.19685275852680206, -0.005891626235097647, -0.05427439883351326, -0.05543044954538345, -0.03029966540634632, -0.013989759609103203, -0.03531048074364662, -0.039636049419641495, 0.18146352469921112, -0.00269883731380105, 0.03844260051846504, -0.0931776762008667, -0.09731920808553696, -0.016542715951800346, -0.0027744797989726067, -0.055365052074193954, -0.028938261792063713, 0.010310227982699871, -0.06056603416800499, -0.030924150720238686, -0.1519249528646469, -0.12125947326421738, -0.025258855894207954, -0.016830982640385628, -0.07893967628479004, -0.03674772381782532, -0.11072339862585068, 0.07288488745689392, 0.030330145731568336, -0.05475235357880592, -0.10276351869106293, -0.043952539563179016, 0.10335232317447662, 0.009547858498990536, 0.1911235898733139, -0.12424132972955704, 0.04921739175915718, -0.08955510705709457, 0.016062792390584946, -0.20628397166728973, 0.11016711592674255, 0.00808212161064148, 0.0901673287153244, -0.04392578452825546, -0.02019302174448967, -0.1367190033197403, 0.060736145824193954, 0.00482034171000123, 0.19966158270835876, -0.0831933543086052, -0.05922412499785423, 0.3573629558086395, -0.06108393520116806, -0.06352889537811279, 0.14833605289459229, 0.0021048742346465588, 0.20911110937595367, 0.1596842259168625, 0.30585354566574097, -0.0779532790184021, -0.021690357476472855, 0.08497149497270584, 0.044753581285476685, -0.08390410989522934, 0.0131804458796978, 0.00941512081772089, 0.0035353738348931074, -0.05399467796087265, 0.05932256206870079, 0.17132900655269623, 0.04270431399345398, -0.03538627550005913, -0.017750151455402374, 0.0046393852680921555, 0.00019133022578898817, 0.10934030264616013, -0.04320478066802025, 0.06563055515289307, -0.0832568109035492, -0.08391620963811874, 0.03446530923247337, 0.03278733417391777, -0.06398021429777145, 0.05231013149023056, -0.1878482848405838, 0.06270858645439148, 0.048656150698661804, 0.08462114632129669, -0.03720426931977272, -0.0058146589435637, -0.02758008800446987, 0.11680710315704346, 0.10282633453607559, -0.02080412395298481, 0.046175722032785416, -0.1116672158241272, -0.03268701583147049, 0.07669106125831604, 0.15563197433948517, -0.03019598126411438, -0.028775546699762344, -0.06468462944030762, 0.11923941969871521, -0.03393463417887688, 0.019042450934648514, 0.026368997991085052, 0.022296767681837082, 0.024253711104393005, 0.07633809745311737, -0.049505822360515594, 0.010014748200774193, 0.047421008348464966, 0.019342323765158653, -0.009323555044829845, -0.025090891867876053, 0.055352821946144104, 0.03594653308391571, -0.09490305930376053, 0.22602103650569916, -0.14213865995407104, 0.09001695364713669, 0.1834074854850769, -0.08411287516355515, -0.04535924643278122, -0.12007595598697662, 0.049652181565761566, -0.024026133120059967, 0.1084383949637413, -0.06807935237884521, 0.2516673505306244, -0.05749089643359184, 0.14332959055900574, -0.062439389526844025, 0.003947106655687094, 0.010499322786927223, -0.11339958012104034, 0.017664877697825432, 0.045868828892707825, 0.02338174544274807, -0.09653876721858978, 0.13709700107574463, -0.017595212906599045, 0.08766188472509384, 0.2471146583557129, 0.020757406949996948, 0.021878229454159737, -0.019240569323301315, 0.05178194120526314, -0.03570019081234932, -0.03601427748799324, -0.4333023726940155, -0.08372975140810013, 0.037747595459222794, 0.03586192801594734, 0.10764218121767044, -0.0606471486389637, -0.025882694870233536, -0.025735696777701378, -0.027835069224238396, -0.0022456778679043055, 0.13230255246162415, -0.008712871931493282, 0.11051096022129059, 0.047898877412080765, -0.01003281306475401, 0.029775483533740044, 0.03650730103254318, -0.0573364719748497, 0.12377768009901047, -0.09995037317276001, -0.39467260241508484, -0.13181088864803314, -0.24269701540470123, -0.017481479793787003, 0.050483182072639465, 0.12747949361801147, -0.18012113869190216, -0.008891768753528595, 0.029717599973082542, 0.13182863593101501, -0.0004357937432359904, 0.008572336286306381, 0.03466344252228737, -0.05861055478453636, -0.09154446423053741, -0.1116570308804512, -0.07856516540050507, -0.030111299827694893, -0.07423700392246246, 0.10329842567443848, -0.2123572677373886, 0.019536195322871208, 0.28078848123550415, 0.024928005412220955, 0.07162739336490631, -0.055668242275714874, 0.14414256811141968, -0.11250102519989014, -0.019763542339205742, 0.1377466470003128, -0.032624002546072006, 0.04570090025663376, 0.13974720239639282, -0.0035390511620789766, -0.11083737760782242, 0.049994125962257385, -0.0410492867231369, -0.10298851877450943, -0.18359743058681488, -0.08603978157043457, -0.04305090382695198, 0.06130507215857506, -0.06134306639432907, 0.03620791807770729, 0.17623235285282135, 0.06505702435970306, -0.008648903109133244, 0.01913084276020527, 0.003319279057905078, 0.098663330078125, 0.2493891566991806, -0.113341324031353, 0.1278422772884369, -0.005358412861824036, -0.12050073593854904, 0.06638075411319733, 0.06651803851127625, 0.06532451510429382, 0.09219320118427277, 0.11215053498744965, 0.005915138404816389, 0.13973428308963776, 0.11753114312887192, 0.013969285413622856, -0.034594450145959854, -0.05915116146206856, -0.026374446228146553, -0.030346034094691277, -0.033875346183776855, 0.03780652582645416, 0.007439277600497007, -0.12516042590141296, 0.04105348512530327, -0.030625438317656517, 0.10269434750080109, 0.10219895094633102, 0.026763131842017174, -0.12883037328720093, -0.04281027987599373, 0.02631293423473835, -0.025825338438153267, -0.09932912141084671, 0.08778072148561478, 0.05658886954188347, -0.11271510273218155, 0.03193680942058563, 0.036720987409353256, 0.0797300785779953, -0.051196713000535965, 0.06701922416687012, -0.14511695504188538, -0.058718789368867874, -0.028496956452727318, 0.0658743754029274, -0.24874387681484222, 0.1444169282913208, -0.015851670876145363, -0.05480450019240379, -0.134344682097435, -0.03784431517124176, -0.0327942855656147, 0.01672257110476494, 0.10549966245889664, 0.0162038616836071, 0.051332809031009674, -0.09501062333583832, -0.06337940692901611, 0.03144259378314018, 0.113514244556427, 0.052358873188495636, -0.0002590309304650873, -0.017854832112789154, 0.020983820781111717, 0.0023917686194181442, -0.06581966578960419, 0.0337425135076046, -0.07997647672891617, 0.04398953169584274, 0.1485455185174942, 0.09058475494384766, 0.03490939363837242, -0.01453388761729002, -0.03848641365766525, 0.2626515030860901, 0.06859652698040009, -0.10849277675151825, -0.07054793834686279, -0.07284216582775116, 0.06728845089673996, -0.06291760504245758, 0.06674426794052124, -0.03967202082276344, 0.01116746012121439, 0.0010994432959705591, -0.11334773153066635, 0.17679354548454285, -0.028891336172819138, -0.012975658290088177, 0.00913013331592083, 0.21756285429000854, 0.025471502915024757, 0.017647020518779755, 0.07066883146762848, -0.06785011291503906, -0.01271657831966877, -0.09903772175312042, -0.015813879668712616, 0.04116467386484146, -0.022480905055999756, -0.0032199027482420206, -0.02115771174430847, -0.0741787701845169, -0.10596677660942078, -0.09520988911390305, 0.26215940713882446, 0.10458087921142578, -0.023971829563379288, 0.13304279744625092, 0.16938550770282745, -0.0699157640337944, -0.20911146700382233, -0.11100610345602036, -0.05482923611998558, -0.029945170506834984, -0.10831590741872787, -0.13776177167892456, 0.007639265153557062, 0.02427593804895878, -0.0014209733344614506, 0.0464499369263649, -0.342896968126297, -0.1295640766620636, 0.1397733986377716, -0.022400738671422005, 0.49371349811553955, -0.11797525733709335, -0.10284527391195297, -0.05345381796360016, -0.19377116858959198, 0.05329021066427231, -0.024170193821191788, 0.12926149368286133, -0.007783268112689257, 0.20425568521022797, 0.06672418862581253, 0.034965742379426956, 0.06230826675891876, -0.018072303384542465, -0.04950714483857155, -0.13649968802928925, -0.12704885005950928, 0.03942421078681946, 0.02753172069787979, 0.12093245983123779, -0.06983959674835205, -0.028955869376659393, -0.05913905054330826, -0.10454433411359787, -0.0737813264131546, -0.03236525505781174, 0.036638855934143066, -0.10088227689266205, -0.03467997908592224, -0.04328588396310806, -0.012819916941225529, 0.01054319366812706, 0.08185131847858429, -0.1544310748577118, 0.12221252173185349, 0.12590794265270233, 0.08238087594509125, -0.18435914814472198, 0.005085410550236702, -0.02997220680117607, -0.059818703681230545, 0.06106435880064964, -0.12195173650979996, 0.03346628323197365, 0.06820018589496613, -0.003469286020845175, 0.11438643932342529, 0.03234852850437164, -0.04162263497710228, 0.09296325594186783, 0.1125577911734581, -0.08991344273090363, -0.1462550163269043, -0.02949637547135353, 0.00819358415901661, 0.09029287844896317, 0.13019251823425293, 0.14846839010715485, -0.056357208639383316, -0.04229879751801491, -0.0110990721732378, -0.012434359639883041, -0.09356066584587097, 0.09308813512325287, -0.02399405464529991, -0.0029632607474923134, -0.14092513918876648, 0.04911719635128975, 0.014650380238890648, -0.0017851238371804357, 0.07625965774059296, 0.14633940160274506, -0.07026063650846481, -0.07439268380403519, -0.06143736466765404, 0.04068100452423096, -0.06497107446193695, -0.03648938238620758, -0.03320073336362839, -0.09436265379190445, 0.011392825283110142, 0.15066543221473694, 0.02551417611539364, 0.06875356286764145, -0.042386941611766815, 0.03205294534564018, 0.0185985267162323, -0.05283253639936447, 0.06848110258579254, -0.00391615554690361, -0.05651028826832771, 0.1349705308675766, 0.03503970429301262, 0.07009155303239822, -0.10302083939313889, -0.12550660967826843, -0.20072562992572784, 0.06131051108241081, -0.1753096878528595, -0.06903845816850662, -0.12281129509210587, -0.04395134374499321, 0.0059111155569553375, -0.05360264703631401, -0.04282402619719505, -0.015435607172548771, -0.08409091830253601, 0.0512218102812767, -0.0049322848208248615, -0.0027238288894295692, -0.045130982995033264, 0.03515645116567612, 0.03989044576883316, 0.024907436221837997, 0.1627012938261032, 0.13330891728401184, -0.14627809822559357, 0.11182424426078796, -0.12501923739910126, 0.03177975118160248, 0.16193196177482605, 0.00526501564309001, 0.0016569967847317457, 0.07652024179697037, -0.017842374742031097, 0.05449341610074043, 0.09666693955659866, 0.06596720963716507, 0.0683036595582962, -0.12066382169723511, 0.04804063215851784, -0.06111254543066025, -0.11262556165456772, -0.00984582956880331, -0.017662176862359047, 0.06855504959821701, 0.04767197370529175, 0.01813833974301815, -0.07432875782251358, 0.02666114643216133, -0.03767376393079758, 0.014111881144344807, 0.037081003189086914, -0.08953166007995605, 0.08964177966117859, -0.09082107245922089, 0.005059151444584131, 0.022746501490473747, 0.21062707901000977, 0.02290964126586914, -0.07753519713878632, -0.01965111494064331, 0.02273525297641754, 0.0009572574635967612, -0.04015079140663147, 0.1935957670211792, 0.10885211080312729, -0.0499398298561573, -0.09179763495922089, 0.07535550743341446, 0.038340888917446136, -0.004554651211947203, 0.023607924580574036, -0.044936858117580414, 0.07520464807748795, 0.07679909467697144, 0.057445600628852844, 0.09704224020242691, -0.11277362704277039, -0.0662958100438118, -0.0928039401769638, 0.034087494015693665, -0.09182076156139374, 0.2094673216342926, 0.12190033495426178, -0.003970508463680744, 0.03770967945456505, -0.004391936119645834, -0.05553793907165527, -0.10966256260871887, -0.2684299647808075, -0.06276524066925049, -0.251914918422699, 0.0215406883507967, -0.09086865931749344, 0.06339049339294434, 0.0014172043884173036, 0.08453790098428726, -0.08452001959085464, 0.13045446574687958, -0.04757397621870041, -0.14167596399784088, 0.09214794635772705, -0.020335908979177475, 0.04708994925022125, -0.0832381471991539, 0.00022537304903380573, -0.03394808992743492, 0.08740399777889252, 0.04328691214323044, 0.0810052752494812, -0.027961663901805878, -0.05250345170497894, -0.13254296779632568, -0.01757412776350975, -0.05710429325699806, 0.046869441866874695, 0.003959235735237598, 0.04818039387464523, 0.032885193824768066, -0.06838425248861313, -0.01581670716404915, 0.18598340451717377, -0.05360586568713188, -0.1193251833319664, -0.0940299928188324, 0.1251540333032608, -0.02830209583044052, 0.06555783748626709, -0.08705069869756699, -0.035168614238500595, -0.10250000655651093, 0.36103442311286926, 0.17625853419303894, 0.02290673367679119, -0.004477659706026316, -0.028868841007351875, 0.05864143371582031, 0.016010191291570663, 0.16143181920051575, 0.16480909287929535, 0.23729528486728668, -0.004820948000997305, -0.01438162662088871, 0.009815041907131672, -0.07647491246461868, -0.07292533665895462, -0.020488642156124115, 0.011094258166849613, -0.051055196672677994, -0.06534554809331894, 0.10540679097175598, -0.2597551941871643, 0.041660577058792114, -0.12082885205745697, -0.14848756790161133, -0.06657250225543976, -0.019846154376864433, 0.00252852332778275, 0.06743456423282623, 0.06646864861249924, 0.027347506955266, -0.031448379158973694, 0.03021077997982502, 0.03696194291114807, -0.19281432032585144, -0.09486900269985199, 0.09260652959346771, -0.07803262770175934, -0.00003740707325050607, -0.003865981474518776, 0.05945563316345215, 0.04469712823629379, 0.0640871524810791, 0.04293321818113327, 0.07846006006002426, -0.020527541637420654, -0.04561101272702217, -0.005853334441781044, 0.006407697219401598, 0.013719788752496243, -0.05148680508136749, 0.011706551536917686, -0.06319327652454376, 0.002693349728360772, 0.028684580698609352, 0.07164401561021805, -0.1055927574634552, 0.020599471405148506, -0.11294517666101456, 0.07740907371044159, 0.094901442527771, -0.029995573684573174, 0.003935237880796194, -0.01469742227345705, 0.002207699231803417, -0.017148001119494438, -0.09045255184173584, -0.08814606070518494, -0.2365666776895523, -0.10607331991195679, 0.08755884319543839, -0.003980881068855524, -0.10542482882738113, 0.013386562466621399, -0.08654456585645676, 0.07615899294614792, -0.08969707787036896, 0.09189076721668243, 0.097524493932724, 0.05488273501396179, 0.004913047421723604, -0.1144075021147728, 0.04125674441456795, 0.1389532834291458, -0.1138160228729248, -0.05832810327410698 ]
null
null
transformers
# Tagalog DialoGPT This is an extension of the base Tagalog DialoGPT model (https://huggingface.co/gabtan99/dialogpt-tagalog-medium). This model is trained on 52K original conversations and 52K synthetic conversations, where 20% of tokens in each utterance in the synthetic conversation are machine-generated tokens.
{"language": ["tl"], "tags": ["conversational", "tagalog", "filipino"], "inference": false}
text-generation
gabtan99/dialogpt-tagalog-medium-20
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "tagalog", "filipino", "tl", "autotrain_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "tl" ]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #tagalog #filipino #tl #autotrain_compatible #text-generation-inference #region-us
# Tagalog DialoGPT This is an extension of the base Tagalog DialoGPT model (URL This model is trained on 52K original conversations and 52K synthetic conversations, where 20% of tokens in each utterance in the synthetic conversation are machine-generated tokens.
[ "# Tagalog DialoGPT\n\nThis is an extension of the base Tagalog DialoGPT model (URL \n\nThis model is trained on 52K original conversations and 52K synthetic conversations, where 20% of tokens in each utterance in the synthetic conversation are machine-generated tokens." ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #tagalog #filipino #tl #autotrain_compatible #text-generation-inference #region-us \n", "# Tagalog DialoGPT\n\nThis is an extension of the base Tagalog DialoGPT model (URL \n\nThis model is trained on 52K original conversations and 52K synthetic conversations, where 20% of tokens in each utterance in the synthetic conversation are machine-generated tokens." ]
[ 53, 65 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #tagalog #filipino #tl #autotrain_compatible #text-generation-inference #region-us \n# Tagalog DialoGPT\n\nThis is an extension of the base Tagalog DialoGPT model (URL \n\nThis model is trained on 52K original conversations and 52K synthetic conversations, where 20% of tokens in each utterance in the synthetic conversation are machine-generated tokens." ]
[ 0.02899603731930256, -0.04077323153614998, -0.005138820502907038, 0.033600274473428726, 0.104488804936409, -0.0011734931031242013, 0.15382525324821472, 0.15917038917541504, 0.004302134271711111, -0.03114873543381691, 0.029274532571434975, 0.09935332089662552, 0.06588415801525116, 0.12620078027248383, 0.04040982201695442, -0.25122708082199097, 0.062316589057445526, 0.0012091022217646241, 0.11609528958797455, 0.11831264942884445, 0.04528626427054405, 0.0015394282527267933, 0.05372970551252365, -0.008795189671218395, -0.14825011789798737, 0.07397478073835373, 0.030553100630640984, -0.10834319144487381, 0.06867947429418564, 0.027126379311084747, -0.005992895923554897, -0.0005573481321334839, -0.043414946645498276, -0.1193055808544159, 0.024439573287963867, -0.01338991615921259, -0.015582963824272156, 0.019468283280730247, -0.14156563580036163, -0.011651351116597652, 0.11536357551813126, 0.13967713713645935, 0.039320919662714005, 0.06511908024549484, -0.16904938220977783, -0.1399775594472885, 0.012939107604324818, 0.10681301355361938, -0.01848425902426243, 0.021444842219352722, -0.09066084772348404, 0.16964592039585114, -0.07567227631807327, 0.062052059918642044, 0.12168057262897491, -0.2921353578567505, -0.015023809857666492, 0.162318617105484, 0.05307091400027275, 0.165005624294281, -0.016671888530254364, 0.07231248915195465, 0.04244761914014816, 0.012096677906811237, -0.053073372691869736, -0.09874313324689865, -0.08664866536855698, -0.039271287620067596, -0.08709074556827545, 0.010415780358016491, 0.19459602236747742, -0.00896366685628891, -0.0018428632756695151, -0.08981111645698547, -0.05659884214401245, -0.08912814408540726, -0.07540508359670639, -0.06602510064840317, -0.041914477944374084, 0.09672923386096954, -0.03104165755212307, -0.10723330080509186, -0.08759603649377823, -0.07420288771390915, -0.1452956348657608, 0.15798769891262054, 0.06611232459545135, 0.03226277977228165, -0.173788920044899, 0.07402215152978897, -0.053139228373765945, -0.10499968379735947, 0.02085758186876774, -0.09352204948663712, 0.029797540977597237, 0.09547606110572815, 0.01602182537317276, -0.04790274053812027, 0.11958625912666321, 0.0698726698756218, -0.0010068772826343775, 0.06530575454235077, -0.0713777095079422, 0.0532802976667881, 0.021932637318968773, 0.13722966611385345, -0.022370439022779465, 0.010376328602433205, 0.016556983813643456, -0.07558839023113251, 0.013750565238296986, -0.07020922005176544, -0.18424062430858612, 0.042926836758852005, 0.0007695337990298867, 0.06930840015411377, -0.016642514616250992, 0.19428950548171997, -0.012601071037352085, -0.05121660232543945, -0.05592230707406998, -0.023129725828766823, -0.011465650983154774, -0.0223099272698164, -0.0665077492594719, 0.15980324149131775, 0.015051999129354954, 0.022711487486958504, -0.08639681339263916, -0.14383059740066528, -0.022906148806214333, -0.011708665639162064, -0.05671053007245064, -0.05298984795808792, 0.004201275296509266, -0.015400436706840992, -0.05306011438369751, -0.14868159592151642, -0.09896920621395111, -0.029306907206773758, -0.018851635977625847, -0.07467394322156906, -0.02992885932326317, -0.13003553450107574, 0.09420452266931534, 0.012420481070876122, -0.04818117246031761, -0.11611910909414291, -0.04161524772644043, 0.11957617104053497, -0.010745387524366379, 0.18802426755428314, -0.12429418414831161, 0.056364819407463074, -0.09206592291593552, 0.006732683163136244, -0.1870546191930771, 0.10249204188585281, -0.015847953036427498, 0.08307132124900818, -0.02948760613799095, -0.008960727602243423, -0.11063172668218613, 0.07256507873535156, 0.004774870350956917, 0.19367603957653046, -0.08615522086620331, -0.07302867621183395, 0.3745323419570923, -0.05896024405956268, -0.037103213369846344, 0.16243232786655426, -0.0025565193500369787, 0.2190098911523819, 0.1620674580335617, 0.3259812593460083, -0.09846467524766922, 0.0224843081086874, 0.11487653851509094, 0.0415140837430954, -0.08333040028810501, 0.041547562927007675, 0.006090181414037943, -0.008698957972228527, -0.06265675276517868, 0.05288683623075485, 0.17418257892131805, 0.052902039140462875, -0.04799475520849228, -0.03019474260509014, 0.0209504384547472, 0.0028576392214745283, 0.12782596051692963, -0.054511334747076035, 0.05403481796383858, -0.08850132673978806, -0.08423756062984467, 0.0431353934109211, 0.05206269025802612, -0.06659740954637527, 0.03818272426724434, -0.18589210510253906, 0.050490111112594604, 0.09032660722732544, 0.10913726687431335, -0.016877247020602226, -0.03086780197918415, -0.019483715295791626, 0.0774674043059349, 0.13950560986995697, 0.008274435997009277, 0.03442695736885071, -0.09921961277723312, -0.020126324146986008, 0.07858356088399887, 0.1045604720711708, -0.035938069224357605, 0.0005345282843336463, -0.06793100386857986, 0.13795360922813416, -0.03179803490638733, 0.0417652390897274, 0.03488897159695625, 0.022391416132450104, 0.007072241976857185, 0.06463015824556351, -0.04148337244987488, 0.016277112066745758, 0.06156256049871445, 0.022634893655776978, -0.0034030359238386154, -0.03359513729810715, 0.050533220171928406, 0.0158060435205698, -0.09423291683197021, 0.2003467082977295, -0.11890915036201477, 0.08345775306224823, 0.18364278972148895, -0.07812749594449997, -0.03773944452404976, -0.10926755517721176, 0.056759681552648544, -0.015379616059362888, 0.11288788169622421, -0.07221374660730362, 0.22611986100673676, -0.05568956211209297, 0.13629989326000214, -0.054501429200172424, 0.032255303114652634, -0.0035684602335095406, -0.11751139163970947, 0.026674972847104073, 0.05737832188606262, 0.028635138645768166, -0.08938644826412201, 0.13922269642353058, -0.017044708132743835, 0.08150380104780197, 0.2600274980068207, 0.011304878629744053, 0.03447972610592842, -0.013840833678841591, 0.05619218572974205, -0.04427262768149376, -0.04320240020751953, -0.38401779532432556, -0.08526627719402313, 0.03877732902765274, 0.032638389617204666, 0.10640308260917664, -0.07083974778652191, -0.03533950820565224, -0.039121173322200775, -0.05074450746178627, -0.01934976689517498, 0.11506863683462143, -0.01855779066681862, 0.10949886590242386, 0.03991559520363808, -0.024908937513828278, 0.0276048444211483, 0.03138267621397972, -0.04017600417137146, 0.10195865482091904, -0.11529682576656342, -0.4097432792186737, -0.14368967711925507, -0.21471378207206726, -0.04070231318473816, 0.04762331768870354, 0.12206432223320007, -0.202372744679451, -0.009663172997534275, 0.03171239420771599, 0.16219495236873627, 0.014914052560925484, -0.014002686366438866, 0.014373240061104298, -0.0642768144607544, -0.08100393414497375, -0.13105833530426025, -0.0593760721385479, -0.039731647819280624, -0.0911642536520958, 0.11274424195289612, -0.22532996535301208, 0.04012458398938179, 0.2853965759277344, 0.024192919954657555, 0.06155358999967575, -0.05479394271969795, 0.1377200484275818, -0.10798466205596924, -0.031772222369909286, 0.11946052312850952, -0.02741747349500656, 0.05334635078907013, 0.11894480139017105, 0.007800249848514795, -0.11020776629447937, 0.06728457659482956, -0.03214564919471741, -0.09921245276927948, -0.20770113170146942, -0.08848520368337631, -0.015830814838409424, 0.06909552216529846, -0.06650953739881516, 0.051525942981243134, 0.14742381870746613, 0.04803800210356712, -0.004548357333987951, -0.0003909236693289131, 0.0026066950522363186, 0.06988807767629623, 0.24118560552597046, -0.13419486582279205, 0.1186433881521225, -0.00614424841478467, -0.13011354207992554, 0.08759881556034088, 0.06327281147241592, 0.08384514600038528, 0.0994134321808815, 0.14535179734230042, 0.019951311871409416, 0.11961790174245834, 0.12105009704828262, -0.004375695716589689, -0.008963335305452347, -0.03926946967840195, -0.04740752652287483, -0.03150980547070503, -0.02254439704120159, 0.03207947686314583, 0.02949710376560688, -0.13903453946113586, 0.057480935007333755, -0.022641925141215324, 0.12339143455028534, 0.08764700591564178, 0.006410264875739813, -0.10383948683738708, -0.01346823200583458, 0.022179191932082176, -0.011037778109312057, -0.09232008457183838, 0.08391661196947098, 0.0765312984585762, -0.10700617730617523, -0.0008853371255099773, 0.052684806287288666, 0.0823759138584137, -0.08368190377950668, 0.06648274511098862, -0.16257387399673462, -0.014984998852014542, -0.033806849271059036, 0.07070587575435638, -0.22636187076568604, 0.150893434882164, -0.02100146934390068, -0.08997251093387604, -0.11584287881851196, -0.0454893596470356, -0.012757183983922005, -0.003135203616693616, 0.07315616309642792, 0.02862960658967495, 0.07976630330085754, -0.08512575924396515, -0.07484535872936249, 0.047296471893787384, 0.11479490250349045, 0.04480082541704178, -0.02585703507065773, -0.012081781402230263, 0.029453080147504807, -0.006381866056472063, -0.06731809675693512, 0.022231008857488632, -0.059979528188705444, 0.035659655928611755, 0.12412837147712708, 0.09371016919612885, 0.03339608013629913, -0.022307151928544044, -0.0435764417052269, 0.2573792338371277, 0.11589904874563217, -0.100925512611866, -0.05941612273454666, -0.0658937394618988, 0.06642851233482361, -0.054835375398397446, 0.050535451620817184, -0.041866954416036606, -0.0033530201762914658, -0.00703359255567193, -0.10170353949069977, 0.17667156457901, -0.002891391050070524, -0.024496618658304214, 0.003650693455711007, 0.17697656154632568, 0.032484039664268494, 0.01914181560277939, 0.0437217615544796, -0.07950996607542038, -0.007599389646202326, -0.11053062230348587, 0.004487974569201469, 0.03929530456662178, -0.07353219389915466, 0.02736847847700119, -0.019328266382217407, -0.058309510350227356, -0.11919094622135162, -0.09853202849626541, 0.2497209906578064, 0.12673494219779968, -0.012461445294320583, 0.12610919773578644, 0.21659132838249207, -0.0665694996714592, -0.21842634677886963, -0.12168078869581223, -0.08436398953199387, -0.04097045585513115, -0.08656933158636093, -0.18668590486049652, -0.03900544345378876, 0.04358169063925743, -0.004454441834241152, 0.05106702074408531, -0.3410389721393585, -0.11607404053211212, 0.1530607044696808, -0.003959210589528084, 0.4617762267589569, -0.12018117308616638, -0.10918401926755905, -0.04962191730737686, -0.12117963284254074, 0.05141175165772438, -0.05756504461169243, 0.1375311315059662, -0.0016339077847078443, 0.2116042524576187, 0.06372155249118805, 0.05466773733496666, 0.07484017312526703, -0.028694527223706245, -0.03503815457224846, -0.1373884677886963, -0.12892013788223267, 0.05018644407391548, 0.040380727499723434, 0.11177679151296616, -0.08693642169237137, -0.022628597915172577, -0.011814521625638008, -0.09553196281194687, -0.06394000351428986, -0.0421244353055954, 0.039639636874198914, -0.1088327020406723, -0.05520663782954216, -0.03243520110845566, -0.037817683070898056, 0.019083822146058083, 0.059512607753276825, -0.13992391526699066, 0.10911473631858826, 0.15734094381332397, 0.027519112452864647, -0.15687188506126404, 0.019241292029619217, -0.03828221186995506, -0.0594114251434803, 0.07694751024246216, -0.13134461641311646, 0.02405897155404091, 0.059453438967466354, -0.005443866364657879, 0.14023087918758392, 0.012775162234902382, -0.06958453357219696, 0.09644483774900436, 0.11471094191074371, -0.0951838567852974, -0.16556760668754578, -0.031230486929416656, -0.0010930944699794054, 0.09742780774831772, 0.10956407338380814, 0.15612539649009705, -0.04471895471215248, -0.03852134197950363, 0.0018703864188864827, -0.008773543871939182, -0.07911797612905502, 0.09760338813066483, -0.004513515625149012, -0.003653033636510372, -0.14370082318782806, 0.07343106716871262, 0.04615470767021179, 0.03501352667808533, 0.06190115585923195, 0.14734520018100739, -0.061003729701042175, -0.07684572041034698, -0.07004011422395706, 0.070125050842762, -0.052547622472047806, -0.038534145802259445, -0.05218220129609108, -0.08842837065458298, -0.011633708141744137, 0.09197758883237839, 0.017021983861923218, 0.07996339350938797, -0.0379863902926445, 0.03963275998830795, 0.05702178180217743, -0.05103760212659836, 0.07508993148803711, -0.019930725917220116, -0.04628317430615425, 0.12553878128528595, 0.05431777611374855, 0.0694725289940834, -0.10659456253051758, -0.14774154126644135, -0.21607598662376404, 0.06496736407279968, -0.18291424214839935, -0.07018592953681946, -0.08779577910900116, -0.02896830067038536, 0.001085539348423481, -0.04934670403599739, -0.03566049039363861, -0.025985412299633026, -0.09303100407123566, 0.03541979193687439, 0.011650304310023785, 0.006346986163407564, -0.05998916178941727, 0.03823913261294365, 0.040782444179058075, 0.03715154156088829, 0.17360593378543854, 0.10590009391307831, -0.14877450466156006, 0.11473993211984634, -0.12508970499038696, 0.022740552201867104, 0.16839532554149628, 0.01804446056485176, -0.0011277742451056838, 0.10033079236745834, -0.029884422197937965, 0.02456154301762581, 0.09504573047161102, 0.059842851012945175, 0.06149660795927048, -0.12157624959945679, 0.05947675555944443, -0.06945550441741943, -0.11820127815008163, -0.004780399147421122, -0.014523263089358807, 0.060949359089136124, 0.05732514709234238, 0.009216354228556156, -0.07491730153560638, 0.01974467560648918, -0.03761032968759537, 0.012662279419600964, 0.05127812922000885, -0.0928899422287941, 0.07366879284381866, -0.09951433539390564, -0.01009549479931593, 0.014449641108512878, 0.22008712589740753, 0.04770021140575409, -0.09993480145931244, -0.030829261988401413, 0.048613741993904114, -0.023269226774573326, -0.047024860978126526, 0.14198943972587585, 0.11576028913259506, -0.04541701450943947, -0.10998828709125519, 0.05650263652205467, 0.031012801453471184, -0.03438972309231758, 0.01608402468264103, -0.0630815327167511, 0.09852536022663116, 0.049111682921648026, 0.08496061712503433, 0.10473468899726868, -0.08289564400911331, -0.04726506024599075, -0.06105753779411316, 0.03891226276755333, -0.09595507383346558, 0.17741183936595917, 0.10690639168024063, 0.026850832626223564, 0.04329155385494232, 0.017746809870004654, -0.05830681696534157, -0.10733035206794739, -0.24226900935173035, -0.06336566805839539, -0.2706245481967926, 0.014039132744073868, -0.07048314809799194, 0.058012835681438446, 0.004938336554914713, 0.08312905579805374, -0.09681965410709381, 0.10158073902130127, -0.031673211604356766, -0.15607570111751556, 0.075081966817379, -0.031424567103385925, 0.023964842781424522, -0.10649421811103821, -0.03702797740697861, -0.008277812972664833, 0.14310802519321442, 0.0386665016412735, 0.0946437194943428, -0.01780383102595806, -0.04702502489089966, -0.1603972613811493, -0.009824010543525219, -0.06432979553937912, 0.03578249737620354, -0.017423981800675392, 0.03739004209637642, 0.039436034858226776, -0.08976402133703232, -0.0007803807966411114, 0.17238099873065948, -0.036322299391031265, -0.11480681598186493, -0.08250120282173157, 0.15392813086509705, -0.07141133397817612, 0.07808393239974976, -0.0863567441701889, -0.03556742146611214, -0.140427827835083, 0.33620965480804443, 0.21859411895275116, -0.00957990437746048, -0.010541691444814205, -0.03904014453291893, 0.054804641753435135, -0.016628263518214226, 0.18013380467891693, 0.1484973132610321, 0.19001975655555725, 0.015658216550946236, -0.007134484127163887, 0.0025844560004770756, -0.08151115477085114, -0.057419538497924805, -0.03400985524058342, 0.03677432984113693, -0.03343517705798149, -0.08736424148082733, 0.1064218133687973, -0.2857897877693176, 0.0344088040292263, -0.13751277327537537, -0.1263725459575653, -0.08458946645259857, -0.02714383788406849, 0.001440147403627634, 0.09727746993303299, 0.06770750880241394, 0.015055523253977299, -0.030356906354427338, 0.003155845683068037, 0.02913772128522396, -0.19453121721744537, -0.12469961494207382, 0.10292644053697586, -0.10454817861318588, 0.013325816951692104, -0.0020481212995946407, 0.0546213835477829, 0.02805270440876484, 0.04718833416700363, 0.05939166992902756, 0.10529490560293198, -0.02268989570438862, -0.03452480211853981, -0.021809644997119904, -0.020600004121661186, 0.016212545335292816, -0.04557165876030922, 0.01916721649467945, -0.02270597033202648, -0.013907619751989841, 0.04236563667654991, 0.06313671916723251, -0.1052619144320488, 0.00392161775380373, -0.11831607669591904, 0.052471060305833817, 0.10145047307014465, -0.019453980028629303, 0.02770070731639862, -0.012672009877860546, 0.006073453463613987, -0.027728063985705376, -0.12162637710571289, -0.0864439308643341, -0.20901763439178467, -0.12080676108598709, 0.09755493700504303, -0.003598308190703392, -0.10284484922885895, 0.029595816507935524, -0.10793940722942352, 0.09660881757736206, -0.06325024366378784, 0.10270161926746368, 0.10096950083971024, 0.0647418349981308, 0.010300803929567337, -0.11157505959272385, 0.03736059367656708, 0.12436128407716751, -0.11441915482282639, -0.04081403836607933 ]
null
null
transformers
# Tagalog DialoGPT This is an extension of the base Tagalog DialoGPT model (https://huggingface.co/gabtan99/dialogpt-tagalog-medium). This model is trained on 52K original conversations and 52K synthetic conversations, where 30% of tokens in each utterance in the synthetic conversation are machine-generated tokens.
{"language": ["tl"], "tags": ["conversational", "tagalog", "filipino"], "inference": false}
text-generation
gabtan99/dialogpt-tagalog-medium-30
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "tagalog", "filipino", "tl", "autotrain_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "tl" ]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #tagalog #filipino #tl #autotrain_compatible #text-generation-inference #region-us
# Tagalog DialoGPT This is an extension of the base Tagalog DialoGPT model (URL This model is trained on 52K original conversations and 52K synthetic conversations, where 30% of tokens in each utterance in the synthetic conversation are machine-generated tokens.
[ "# Tagalog DialoGPT\n\nThis is an extension of the base Tagalog DialoGPT model (URL \n\nThis model is trained on 52K original conversations and 52K synthetic conversations, where 30% of tokens in each utterance in the synthetic conversation are machine-generated tokens." ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #tagalog #filipino #tl #autotrain_compatible #text-generation-inference #region-us \n", "# Tagalog DialoGPT\n\nThis is an extension of the base Tagalog DialoGPT model (URL \n\nThis model is trained on 52K original conversations and 52K synthetic conversations, where 30% of tokens in each utterance in the synthetic conversation are machine-generated tokens." ]
[ 53, 65 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #tagalog #filipino #tl #autotrain_compatible #text-generation-inference #region-us \n# Tagalog DialoGPT\n\nThis is an extension of the base Tagalog DialoGPT model (URL \n\nThis model is trained on 52K original conversations and 52K synthetic conversations, where 30% of tokens in each utterance in the synthetic conversation are machine-generated tokens." ]
[ 0.03058822825551033, -0.04383495822548866, -0.005014630500227213, 0.033478766679763794, 0.1056985855102539, -0.00009374295768793672, 0.15552373230457306, 0.15727713704109192, 0.002039483515545726, -0.0347898006439209, 0.028746789321303368, 0.09968236088752747, 0.06534697115421295, 0.11847849935293198, 0.039009157568216324, -0.24773149192333221, 0.062236104160547256, 0.002727445214986801, 0.11181043088436127, 0.12053205817937851, 0.04442744702100754, 0.0031331328209489584, 0.0532960407435894, -0.009375572204589844, -0.14389902353286743, 0.07305645197629929, 0.02738356404006481, -0.10570941120386124, 0.06828263401985168, 0.02684428170323372, -0.004075523465871811, -0.002872719895094633, -0.040230944752693176, -0.12282445281744003, 0.025400275364518166, -0.015168680809438229, -0.012188308872282505, 0.019202405586838722, -0.14449110627174377, -0.009206090122461319, 0.10905201733112335, 0.14148786664009094, 0.03703780472278595, 0.0661766305565834, -0.16989336907863617, -0.13047924637794495, 0.013851695694029331, 0.10135988891124725, -0.018502019345760345, 0.024358680471777916, -0.08956387639045715, 0.17694821953773499, -0.07859497517347336, 0.06064275652170181, 0.12181498855352402, -0.2934362590312958, -0.011640116572380066, 0.16293425858020782, 0.05163544416427612, 0.15932147204875946, -0.01866985112428665, 0.07012949138879776, 0.042337290942668915, 0.011320676654577255, -0.0500662624835968, -0.09940564632415771, -0.0816924199461937, -0.04197363927960396, -0.08707091957330704, 0.011334311217069626, 0.18561233580112457, -0.0039382572285830975, -0.003300152951851487, -0.09100121259689331, -0.05603819712996483, -0.08615963160991669, -0.07525163888931274, -0.06277143955230713, -0.043042466044425964, 0.09589134901762009, -0.0312141515314579, -0.10830523073673248, -0.08760931342840195, -0.07553843408823013, -0.14203423261642456, 0.1700785756111145, 0.06706222146749496, 0.03294345736503601, -0.17116108536720276, 0.07466767728328705, -0.04570123553276062, -0.10843657702207565, 0.018853699788451195, -0.09318186342716217, 0.029419003054499626, 0.09642130881547928, 0.01801924966275692, -0.04530433565378189, 0.11953049153089523, 0.06681656092405319, -0.006884618662297726, 0.06311433762311935, -0.06677735596895218, 0.05126069858670235, 0.02481323853135109, 0.13852614164352417, -0.020101109519600868, 0.007630542851984501, 0.013524140231311321, -0.07105983793735504, 0.01283580157905817, -0.06763457506895065, -0.18296906352043152, 0.0425269715487957, 0.0002329425187781453, 0.07161034643650055, -0.020631464198231697, 0.19372420012950897, -0.012144903652369976, -0.0522209070622921, -0.047543395310640335, -0.02213362604379654, -0.015324302949011326, -0.02165021188557148, -0.06796420365571976, 0.16008588671684265, 0.011967504397034645, 0.02314114384353161, -0.08871559798717499, -0.1438731551170349, -0.020946573466062546, -0.011988609097898006, -0.05716115981340408, -0.05461813136935234, 0.0012140229810029268, -0.018763430416584015, -0.05210721865296364, -0.14929668605327606, -0.10269519686698914, -0.030381441116333008, -0.018795279785990715, -0.07578461617231369, -0.02662753313779831, -0.12774842977523804, 0.09528939425945282, 0.013173064216971397, -0.047479912638664246, -0.11817993968725204, -0.04257390648126602, 0.12055562436580658, -0.007074592635035515, 0.18507200479507446, -0.12716203927993774, 0.057065073400735855, -0.09703928232192993, 0.0038035765755921602, -0.18086424469947815, 0.10281737893819809, -0.01242555771023035, 0.08562472462654114, -0.03199872747063637, -0.011789128184318542, -0.11502299457788467, 0.07237357646226883, 0.0049362038262188435, 0.19357724487781525, -0.07849057018756866, -0.07004864513874054, 0.3656105399131775, -0.057302284985780716, -0.037061113864183426, 0.16020818054676056, -0.003873460693284869, 0.216835156083107, 0.159353107213974, 0.3222540616989136, -0.09861766546964645, 0.02114476077258587, 0.11504755169153214, 0.04351358488202095, -0.08281686156988144, 0.03786802664399147, 0.008089954033493996, -0.00905518140643835, -0.0558411106467247, 0.04971490427851677, 0.17490822076797485, 0.05667667090892792, -0.04805585741996765, -0.030636126175522804, 0.02257399819791317, -0.0011135920649394393, 0.12272189557552338, -0.054246798157691956, 0.055964142084121704, -0.08685116469860077, -0.08516628295183182, 0.04189729318022728, 0.05262700095772743, -0.0677994042634964, 0.03990448638796806, -0.184396892786026, 0.049814920872449875, 0.08733528107404709, 0.10815483331680298, -0.018802758306264877, -0.03247321769595146, -0.018665067851543427, 0.07882817089557648, 0.1398637592792511, 0.007329858839511871, 0.03407964110374451, -0.10044372826814651, -0.020964117720723152, 0.07735168188810349, 0.10203604400157928, -0.038181621581315994, 0.0037790914066135883, -0.06526529788970947, 0.1358116865158081, -0.03363576903939247, 0.03178904950618744, 0.04304727911949158, 0.020802050828933716, 0.0017831026343628764, 0.061549779027700424, -0.04181215167045593, 0.015589239075779915, 0.06375116109848022, 0.023591261357069016, -0.005055946297943592, -0.0320812463760376, 0.05226489156484604, 0.017361531034111977, -0.09400200098752975, 0.20250430703163147, -0.11916555464267731, 0.08074326813220978, 0.18288616836071014, -0.0784117802977562, -0.039087966084480286, -0.11233135312795639, 0.0564785972237587, -0.013348669745028019, 0.11161898076534271, -0.07168169319629669, 0.2253928929567337, -0.05515202879905701, 0.13845324516296387, -0.054960135370492935, 0.03154972195625305, -0.005672654137015343, -0.11748076975345612, 0.028873683884739876, 0.0547286719083786, 0.02885640785098076, -0.09251288324594498, 0.13893766701221466, -0.01678943634033203, 0.08619974553585052, 0.2641070485115051, 0.010280854068696499, 0.032414767891168594, -0.013953945599496365, 0.053886380046606064, -0.045682862401008606, -0.04098309576511383, -0.3759814202785492, -0.08448231220245361, 0.038199733942747116, 0.03540729731321335, 0.10527773201465607, -0.07037416845560074, -0.03540351241827011, -0.03748568147420883, -0.05172690004110336, -0.016133936122059822, 0.11425692588090897, -0.020242467522621155, 0.10851933062076569, 0.042630065232515335, -0.025219079107046127, 0.02430419810116291, 0.029119987040758133, -0.03921610489487648, 0.10306030511856079, -0.10973118990659714, -0.40839359164237976, -0.14623260498046875, -0.21791015565395355, -0.03979726880788803, 0.04646318405866623, 0.11960542947053909, -0.20421645045280457, -0.011105768382549286, 0.033177897334098816, 0.1649695187807083, 0.020009279251098633, -0.015369368717074394, 0.01840793341398239, -0.06336572766304016, -0.08081299066543579, -0.13238587975502014, -0.06007374823093414, -0.04298834502696991, -0.08707502484321594, 0.10895541310310364, -0.22641439735889435, 0.0418865941464901, 0.2839983403682709, 0.024036554619669914, 0.061083413660526276, -0.055416882038116455, 0.14148558676242828, -0.10844726115465164, -0.03285679593682289, 0.12310245633125305, -0.027439691126346588, 0.052246082574129105, 0.11663689464330673, 0.010696562007069588, -0.11158069968223572, 0.07039327919483185, -0.029606012627482414, -0.10088644921779633, -0.20481471717357635, -0.08987925946712494, -0.016491087153553963, 0.06255646795034409, -0.06685188412666321, 0.05139961838722229, 0.14420244097709656, 0.04748421907424927, -0.005063384771347046, -0.008703745901584625, -0.0001147493821918033, 0.07009623944759369, 0.23676878213882446, -0.12792722880840302, 0.11936637759208679, -0.007629664149135351, -0.13254325091838837, 0.08636496216058731, 0.06605876982212067, 0.08135276287794113, 0.09666231274604797, 0.142465278506279, 0.018014295026659966, 0.11448056250810623, 0.12006472051143646, -0.0041230241768062115, -0.012278764508664608, -0.039955753833055496, -0.04811737313866615, -0.0315549261868, -0.022333290427923203, 0.029917042702436447, 0.027202729135751724, -0.1372978240251541, 0.05974574014544487, -0.020348530262708664, 0.12351176142692566, 0.0944320484995842, 0.013193687424063683, -0.10222286731004715, -0.014904545620083809, 0.02143746428191662, -0.012276850640773773, -0.0886574238538742, 0.08505135774612427, 0.0680558905005455, -0.10454972833395004, -0.0006772105116397142, 0.053752653300762177, 0.08427579700946808, -0.08687464892864227, 0.06877722591161728, -0.1592368334531784, -0.017706280574202538, -0.03481367975473404, 0.06885705888271332, -0.2258078008890152, 0.15483655035495758, -0.02095145545899868, -0.08959852159023285, -0.11273796856403351, -0.045287877321243286, -0.012261816300451756, -0.006036489736288786, 0.07405854016542435, 0.02798331156373024, 0.10195950418710709, -0.08347003906965256, -0.07487517595291138, 0.04568960517644882, 0.1140182688832283, 0.05161512270569801, -0.02262418158352375, -0.013678171671926975, 0.029176775366067886, -0.005042554344981909, -0.07253032922744751, 0.018782129511237144, -0.05907671898603439, 0.032920219004154205, 0.1224832832813263, 0.09097190201282501, 0.03680778667330742, -0.022033071145415306, -0.0479406975209713, 0.2560168206691742, 0.12559370696544647, -0.10133866965770721, -0.05813079699873924, -0.06624149531126022, 0.06315036863088608, -0.052798040211200714, 0.04743748530745506, -0.044950876384973526, -0.006703940220177174, -0.0053678457625210285, -0.09935399889945984, 0.17435190081596375, -0.0035537865478545427, -0.020958617329597473, 0.004558331798762083, 0.17531710863113403, 0.032499928027391434, 0.01936066336929798, 0.04343400150537491, -0.08164796978235245, -0.006882686633616686, -0.11061763763427734, 0.0035681084264069796, 0.024821192026138306, -0.07392378151416779, 0.03514420613646507, -0.02163257822394371, -0.06482347846031189, -0.12072817981243134, -0.09856738895177841, 0.2510247826576233, 0.1259295791387558, -0.010040775872766972, 0.12270809710025787, 0.2241901308298111, -0.06480077654123306, -0.21999289095401764, -0.12297197431325912, -0.08612225204706192, -0.03873125836253166, -0.08756455779075623, -0.18584980070590973, -0.04292745515704155, 0.03845575824379921, -0.0039449273608624935, 0.03884994611144066, -0.33928173780441284, -0.11644189059734344, 0.152308389544487, -0.002056061290204525, 0.4626137912273407, -0.1202680841088295, -0.10862346738576889, -0.051675982773303986, -0.13034801185131073, 0.05251859128475189, -0.0447300560772419, 0.1389831006526947, -0.004045173525810242, 0.20815613865852356, 0.06319672614336014, 0.0522448867559433, 0.07470330595970154, -0.02945113554596901, -0.03360331803560257, -0.13671685755252838, -0.12689101696014404, 0.05367035046219826, 0.0407036691904068, 0.11279168725013733, -0.08895863592624664, -0.021814873442053795, -0.013095797039568424, -0.09721241891384125, -0.06545963883399963, -0.043911587446928024, 0.0413326658308506, -0.11210714280605316, -0.05388995632529259, -0.02951383963227272, -0.037544142454862595, 0.019195975735783577, 0.06304056942462921, -0.14217109978199005, 0.10769587010145187, 0.15717041492462158, 0.029928285628557205, -0.1545172482728958, 0.018300555646419525, -0.03841110318899155, -0.05783431977033615, 0.07567701488733292, -0.13220787048339844, 0.024076193571090698, 0.06110259145498276, -0.004517856054008007, 0.14037157595157623, 0.01267704926431179, -0.06607937067747116, 0.09689099341630936, 0.11418610066175461, -0.08993613719940186, -0.16120852530002594, -0.029173504561185837, -0.008246914483606815, 0.09744684398174286, 0.11007047444581985, 0.15714944899082184, -0.04224815219640732, -0.03755743056535721, 0.005022476892918348, -0.011686479672789574, -0.0808224081993103, 0.09826216846704483, -0.0038713973481208086, -0.00373460678383708, -0.14291086792945862, 0.07412680983543396, 0.04360233247280121, 0.044160231947898865, 0.06076997518539429, 0.14934730529785156, -0.06249347701668739, -0.07623983174562454, -0.06956023722887039, 0.06747441738843918, -0.059256792068481445, -0.03586744889616966, -0.05822194367647171, -0.08749812096357346, -0.009900493547320366, 0.09164556860923767, 0.020006461068987846, 0.07801813632249832, -0.041471175849437714, 0.03908012434840202, 0.0557321198284626, -0.05206766724586487, 0.07315915077924728, -0.021241748705506325, -0.04598189890384674, 0.11513437330722809, 0.053704891353845596, 0.0687074214220047, -0.1071772500872612, -0.14858336746692657, -0.21735382080078125, 0.06645168364048004, -0.19058653712272644, -0.0730820819735527, -0.08791828155517578, -0.029345806688070297, -0.002510206075385213, -0.04971590265631676, -0.03549263998866081, -0.023821596056222916, -0.0910051092505455, 0.03511856496334076, 0.012869207188487053, 0.004952423274517059, -0.05835879594087601, 0.03920809179544449, 0.0393160842359066, 0.039031606167554855, 0.173774853348732, 0.10224633663892746, -0.1488495171070099, 0.11239935457706451, -0.12027823179960251, 0.022002311423420906, 0.17093637585639954, 0.018979258835315704, -0.004552137106657028, 0.10091845691204071, -0.029088309034705162, 0.02571181394159794, 0.09644908457994461, 0.0596768818795681, 0.07136417180299759, -0.12062377482652664, 0.06415046751499176, -0.07111187279224396, -0.11650864034891129, -0.0026051602326333523, -0.012400512583553791, 0.0628633052110672, 0.059111420065164566, 0.011643842794001102, -0.0719701424241066, 0.019275017082691193, -0.03647741675376892, 0.013253315351903439, 0.05309722200036049, -0.09309419989585876, 0.0625007376074791, -0.1001693457365036, -0.01173337735235691, 0.017126107588410378, 0.219380185008049, 0.04715661704540253, -0.0989660918712616, -0.02869831770658493, 0.0555335097014904, -0.02232193388044834, -0.04815635085105896, 0.14091436564922333, 0.11573789268732071, -0.046961039304733276, -0.10584709048271179, 0.05689861997961998, 0.03246960788965225, -0.0351080596446991, 0.01736387237906456, -0.058917418122291565, 0.10738559067249298, 0.04945773631334305, 0.08537707477807999, 0.1040673702955246, -0.07967091351747513, -0.04995207488536835, -0.05837250128388405, 0.03674735501408577, -0.09327123314142227, 0.17818963527679443, 0.10338497161865234, 0.02994465082883835, 0.04471870884299278, 0.017836906015872955, -0.0572773702442646, -0.1068369448184967, -0.23481668531894684, -0.06076874956488609, -0.26787230372428894, 0.016350576654076576, -0.07006313651800156, 0.057082388550043106, 0.00865968782454729, 0.08309553563594818, -0.0931737944483757, 0.10462047159671783, -0.0343654528260231, -0.15643221139907837, 0.0765305757522583, -0.033522266894578934, 0.02471216395497322, -0.10517802834510803, -0.033859699964523315, -0.007570345886051655, 0.14283525943756104, 0.03931865468621254, 0.09538811445236206, -0.0201384536921978, -0.047494422644376755, -0.15847860276699066, -0.009636351838707924, -0.0643884539604187, 0.036867544054985046, -0.016220442950725555, 0.049337077885866165, 0.03805724158883095, -0.08537539839744568, -0.0013495557941496372, 0.17658938467502594, -0.037319786846637726, -0.12016144394874573, -0.08236528187990189, 0.15102939307689667, -0.06989707052707672, 0.08057703822851181, -0.08710784465074539, -0.03468073531985283, -0.14039041101932526, 0.33637839555740356, 0.223253533244133, -0.015952441841363907, -0.009005704894661903, -0.03758938983082771, 0.055054694414138794, -0.014298629015684128, 0.18052753806114197, 0.14841216802597046, 0.18747015297412872, 0.015442004427313805, -0.010402360931038857, 0.0025995506439357996, -0.08323563635349274, -0.05244867876172066, -0.036220043897628784, 0.03889455646276474, -0.033403050154447556, -0.08991877734661102, 0.10672130435705185, -0.2844321131706238, 0.035635292530059814, -0.13988912105560303, -0.13092945516109467, -0.08537445962429047, -0.027599792927503586, -0.0032091138418763876, 0.09699629247188568, 0.06784532219171524, 0.015526041388511658, -0.028486886993050575, 0.003312452230602503, 0.02901773527264595, -0.19360476732254028, -0.12731888890266418, 0.10419207066297531, -0.10710938274860382, 0.009232496842741966, -0.0002444274432491511, 0.052976809442043304, 0.028577160090208054, 0.044852983206510544, 0.05869338661432266, 0.1068657636642456, -0.023219719529151917, -0.03401503711938858, -0.023068692535161972, -0.024059182032942772, 0.016591748222708702, -0.04024447873234749, 0.015715263783931732, -0.02731209062039852, -0.014216816052794456, 0.04868491739034653, 0.05939101427793503, -0.10514651238918304, 0.0033701774664223194, -0.1179482638835907, 0.05431809648871422, 0.10141105204820633, -0.017448795959353447, 0.027013259008526802, -0.009510166943073273, 0.0071176751516759396, -0.025997592136263847, -0.12394446134567261, -0.08923587948083878, -0.21010199189186096, -0.11999423056840897, 0.09350947290658951, -0.0032800177577883005, -0.11068188399076462, 0.029177015647292137, -0.10716219246387482, 0.09692387282848358, -0.06248395889997482, 0.10212694853544235, 0.09709987044334412, 0.06803819537162781, 0.011607633903622627, -0.11516279727220535, 0.03792379051446915, 0.12224683910608292, -0.11574938148260117, -0.04391450434923172 ]
null
null
transformers
# Tagalog DialoGPT A DialoGPT-medium model fine-tuned on Tagalog conversational data scraped from the web. This model is an output of a research on RoBERTa-based data augmentation for low resource languages. This is the baseline model which did not use any synthetic data in training. # Latest release: July 25, 2021 * The model is currently only able to respond based on the history of 3 previous utterances before being limited. This is a result of the scarce amount of Tagalog conversations in our dataset. # Dataset [PEx Conversations Dataset](https://huggingface.co/datasets/gabtan99/pex-conversations) # Usage Here is an example of using beam search for model inference. ``` for step in range(2): # encode the new user input, add the eos_token and return a tensor in Pytorch new_user_input_ids = tokenizer.encode(input(">> User:") + tokenizer.eos_token, return_tensors='pt') # append the new user input tokens to the chat history bot_input_ids = torch.cat([chat_history_ids, new_user_input_ids], dim=-1) if step > 0 else new_user_input_ids # we limit the generation to 512 tokens, each utterance in training had a maximum of 128 tokens chat_history_ids = model.generate( bot_input_ids, max_length=512, pad_token_id=tokenizer.eos_token_id, num_beams=5, no_repeat_ngram_size=3 ) # pretty print last ouput tokens from bot print("DialoGPT: {}".format(tokenizer.decode(chat_history_ids[:, bot_input_ids.shape[-1]:][0], skip_special_tokens=True))) ``` # Training Script [Fine-tuning script adapted from Spanish DialoGPT](https://colab.research.google.com/github/ncoop57/i-am-a-nerd/blob/master/_notebooks/2020-05-12-chatbot-part-1.ipynb) # Research by * [tyadrianpaule](https://huggingface.co/tyadrianpaule) * [schuylerng](https://huggingface.co/schuylerng) * [dcl127](https://huggingface.co/dcl127)
{"language": ["tl"], "tags": ["conversational", "tagalog", "filipino"], "datasets": ["gabtan99/pex-conversations"], "inference": false}
text-generation
gabtan99/dialogpt-tagalog-medium
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "tagalog", "filipino", "tl", "dataset:gabtan99/pex-conversations", "autotrain_compatible", "has_space", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "tl" ]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #tagalog #filipino #tl #dataset-gabtan99/pex-conversations #autotrain_compatible #has_space #text-generation-inference #region-us
# Tagalog DialoGPT A DialoGPT-medium model fine-tuned on Tagalog conversational data scraped from the web. This model is an output of a research on RoBERTa-based data augmentation for low resource languages. This is the baseline model which did not use any synthetic data in training. # Latest release: July 25, 2021 * The model is currently only able to respond based on the history of 3 previous utterances before being limited. This is a result of the scarce amount of Tagalog conversations in our dataset. # Dataset PEx Conversations Dataset # Usage Here is an example of using beam search for model inference. # Training Script Fine-tuning script adapted from Spanish DialoGPT # Research by * tyadrianpaule * schuylerng * dcl127
[ "# Tagalog DialoGPT\nA DialoGPT-medium model fine-tuned on Tagalog conversational data scraped from the web. This model is an output of a research on RoBERTa-based data augmentation for low resource languages. This is the baseline model which did not use any synthetic data in training.", "# Latest release: July 25, 2021\n* The model is currently only able to respond based on the history of 3 previous utterances before being limited. This is a result of the scarce amount of Tagalog conversations in our dataset.", "# Dataset\nPEx Conversations Dataset", "# Usage\nHere is an example of using beam search for model inference.", "# Training Script\nFine-tuning script adapted from Spanish DialoGPT", "# Research by\n* tyadrianpaule\n* schuylerng\n* dcl127" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #tagalog #filipino #tl #dataset-gabtan99/pex-conversations #autotrain_compatible #has_space #text-generation-inference #region-us \n", "# Tagalog DialoGPT\nA DialoGPT-medium model fine-tuned on Tagalog conversational data scraped from the web. This model is an output of a research on RoBERTa-based data augmentation for low resource languages. This is the baseline model which did not use any synthetic data in training.", "# Latest release: July 25, 2021\n* The model is currently only able to respond based on the history of 3 previous utterances before being limited. This is a result of the scarce amount of Tagalog conversations in our dataset.", "# Dataset\nPEx Conversations Dataset", "# Usage\nHere is an example of using beam search for model inference.", "# Training Script\nFine-tuning script adapted from Spanish DialoGPT", "# Research by\n* tyadrianpaule\n* schuylerng\n* dcl127" ]
[ 71, 71, 49, 10, 17, 16, 19 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #tagalog #filipino #tl #dataset-gabtan99/pex-conversations #autotrain_compatible #has_space #text-generation-inference #region-us \n# Tagalog DialoGPT\nA DialoGPT-medium model fine-tuned on Tagalog conversational data scraped from the web. This model is an output of a research on RoBERTa-based data augmentation for low resource languages. This is the baseline model which did not use any synthetic data in training.# Latest release: July 25, 2021\n* The model is currently only able to respond based on the history of 3 previous utterances before being limited. This is a result of the scarce amount of Tagalog conversations in our dataset.# Dataset\nPEx Conversations Dataset# Usage\nHere is an example of using beam search for model inference.# Training Script\nFine-tuning script adapted from Spanish DialoGPT# Research by\n* tyadrianpaule\n* schuylerng\n* dcl127" ]
[ -0.10948849469423294, 0.04756765067577362, -0.0013239843538030982, 0.04513074830174446, 0.12073249369859695, -0.04256940260529518, 0.09715045243501663, 0.1413862705230713, -0.02418329007923603, 0.07064968347549438, 0.03404878452420235, 0.0195822361856699, 0.05304250493645668, 0.13499747216701508, 0.023869572207331657, -0.25297215580940247, 0.03568092733621597, -0.03374338522553444, 0.08311137557029724, 0.11527923494577408, 0.1030326783657074, -0.01972632110118866, 0.06807636469602585, 0.02262343280017376, -0.1159864217042923, 0.05151544138789177, 0.003912475425750017, -0.12315008044242859, 0.09917058050632477, 0.0767144188284874, 0.06014769896864891, 0.0070919375866651535, 0.030215421691536903, -0.10251237452030182, 0.024195516481995583, 0.02422466315329075, 0.015602197498083115, 0.056755490601062775, 0.06723817437887192, -0.07128087431192398, 0.22163772583007812, 0.0503951720893383, 0.05517774075269699, 0.08380527049303055, -0.1633378118276596, -0.06638458371162415, -0.06240665912628174, 0.002652072114869952, 0.03599638119339943, 0.13695979118347168, -0.055835261940956116, 0.12401170283555984, -0.06266482174396515, 0.0077636404894292355, 0.11094632744789124, -0.26602935791015625, -0.014353649690747261, 0.11794397979974747, 0.10274828970432281, 0.028443727642297745, -0.04524639621376991, 0.03548211604356766, 0.007111567072570324, 0.007158046122640371, 0.043651774525642395, -0.012013445608317852, -0.08936816453933716, -0.06810630112886429, -0.11442884802818298, -0.008789501152932644, 0.2145131379365921, -0.016252079978585243, -0.04916689544916153, -0.13260143995285034, -0.01609113998711109, 0.04811911657452583, -0.024099325761198997, -0.06391323357820511, -0.01593036949634552, 0.04160866141319275, 0.06721900403499603, -0.11645112186670303, -0.08672083169221878, -0.09472298622131348, -0.10605308413505554, 0.12998633086681366, 0.03380856290459633, -0.0004010631237179041, -0.09061800688505173, 0.09347693622112274, -0.06985650211572647, -0.05896594747900963, -0.03088485635817051, -0.014501556754112244, -0.005865123122930527, -0.018474269658327103, -0.03241242840886116, -0.0702868103981018, 0.004141254350543022, 0.15262874960899353, 0.020581120625138283, 0.03551478311419487, -0.01924668252468109, 0.023512303829193115, -0.006190169136971235, 0.10306265950202942, -0.07930067181587219, -0.0028521264903247356, 0.04224396497011185, 0.060826607048511505, 0.008304157294332981, -0.004485161975026131, -0.11252632737159729, -0.06188972294330597, 0.015168514102697372, 0.08412815630435944, 0.06500490009784698, 0.06965938210487366, -0.05210736021399498, -0.06676807254552841, 0.11777176707983017, -0.12280149757862091, -0.05532205477356911, 0.0038485887926071882, -0.13966847956180573, 0.045539967715740204, 0.059239260852336884, 0.03684420511126518, -0.07910660654306412, -0.12213723361492157, -0.08408275246620178, -0.004561776295304298, -0.03917786851525307, -0.08091700822114944, 0.019441189244389534, 0.002916989615187049, -0.06995078921318054, -0.14890463650226593, -0.23636452853679657, -0.05965612456202507, -0.05089326202869415, -0.09482455998659134, 0.028288649395108223, -0.11135713011026382, 0.006612231023609638, -0.010505103506147861, -0.020274605602025986, -0.024559833109378815, -0.03757074102759361, 0.0368172712624073, 0.007510042283684015, 0.05554952099919319, -0.062498509883880615, 0.05611024051904678, -0.11665132641792297, -0.02499454654753208, -0.15601496398448944, 0.2161930650472641, -0.024173002690076828, -0.016461629420518875, -0.17871861159801483, -0.07817640155553818, -0.10493385791778564, 0.06789792329072952, 0.04240509122610092, 0.19812214374542236, -0.16871123015880585, -0.06055620312690735, 0.26888570189476013, -0.08719117194414139, -0.0683189332485199, 0.1044149398803711, -0.005957666784524918, 0.14298835396766663, 0.1383681297302246, 0.20511120557785034, -0.0015710117295384407, -0.09240826964378357, 0.0028796710539609194, -0.023500455543398857, -0.046930860728025436, -0.0008057748782448471, 0.10457970947027206, -0.09692943096160889, -0.0013337245909497142, 0.019549861550331116, -0.09232785552740097, 0.07191191613674164, -0.019943207502365112, -0.04701463133096695, 0.03808855637907982, -0.0314490869641304, 0.04441608116030693, -0.02382715791463852, 0.09382043033838272, -0.0034353001974523067, -0.11540576070547104, 0.053274478763341904, 0.10109597444534302, -0.03523784875869751, 0.0030584861524403095, -0.16049540042877197, 0.11111563444137573, -0.00802257377654314, 0.01388950739055872, -0.15008625388145447, -0.08474567532539368, 0.022318661212921143, 0.05048701539635658, 0.06465165317058563, -0.07802558690309525, 0.04811559617519379, 0.0242520309984684, -0.023672189563512802, 0.014928572811186314, -0.0013720659771934152, -0.03522869199514389, -0.046356815844774246, -0.08224482834339142, -0.006623739842325449, -0.039341554045677185, 0.1013806089758873, -0.16112111508846283, 0.030564401298761368, -0.006943551357835531, 0.06770259141921997, -0.017033100128173828, -0.06016194820404053, 0.08400163799524307, -0.0030111323576420546, -0.01256392989307642, -0.08312581479549408, 0.06400281190872192, 0.002539471024647355, -0.1083773598074913, 0.1181483194231987, -0.19143790006637573, -0.03399958088994026, 0.08550160378217697, 0.00636983523145318, -0.05603891983628273, -0.12402904033660889, -0.025859827175736427, -0.009995377622544765, -0.01169864647090435, -0.08439815789461136, 0.23577317595481873, -0.014183544553816319, 0.1378849297761917, -0.12768623232841492, 0.022733204066753387, -0.02329793944954872, -0.02929455228149891, -0.014059484004974365, 0.06874681264162064, -0.00033860167604871094, -0.05969373136758804, 0.06469365954399109, -0.043168291449546814, 0.024652840569615364, 0.1968800574541092, 0.01131501980125904, -0.052585404366254807, 0.01070482935756445, 0.07005973160266876, 0.015460208989679813, -0.01126446295529604, -0.09766548126935959, -0.04949469864368439, 0.04755508527159691, 0.0491781160235405, 0.08815917372703552, -0.08595146238803864, -0.01548755168914795, -0.0041631911881268024, -0.059505753219127655, -0.0027018184773623943, 0.09142526239156723, -0.039101213216781616, 0.11184021830558777, 0.04825734719634056, -0.012246652506291866, -0.027154870331287384, -0.029099322855472565, -0.11486238986253738, 0.18486520648002625, -0.11171562224626541, -0.22564928233623505, -0.11133704334497452, 0.06721403449773788, -0.01498947013169527, 0.07323186844587326, 0.024461159482598305, -0.10964140295982361, 0.0007352581014856696, -0.1023898720741272, 0.10803627967834473, -0.01567106693983078, -0.05642092600464821, -0.05699769780039787, -0.030683154240250587, -0.004705757834017277, -0.139176145195961, -0.015579947270452976, -0.03260745108127594, -0.17403939366340637, -0.019127504900097847, -0.14857609570026398, 0.04164774343371391, 0.1824125349521637, 0.04562382400035858, 0.006098708137869835, -0.08656476438045502, 0.16858826577663422, -0.06308209896087646, -0.004756203852593899, 0.2530127763748169, 0.0536997951567173, -0.009133885614573956, 0.025880679488182068, -0.00449112243950367, -0.11422578245401382, 0.03971507027745247, -0.008566745556890965, -0.07866895198822021, -0.2157880812883377, -0.1332414299249649, -0.02898786962032318, -0.02296607941389084, 0.0622890330851078, 0.037816550582647324, 0.11602035164833069, 0.04454623907804489, -0.05171686038374901, -0.034371472895145416, 0.0664316713809967, 0.06640862673521042, 0.1390470713376999, 0.0039605130441486835, 0.11099174618721008, 0.0002737916074693203, 0.002625502413138747, 0.07599122822284698, 0.0013980745570734143, 0.13084381818771362, -0.011361329816281796, 0.15897074341773987, 0.06038367748260498, 0.0684230849146843, 0.024905776605010033, -0.010583002120256424, 0.010105426423251629, 0.012572276405990124, -0.030599510297179222, -0.08018095791339874, -0.05570318549871445, 0.030967675149440765, 0.011230194009840488, -0.04168543964624405, -0.09315843135118484, 0.031471122056245804, 0.05615656450390816, 0.054285645484924316, 0.02497047558426857, -0.28452831506729126, -0.09207416325807571, -0.02331005595624447, -0.001590452971868217, -0.04199565574526787, 0.05229644477367401, 0.058384958654642105, -0.1368103325366974, 0.0002558128035161644, 0.004985700827091932, 0.0926596149802208, -0.053233884274959564, -0.0043213884346187115, -0.07200782746076584, 0.08827520906925201, -0.02034575864672661, 0.08143244683742523, -0.1839408427476883, 0.1435929536819458, 0.010464943014085293, 0.12226749956607819, -0.056231144815683365, -0.011819667182862759, 0.04496442526578903, -0.03580666705965996, 0.13732799887657166, 0.010714169591665268, 0.10427822172641754, -0.06279570609331131, -0.12662015855312347, 0.033482179045677185, 0.036675479263067245, -0.08088445663452148, 0.05187622085213661, -0.0314374715089798, 0.04957643523812294, 0.01582920551300049, -0.10121644288301468, -0.12439171224832535, -0.09864988923072815, 0.06853584200143814, 0.06207925081253052, 0.06827898323535919, -0.039632320404052734, -0.061235956847667694, 0.007312094792723656, 0.09977048635482788, -0.06908153742551804, -0.08608709275722504, -0.12035723030567169, 0.1278788149356842, 0.14307494461536407, -0.06822332739830017, 0.0899655818939209, 0.013005880638957024, 0.038103535771369934, 0.029691455885767937, -0.09579414129257202, 0.10987768322229385, -0.11006200313568115, -0.10175896435976028, -0.009673305787146091, 0.16181467473506927, 0.10361208766698837, 0.027338692918419838, 0.05249432101845741, 0.008046506904065609, 0.016329960897564888, -0.08195435255765915, -0.014236027374863625, 0.08392103016376495, 0.03591472655534744, 0.07515136152505875, -0.05537230148911476, -0.06847304850816727, -0.10915187746286392, -0.048299528658390045, 0.15262658894062042, 0.08051268756389618, -0.03862431272864342, 0.09425772726535797, 0.11646414548158646, -0.07332850247621536, -0.18570257723331451, -0.012061042711138725, 0.08267448842525482, 0.018236462026834488, -0.016149060800671577, -0.23622193932533264, 0.03962477669119835, 0.09106552600860596, -0.02236461453139782, 0.0037221673410385847, -0.4270498752593994, -0.1319032460451126, 0.10079972445964813, 0.0034515419974923134, 0.17858119308948517, -0.1330164521932602, -0.0749584510922432, 0.032228585332632065, -0.0702739953994751, 0.08498962223529816, -0.056827884167432785, 0.08572329580783844, 0.008055704645812511, 0.058344610035419464, 0.04386933520436287, 0.008325953036546707, 0.10607296973466873, 0.038972266018390656, 0.002925364300608635, -0.05446404218673706, -0.026802921667695045, 0.05053571239113808, -0.006768080405890942, 0.1378553956747055, -0.019091052934527397, 0.049350839108228683, -0.12581010162830353, -0.08157653361558914, -0.115484818816185, 0.09442051500082016, 0.025440426543354988, -0.043301861733198166, -0.04728211835026741, 0.03396642580628395, 0.03213553875684738, 0.02678230218589306, -0.042461782693862915, -0.10476173460483551, 0.12025187909603119, 0.013558972626924515, 0.1589246243238449, 0.06338848918676376, -0.08791393041610718, 0.028856836259365082, 0.01098617259413004, 0.09498845785856247, -0.11838503181934357, -0.011458545923233032, 0.10596417635679245, 0.011495460756123066, 0.13855060935020447, 0.04910685867071152, -0.11964314430952072, 0.06741119921207428, 0.08723101019859314, -0.03959076106548309, -0.15359336137771606, -0.03518335893750191, 0.10502275079488754, -0.016958341002464294, 0.005280556622892618, 0.13101978600025177, -0.06617952138185501, -0.05116025358438492, 0.0077172815799713135, 0.02983933873474598, -0.016447357833385468, 0.15111899375915527, 0.022352954372763634, -0.0002318232145626098, -0.0851946696639061, 0.13681179285049438, 0.04396580904722214, -0.029300624504685402, 0.1037713885307312, 0.09061365574598312, -0.10418602079153061, -0.05435856804251671, -0.08717533946037292, 0.07263465970754623, -0.1439344584941864, -0.03875849395990372, -0.03252242133021355, -0.05372465029358864, 0.01571003533899784, 0.0697731152176857, 0.02320072427392006, 0.029053345322608948, -0.06928814202547073, 0.04214540496468544, -0.03299074247479439, 0.01091044396162033, 0.0861276388168335, -0.03308353200554848, -0.04399355128407478, 0.15284284949302673, 0.05403304845094681, -0.030550824478268623, -0.0489492304623127, -0.06593625247478485, -0.0985737070441246, 0.018427180126309395, -0.07970497757196426, -0.030734268948435783, -0.10743378102779388, -0.040203437209129333, -0.041392117738723755, -0.05458156764507294, 0.001070674741640687, 0.015724103897809982, -0.05795086547732353, -0.034388575702905655, -0.05069991573691368, 0.06553175300359726, -0.10436959564685822, 0.013773580081760883, 0.020486824214458466, -0.03178197890520096, 0.12308480590581894, 0.12853743135929108, -0.002641604281961918, 0.08557483553886414, -0.16350288689136505, 0.016487348824739456, 0.037389930337667465, 0.013005836866796017, -0.023698993027210236, -0.029693713411688805, 0.004143295343965292, 0.013850513845682144, 0.03905680775642395, 0.005103773437440395, 0.06364933401346207, -0.090123251080513, -0.010068110190331936, -0.010034295730292797, 0.0240987166762352, -0.05891009420156479, 0.07219786942005157, 0.08629747480154037, 0.10180068016052246, 0.04110237583518028, -0.08514633029699326, 0.029871154576539993, -0.11918200552463531, 0.0008461312972940505, -0.05576709657907486, -0.019715556874871254, -0.03240232914686203, -0.026046432554721832, 0.06901811063289642, 0.01188463345170021, 0.10769841074943542, 0.030451705679297447, -0.019928863272070885, 0.004459300544112921, 0.03941573575139046, 0.06626423448324203, -0.004695456009358168, 0.139093816280365, 0.10813093185424805, -0.006319062318652868, 0.020650899037718773, 0.05457637459039688, 0.03786103054881096, 0.007925859652459621, 0.11461716890335083, 0.1256495863199234, -0.01786964200437069, 0.08358527719974518, 0.04023206606507301, -0.10457023978233337, -0.08957631886005402, -0.09071052819490433, -0.03060612455010414, 0.03966890648007393, -0.04544052854180336, 0.11846446990966797, 0.14310692250728607, -0.059071771800518036, 0.05714986100792885, 0.02329174056649208, -0.07640723884105682, -0.13325099647045135, -0.14839832484722137, -0.00943796057254076, -0.08173403143882751, -0.036547448486089706, -0.11095066368579865, 0.013205183669924736, 0.03726743161678314, 0.05247776582837105, -0.032657038420438766, 0.20964403450489044, -0.05719400942325592, -0.17461663484573364, 0.08534138649702072, -0.04235142469406128, 0.0434732660651207, -0.051763780415058136, -0.012960776686668396, 0.0646897628903389, 0.07511653751134872, 0.07466170936822891, 0.043609898537397385, -0.012709192000329494, 0.02059878036379814, -0.09353899955749512, -0.06980632245540619, -0.030133895576000214, 0.05302860215306282, -0.01980254054069519, 0.10719677805900574, 0.052991319447755814, -0.08004416525363922, 0.009627964347600937, 0.25568655133247375, 0.0035307046491652727, -0.01502920500934124, -0.13286122679710388, 0.20045357942581177, 0.005707064177840948, -0.015985090285539627, -0.08376608043909073, -0.07834014296531677, -0.009184564463794231, 0.24417898058891296, 0.18200910091400146, -0.0446309894323349, -0.0378606915473938, -0.028113706037402153, 0.006997633725404739, 0.04921784996986389, 0.09424888342618942, 0.02836155891418457, 0.24276530742645264, -0.04291103407740593, 0.06775452196598053, 0.000977582996711135, -0.03977963328361511, -0.013561083935201168, 0.0336238294839859, -0.0014254511334002018, 0.018271898850798607, -0.047604359686374664, 0.09068021178245544, -0.10139860212802887, -0.14568881690502167, -0.08089704811573029, -0.0863579511642456, -0.09611962735652924, -0.06571509689092636, -0.10456150025129318, 0.08948342502117157, 0.1073082983493805, 0.006720183417201042, 0.008730965666472912, -0.008843157440423965, 0.026508117094635963, -0.14693240821361542, -0.09494133293628693, 0.09677901864051819, 0.05255138501524925, 0.13035114109516144, -0.059374138712882996, 0.10437469929456711, 0.06200307980179787, 0.04651142656803131, -0.11486257612705231, 0.041795987635850906, -0.014519240707159042, 0.07674315571784973, 0.06097272410988808, 0.05523703619837761, 0.013165457174181938, 0.007776548620313406, 0.055457815527915955, -0.1100275069475174, 0.02382359653711319, 0.025007613003253937, -0.06291207671165466, -0.1039976254105568, 0.02251521125435829, -0.08574359118938446, 0.14434227347373962, 0.13088656961917877, -0.06788275390863419, 0.0033469628542661667, -0.02902970276772976, 0.03073420748114586, 0.03170675039291382, 0.07647336274385452, -0.03073514625430107, -0.23864205181598663, -0.02502680942416191, -0.06503652781248093, 0.004446420352905989, -0.23672638833522797, 0.012233834713697433, -0.06978204846382141, -0.007800647523254156, 0.003595454152673483, 0.12573063373565674, 0.047294557094573975, 0.03898344933986664, -0.01346175093203783, -0.0060662804171442986, 0.013675264082849026, 0.13103586435317993, -0.16031625866889954, -0.044152047485113144 ]
null
null
null
I am adding my first README in order to test the interface. How good is it really?
{}
null
gael1130/gael_first_model
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
I am adding my first README in order to test the interface. How good is it really?
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
transformers
This model is used in the paper **Generative Relation Linking for Question Answering over Knowledge Bases**. [ArXiv](https://arxiv.org/abs/2108.07337), [GitHub](https://github.com/IBM/kbqa-relation-linking) ## Citation ```bibtex @inproceedings{rossiello-genrl-2021, title={Generative relation linking for question answering over knowledge bases}, author={Rossiello, Gaetano and Mihindukulasooriya, Nandana and Abdelaziz, Ibrahim and Bornea, Mihaela and Gliozzo, Alfio and Naseem, Tahira and Kapanipathi, Pavan}, booktitle={International Semantic Web Conference}, pages={321--337}, year={2021}, organization={Springer}, url = "https://link.springer.com/chapter/10.1007/978-3-030-88361-4_19", doi = "10.1007/978-3-030-88361-4_19" } ```
{"license": "apache-2.0"}
text2text-generation
gaetangate/bart-large_genrl_lcquad1
[ "transformers", "pytorch", "bart", "text2text-generation", "arxiv:2108.07337", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2108.07337" ]
[]
TAGS #transformers #pytorch #bart #text2text-generation #arxiv-2108.07337 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
This model is used in the paper Generative Relation Linking for Question Answering over Knowledge Bases. ArXiv, GitHub
[]
[ "TAGS\n#transformers #pytorch #bart #text2text-generation #arxiv-2108.07337 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #bart #text2text-generation #arxiv-2108.07337 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.05482131987810135, 0.07781752198934555, -0.005638949107378721, 0.022345583885908127, 0.08416979014873505, 0.00895537156611681, 0.1526346057653427, 0.13391147553920746, -0.0098402239382267, -0.04337620735168457, 0.1866959184408188, 0.21171580255031586, 0.0014640424633398652, 0.08603023737668991, -0.0680188313126564, -0.17926882207393646, 0.07028527557849884, 0.06794054061174393, 0.03506860509514809, 0.11241799592971802, 0.11079103499650955, -0.028155889362096786, 0.05943622440099716, 0.007174516562372446, -0.10733117908239365, 0.030625032261013985, 0.052361808717250824, -0.10477364808320999, 0.10313612967729568, 0.04243432357907295, 0.08389689028263092, 0.046983979642391205, -0.030626261606812477, -0.15681667625904083, 0.012176413089036942, 0.0015603452920913696, -0.058309849351644516, 0.0642048567533493, 0.07419206202030182, -0.04541759938001633, 0.08018051087856293, 0.05972139164805412, -0.036842163652181625, 0.059354767203330994, -0.0846671313047409, -0.163845032453537, -0.11407594382762909, 0.09986864775419235, 0.06210153549909592, 0.11018134653568268, 0.05159074440598488, 0.14422264695167542, -0.051009975373744965, 0.07602405548095703, 0.17427119612693787, -0.3681010901927948, 0.004177485127002001, 0.02464538998901844, 0.04596325010061264, 0.006552176084369421, -0.0007694204105064273, 0.020827608183026314, 0.06091023609042168, 0.035478681325912476, 0.003229860682040453, -0.06632944941520691, -0.14235669374465942, 0.02472204528748989, -0.03857053443789482, -0.08070261031389236, 0.24721753597259521, -0.024514835327863693, 0.012871340848505497, -0.0017009302973747253, -0.07462748140096664, -0.0014764306833967566, -0.013516562059521675, 0.04367212578654289, -0.0050034900195896626, 0.09492382407188416, 0.04887858033180237, -0.04823123291134834, -0.15984714031219482, 0.00207934295758605, -0.20301559567451477, 0.08240332454442978, -0.0022943257354199886, 0.09173347055912018, -0.1661391705274582, 0.06920570880174637, 0.0059766024351119995, -0.12174828350543976, 0.010891247540712357, -0.0597551204264164, 0.15621905028820038, 0.05395321547985077, -0.0691797286272049, -0.0048287902027368546, 0.09994350373744965, 0.23596175014972687, 0.04119456186890602, 0.01503798644989729, -0.06155439466238022, 0.09958167374134064, -0.07166501879692078, 0.05998682230710983, 0.05347549170255661, -0.05067494511604309, 0.10866204649209976, -0.1072971373796463, 0.09862447530031204, -0.033908989280462265, -0.16214343905448914, -0.0692695900797844, 0.002642169129103422, 0.10096638649702072, 0.08890148997306824, 0.028074882924556732, -0.04016926512122154, 0.01901845633983612, 0.16369247436523438, -0.060025133192539215, 0.0010842434130609035, -0.011274623684585094, 0.019521581009030342, 0.0889275074005127, 0.06706152856349945, 0.03201710432767868, -0.07721319049596786, 0.07256565243005753, -0.04542548581957817, -0.0067802430130541325, -0.03321782127022743, -0.01449374109506607, 0.0770699754357338, -0.07354425638914108, 0.039060868322849274, -0.16507947444915771, -0.18075479567050934, 0.02805604226887226, 0.044291190803050995, -0.005039107520133257, -0.07841252535581589, 0.031263597309589386, -0.016256261616945267, 0.06746867299079895, -0.08927186578512192, 0.023773107677698135, -0.06050211563706398, 0.07617700845003128, -0.03449065983295441, 0.03464970737695694, -0.1939922422170639, 0.05469327047467232, -0.11378049850463867, -0.017431724816560745, -0.025046037510037422, -0.04177875444293022, -0.05100778117775917, 0.16310332715511322, -0.05167950689792633, -0.023722611367702484, -0.032557837665081024, 0.0038679158315062523, 0.013324900530278683, 0.13822630047798157, -0.09300900250673294, -0.05209609121084213, 0.20359504222869873, -0.10541636496782303, -0.19766826927661896, 0.05312618613243103, 0.03264372795820236, 0.016446685418486595, 0.06077985465526581, 0.14917801320552826, 0.04373243823647499, -0.04614674672484398, 0.06058279424905777, 0.12388131022453308, -0.07293485105037689, -0.2121901661157608, 0.041465435177087784, -0.06329382956027985, -0.110641710460186, 0.06406300514936447, 0.004137752577662468, 0.09876012057065964, 0.0005344409146346152, -0.06045414134860039, -0.08519785851240158, -0.04539100080728531, 0.002937682904303074, -0.007939782924950123, 0.07327837496995926, -0.08597764372825623, -0.01201961562037468, -0.06457104533910751, 0.03227653354406357, 0.047538693994283676, 0.06602434813976288, -0.038587696850299835, 0.07384616136550903, 0.0036276394966989756, 0.025109659880399704, -0.1279795616865158, 0.03790568932890892, -0.01887362077832222, 0.05401679128408432, 0.0003365709853824228, 0.04780033603310585, 0.052929893136024475, -0.07011021673679352, 0.007165586110204458, -0.016639281064271927, 0.13530054688453674, 0.024718094617128372, -0.04768262803554535, -0.09143251925706863, 0.05007249489426613, -0.03446045145392418, 0.029898418113589287, -0.018560517579317093, 0.021972866728901863, 0.002617587335407734, 0.10428658872842789, -0.04944944009184837, 0.07664937525987625, -0.03424760326743126, -0.008887539617717266, -0.07868296653032303, 0.004296313505619764, 0.11207737028598785, 0.06150135025382042, -0.06777924299240112, 0.23034334182739258, -0.09824039787054062, 0.23176263272762299, 0.21014845371246338, -0.16900676488876343, 0.062023747712373734, -0.006118505261838436, -0.026152264326810837, -0.022883309051394463, 0.06577979773283005, 0.013145336881279945, 0.018537122756242752, 0.019990185275673866, 0.17416749894618988, -0.05235980451107025, -0.027067814022302628, -0.011080794967710972, -0.07247885316610336, -0.00978275015950203, 0.04679108038544655, 0.08780468255281448, -0.1469167172908783, 0.16849613189697266, 0.31132781505584717, -0.013906165026128292, 0.07057876884937286, -0.03862464800477028, -0.00969420000910759, 0.03858736529946327, -0.02555689960718155, -0.01832895167171955, -0.00753486854955554, -0.12924771010875702, 0.009394427761435509, 0.10876920074224472, 0.015896691009402275, 0.07074429839849472, -0.13710680603981018, -0.045919403433799744, 0.003918658941984177, -0.01009833998978138, -0.032725732773542404, 0.07581792026758194, 0.008631033822894096, 0.10835693031549454, -0.03264651447534561, -0.06437046080827713, 0.123822420835495, 0.029191283509135246, -0.09418227523565292, 0.1392623484134674, -0.17335090041160583, -0.26775309443473816, -0.1764364242553711, -0.12391039729118347, -0.030479945242404938, 0.015311352908611298, 0.15357957780361176, -0.03749791532754898, -0.048915065824985504, -0.00007263442239491269, -0.08474317193031311, -0.031013313680887222, 0.005559484474360943, 0.005002529360353947, 0.03174421191215515, 0.024705540388822556, -0.12703348696231842, -0.05319864675402641, 0.02746308036148548, -0.023468509316444397, 0.09643476456403732, -0.07996387779712677, 0.08806337416172028, 0.0958549901843071, 0.020643893629312515, 0.02932734414935112, -0.00825952086597681, 0.13522061705589294, -0.014479542151093483, 0.0013834310229867697, 0.25092613697052, -0.013742741197347641, 0.08801493793725967, 0.11183018237352371, 0.010136323049664497, -0.04858143627643585, 0.009055139496922493, -0.06658154726028442, -0.07651500403881073, -0.2714214622974396, -0.09263240545988083, -0.10003478080034256, 0.06438066810369492, 0.061698928475379944, 0.07818742096424103, 0.13583698868751526, 0.08747946470975876, -0.03248957544565201, 0.030506648123264313, -0.03211503103375435, 0.09743855148553848, 0.2555447220802307, -0.025176601484417915, 0.12891273200511932, -0.1269153654575348, -0.030954653397202492, 0.12694242596626282, 0.08371598273515701, 0.10590073466300964, 0.10160430520772934, 0.051557447761297226, 0.05325188860297203, 0.17843466997146606, 0.07512391358613968, 0.13275974988937378, 0.03223362937569618, -0.0218141358345747, -0.04455477371811867, -0.029173532500863075, -0.0705365464091301, 0.05708703026175499, -0.07295859605073929, -0.11087924987077713, -0.026086905971169472, -0.12267530709505081, 0.04912329465150833, 0.15167374908924103, 0.03260583057999611, -0.17569538950920105, 0.004984318278729916, 0.0684858188033104, -0.010155906900763512, -0.07671070843935013, 0.07608667016029358, -0.0805806964635849, -0.10683809220790863, 0.11846431344747543, -0.025947168469429016, 0.12722750008106232, 0.034859783947467804, 0.05635649710893631, -0.05809636041522026, -0.10959172993898392, 0.06383580714464188, 0.13386929035186768, -0.34036576747894287, 0.19154304265975952, -0.013072090223431587, -0.026505006477236748, -0.09062088280916214, 0.007474956102669239, 0.033928144723176956, 0.1653033196926117, 0.07942576706409454, 0.02359463833272457, -0.0978393405675888, -0.013540403917431831, -0.04927901178598404, 0.040264792740345, 0.026391852647066116, 0.0354730524122715, -0.0479319766163826, -0.060511134564876556, -0.017747119069099426, -0.0016101868823170662, 0.056854475289583206, -0.043094977736473083, -0.15984880924224854, 0.07390711456537247, 0.1282113492488861, 0.03012380562722683, -0.056982096284627914, -0.006998253054916859, -0.05290612950921059, 0.19696283340454102, -0.08011247217655182, -0.07158086448907852, -0.09493527561426163, -0.12436679005622864, 0.05385299399495125, -0.07199086993932724, 0.06745593994855881, -0.09251400828361511, -0.011777443811297417, -0.07750517874956131, -0.19208881258964539, 0.09629534929990768, -0.13449956476688385, -0.04357421398162842, -0.04016205295920372, 0.1267463117837906, -0.10915971547365189, 0.01363675482571125, 0.031107960268855095, 0.0045674326829612255, -0.1442185789346695, -0.10512310266494751, -0.01466408558189869, 0.01390861440449953, 0.07032927870750427, -0.06855132430791855, -0.059206217527389526, -0.04210440069437027, 0.0002269123069709167, -0.07095591723918915, 0.23760126531124115, 0.1743471324443817, -0.09102169424295425, 0.18125031888484955, 0.16369777917861938, -0.08062667399644852, -0.27284666895866394, -0.17046639323234558, -0.1256026327610016, -0.07379941642284393, -0.02152799814939499, -0.13010862469673157, 0.0928943082690239, 0.03854353725910187, -0.08531101793050766, 0.13309340178966522, -0.20278358459472656, -0.08868534117937088, 0.19920559227466583, -0.029639476910233498, 0.3354334533214569, -0.14678631722927094, -0.09665258973836899, -0.09883969277143478, -0.23874326050281525, 0.12705844640731812, -0.05784017592668533, 0.05659428983926773, -0.05013212561607361, 0.09030452370643616, -0.0055581978522241116, -0.07001134753227234, 0.1165020614862442, -0.019615652039647102, 0.011899770237505436, -0.12998872995376587, 0.015354842878878117, 0.05959603562951088, -0.025507992133498192, 0.0682363361120224, -0.16799941658973694, 0.04249343276023865, -0.1297016143798828, -0.00840586144477129, -0.08478473871946335, 0.0836782157421112, -0.00001981135028472636, -0.03807370737195015, -0.020114202052354813, -0.0809778943657875, 0.00864352472126484, -0.005378738511353731, 0.21508856117725372, 0.021311039105057716, 0.1004076674580574, 0.1621607393026352, 0.0570119246840477, -0.19071568548679352, 0.018602002412080765, -0.09594377875328064, -0.09199532866477966, 0.04919834062457085, -0.14196406304836273, 0.03942728415131569, 0.09658144414424896, -0.03966351971030235, 0.04124578833580017, 0.062003325670957565, 0.002129205269739032, -0.038048967719078064, 0.13631731271743774, -0.1981479525566101, 0.06881100684404373, -0.020387141034007072, 0.16533978283405304, 0.06280216574668884, 0.04884500801563263, 0.14566610753536224, 0.014593229629099369, -0.046061012893915176, 0.005350813735276461, 0.019779937341809273, -0.07924901694059372, 0.04157470911741257, 0.05258270353078842, 0.002496576402336359, -0.11633665859699249, 0.08739665150642395, -0.0006512592080980539, -0.1388893723487854, 0.00044502574019134045, 0.13315072655677795, -0.13069625198841095, -0.11353246122598648, 0.013988176360726357, 0.05645694583654404, -0.21390998363494873, -0.11363053321838379, -0.043982721865177155, -0.10140077024698257, 0.07960906624794006, 0.10171038657426834, 0.077805295586586, 0.0663294568657875, -0.011303826235234737, -0.07232298702001572, 0.023237422108650208, -0.023759078234434128, -0.048108603805303574, 0.04942779242992401, -0.07055307179689407, -0.015623528510332108, 0.00967564806342125, 0.09083372354507446, -0.036806222051382065, 0.0076259043999016285, -0.07885701209306717, 0.024485094472765923, -0.18741877377033234, -0.006544181611388922, -0.09549658745527267, -0.034454576671123505, 0.007622649893164635, -0.07942161709070206, -0.04239603504538536, 0.012974384240806103, -0.11313555389642715, -0.03135130554437637, -0.03334683179855347, 0.03711055591702461, -0.11675328016281128, -0.03606405481696129, 0.08109122514724731, -0.007913854904472828, 0.08382098376750946, 0.1311119794845581, -0.06978994607925415, 0.06548008322715759, -0.152426078915596, -0.09907804429531097, 0.07366190105676651, 0.03581872954964638, 0.014548269100487232, 0.009208097122609615, -0.005308505147695541, 0.14182445406913757, -0.005868135020136833, 0.02698882669210434, 0.06259270012378693, -0.14154714345932007, -0.036849476397037506, -0.002217935398221016, -0.1273764669895172, -0.0359746553003788, -0.08520586043596268, 0.11702703684568405, 0.02218184620141983, 0.15673206746578217, -0.0330800786614418, 0.04871964454650879, -0.052613042294979095, 0.030562758445739746, -0.03907517343759537, -0.14916402101516724, -0.1273747682571411, -0.06134314462542534, -0.0387486107647419, 0.002467484911903739, 0.21433435380458832, -0.029699983075261116, -0.03970152512192726, 0.059296615421772, 0.08205589652061462, -0.011385519988834858, -0.02196643315255642, 0.24418558180332184, 0.04766358062624931, -0.009007682092487812, -0.08916512131690979, 0.046584710478782654, 0.0018661929061636329, -0.051556818187236786, 0.08721576631069183, 0.08817291259765625, 0.067312091588974, 0.0969495177268982, 0.014695269986987114, -0.0016428417293354869, -0.13492125272750854, -0.17150603234767914, 0.002291272394359112, 0.0770316794514656, -0.0029846064280718565, 0.16163335740566254, 0.16227170825004578, -0.028343440964818, -0.005498356651514769, -0.04988420009613037, -0.003568929387256503, -0.15064309537410736, -0.1273166984319687, -0.06956228613853455, -0.12177696079015732, 0.0006072228425182402, -0.036003220826387405, 0.07679333537817001, 0.07427064329385757, 0.036336399614810944, -0.06808611750602722, -0.00641190679743886, 0.012408782728016376, -0.07595568895339966, 0.020203862339258194, -0.01782633550465107, 0.005795134697109461, -0.028218304738402367, -0.02324475534260273, -0.08802057057619095, -0.028519567102193832, -0.019917292520403862, 0.051076699048280716, 0.016268715262413025, 0.03196297585964203, -0.07994981110095978, -0.06839286535978317, -0.050182320177555084, 0.05162300914525986, 0.0382181853055954, 0.19898340106010437, 0.010192176327109337, 0.05114158242940903, 0.06887320429086685, 0.1854466199874878, -0.08363981544971466, -0.1572999656200409, -0.026065431535243988, 0.16500923037528992, 0.03500135987997055, 0.04705958440899849, -0.0141305448487401, 0.03860333189368248, -0.05177157372236252, 0.3302074074745178, 0.2516878843307495, -0.03691922873258591, 0.022354472428560257, -0.015065383166074753, 0.029254477471113205, 0.07383154332637787, 0.10979072004556656, 0.14805249869823456, 0.2221984714269638, -0.07927544414997101, -0.027760548517107964, -0.0504816398024559, 0.0109883863478899, -0.15811172127723694, 0.10640375316143036, -0.05229863524436951, -0.09875480830669403, 0.0020930017344653606, 0.06957528740167618, -0.0756518617272377, 0.12091214954853058, -0.05045023560523987, -0.12030607461929321, -0.031558405607938766, 0.021446725353598595, 0.21504542231559753, 0.007502072025090456, 0.039664313197135925, -0.027888575568795204, -0.03702016547322273, 0.1234923005104065, -0.027337750419974327, -0.21466319262981415, -0.029334696009755135, 0.06843926757574081, -0.11448980122804642, 0.09837319701910019, 0.003119915723800659, 0.06310181319713593, 0.09492339938879013, 0.09807661175727844, -0.11998645216226578, 0.0734376385807991, 0.00833623856306076, -0.043060753494501114, 0.019821465015411377, -0.0845298171043396, -0.009848442859947681, -0.08281207829713821, 0.0405195876955986, -0.09080923348665237, 0.05419011786580086, 0.075287364423275, -0.02639298513531685, -0.02869352139532566, 0.024383092299103737, -0.0703892856836319, 0.0735979676246643, -0.0008474259520880878, -0.039789699018001556, -0.0693918988108635, -0.05415460467338562, -0.028567655012011528, 0.02583366259932518, -0.13945288956165314, -0.06117057800292969, -0.04122313857078552, -0.0348254069685936, 0.10306832194328308, 0.03450657054781914, -0.15002690255641937, -0.024332381784915924, -0.10277141630649567, -0.0017641499871388078, -0.1554490476846695, 0.04002204164862633, 0.0484294556081295, -0.012734971940517426, 0.007129951845854521, -0.05800299718976021, 0.019703486934304237, 0.046860262751579285, -0.07697559148073196, -0.10241975635290146 ]
null
null
transformers
This model is used in the paper **Generative Relation Linking for Question Answering over Knowledge Bases**. [ArXiv](https://arxiv.org/abs/2108.07337), [GitHub](https://github.com/IBM/kbqa-relation-linking) ## Citation ```bibtex @inproceedings{rossiello-genrl-2021, title={Generative relation linking for question answering over knowledge bases}, author={Rossiello, Gaetano and Mihindukulasooriya, Nandana and Abdelaziz, Ibrahim and Bornea, Mihaela and Gliozzo, Alfio and Naseem, Tahira and Kapanipathi, Pavan}, booktitle={International Semantic Web Conference}, pages={321--337}, year={2021}, organization={Springer}, url = "https://link.springer.com/chapter/10.1007/978-3-030-88361-4_19", doi = "10.1007/978-3-030-88361-4_19" } ```
{"license": "apache-2.0"}
text2text-generation
gaetangate/bart-large_genrl_lcquad2
[ "transformers", "pytorch", "bart", "text2text-generation", "arxiv:2108.07337", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2108.07337" ]
[]
TAGS #transformers #pytorch #bart #text2text-generation #arxiv-2108.07337 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
This model is used in the paper Generative Relation Linking for Question Answering over Knowledge Bases. ArXiv, GitHub
[]
[ "TAGS\n#transformers #pytorch #bart #text2text-generation #arxiv-2108.07337 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #bart #text2text-generation #arxiv-2108.07337 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.05482131987810135, 0.07781752198934555, -0.005638949107378721, 0.022345583885908127, 0.08416979014873505, 0.00895537156611681, 0.1526346057653427, 0.13391147553920746, -0.0098402239382267, -0.04337620735168457, 0.1866959184408188, 0.21171580255031586, 0.0014640424633398652, 0.08603023737668991, -0.0680188313126564, -0.17926882207393646, 0.07028527557849884, 0.06794054061174393, 0.03506860509514809, 0.11241799592971802, 0.11079103499650955, -0.028155889362096786, 0.05943622440099716, 0.007174516562372446, -0.10733117908239365, 0.030625032261013985, 0.052361808717250824, -0.10477364808320999, 0.10313612967729568, 0.04243432357907295, 0.08389689028263092, 0.046983979642391205, -0.030626261606812477, -0.15681667625904083, 0.012176413089036942, 0.0015603452920913696, -0.058309849351644516, 0.0642048567533493, 0.07419206202030182, -0.04541759938001633, 0.08018051087856293, 0.05972139164805412, -0.036842163652181625, 0.059354767203330994, -0.0846671313047409, -0.163845032453537, -0.11407594382762909, 0.09986864775419235, 0.06210153549909592, 0.11018134653568268, 0.05159074440598488, 0.14422264695167542, -0.051009975373744965, 0.07602405548095703, 0.17427119612693787, -0.3681010901927948, 0.004177485127002001, 0.02464538998901844, 0.04596325010061264, 0.006552176084369421, -0.0007694204105064273, 0.020827608183026314, 0.06091023609042168, 0.035478681325912476, 0.003229860682040453, -0.06632944941520691, -0.14235669374465942, 0.02472204528748989, -0.03857053443789482, -0.08070261031389236, 0.24721753597259521, -0.024514835327863693, 0.012871340848505497, -0.0017009302973747253, -0.07462748140096664, -0.0014764306833967566, -0.013516562059521675, 0.04367212578654289, -0.0050034900195896626, 0.09492382407188416, 0.04887858033180237, -0.04823123291134834, -0.15984714031219482, 0.00207934295758605, -0.20301559567451477, 0.08240332454442978, -0.0022943257354199886, 0.09173347055912018, -0.1661391705274582, 0.06920570880174637, 0.0059766024351119995, -0.12174828350543976, 0.010891247540712357, -0.0597551204264164, 0.15621905028820038, 0.05395321547985077, -0.0691797286272049, -0.0048287902027368546, 0.09994350373744965, 0.23596175014972687, 0.04119456186890602, 0.01503798644989729, -0.06155439466238022, 0.09958167374134064, -0.07166501879692078, 0.05998682230710983, 0.05347549170255661, -0.05067494511604309, 0.10866204649209976, -0.1072971373796463, 0.09862447530031204, -0.033908989280462265, -0.16214343905448914, -0.0692695900797844, 0.002642169129103422, 0.10096638649702072, 0.08890148997306824, 0.028074882924556732, -0.04016926512122154, 0.01901845633983612, 0.16369247436523438, -0.060025133192539215, 0.0010842434130609035, -0.011274623684585094, 0.019521581009030342, 0.0889275074005127, 0.06706152856349945, 0.03201710432767868, -0.07721319049596786, 0.07256565243005753, -0.04542548581957817, -0.0067802430130541325, -0.03321782127022743, -0.01449374109506607, 0.0770699754357338, -0.07354425638914108, 0.039060868322849274, -0.16507947444915771, -0.18075479567050934, 0.02805604226887226, 0.044291190803050995, -0.005039107520133257, -0.07841252535581589, 0.031263597309589386, -0.016256261616945267, 0.06746867299079895, -0.08927186578512192, 0.023773107677698135, -0.06050211563706398, 0.07617700845003128, -0.03449065983295441, 0.03464970737695694, -0.1939922422170639, 0.05469327047467232, -0.11378049850463867, -0.017431724816560745, -0.025046037510037422, -0.04177875444293022, -0.05100778117775917, 0.16310332715511322, -0.05167950689792633, -0.023722611367702484, -0.032557837665081024, 0.0038679158315062523, 0.013324900530278683, 0.13822630047798157, -0.09300900250673294, -0.05209609121084213, 0.20359504222869873, -0.10541636496782303, -0.19766826927661896, 0.05312618613243103, 0.03264372795820236, 0.016446685418486595, 0.06077985465526581, 0.14917801320552826, 0.04373243823647499, -0.04614674672484398, 0.06058279424905777, 0.12388131022453308, -0.07293485105037689, -0.2121901661157608, 0.041465435177087784, -0.06329382956027985, -0.110641710460186, 0.06406300514936447, 0.004137752577662468, 0.09876012057065964, 0.0005344409146346152, -0.06045414134860039, -0.08519785851240158, -0.04539100080728531, 0.002937682904303074, -0.007939782924950123, 0.07327837496995926, -0.08597764372825623, -0.01201961562037468, -0.06457104533910751, 0.03227653354406357, 0.047538693994283676, 0.06602434813976288, -0.038587696850299835, 0.07384616136550903, 0.0036276394966989756, 0.025109659880399704, -0.1279795616865158, 0.03790568932890892, -0.01887362077832222, 0.05401679128408432, 0.0003365709853824228, 0.04780033603310585, 0.052929893136024475, -0.07011021673679352, 0.007165586110204458, -0.016639281064271927, 0.13530054688453674, 0.024718094617128372, -0.04768262803554535, -0.09143251925706863, 0.05007249489426613, -0.03446045145392418, 0.029898418113589287, -0.018560517579317093, 0.021972866728901863, 0.002617587335407734, 0.10428658872842789, -0.04944944009184837, 0.07664937525987625, -0.03424760326743126, -0.008887539617717266, -0.07868296653032303, 0.004296313505619764, 0.11207737028598785, 0.06150135025382042, -0.06777924299240112, 0.23034334182739258, -0.09824039787054062, 0.23176263272762299, 0.21014845371246338, -0.16900676488876343, 0.062023747712373734, -0.006118505261838436, -0.026152264326810837, -0.022883309051394463, 0.06577979773283005, 0.013145336881279945, 0.018537122756242752, 0.019990185275673866, 0.17416749894618988, -0.05235980451107025, -0.027067814022302628, -0.011080794967710972, -0.07247885316610336, -0.00978275015950203, 0.04679108038544655, 0.08780468255281448, -0.1469167172908783, 0.16849613189697266, 0.31132781505584717, -0.013906165026128292, 0.07057876884937286, -0.03862464800477028, -0.00969420000910759, 0.03858736529946327, -0.02555689960718155, -0.01832895167171955, -0.00753486854955554, -0.12924771010875702, 0.009394427761435509, 0.10876920074224472, 0.015896691009402275, 0.07074429839849472, -0.13710680603981018, -0.045919403433799744, 0.003918658941984177, -0.01009833998978138, -0.032725732773542404, 0.07581792026758194, 0.008631033822894096, 0.10835693031549454, -0.03264651447534561, -0.06437046080827713, 0.123822420835495, 0.029191283509135246, -0.09418227523565292, 0.1392623484134674, -0.17335090041160583, -0.26775309443473816, -0.1764364242553711, -0.12391039729118347, -0.030479945242404938, 0.015311352908611298, 0.15357957780361176, -0.03749791532754898, -0.048915065824985504, -0.00007263442239491269, -0.08474317193031311, -0.031013313680887222, 0.005559484474360943, 0.005002529360353947, 0.03174421191215515, 0.024705540388822556, -0.12703348696231842, -0.05319864675402641, 0.02746308036148548, -0.023468509316444397, 0.09643476456403732, -0.07996387779712677, 0.08806337416172028, 0.0958549901843071, 0.020643893629312515, 0.02932734414935112, -0.00825952086597681, 0.13522061705589294, -0.014479542151093483, 0.0013834310229867697, 0.25092613697052, -0.013742741197347641, 0.08801493793725967, 0.11183018237352371, 0.010136323049664497, -0.04858143627643585, 0.009055139496922493, -0.06658154726028442, -0.07651500403881073, -0.2714214622974396, -0.09263240545988083, -0.10003478080034256, 0.06438066810369492, 0.061698928475379944, 0.07818742096424103, 0.13583698868751526, 0.08747946470975876, -0.03248957544565201, 0.030506648123264313, -0.03211503103375435, 0.09743855148553848, 0.2555447220802307, -0.025176601484417915, 0.12891273200511932, -0.1269153654575348, -0.030954653397202492, 0.12694242596626282, 0.08371598273515701, 0.10590073466300964, 0.10160430520772934, 0.051557447761297226, 0.05325188860297203, 0.17843466997146606, 0.07512391358613968, 0.13275974988937378, 0.03223362937569618, -0.0218141358345747, -0.04455477371811867, -0.029173532500863075, -0.0705365464091301, 0.05708703026175499, -0.07295859605073929, -0.11087924987077713, -0.026086905971169472, -0.12267530709505081, 0.04912329465150833, 0.15167374908924103, 0.03260583057999611, -0.17569538950920105, 0.004984318278729916, 0.0684858188033104, -0.010155906900763512, -0.07671070843935013, 0.07608667016029358, -0.0805806964635849, -0.10683809220790863, 0.11846431344747543, -0.025947168469429016, 0.12722750008106232, 0.034859783947467804, 0.05635649710893631, -0.05809636041522026, -0.10959172993898392, 0.06383580714464188, 0.13386929035186768, -0.34036576747894287, 0.19154304265975952, -0.013072090223431587, -0.026505006477236748, -0.09062088280916214, 0.007474956102669239, 0.033928144723176956, 0.1653033196926117, 0.07942576706409454, 0.02359463833272457, -0.0978393405675888, -0.013540403917431831, -0.04927901178598404, 0.040264792740345, 0.026391852647066116, 0.0354730524122715, -0.0479319766163826, -0.060511134564876556, -0.017747119069099426, -0.0016101868823170662, 0.056854475289583206, -0.043094977736473083, -0.15984880924224854, 0.07390711456537247, 0.1282113492488861, 0.03012380562722683, -0.056982096284627914, -0.006998253054916859, -0.05290612950921059, 0.19696283340454102, -0.08011247217655182, -0.07158086448907852, -0.09493527561426163, -0.12436679005622864, 0.05385299399495125, -0.07199086993932724, 0.06745593994855881, -0.09251400828361511, -0.011777443811297417, -0.07750517874956131, -0.19208881258964539, 0.09629534929990768, -0.13449956476688385, -0.04357421398162842, -0.04016205295920372, 0.1267463117837906, -0.10915971547365189, 0.01363675482571125, 0.031107960268855095, 0.0045674326829612255, -0.1442185789346695, -0.10512310266494751, -0.01466408558189869, 0.01390861440449953, 0.07032927870750427, -0.06855132430791855, -0.059206217527389526, -0.04210440069437027, 0.0002269123069709167, -0.07095591723918915, 0.23760126531124115, 0.1743471324443817, -0.09102169424295425, 0.18125031888484955, 0.16369777917861938, -0.08062667399644852, -0.27284666895866394, -0.17046639323234558, -0.1256026327610016, -0.07379941642284393, -0.02152799814939499, -0.13010862469673157, 0.0928943082690239, 0.03854353725910187, -0.08531101793050766, 0.13309340178966522, -0.20278358459472656, -0.08868534117937088, 0.19920559227466583, -0.029639476910233498, 0.3354334533214569, -0.14678631722927094, -0.09665258973836899, -0.09883969277143478, -0.23874326050281525, 0.12705844640731812, -0.05784017592668533, 0.05659428983926773, -0.05013212561607361, 0.09030452370643616, -0.0055581978522241116, -0.07001134753227234, 0.1165020614862442, -0.019615652039647102, 0.011899770237505436, -0.12998872995376587, 0.015354842878878117, 0.05959603562951088, -0.025507992133498192, 0.0682363361120224, -0.16799941658973694, 0.04249343276023865, -0.1297016143798828, -0.00840586144477129, -0.08478473871946335, 0.0836782157421112, -0.00001981135028472636, -0.03807370737195015, -0.020114202052354813, -0.0809778943657875, 0.00864352472126484, -0.005378738511353731, 0.21508856117725372, 0.021311039105057716, 0.1004076674580574, 0.1621607393026352, 0.0570119246840477, -0.19071568548679352, 0.018602002412080765, -0.09594377875328064, -0.09199532866477966, 0.04919834062457085, -0.14196406304836273, 0.03942728415131569, 0.09658144414424896, -0.03966351971030235, 0.04124578833580017, 0.062003325670957565, 0.002129205269739032, -0.038048967719078064, 0.13631731271743774, -0.1981479525566101, 0.06881100684404373, -0.020387141034007072, 0.16533978283405304, 0.06280216574668884, 0.04884500801563263, 0.14566610753536224, 0.014593229629099369, -0.046061012893915176, 0.005350813735276461, 0.019779937341809273, -0.07924901694059372, 0.04157470911741257, 0.05258270353078842, 0.002496576402336359, -0.11633665859699249, 0.08739665150642395, -0.0006512592080980539, -0.1388893723487854, 0.00044502574019134045, 0.13315072655677795, -0.13069625198841095, -0.11353246122598648, 0.013988176360726357, 0.05645694583654404, -0.21390998363494873, -0.11363053321838379, -0.043982721865177155, -0.10140077024698257, 0.07960906624794006, 0.10171038657426834, 0.077805295586586, 0.0663294568657875, -0.011303826235234737, -0.07232298702001572, 0.023237422108650208, -0.023759078234434128, -0.048108603805303574, 0.04942779242992401, -0.07055307179689407, -0.015623528510332108, 0.00967564806342125, 0.09083372354507446, -0.036806222051382065, 0.0076259043999016285, -0.07885701209306717, 0.024485094472765923, -0.18741877377033234, -0.006544181611388922, -0.09549658745527267, -0.034454576671123505, 0.007622649893164635, -0.07942161709070206, -0.04239603504538536, 0.012974384240806103, -0.11313555389642715, -0.03135130554437637, -0.03334683179855347, 0.03711055591702461, -0.11675328016281128, -0.03606405481696129, 0.08109122514724731, -0.007913854904472828, 0.08382098376750946, 0.1311119794845581, -0.06978994607925415, 0.06548008322715759, -0.152426078915596, -0.09907804429531097, 0.07366190105676651, 0.03581872954964638, 0.014548269100487232, 0.009208097122609615, -0.005308505147695541, 0.14182445406913757, -0.005868135020136833, 0.02698882669210434, 0.06259270012378693, -0.14154714345932007, -0.036849476397037506, -0.002217935398221016, -0.1273764669895172, -0.0359746553003788, -0.08520586043596268, 0.11702703684568405, 0.02218184620141983, 0.15673206746578217, -0.0330800786614418, 0.04871964454650879, -0.052613042294979095, 0.030562758445739746, -0.03907517343759537, -0.14916402101516724, -0.1273747682571411, -0.06134314462542534, -0.0387486107647419, 0.002467484911903739, 0.21433435380458832, -0.029699983075261116, -0.03970152512192726, 0.059296615421772, 0.08205589652061462, -0.011385519988834858, -0.02196643315255642, 0.24418558180332184, 0.04766358062624931, -0.009007682092487812, -0.08916512131690979, 0.046584710478782654, 0.0018661929061636329, -0.051556818187236786, 0.08721576631069183, 0.08817291259765625, 0.067312091588974, 0.0969495177268982, 0.014695269986987114, -0.0016428417293354869, -0.13492125272750854, -0.17150603234767914, 0.002291272394359112, 0.0770316794514656, -0.0029846064280718565, 0.16163335740566254, 0.16227170825004578, -0.028343440964818, -0.005498356651514769, -0.04988420009613037, -0.003568929387256503, -0.15064309537410736, -0.1273166984319687, -0.06956228613853455, -0.12177696079015732, 0.0006072228425182402, -0.036003220826387405, 0.07679333537817001, 0.07427064329385757, 0.036336399614810944, -0.06808611750602722, -0.00641190679743886, 0.012408782728016376, -0.07595568895339966, 0.020203862339258194, -0.01782633550465107, 0.005795134697109461, -0.028218304738402367, -0.02324475534260273, -0.08802057057619095, -0.028519567102193832, -0.019917292520403862, 0.051076699048280716, 0.016268715262413025, 0.03196297585964203, -0.07994981110095978, -0.06839286535978317, -0.050182320177555084, 0.05162300914525986, 0.0382181853055954, 0.19898340106010437, 0.010192176327109337, 0.05114158242940903, 0.06887320429086685, 0.1854466199874878, -0.08363981544971466, -0.1572999656200409, -0.026065431535243988, 0.16500923037528992, 0.03500135987997055, 0.04705958440899849, -0.0141305448487401, 0.03860333189368248, -0.05177157372236252, 0.3302074074745178, 0.2516878843307495, -0.03691922873258591, 0.022354472428560257, -0.015065383166074753, 0.029254477471113205, 0.07383154332637787, 0.10979072004556656, 0.14805249869823456, 0.2221984714269638, -0.07927544414997101, -0.027760548517107964, -0.0504816398024559, 0.0109883863478899, -0.15811172127723694, 0.10640375316143036, -0.05229863524436951, -0.09875480830669403, 0.0020930017344653606, 0.06957528740167618, -0.0756518617272377, 0.12091214954853058, -0.05045023560523987, -0.12030607461929321, -0.031558405607938766, 0.021446725353598595, 0.21504542231559753, 0.007502072025090456, 0.039664313197135925, -0.027888575568795204, -0.03702016547322273, 0.1234923005104065, -0.027337750419974327, -0.21466319262981415, -0.029334696009755135, 0.06843926757574081, -0.11448980122804642, 0.09837319701910019, 0.003119915723800659, 0.06310181319713593, 0.09492339938879013, 0.09807661175727844, -0.11998645216226578, 0.0734376385807991, 0.00833623856306076, -0.043060753494501114, 0.019821465015411377, -0.0845298171043396, -0.009848442859947681, -0.08281207829713821, 0.0405195876955986, -0.09080923348665237, 0.05419011786580086, 0.075287364423275, -0.02639298513531685, -0.02869352139532566, 0.024383092299103737, -0.0703892856836319, 0.0735979676246643, -0.0008474259520880878, -0.039789699018001556, -0.0693918988108635, -0.05415460467338562, -0.028567655012011528, 0.02583366259932518, -0.13945288956165314, -0.06117057800292969, -0.04122313857078552, -0.0348254069685936, 0.10306832194328308, 0.03450657054781914, -0.15002690255641937, -0.024332381784915924, -0.10277141630649567, -0.0017641499871388078, -0.1554490476846695, 0.04002204164862633, 0.0484294556081295, -0.012734971940517426, 0.007129951845854521, -0.05800299718976021, 0.019703486934304237, 0.046860262751579285, -0.07697559148073196, -0.10241975635290146 ]
null
null
transformers
This model is used in the paper **Generative Relation Linking for Question Answering over Knowledge Bases**. [ArXiv](https://arxiv.org/abs/2108.07337), [GitHub](https://github.com/IBM/kbqa-relation-linking) ## Citation ```bibtex @inproceedings{rossiello-genrl-2021, title={Generative relation linking for question answering over knowledge bases}, author={Rossiello, Gaetano and Mihindukulasooriya, Nandana and Abdelaziz, Ibrahim and Bornea, Mihaela and Gliozzo, Alfio and Naseem, Tahira and Kapanipathi, Pavan}, booktitle={International Semantic Web Conference}, pages={321--337}, year={2021}, organization={Springer}, url = "https://link.springer.com/chapter/10.1007/978-3-030-88361-4_19", doi = "10.1007/978-3-030-88361-4_19" } ```
{"license": "apache-2.0"}
text2text-generation
gaetangate/bart-large_genrl_qald9
[ "transformers", "pytorch", "bart", "text2text-generation", "arxiv:2108.07337", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2108.07337" ]
[]
TAGS #transformers #pytorch #bart #text2text-generation #arxiv-2108.07337 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
This model is used in the paper Generative Relation Linking for Question Answering over Knowledge Bases. ArXiv, GitHub
[]
[ "TAGS\n#transformers #pytorch #bart #text2text-generation #arxiv-2108.07337 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #bart #text2text-generation #arxiv-2108.07337 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.05482131987810135, 0.07781752198934555, -0.005638949107378721, 0.022345583885908127, 0.08416979014873505, 0.00895537156611681, 0.1526346057653427, 0.13391147553920746, -0.0098402239382267, -0.04337620735168457, 0.1866959184408188, 0.21171580255031586, 0.0014640424633398652, 0.08603023737668991, -0.0680188313126564, -0.17926882207393646, 0.07028527557849884, 0.06794054061174393, 0.03506860509514809, 0.11241799592971802, 0.11079103499650955, -0.028155889362096786, 0.05943622440099716, 0.007174516562372446, -0.10733117908239365, 0.030625032261013985, 0.052361808717250824, -0.10477364808320999, 0.10313612967729568, 0.04243432357907295, 0.08389689028263092, 0.046983979642391205, -0.030626261606812477, -0.15681667625904083, 0.012176413089036942, 0.0015603452920913696, -0.058309849351644516, 0.0642048567533493, 0.07419206202030182, -0.04541759938001633, 0.08018051087856293, 0.05972139164805412, -0.036842163652181625, 0.059354767203330994, -0.0846671313047409, -0.163845032453537, -0.11407594382762909, 0.09986864775419235, 0.06210153549909592, 0.11018134653568268, 0.05159074440598488, 0.14422264695167542, -0.051009975373744965, 0.07602405548095703, 0.17427119612693787, -0.3681010901927948, 0.004177485127002001, 0.02464538998901844, 0.04596325010061264, 0.006552176084369421, -0.0007694204105064273, 0.020827608183026314, 0.06091023609042168, 0.035478681325912476, 0.003229860682040453, -0.06632944941520691, -0.14235669374465942, 0.02472204528748989, -0.03857053443789482, -0.08070261031389236, 0.24721753597259521, -0.024514835327863693, 0.012871340848505497, -0.0017009302973747253, -0.07462748140096664, -0.0014764306833967566, -0.013516562059521675, 0.04367212578654289, -0.0050034900195896626, 0.09492382407188416, 0.04887858033180237, -0.04823123291134834, -0.15984714031219482, 0.00207934295758605, -0.20301559567451477, 0.08240332454442978, -0.0022943257354199886, 0.09173347055912018, -0.1661391705274582, 0.06920570880174637, 0.0059766024351119995, -0.12174828350543976, 0.010891247540712357, -0.0597551204264164, 0.15621905028820038, 0.05395321547985077, -0.0691797286272049, -0.0048287902027368546, 0.09994350373744965, 0.23596175014972687, 0.04119456186890602, 0.01503798644989729, -0.06155439466238022, 0.09958167374134064, -0.07166501879692078, 0.05998682230710983, 0.05347549170255661, -0.05067494511604309, 0.10866204649209976, -0.1072971373796463, 0.09862447530031204, -0.033908989280462265, -0.16214343905448914, -0.0692695900797844, 0.002642169129103422, 0.10096638649702072, 0.08890148997306824, 0.028074882924556732, -0.04016926512122154, 0.01901845633983612, 0.16369247436523438, -0.060025133192539215, 0.0010842434130609035, -0.011274623684585094, 0.019521581009030342, 0.0889275074005127, 0.06706152856349945, 0.03201710432767868, -0.07721319049596786, 0.07256565243005753, -0.04542548581957817, -0.0067802430130541325, -0.03321782127022743, -0.01449374109506607, 0.0770699754357338, -0.07354425638914108, 0.039060868322849274, -0.16507947444915771, -0.18075479567050934, 0.02805604226887226, 0.044291190803050995, -0.005039107520133257, -0.07841252535581589, 0.031263597309589386, -0.016256261616945267, 0.06746867299079895, -0.08927186578512192, 0.023773107677698135, -0.06050211563706398, 0.07617700845003128, -0.03449065983295441, 0.03464970737695694, -0.1939922422170639, 0.05469327047467232, -0.11378049850463867, -0.017431724816560745, -0.025046037510037422, -0.04177875444293022, -0.05100778117775917, 0.16310332715511322, -0.05167950689792633, -0.023722611367702484, -0.032557837665081024, 0.0038679158315062523, 0.013324900530278683, 0.13822630047798157, -0.09300900250673294, -0.05209609121084213, 0.20359504222869873, -0.10541636496782303, -0.19766826927661896, 0.05312618613243103, 0.03264372795820236, 0.016446685418486595, 0.06077985465526581, 0.14917801320552826, 0.04373243823647499, -0.04614674672484398, 0.06058279424905777, 0.12388131022453308, -0.07293485105037689, -0.2121901661157608, 0.041465435177087784, -0.06329382956027985, -0.110641710460186, 0.06406300514936447, 0.004137752577662468, 0.09876012057065964, 0.0005344409146346152, -0.06045414134860039, -0.08519785851240158, -0.04539100080728531, 0.002937682904303074, -0.007939782924950123, 0.07327837496995926, -0.08597764372825623, -0.01201961562037468, -0.06457104533910751, 0.03227653354406357, 0.047538693994283676, 0.06602434813976288, -0.038587696850299835, 0.07384616136550903, 0.0036276394966989756, 0.025109659880399704, -0.1279795616865158, 0.03790568932890892, -0.01887362077832222, 0.05401679128408432, 0.0003365709853824228, 0.04780033603310585, 0.052929893136024475, -0.07011021673679352, 0.007165586110204458, -0.016639281064271927, 0.13530054688453674, 0.024718094617128372, -0.04768262803554535, -0.09143251925706863, 0.05007249489426613, -0.03446045145392418, 0.029898418113589287, -0.018560517579317093, 0.021972866728901863, 0.002617587335407734, 0.10428658872842789, -0.04944944009184837, 0.07664937525987625, -0.03424760326743126, -0.008887539617717266, -0.07868296653032303, 0.004296313505619764, 0.11207737028598785, 0.06150135025382042, -0.06777924299240112, 0.23034334182739258, -0.09824039787054062, 0.23176263272762299, 0.21014845371246338, -0.16900676488876343, 0.062023747712373734, -0.006118505261838436, -0.026152264326810837, -0.022883309051394463, 0.06577979773283005, 0.013145336881279945, 0.018537122756242752, 0.019990185275673866, 0.17416749894618988, -0.05235980451107025, -0.027067814022302628, -0.011080794967710972, -0.07247885316610336, -0.00978275015950203, 0.04679108038544655, 0.08780468255281448, -0.1469167172908783, 0.16849613189697266, 0.31132781505584717, -0.013906165026128292, 0.07057876884937286, -0.03862464800477028, -0.00969420000910759, 0.03858736529946327, -0.02555689960718155, -0.01832895167171955, -0.00753486854955554, -0.12924771010875702, 0.009394427761435509, 0.10876920074224472, 0.015896691009402275, 0.07074429839849472, -0.13710680603981018, -0.045919403433799744, 0.003918658941984177, -0.01009833998978138, -0.032725732773542404, 0.07581792026758194, 0.008631033822894096, 0.10835693031549454, -0.03264651447534561, -0.06437046080827713, 0.123822420835495, 0.029191283509135246, -0.09418227523565292, 0.1392623484134674, -0.17335090041160583, -0.26775309443473816, -0.1764364242553711, -0.12391039729118347, -0.030479945242404938, 0.015311352908611298, 0.15357957780361176, -0.03749791532754898, -0.048915065824985504, -0.00007263442239491269, -0.08474317193031311, -0.031013313680887222, 0.005559484474360943, 0.005002529360353947, 0.03174421191215515, 0.024705540388822556, -0.12703348696231842, -0.05319864675402641, 0.02746308036148548, -0.023468509316444397, 0.09643476456403732, -0.07996387779712677, 0.08806337416172028, 0.0958549901843071, 0.020643893629312515, 0.02932734414935112, -0.00825952086597681, 0.13522061705589294, -0.014479542151093483, 0.0013834310229867697, 0.25092613697052, -0.013742741197347641, 0.08801493793725967, 0.11183018237352371, 0.010136323049664497, -0.04858143627643585, 0.009055139496922493, -0.06658154726028442, -0.07651500403881073, -0.2714214622974396, -0.09263240545988083, -0.10003478080034256, 0.06438066810369492, 0.061698928475379944, 0.07818742096424103, 0.13583698868751526, 0.08747946470975876, -0.03248957544565201, 0.030506648123264313, -0.03211503103375435, 0.09743855148553848, 0.2555447220802307, -0.025176601484417915, 0.12891273200511932, -0.1269153654575348, -0.030954653397202492, 0.12694242596626282, 0.08371598273515701, 0.10590073466300964, 0.10160430520772934, 0.051557447761297226, 0.05325188860297203, 0.17843466997146606, 0.07512391358613968, 0.13275974988937378, 0.03223362937569618, -0.0218141358345747, -0.04455477371811867, -0.029173532500863075, -0.0705365464091301, 0.05708703026175499, -0.07295859605073929, -0.11087924987077713, -0.026086905971169472, -0.12267530709505081, 0.04912329465150833, 0.15167374908924103, 0.03260583057999611, -0.17569538950920105, 0.004984318278729916, 0.0684858188033104, -0.010155906900763512, -0.07671070843935013, 0.07608667016029358, -0.0805806964635849, -0.10683809220790863, 0.11846431344747543, -0.025947168469429016, 0.12722750008106232, 0.034859783947467804, 0.05635649710893631, -0.05809636041522026, -0.10959172993898392, 0.06383580714464188, 0.13386929035186768, -0.34036576747894287, 0.19154304265975952, -0.013072090223431587, -0.026505006477236748, -0.09062088280916214, 0.007474956102669239, 0.033928144723176956, 0.1653033196926117, 0.07942576706409454, 0.02359463833272457, -0.0978393405675888, -0.013540403917431831, -0.04927901178598404, 0.040264792740345, 0.026391852647066116, 0.0354730524122715, -0.0479319766163826, -0.060511134564876556, -0.017747119069099426, -0.0016101868823170662, 0.056854475289583206, -0.043094977736473083, -0.15984880924224854, 0.07390711456537247, 0.1282113492488861, 0.03012380562722683, -0.056982096284627914, -0.006998253054916859, -0.05290612950921059, 0.19696283340454102, -0.08011247217655182, -0.07158086448907852, -0.09493527561426163, -0.12436679005622864, 0.05385299399495125, -0.07199086993932724, 0.06745593994855881, -0.09251400828361511, -0.011777443811297417, -0.07750517874956131, -0.19208881258964539, 0.09629534929990768, -0.13449956476688385, -0.04357421398162842, -0.04016205295920372, 0.1267463117837906, -0.10915971547365189, 0.01363675482571125, 0.031107960268855095, 0.0045674326829612255, -0.1442185789346695, -0.10512310266494751, -0.01466408558189869, 0.01390861440449953, 0.07032927870750427, -0.06855132430791855, -0.059206217527389526, -0.04210440069437027, 0.0002269123069709167, -0.07095591723918915, 0.23760126531124115, 0.1743471324443817, -0.09102169424295425, 0.18125031888484955, 0.16369777917861938, -0.08062667399644852, -0.27284666895866394, -0.17046639323234558, -0.1256026327610016, -0.07379941642284393, -0.02152799814939499, -0.13010862469673157, 0.0928943082690239, 0.03854353725910187, -0.08531101793050766, 0.13309340178966522, -0.20278358459472656, -0.08868534117937088, 0.19920559227466583, -0.029639476910233498, 0.3354334533214569, -0.14678631722927094, -0.09665258973836899, -0.09883969277143478, -0.23874326050281525, 0.12705844640731812, -0.05784017592668533, 0.05659428983926773, -0.05013212561607361, 0.09030452370643616, -0.0055581978522241116, -0.07001134753227234, 0.1165020614862442, -0.019615652039647102, 0.011899770237505436, -0.12998872995376587, 0.015354842878878117, 0.05959603562951088, -0.025507992133498192, 0.0682363361120224, -0.16799941658973694, 0.04249343276023865, -0.1297016143798828, -0.00840586144477129, -0.08478473871946335, 0.0836782157421112, -0.00001981135028472636, -0.03807370737195015, -0.020114202052354813, -0.0809778943657875, 0.00864352472126484, -0.005378738511353731, 0.21508856117725372, 0.021311039105057716, 0.1004076674580574, 0.1621607393026352, 0.0570119246840477, -0.19071568548679352, 0.018602002412080765, -0.09594377875328064, -0.09199532866477966, 0.04919834062457085, -0.14196406304836273, 0.03942728415131569, 0.09658144414424896, -0.03966351971030235, 0.04124578833580017, 0.062003325670957565, 0.002129205269739032, -0.038048967719078064, 0.13631731271743774, -0.1981479525566101, 0.06881100684404373, -0.020387141034007072, 0.16533978283405304, 0.06280216574668884, 0.04884500801563263, 0.14566610753536224, 0.014593229629099369, -0.046061012893915176, 0.005350813735276461, 0.019779937341809273, -0.07924901694059372, 0.04157470911741257, 0.05258270353078842, 0.002496576402336359, -0.11633665859699249, 0.08739665150642395, -0.0006512592080980539, -0.1388893723487854, 0.00044502574019134045, 0.13315072655677795, -0.13069625198841095, -0.11353246122598648, 0.013988176360726357, 0.05645694583654404, -0.21390998363494873, -0.11363053321838379, -0.043982721865177155, -0.10140077024698257, 0.07960906624794006, 0.10171038657426834, 0.077805295586586, 0.0663294568657875, -0.011303826235234737, -0.07232298702001572, 0.023237422108650208, -0.023759078234434128, -0.048108603805303574, 0.04942779242992401, -0.07055307179689407, -0.015623528510332108, 0.00967564806342125, 0.09083372354507446, -0.036806222051382065, 0.0076259043999016285, -0.07885701209306717, 0.024485094472765923, -0.18741877377033234, -0.006544181611388922, -0.09549658745527267, -0.034454576671123505, 0.007622649893164635, -0.07942161709070206, -0.04239603504538536, 0.012974384240806103, -0.11313555389642715, -0.03135130554437637, -0.03334683179855347, 0.03711055591702461, -0.11675328016281128, -0.03606405481696129, 0.08109122514724731, -0.007913854904472828, 0.08382098376750946, 0.1311119794845581, -0.06978994607925415, 0.06548008322715759, -0.152426078915596, -0.09907804429531097, 0.07366190105676651, 0.03581872954964638, 0.014548269100487232, 0.009208097122609615, -0.005308505147695541, 0.14182445406913757, -0.005868135020136833, 0.02698882669210434, 0.06259270012378693, -0.14154714345932007, -0.036849476397037506, -0.002217935398221016, -0.1273764669895172, -0.0359746553003788, -0.08520586043596268, 0.11702703684568405, 0.02218184620141983, 0.15673206746578217, -0.0330800786614418, 0.04871964454650879, -0.052613042294979095, 0.030562758445739746, -0.03907517343759537, -0.14916402101516724, -0.1273747682571411, -0.06134314462542534, -0.0387486107647419, 0.002467484911903739, 0.21433435380458832, -0.029699983075261116, -0.03970152512192726, 0.059296615421772, 0.08205589652061462, -0.011385519988834858, -0.02196643315255642, 0.24418558180332184, 0.04766358062624931, -0.009007682092487812, -0.08916512131690979, 0.046584710478782654, 0.0018661929061636329, -0.051556818187236786, 0.08721576631069183, 0.08817291259765625, 0.067312091588974, 0.0969495177268982, 0.014695269986987114, -0.0016428417293354869, -0.13492125272750854, -0.17150603234767914, 0.002291272394359112, 0.0770316794514656, -0.0029846064280718565, 0.16163335740566254, 0.16227170825004578, -0.028343440964818, -0.005498356651514769, -0.04988420009613037, -0.003568929387256503, -0.15064309537410736, -0.1273166984319687, -0.06956228613853455, -0.12177696079015732, 0.0006072228425182402, -0.036003220826387405, 0.07679333537817001, 0.07427064329385757, 0.036336399614810944, -0.06808611750602722, -0.00641190679743886, 0.012408782728016376, -0.07595568895339966, 0.020203862339258194, -0.01782633550465107, 0.005795134697109461, -0.028218304738402367, -0.02324475534260273, -0.08802057057619095, -0.028519567102193832, -0.019917292520403862, 0.051076699048280716, 0.016268715262413025, 0.03196297585964203, -0.07994981110095978, -0.06839286535978317, -0.050182320177555084, 0.05162300914525986, 0.0382181853055954, 0.19898340106010437, 0.010192176327109337, 0.05114158242940903, 0.06887320429086685, 0.1854466199874878, -0.08363981544971466, -0.1572999656200409, -0.026065431535243988, 0.16500923037528992, 0.03500135987997055, 0.04705958440899849, -0.0141305448487401, 0.03860333189368248, -0.05177157372236252, 0.3302074074745178, 0.2516878843307495, -0.03691922873258591, 0.022354472428560257, -0.015065383166074753, 0.029254477471113205, 0.07383154332637787, 0.10979072004556656, 0.14805249869823456, 0.2221984714269638, -0.07927544414997101, -0.027760548517107964, -0.0504816398024559, 0.0109883863478899, -0.15811172127723694, 0.10640375316143036, -0.05229863524436951, -0.09875480830669403, 0.0020930017344653606, 0.06957528740167618, -0.0756518617272377, 0.12091214954853058, -0.05045023560523987, -0.12030607461929321, -0.031558405607938766, 0.021446725353598595, 0.21504542231559753, 0.007502072025090456, 0.039664313197135925, -0.027888575568795204, -0.03702016547322273, 0.1234923005104065, -0.027337750419974327, -0.21466319262981415, -0.029334696009755135, 0.06843926757574081, -0.11448980122804642, 0.09837319701910019, 0.003119915723800659, 0.06310181319713593, 0.09492339938879013, 0.09807661175727844, -0.11998645216226578, 0.0734376385807991, 0.00833623856306076, -0.043060753494501114, 0.019821465015411377, -0.0845298171043396, -0.009848442859947681, -0.08281207829713821, 0.0405195876955986, -0.09080923348665237, 0.05419011786580086, 0.075287364423275, -0.02639298513531685, -0.02869352139532566, 0.024383092299103737, -0.0703892856836319, 0.0735979676246643, -0.0008474259520880878, -0.039789699018001556, -0.0693918988108635, -0.05415460467338562, -0.028567655012011528, 0.02583366259932518, -0.13945288956165314, -0.06117057800292969, -0.04122313857078552, -0.0348254069685936, 0.10306832194328308, 0.03450657054781914, -0.15002690255641937, -0.024332381784915924, -0.10277141630649567, -0.0017641499871388078, -0.1554490476846695, 0.04002204164862633, 0.0484294556081295, -0.012734971940517426, 0.007129951845854521, -0.05800299718976021, 0.019703486934304237, 0.046860262751579285, -0.07697559148073196, -0.10241975635290146 ]
null
null
transformers
This model is used in the paper **Generative Relation Linking for Question Answering over Knowledge Bases**. [ArXiv](https://arxiv.org/abs/2108.07337), [GitHub](https://github.com/IBM/kbqa-relation-linking) ## Citation ```bibtex @inproceedings{rossiello-genrl-2021, title={Generative relation linking for question answering over knowledge bases}, author={Rossiello, Gaetano and Mihindukulasooriya, Nandana and Abdelaziz, Ibrahim and Bornea, Mihaela and Gliozzo, Alfio and Naseem, Tahira and Kapanipathi, Pavan}, booktitle={International Semantic Web Conference}, pages={321--337}, year={2021}, organization={Springer}, url = "https://link.springer.com/chapter/10.1007/978-3-030-88361-4_19", doi = "10.1007/978-3-030-88361-4_19" } ```
{"license": "apache-2.0"}
text2text-generation
gaetangate/bart-large_genrl_simpleq
[ "transformers", "pytorch", "bart", "text2text-generation", "arxiv:2108.07337", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2108.07337" ]
[]
TAGS #transformers #pytorch #bart #text2text-generation #arxiv-2108.07337 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
This model is used in the paper Generative Relation Linking for Question Answering over Knowledge Bases. ArXiv, GitHub
[]
[ "TAGS\n#transformers #pytorch #bart #text2text-generation #arxiv-2108.07337 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #bart #text2text-generation #arxiv-2108.07337 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.05482131987810135, 0.07781752198934555, -0.005638949107378721, 0.022345583885908127, 0.08416979014873505, 0.00895537156611681, 0.1526346057653427, 0.13391147553920746, -0.0098402239382267, -0.04337620735168457, 0.1866959184408188, 0.21171580255031586, 0.0014640424633398652, 0.08603023737668991, -0.0680188313126564, -0.17926882207393646, 0.07028527557849884, 0.06794054061174393, 0.03506860509514809, 0.11241799592971802, 0.11079103499650955, -0.028155889362096786, 0.05943622440099716, 0.007174516562372446, -0.10733117908239365, 0.030625032261013985, 0.052361808717250824, -0.10477364808320999, 0.10313612967729568, 0.04243432357907295, 0.08389689028263092, 0.046983979642391205, -0.030626261606812477, -0.15681667625904083, 0.012176413089036942, 0.0015603452920913696, -0.058309849351644516, 0.0642048567533493, 0.07419206202030182, -0.04541759938001633, 0.08018051087856293, 0.05972139164805412, -0.036842163652181625, 0.059354767203330994, -0.0846671313047409, -0.163845032453537, -0.11407594382762909, 0.09986864775419235, 0.06210153549909592, 0.11018134653568268, 0.05159074440598488, 0.14422264695167542, -0.051009975373744965, 0.07602405548095703, 0.17427119612693787, -0.3681010901927948, 0.004177485127002001, 0.02464538998901844, 0.04596325010061264, 0.006552176084369421, -0.0007694204105064273, 0.020827608183026314, 0.06091023609042168, 0.035478681325912476, 0.003229860682040453, -0.06632944941520691, -0.14235669374465942, 0.02472204528748989, -0.03857053443789482, -0.08070261031389236, 0.24721753597259521, -0.024514835327863693, 0.012871340848505497, -0.0017009302973747253, -0.07462748140096664, -0.0014764306833967566, -0.013516562059521675, 0.04367212578654289, -0.0050034900195896626, 0.09492382407188416, 0.04887858033180237, -0.04823123291134834, -0.15984714031219482, 0.00207934295758605, -0.20301559567451477, 0.08240332454442978, -0.0022943257354199886, 0.09173347055912018, -0.1661391705274582, 0.06920570880174637, 0.0059766024351119995, -0.12174828350543976, 0.010891247540712357, -0.0597551204264164, 0.15621905028820038, 0.05395321547985077, -0.0691797286272049, -0.0048287902027368546, 0.09994350373744965, 0.23596175014972687, 0.04119456186890602, 0.01503798644989729, -0.06155439466238022, 0.09958167374134064, -0.07166501879692078, 0.05998682230710983, 0.05347549170255661, -0.05067494511604309, 0.10866204649209976, -0.1072971373796463, 0.09862447530031204, -0.033908989280462265, -0.16214343905448914, -0.0692695900797844, 0.002642169129103422, 0.10096638649702072, 0.08890148997306824, 0.028074882924556732, -0.04016926512122154, 0.01901845633983612, 0.16369247436523438, -0.060025133192539215, 0.0010842434130609035, -0.011274623684585094, 0.019521581009030342, 0.0889275074005127, 0.06706152856349945, 0.03201710432767868, -0.07721319049596786, 0.07256565243005753, -0.04542548581957817, -0.0067802430130541325, -0.03321782127022743, -0.01449374109506607, 0.0770699754357338, -0.07354425638914108, 0.039060868322849274, -0.16507947444915771, -0.18075479567050934, 0.02805604226887226, 0.044291190803050995, -0.005039107520133257, -0.07841252535581589, 0.031263597309589386, -0.016256261616945267, 0.06746867299079895, -0.08927186578512192, 0.023773107677698135, -0.06050211563706398, 0.07617700845003128, -0.03449065983295441, 0.03464970737695694, -0.1939922422170639, 0.05469327047467232, -0.11378049850463867, -0.017431724816560745, -0.025046037510037422, -0.04177875444293022, -0.05100778117775917, 0.16310332715511322, -0.05167950689792633, -0.023722611367702484, -0.032557837665081024, 0.0038679158315062523, 0.013324900530278683, 0.13822630047798157, -0.09300900250673294, -0.05209609121084213, 0.20359504222869873, -0.10541636496782303, -0.19766826927661896, 0.05312618613243103, 0.03264372795820236, 0.016446685418486595, 0.06077985465526581, 0.14917801320552826, 0.04373243823647499, -0.04614674672484398, 0.06058279424905777, 0.12388131022453308, -0.07293485105037689, -0.2121901661157608, 0.041465435177087784, -0.06329382956027985, -0.110641710460186, 0.06406300514936447, 0.004137752577662468, 0.09876012057065964, 0.0005344409146346152, -0.06045414134860039, -0.08519785851240158, -0.04539100080728531, 0.002937682904303074, -0.007939782924950123, 0.07327837496995926, -0.08597764372825623, -0.01201961562037468, -0.06457104533910751, 0.03227653354406357, 0.047538693994283676, 0.06602434813976288, -0.038587696850299835, 0.07384616136550903, 0.0036276394966989756, 0.025109659880399704, -0.1279795616865158, 0.03790568932890892, -0.01887362077832222, 0.05401679128408432, 0.0003365709853824228, 0.04780033603310585, 0.052929893136024475, -0.07011021673679352, 0.007165586110204458, -0.016639281064271927, 0.13530054688453674, 0.024718094617128372, -0.04768262803554535, -0.09143251925706863, 0.05007249489426613, -0.03446045145392418, 0.029898418113589287, -0.018560517579317093, 0.021972866728901863, 0.002617587335407734, 0.10428658872842789, -0.04944944009184837, 0.07664937525987625, -0.03424760326743126, -0.008887539617717266, -0.07868296653032303, 0.004296313505619764, 0.11207737028598785, 0.06150135025382042, -0.06777924299240112, 0.23034334182739258, -0.09824039787054062, 0.23176263272762299, 0.21014845371246338, -0.16900676488876343, 0.062023747712373734, -0.006118505261838436, -0.026152264326810837, -0.022883309051394463, 0.06577979773283005, 0.013145336881279945, 0.018537122756242752, 0.019990185275673866, 0.17416749894618988, -0.05235980451107025, -0.027067814022302628, -0.011080794967710972, -0.07247885316610336, -0.00978275015950203, 0.04679108038544655, 0.08780468255281448, -0.1469167172908783, 0.16849613189697266, 0.31132781505584717, -0.013906165026128292, 0.07057876884937286, -0.03862464800477028, -0.00969420000910759, 0.03858736529946327, -0.02555689960718155, -0.01832895167171955, -0.00753486854955554, -0.12924771010875702, 0.009394427761435509, 0.10876920074224472, 0.015896691009402275, 0.07074429839849472, -0.13710680603981018, -0.045919403433799744, 0.003918658941984177, -0.01009833998978138, -0.032725732773542404, 0.07581792026758194, 0.008631033822894096, 0.10835693031549454, -0.03264651447534561, -0.06437046080827713, 0.123822420835495, 0.029191283509135246, -0.09418227523565292, 0.1392623484134674, -0.17335090041160583, -0.26775309443473816, -0.1764364242553711, -0.12391039729118347, -0.030479945242404938, 0.015311352908611298, 0.15357957780361176, -0.03749791532754898, -0.048915065824985504, -0.00007263442239491269, -0.08474317193031311, -0.031013313680887222, 0.005559484474360943, 0.005002529360353947, 0.03174421191215515, 0.024705540388822556, -0.12703348696231842, -0.05319864675402641, 0.02746308036148548, -0.023468509316444397, 0.09643476456403732, -0.07996387779712677, 0.08806337416172028, 0.0958549901843071, 0.020643893629312515, 0.02932734414935112, -0.00825952086597681, 0.13522061705589294, -0.014479542151093483, 0.0013834310229867697, 0.25092613697052, -0.013742741197347641, 0.08801493793725967, 0.11183018237352371, 0.010136323049664497, -0.04858143627643585, 0.009055139496922493, -0.06658154726028442, -0.07651500403881073, -0.2714214622974396, -0.09263240545988083, -0.10003478080034256, 0.06438066810369492, 0.061698928475379944, 0.07818742096424103, 0.13583698868751526, 0.08747946470975876, -0.03248957544565201, 0.030506648123264313, -0.03211503103375435, 0.09743855148553848, 0.2555447220802307, -0.025176601484417915, 0.12891273200511932, -0.1269153654575348, -0.030954653397202492, 0.12694242596626282, 0.08371598273515701, 0.10590073466300964, 0.10160430520772934, 0.051557447761297226, 0.05325188860297203, 0.17843466997146606, 0.07512391358613968, 0.13275974988937378, 0.03223362937569618, -0.0218141358345747, -0.04455477371811867, -0.029173532500863075, -0.0705365464091301, 0.05708703026175499, -0.07295859605073929, -0.11087924987077713, -0.026086905971169472, -0.12267530709505081, 0.04912329465150833, 0.15167374908924103, 0.03260583057999611, -0.17569538950920105, 0.004984318278729916, 0.0684858188033104, -0.010155906900763512, -0.07671070843935013, 0.07608667016029358, -0.0805806964635849, -0.10683809220790863, 0.11846431344747543, -0.025947168469429016, 0.12722750008106232, 0.034859783947467804, 0.05635649710893631, -0.05809636041522026, -0.10959172993898392, 0.06383580714464188, 0.13386929035186768, -0.34036576747894287, 0.19154304265975952, -0.013072090223431587, -0.026505006477236748, -0.09062088280916214, 0.007474956102669239, 0.033928144723176956, 0.1653033196926117, 0.07942576706409454, 0.02359463833272457, -0.0978393405675888, -0.013540403917431831, -0.04927901178598404, 0.040264792740345, 0.026391852647066116, 0.0354730524122715, -0.0479319766163826, -0.060511134564876556, -0.017747119069099426, -0.0016101868823170662, 0.056854475289583206, -0.043094977736473083, -0.15984880924224854, 0.07390711456537247, 0.1282113492488861, 0.03012380562722683, -0.056982096284627914, -0.006998253054916859, -0.05290612950921059, 0.19696283340454102, -0.08011247217655182, -0.07158086448907852, -0.09493527561426163, -0.12436679005622864, 0.05385299399495125, -0.07199086993932724, 0.06745593994855881, -0.09251400828361511, -0.011777443811297417, -0.07750517874956131, -0.19208881258964539, 0.09629534929990768, -0.13449956476688385, -0.04357421398162842, -0.04016205295920372, 0.1267463117837906, -0.10915971547365189, 0.01363675482571125, 0.031107960268855095, 0.0045674326829612255, -0.1442185789346695, -0.10512310266494751, -0.01466408558189869, 0.01390861440449953, 0.07032927870750427, -0.06855132430791855, -0.059206217527389526, -0.04210440069437027, 0.0002269123069709167, -0.07095591723918915, 0.23760126531124115, 0.1743471324443817, -0.09102169424295425, 0.18125031888484955, 0.16369777917861938, -0.08062667399644852, -0.27284666895866394, -0.17046639323234558, -0.1256026327610016, -0.07379941642284393, -0.02152799814939499, -0.13010862469673157, 0.0928943082690239, 0.03854353725910187, -0.08531101793050766, 0.13309340178966522, -0.20278358459472656, -0.08868534117937088, 0.19920559227466583, -0.029639476910233498, 0.3354334533214569, -0.14678631722927094, -0.09665258973836899, -0.09883969277143478, -0.23874326050281525, 0.12705844640731812, -0.05784017592668533, 0.05659428983926773, -0.05013212561607361, 0.09030452370643616, -0.0055581978522241116, -0.07001134753227234, 0.1165020614862442, -0.019615652039647102, 0.011899770237505436, -0.12998872995376587, 0.015354842878878117, 0.05959603562951088, -0.025507992133498192, 0.0682363361120224, -0.16799941658973694, 0.04249343276023865, -0.1297016143798828, -0.00840586144477129, -0.08478473871946335, 0.0836782157421112, -0.00001981135028472636, -0.03807370737195015, -0.020114202052354813, -0.0809778943657875, 0.00864352472126484, -0.005378738511353731, 0.21508856117725372, 0.021311039105057716, 0.1004076674580574, 0.1621607393026352, 0.0570119246840477, -0.19071568548679352, 0.018602002412080765, -0.09594377875328064, -0.09199532866477966, 0.04919834062457085, -0.14196406304836273, 0.03942728415131569, 0.09658144414424896, -0.03966351971030235, 0.04124578833580017, 0.062003325670957565, 0.002129205269739032, -0.038048967719078064, 0.13631731271743774, -0.1981479525566101, 0.06881100684404373, -0.020387141034007072, 0.16533978283405304, 0.06280216574668884, 0.04884500801563263, 0.14566610753536224, 0.014593229629099369, -0.046061012893915176, 0.005350813735276461, 0.019779937341809273, -0.07924901694059372, 0.04157470911741257, 0.05258270353078842, 0.002496576402336359, -0.11633665859699249, 0.08739665150642395, -0.0006512592080980539, -0.1388893723487854, 0.00044502574019134045, 0.13315072655677795, -0.13069625198841095, -0.11353246122598648, 0.013988176360726357, 0.05645694583654404, -0.21390998363494873, -0.11363053321838379, -0.043982721865177155, -0.10140077024698257, 0.07960906624794006, 0.10171038657426834, 0.077805295586586, 0.0663294568657875, -0.011303826235234737, -0.07232298702001572, 0.023237422108650208, -0.023759078234434128, -0.048108603805303574, 0.04942779242992401, -0.07055307179689407, -0.015623528510332108, 0.00967564806342125, 0.09083372354507446, -0.036806222051382065, 0.0076259043999016285, -0.07885701209306717, 0.024485094472765923, -0.18741877377033234, -0.006544181611388922, -0.09549658745527267, -0.034454576671123505, 0.007622649893164635, -0.07942161709070206, -0.04239603504538536, 0.012974384240806103, -0.11313555389642715, -0.03135130554437637, -0.03334683179855347, 0.03711055591702461, -0.11675328016281128, -0.03606405481696129, 0.08109122514724731, -0.007913854904472828, 0.08382098376750946, 0.1311119794845581, -0.06978994607925415, 0.06548008322715759, -0.152426078915596, -0.09907804429531097, 0.07366190105676651, 0.03581872954964638, 0.014548269100487232, 0.009208097122609615, -0.005308505147695541, 0.14182445406913757, -0.005868135020136833, 0.02698882669210434, 0.06259270012378693, -0.14154714345932007, -0.036849476397037506, -0.002217935398221016, -0.1273764669895172, -0.0359746553003788, -0.08520586043596268, 0.11702703684568405, 0.02218184620141983, 0.15673206746578217, -0.0330800786614418, 0.04871964454650879, -0.052613042294979095, 0.030562758445739746, -0.03907517343759537, -0.14916402101516724, -0.1273747682571411, -0.06134314462542534, -0.0387486107647419, 0.002467484911903739, 0.21433435380458832, -0.029699983075261116, -0.03970152512192726, 0.059296615421772, 0.08205589652061462, -0.011385519988834858, -0.02196643315255642, 0.24418558180332184, 0.04766358062624931, -0.009007682092487812, -0.08916512131690979, 0.046584710478782654, 0.0018661929061636329, -0.051556818187236786, 0.08721576631069183, 0.08817291259765625, 0.067312091588974, 0.0969495177268982, 0.014695269986987114, -0.0016428417293354869, -0.13492125272750854, -0.17150603234767914, 0.002291272394359112, 0.0770316794514656, -0.0029846064280718565, 0.16163335740566254, 0.16227170825004578, -0.028343440964818, -0.005498356651514769, -0.04988420009613037, -0.003568929387256503, -0.15064309537410736, -0.1273166984319687, -0.06956228613853455, -0.12177696079015732, 0.0006072228425182402, -0.036003220826387405, 0.07679333537817001, 0.07427064329385757, 0.036336399614810944, -0.06808611750602722, -0.00641190679743886, 0.012408782728016376, -0.07595568895339966, 0.020203862339258194, -0.01782633550465107, 0.005795134697109461, -0.028218304738402367, -0.02324475534260273, -0.08802057057619095, -0.028519567102193832, -0.019917292520403862, 0.051076699048280716, 0.016268715262413025, 0.03196297585964203, -0.07994981110095978, -0.06839286535978317, -0.050182320177555084, 0.05162300914525986, 0.0382181853055954, 0.19898340106010437, 0.010192176327109337, 0.05114158242940903, 0.06887320429086685, 0.1854466199874878, -0.08363981544971466, -0.1572999656200409, -0.026065431535243988, 0.16500923037528992, 0.03500135987997055, 0.04705958440899849, -0.0141305448487401, 0.03860333189368248, -0.05177157372236252, 0.3302074074745178, 0.2516878843307495, -0.03691922873258591, 0.022354472428560257, -0.015065383166074753, 0.029254477471113205, 0.07383154332637787, 0.10979072004556656, 0.14805249869823456, 0.2221984714269638, -0.07927544414997101, -0.027760548517107964, -0.0504816398024559, 0.0109883863478899, -0.15811172127723694, 0.10640375316143036, -0.05229863524436951, -0.09875480830669403, 0.0020930017344653606, 0.06957528740167618, -0.0756518617272377, 0.12091214954853058, -0.05045023560523987, -0.12030607461929321, -0.031558405607938766, 0.021446725353598595, 0.21504542231559753, 0.007502072025090456, 0.039664313197135925, -0.027888575568795204, -0.03702016547322273, 0.1234923005104065, -0.027337750419974327, -0.21466319262981415, -0.029334696009755135, 0.06843926757574081, -0.11448980122804642, 0.09837319701910019, 0.003119915723800659, 0.06310181319713593, 0.09492339938879013, 0.09807661175727844, -0.11998645216226578, 0.0734376385807991, 0.00833623856306076, -0.043060753494501114, 0.019821465015411377, -0.0845298171043396, -0.009848442859947681, -0.08281207829713821, 0.0405195876955986, -0.09080923348665237, 0.05419011786580086, 0.075287364423275, -0.02639298513531685, -0.02869352139532566, 0.024383092299103737, -0.0703892856836319, 0.0735979676246643, -0.0008474259520880878, -0.039789699018001556, -0.0693918988108635, -0.05415460467338562, -0.028567655012011528, 0.02583366259932518, -0.13945288956165314, -0.06117057800292969, -0.04122313857078552, -0.0348254069685936, 0.10306832194328308, 0.03450657054781914, -0.15002690255641937, -0.024332381784915924, -0.10277141630649567, -0.0017641499871388078, -0.1554490476846695, 0.04002204164862633, 0.0484294556081295, -0.012734971940517426, 0.007129951845854521, -0.05800299718976021, 0.019703486934304237, 0.046860262751579285, -0.07697559148073196, -0.10241975635290146 ]
null
null
null
test 123
{}
null
gaga42gaga42/test
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
test 123
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
transformers
# Generating Right Wing News Using GPT2 ### I have built a custom model for it using data from Kaggle Creating a new finetuned model using data from FOX news ### My model can be accessed at gagan3012/Fox-News-Generator Check the [BenchmarkTest](https://github.com/gagan3012/Fox-News-Generator/blob/master/BenchmarkTest.ipynb) notebook for results Find the model at [gagan3012/Fox-News-Generator](https://huggingface.co/gagan3012/Fox-News-Generator) ``` from transformers import AutoTokenizer, AutoModelWithLMHead tokenizer = AutoTokenizer.from_pretrained("gagan3012/Fox-News-Generator") model = AutoModelWithLMHead.from_pretrained("gagan3012/Fox-News-Generator") ```
{}
text-generation
gagan3012/Fox-News-Generator
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Generating Right Wing News Using GPT2 ### I have built a custom model for it using data from Kaggle Creating a new finetuned model using data from FOX news ### My model can be accessed at gagan3012/Fox-News-Generator Check the BenchmarkTest notebook for results Find the model at gagan3012/Fox-News-Generator
[ "# Generating Right Wing News Using GPT2", "### I have built a custom model for it using data from Kaggle \n\nCreating a new finetuned model using data from FOX news", "### My model can be accessed at gagan3012/Fox-News-Generator\n\nCheck the BenchmarkTest notebook for results\n\nFind the model at gagan3012/Fox-News-Generator" ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Generating Right Wing News Using GPT2", "### I have built a custom model for it using data from Kaggle \n\nCreating a new finetuned model using data from FOX news", "### My model can be accessed at gagan3012/Fox-News-Generator\n\nCheck the BenchmarkTest notebook for results\n\nFind the model at gagan3012/Fox-News-Generator" ]
[ 50, 12, 29, 46 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Generating Right Wing News Using GPT2### I have built a custom model for it using data from Kaggle \n\nCreating a new finetuned model using data from FOX news### My model can be accessed at gagan3012/Fox-News-Generator\n\nCheck the BenchmarkTest notebook for results\n\nFind the model at gagan3012/Fox-News-Generator" ]
[ -0.11963313072919846, 0.09185560792684555, -0.0007779410225339234, 0.056687261909246445, 0.07153692841529846, 0.008634492754936218, 0.14654123783111572, 0.07984482496976852, -0.10457547008991241, -0.07969829440116882, 0.13334979116916656, 0.1392175555229187, 0.018882010132074356, 0.19939996302127838, 0.03721965476870537, -0.1377359926700592, 0.053804636001586914, 0.0356033556163311, 0.02349298819899559, 0.14292672276496887, 0.01944834552705288, 0.00940331444144249, 0.0525616779923439, 0.03313099592924118, -0.2396896928548813, 0.011233089491724968, 0.022133292630314827, -0.0508519783616066, 0.1268540620803833, 0.07213041186332703, -0.042629193514585495, 0.037750907242298126, 0.0008231369429267943, -0.0027552444953471422, 0.06744658946990967, -0.0029561296105384827, -0.06271353363990784, 0.09157269448041916, 0.05399633198976517, -0.04542125388979912, 0.18348442018032074, -0.010591171681880951, -0.09587015956640244, 0.013170800171792507, -0.17045336961746216, -0.17049364745616913, -0.10600676387548447, 0.05750017985701561, 0.04887053370475769, 0.12313518673181534, -0.05186554417014122, 0.09093167632818222, -0.011420776136219501, 0.07230140268802643, 0.15673837065696716, -0.3422490060329437, -0.06411666423082352, 0.07308842241764069, -0.012263291515409946, -0.047064587473869324, 0.01644984818994999, 0.10174783319234848, 0.06854811310768127, 0.016979610547423363, 0.025966836139559746, -0.07499351352453232, -0.08640432357788086, -0.00979616679251194, -0.1098104789853096, -0.04183831065893173, 0.1273588240146637, -0.014153888449072838, -0.013123038224875927, -0.13022573292255402, -0.022730911150574684, 0.10363028943538666, -0.023363400250673294, 0.009213241748511791, -0.10641094297170639, 0.018117986619472504, 0.03745454177260399, -0.17921961843967438, -0.13269676268100739, -0.11593399941921234, -0.1465187519788742, 0.2442339062690735, 0.005838141310960054, 0.05375993996858597, -0.11838971823453903, 0.16646289825439453, -0.04125133901834488, -0.08961068838834763, 0.03090461529791355, -0.0707424134016037, 0.006054341327399015, 0.030305495485663414, -0.05585462227463722, -0.09038921445608139, 0.0248632300645113, 0.13388657569885254, -0.02759626694023609, -0.016361132264137268, 0.004200452473014593, 0.04958304017782211, 0.0372152104973793, 0.04481591284275055, -0.18004724383354187, 0.017433295026421547, 0.06492245942354202, -0.05286163091659546, 0.04489227011799812, -0.021434321999549866, -0.04386849328875542, -0.023559927940368652, -0.06384722888469696, 0.027774829417467117, 0.10184106230735779, 0.10216891020536423, -0.06795535981655121, -0.02885945700109005, 0.14824004471302032, -0.06785093247890472, -0.04311405122280121, -0.046692561358213425, 0.009089943021535873, 0.018465964123606682, 0.017169207334518433, 0.025134118273854256, -0.03672146797180176, -0.03445001691579819, -0.03987634927034378, -0.0067271157167851925, -0.0043208301067352295, -0.031231064349412918, -0.011741042137145996, 0.019588373601436615, -0.01425351481884718, -0.10622724145650864, -0.2641839385032654, -0.0360720157623291, -0.014456450939178467, -0.0504150316119194, -0.0803920328617096, -0.06649478524923325, 0.002349461428821087, 0.02908972091972828, -0.016416069120168686, 0.0827602669596672, -0.04351403936743736, 0.0701589435338974, 0.06943699717521667, 0.08005276322364807, -0.06870860606431961, 0.009264823980629444, -0.10450159758329391, -0.01116529293358326, -0.07158894091844559, 0.17239484190940857, -0.04133785143494606, 0.04651764780282974, -0.054202064871788025, 0.012361348606646061, -0.10219766944646835, 0.07441520690917969, 0.07101959735155106, 0.17552201449871063, -0.13013654947280884, -0.08638904243707657, 0.17197735607624054, -0.07053188979625702, -0.006497440859675407, 0.03500572592020035, -0.02615072950720787, 0.1386014223098755, 0.14810580015182495, 0.16948314011096954, 0.055226147174835205, 0.01503013540059328, 0.0035995198413729668, -0.017645366489887238, -0.190084308385849, -0.02883078157901764, 0.07946574687957764, 0.08002862334251404, -0.20724184811115265, 0.06344485282897949, -0.11553864181041718, 0.07189928740262985, -0.021057773381471634, -0.037304285913705826, 0.008354824036359787, -0.03263648599386215, 0.10963672399520874, -0.03766978532075882, 0.08428595215082169, -0.04396135360002518, -0.09834393858909607, -0.14079709351062775, 0.0535859577357769, 0.0048417700454592705, -0.007030096836388111, -0.07639797031879425, 0.1436328887939453, -0.19613933563232422, 0.08771270513534546, -0.1533072143793106, -0.10311370342969894, -0.04651936516165733, 0.03385970741510391, 0.12602059543132782, 0.06005803868174553, 0.02816140651702881, -0.05586445331573486, -0.0278981514275074, -0.018663544207811356, 0.06271453201770782, -0.046089138835668564, -0.02814383991062641, -0.10214252024888992, 0.07928342372179031, -0.054243702441453934, 0.06658957153558731, -0.12071969360113144, 0.03627653419971466, -0.07681281119585037, 0.025966180488467216, -0.004704627208411694, 0.014482072554528713, 0.10256630927324295, -0.01309163961559534, 0.011635185219347477, -0.012949538417160511, 0.08466006070375443, 0.06216326355934143, -0.040794771164655685, 0.16848452389240265, -0.04344593361020088, 0.08064693212509155, 0.09210673719644547, -0.09941785782575607, -0.06702243536710739, 0.1133112981915474, -0.04350193217396736, 0.011789076030254364, -0.07665006816387177, 0.033264804631471634, 0.1362626552581787, -0.03259379416704178, 0.07616822421550751, -0.032667260617017746, -0.06681104004383087, -0.003594750538468361, -0.010794749483466148, -0.014449875801801682, 0.09808992594480515, 0.042874883860349655, -0.09535378962755203, 0.0517885684967041, 0.04994134232401848, 0.055848073214292526, 0.11658140271902084, 0.08423137664794922, -0.005353647284209728, -0.04241999611258507, -0.06679219752550125, -0.022676575928926468, -0.000477250519907102, 0.047381363809108734, 0.03159499168395996, 0.07095428556203842, -0.0006757140508852899, 0.08258242905139923, -0.09208519011735916, -0.0926123708486557, 0.0008679701131768525, -0.04090952128171921, -0.05114893615245819, 0.11485526710748672, -0.06683196127414703, 0.07366689294576645, 0.005055360030382872, 0.04004540666937828, 0.065222829580307, 0.03627939894795418, -0.08491905778646469, 0.0957142636179924, -0.03485890105366707, -0.3119567334651947, -0.16898375749588013, -0.007113758474588394, -0.06006081774830818, 0.0073811630718410015, 0.07326926290988922, -0.08452911674976349, 0.021526483818888664, -0.03364930301904678, -0.01985837332904339, 0.0642232671380043, 0.0037513584829866886, 0.05094210058450699, -0.055816370993852615, 0.008028039708733559, -0.0855308249592781, -0.015051721595227718, -0.05417405068874359, -0.05079547315835953, 0.07685249298810959, -0.05719384923577309, 0.10954104363918304, 0.06342317909002304, 0.06948017328977585, 0.022249380126595497, -0.02200431376695633, 0.24033522605895996, -0.10328827053308487, 0.025475619360804558, 0.1647988110780716, 0.02096061035990715, 0.03547758609056473, 0.09646332263946533, 0.02355942875146866, -0.07709217816591263, 0.024178463965654373, -0.00786310713738203, -0.0808258131146431, -0.21613232791423798, -0.017081547528505325, -0.02507217414677143, 0.029594020918011665, -0.009735394269227982, 0.05923439934849739, 0.022713690996170044, 0.11415477842092514, -0.021124335005879402, 0.06476574391126633, 0.04527465999126434, 0.06124390289187431, 0.1764245629310608, -0.004142447374761105, 0.11481085419654846, -0.08970092236995697, -0.06381931900978088, 0.10625004023313522, 0.051626723259687424, 0.05656663328409195, -0.0019117790507152677, 0.16644346714019775, 0.06261055916547775, 0.022744596004486084, 0.08183231204748154, 0.09487663954496384, -0.00037895701825618744, -0.02362482063472271, 0.0025368398055434227, -0.03384549543261528, 0.01695435866713524, -0.021796781569719315, -0.09890776127576828, -0.05988757684826851, 0.05605943128466606, -0.05405835807323456, 0.02703707665205002, 0.12013412266969681, 0.09128648787736893, -0.25050196051597595, -0.03402001038193703, -0.01068334374576807, 0.00684512872248888, -0.08601708710193634, -0.020511530339717865, -0.00493094976991415, -0.11273691803216934, 0.07725967466831207, 0.010965532623231411, 0.09111742675304413, -0.0811149924993515, 0.001110864570364356, -0.0829337015748024, -0.09157370030879974, -0.04755140468478203, 0.14315162599086761, -0.21976979076862335, 0.16836319863796234, 0.01577574573457241, 0.050333015620708466, -0.021235568448901176, -0.02436191402375698, 0.057165686041116714, -0.04090707749128342, 0.14210940897464752, -0.015192023478448391, -0.04524332284927368, -0.05431042239069939, -0.18371732532978058, 0.03842606768012047, -0.0827118307352066, -0.14671191573143005, 0.028368227183818817, 0.008991114795207977, -0.030071182176470757, -0.05175023153424263, -0.045730095356702805, -0.14011605083942413, -0.0933312475681305, 0.07007939368486404, 0.09841093420982361, 0.005112888291478157, -0.03364376723766327, -0.09863618016242981, -0.11915107816457748, 0.18542258441448212, 0.006131670903414488, -0.12322571128606796, -0.11609165370464325, 0.06683959811925888, -0.0456685870885849, -0.06340110301971436, -0.057603079825639725, 0.023653294891119003, 0.13520315289497375, -0.008391744457185268, -0.15918150544166565, 0.06847252696752548, -0.07202191650867462, -0.11355140060186386, -0.008726971223950386, 0.13985802233219147, 0.10647062957286835, 0.017302019521594048, 0.06420888751745224, 0.05410996451973915, -0.046946894377470016, -0.07792763411998749, 0.05230342224240303, -0.011991363950073719, -0.011536727659404278, 0.023820694535970688, 0.03512980788946152, -0.11088619381189346, -0.07800308614969254, -0.01234576478600502, 0.1706942915916443, 0.08036646246910095, -0.0903887003660202, 0.0882926806807518, 0.16318121552467346, -0.0738358199596405, -0.2469068020582199, -0.07173648476600647, -0.00258798711001873, 0.03759201988577843, 0.04212111979722977, -0.18962080776691437, 0.1603235900402069, 0.06585706025362015, -0.02620832808315754, 0.1346510350704193, -0.2160678207874298, -0.06568162143230438, 0.0693412721157074, 0.0158780999481678, 0.3377305269241333, -0.04844743013381958, -0.09258664399385452, 0.019926786422729492, -0.13167686760425568, 0.12135332822799683, -0.12033242732286453, 0.0729001834988594, -0.04588340222835541, 0.06285984814167023, -0.01352500356733799, -0.02568059042096138, 0.11095230281352997, 0.07756436616182327, 0.0044571952894330025, -0.11498190462589264, -0.02183648943901062, 0.12813088297843933, -0.028525054454803467, 0.12703415751457214, -0.07165564596652985, 0.029649388045072556, -0.19669589400291443, -0.08514564484357834, -0.10011208802461624, -0.007434538099914789, 0.0039260247722268105, 0.02244492620229721, -0.011346939019858837, -0.024474408477544785, -0.08668767660856247, 0.024772915989160538, 0.010446717962622643, -0.07048317790031433, 0.10455670207738876, -0.07680821418762207, 0.08013135939836502, 0.013701271265745163, -0.05608079582452774, -0.10027001053094864, -0.05170460045337677, 0.06309118121862411, -0.09271351993083954, -0.023958293721079826, 0.1186773031949997, -0.0011260207975283265, 0.03265850618481636, 0.056385986506938934, -0.08202958852052689, 0.09171415865421295, 0.07872262597084045, -0.22760388255119324, -0.04247824475169182, -0.07702845335006714, -0.050346311181783676, 0.0015219210181385279, 0.07034580409526825, 0.186017706990242, 0.04494774341583252, -0.08969337493181229, 0.046267516911029816, -0.003337417496368289, -0.023476142436265945, 0.11724121123552322, 0.03481828793883324, 0.013420136645436287, -0.14311257004737854, 0.011926691979169846, 0.02462344989180565, 0.10217288136482239, 0.025991274043917656, 0.10742013156414032, -0.11736655980348587, -0.13448278605937958, -0.01792980171740055, 0.20405448973178864, -0.0719556212425232, -0.05225606635212898, -0.031095048412680626, -0.04402901604771614, 0.09047261625528336, 0.07832647114992142, 0.07119264453649521, 0.03145643323659897, -0.04970800131559372, -0.05782424286007881, 0.0024249833077192307, 0.07367907464504242, 0.11502032727003098, 0.023736031726002693, -0.0324970968067646, -0.07827923446893692, 0.00027351579046808183, 0.17149190604686737, -0.0810568779706955, -0.09052284806966782, -0.039425794035196304, 0.037682898342609406, -0.08233626186847687, -0.008300233632326126, -0.08552197366952896, -0.015446553006768227, -0.04348762333393097, -0.0469161756336689, -0.04387739300727844, 0.005840678699314594, -0.08069156110286713, -0.002483118325471878, -0.017149033024907112, 0.018780726939439774, -0.1354208141565323, 0.005783822853118181, 0.0017344503430649638, -0.01773793436586857, 0.08649665117263794, 0.08399855345487595, -0.007812048774212599, 0.08856938034296036, -0.02520151436328888, -0.048198577016592026, -0.011274170130491257, -0.021992504596710205, 0.011756740510463715, -0.028006339445710182, 0.04343762621283531, 0.01837122067809105, -0.005211230833083391, 0.06530884653329849, -0.030245447531342506, -0.09091369062662125, 0.08041203767061234, -0.007074506022036076, -0.05051514506340027, -0.0216128621250391, -0.009365921840071678, 0.0829114019870758, 0.10114837437868118, 0.13935549557209015, -0.03999766707420349, 0.008469869382679462, -0.09929420799016953, 0.043728068470954895, -0.04863244667649269, -0.1482221484184265, -0.03591323271393776, -0.051396921277046204, 0.05785161256790161, 0.048570409417152405, 0.12023467570543289, -0.01842878758907318, -0.06716111302375793, -0.03301753103733063, 0.0484900027513504, 0.026631653308868408, -0.032816436141729355, 0.05417262390255928, 0.06930471956729889, -0.01026426162570715, -0.05489487200975418, 0.07346247881650925, 0.014494027942419052, -0.034042615443468094, 0.11118939518928528, -0.10998782515525818, 0.10940443724393845, 0.11415580660104752, 0.0065006231889128685, 0.03315933793783188, -0.06432995945215225, -0.06164751201868057, -0.153250589966774, 0.015859130769968033, 0.013977567665278912, 0.06334461271762848, 0.15975432097911835, -0.014185730367898941, -0.013944393023848534, 0.08033843338489532, -0.012705458328127861, -0.06443341076374054, -0.31544747948646545, -0.06218516454100609, -0.0646546259522438, -0.013295923359692097, -0.058941278606653214, -0.07441069930791855, 0.12387427687644958, 0.06874383240938187, -0.04641200229525566, 0.16604235768318176, 0.14824113249778748, -0.04890148341655731, 0.13008330762386322, -0.08340855687856674, -0.01180508267134428, -0.03526785597205162, 0.07496844977140427, 0.02519685961306095, 0.12486397475004196, 0.052319057285785675, 0.035223428159952164, -0.004618968348950148, 0.03567094728350639, -0.10817556828260422, -0.09455855190753937, -0.04440523311495781, 0.08811027556657791, 0.03915686905384064, -0.002668029861524701, 0.06601982563734055, 0.00236289924941957, 0.04893821105360985, 0.15094958245754242, 0.04869777336716652, 0.011921620927751064, -0.10374298691749573, 0.12865060567855835, -0.06546469032764435, 0.039678625762462616, -0.04614073410630226, -0.021706001833081245, 0.02073623239994049, 0.26070183515548706, 0.223290354013443, -0.03471682593226433, 0.011329091154038906, -0.02283775620162487, 0.0050001321360468864, 0.08646666258573532, 0.032427798956632614, 0.028378473594784737, 0.18527460098266602, -0.09188461303710938, -0.07781786471605301, -0.04704492911696434, -0.04978571832180023, 0.005188310518860817, -0.06527682393789291, 0.005734321195632219, -0.0014442817773669958, -0.019551245495676994, 0.17403647303581238, -0.05171119421720505, 0.008902084082365036, -0.1403435915708542, 0.018940772861242294, -0.11098527908325195, -0.005806660745292902, -0.07539937645196915, -0.026050593703985214, 0.07309692353010178, -0.010283500887453556, 0.06163840740919113, 0.07146027684211731, -0.0038115987554192543, -0.05329431965947151, -0.0817437618970871, 0.05182458832859993, -0.03356918320059776, 0.14653250575065613, -0.05552578344941139, 0.07692635804414749, 0.09056795388460159, -0.041513726115226746, -0.1517038345336914, 0.08301891386508942, -0.0760594978928566, 0.06285792589187622, 0.06774493306875229, 0.010217524133622646, 0.05451791733503342, -0.10767218470573425, 0.1100955531001091, -0.020021185278892517, 0.0673796609044075, -0.14725486934185028, 0.007013405207544565, -0.07004983723163605, 0.07595127820968628, -0.049930840730667114, 0.11803462356328964, 0.10259495675563812, -0.04729953780770302, 0.03880723938345909, -0.01845790259540081, 0.03801395744085312, 0.00954238511621952, 0.022375980392098427, -0.02505427598953247, -0.1186317503452301, -0.01793898642063141, -0.0448971726000309, 0.042034659534692764, -0.2521383762359619, -0.007761910557746887, -0.10624484717845917, -0.033991314470767975, 0.04435089975595474, 0.02615496888756752, 0.08615696430206299, -0.007088809739798307, -0.002885175636038184, -0.03885525092482567, -0.018387436866760254, 0.05612585321068764, -0.1268642544746399, -0.07499450445175171 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # ViTGPT2I2A This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the vizwiz dataset. It achieves the following results on the evaluation set: - Loss: 0.0708 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - distributed_type: multi-GPU - num_devices: 2 - total_train_batch_size: 4 - total_eval_batch_size: 4 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 0.1528 | 0.17 | 1000 | 0.0869 | | 0.0899 | 0.34 | 2000 | 0.0817 | | 0.084 | 0.51 | 3000 | 0.0790 | | 0.0814 | 0.68 | 4000 | 0.0773 | | 0.0803 | 0.85 | 5000 | 0.0757 | | 0.077 | 1.02 | 6000 | 0.0745 | | 0.0739 | 1.19 | 7000 | 0.0740 | | 0.0719 | 1.37 | 8000 | 0.0737 | | 0.0717 | 1.54 | 9000 | 0.0730 | | 0.0731 | 1.71 | 10000 | 0.0727 | | 0.0708 | 1.88 | 11000 | 0.0720 | | 0.0697 | 2.05 | 12000 | 0.0717 | | 0.0655 | 2.22 | 13000 | 0.0719 | | 0.0653 | 2.39 | 14000 | 0.0719 | | 0.0657 | 2.56 | 15000 | 0.0712 | | 0.0663 | 2.73 | 16000 | 0.0710 | | 0.0654 | 2.9 | 17000 | 0.0708 | | 0.0645 | 3.07 | 18000 | 0.0716 | | 0.0616 | 3.24 | 19000 | 0.0712 | | 0.0607 | 3.41 | 20000 | 0.0712 | | 0.0611 | 3.58 | 21000 | 0.0711 | | 0.0615 | 3.76 | 22000 | 0.0711 | | 0.0614 | 3.93 | 23000 | 0.0710 | | 0.0594 | 4.1 | 24000 | 0.0716 | | 0.0587 | 4.27 | 25000 | 0.0715 | | 0.0574 | 4.44 | 26000 | 0.0715 | | 0.0579 | 4.61 | 27000 | 0.0715 | | 0.0581 | 4.78 | 28000 | 0.0715 | | 0.0579 | 4.95 | 29000 | 0.0715 | ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.2+cu113 - Datasets 1.18.3 - Tokenizers 0.11.0
{"license": "apache-2.0", "tags": ["image-captioning", "generated_from_trainer"], "model-index": [{"name": "ViTGPT2I2A", "results": []}]}
null
gagan3012/ViTGPT2I2A
[ "transformers", "pytorch", "vision-encoder-decoder", "image-captioning", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #vision-encoder-decoder #image-captioning #generated_from_trainer #license-apache-2.0 #endpoints_compatible #has_space #region-us
ViTGPT2I2A ========== This model is a fine-tuned version of google/vit-base-patch16-224-in21k on the vizwiz dataset. It achieves the following results on the evaluation set: * Loss: 0.0708 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 2 * eval\_batch\_size: 2 * seed: 42 * distributed\_type: multi-GPU * num\_devices: 2 * total\_train\_batch\_size: 4 * total\_eval\_batch\_size: 4 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 5.0 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.16.2 * Pytorch 1.10.2+cu113 * Datasets 1.18.3 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 2\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 2\n* total\\_train\\_batch\\_size: 4\n* total\\_eval\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.2+cu113\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #vision-encoder-decoder #image-captioning #generated_from_trainer #license-apache-2.0 #endpoints_compatible #has_space #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 2\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 2\n* total\\_train\\_batch\\_size: 4\n* total\\_eval\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.2+cu113\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ 54, 162, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #vision-encoder-decoder #image-captioning #generated_from_trainer #license-apache-2.0 #endpoints_compatible #has_space #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 2\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 2\n* total\\_train\\_batch\\_size: 4\n* total\\_eval\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.2+cu113\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ -0.12856616079807281, 0.06857029348611832, -0.0018251098226755857, 0.04902142286300659, 0.1469397097826004, 0.023551752790808678, 0.09244849532842636, 0.13273078203201294, -0.10760407894849777, 0.09026198834180832, 0.09397052228450775, 0.05353083834052086, 0.07332364469766617, 0.11959796398878098, -0.003441691165789962, -0.327334463596344, 0.035638436675071716, 0.0357758104801178, -0.1579083651304245, 0.1114690750837326, 0.1038006842136383, -0.11435340344905853, 0.06103459745645523, 0.03672514483332634, -0.17514993250370026, 0.005576816387474537, -0.03446066752076149, -0.05510198697447777, 0.09966626763343811, 0.062460508197546005, 0.10462162643671036, -0.0030426287557929754, 0.07734812051057816, -0.22001121938228607, 0.0037027308717370033, 0.09538591653108597, 0.01286247093230486, 0.07378855347633362, 0.11101038753986359, 0.0022797249257564545, 0.16280457377433777, -0.08830421417951584, 0.05927694961428642, 0.06475604325532913, -0.11965689063072205, -0.3208426237106323, -0.08502158522605896, 0.0652308389544487, 0.10681473463773727, 0.08182358741760254, -0.017406120896339417, 0.06086991727352142, -0.07402762770652771, 0.07808513939380646, 0.2236102670431137, -0.21949441730976105, -0.09333537518978119, 0.006337581668049097, 0.04679505154490471, 0.03912503644824028, -0.11548709124326706, -0.011776551604270935, 0.05284476652741432, 0.04135752096772194, 0.0818236693739891, 0.030272720381617546, -0.030942898243665695, -0.005740384105592966, -0.15303336083889008, -0.0930464044213295, 0.1763640195131302, 0.09557200223207474, -0.04021938517689705, -0.04992865025997162, -0.03953045606613159, -0.2069891095161438, -0.04893054813146591, 0.03685445338487625, 0.03671165555715561, -0.04822542518377304, -0.12273982167243958, 0.0020284797064960003, -0.0991528183221817, -0.09666956961154938, -0.008007323369383812, 0.14590030908584595, 0.05879300460219383, 0.0027325276751071215, -0.01833193190395832, 0.12758031487464905, 0.011383473873138428, -0.14154070615768433, -0.05147700384259224, 0.022379880771040916, -0.06627310067415237, -0.009883787482976913, -0.054334189742803574, 0.01080882828682661, -0.009926870465278625, 0.1349475383758545, -0.06051890552043915, 0.05966470390558243, 0.01678427867591381, 0.050234176218509674, -0.09860600531101227, 0.20523275434970856, -0.08840158581733704, 0.03877318650484085, -0.040057532489299774, 0.06868312507867813, -0.01636507362127304, -0.008791370317339897, -0.05539736896753311, 0.01475521270185709, 0.10599058121442795, 0.02838376723229885, -0.019519342109560966, 0.030921999365091324, -0.05238984525203705, -0.026359183713793755, 0.028148150071501732, -0.08587580919265747, 0.033863067626953125, 0.009609424509108067, -0.09557947516441345, 0.0335216261446476, 0.0485721081495285, -0.00940032210201025, -0.007119427900761366, 0.09634388238191605, -0.09212712198495865, -0.0006251304876059294, -0.1193406879901886, -0.11162494868040085, 0.045284755527973175, -0.007394679822027683, -0.0016250472981482744, -0.09731009602546692, -0.11633311957120895, -0.021092111244797707, 0.057047367095947266, -0.02930961363017559, -0.05787788704037666, -0.015896176919341087, -0.09271836280822754, 0.04804747924208641, -0.014232590794563293, 0.14441686868667603, -0.0439562126994133, 0.11417125165462494, 0.06412193179130554, 0.05558185651898384, 0.003391091013327241, 0.05184713378548622, -0.056053996086120605, 0.06613557040691376, -0.16515235602855682, 0.06860480457544327, -0.08445880562067032, 0.07126294821500778, -0.11652451753616333, -0.14313355088233948, -0.01269803661853075, -0.008960342034697533, 0.09689917415380478, 0.09521234035491943, -0.1337486058473587, -0.08585399389266968, 0.15969255566596985, -0.09371563792228699, -0.1432458609342575, 0.11340739578008652, -0.0015298143262043595, -0.03379114717245102, 0.027499891817569733, 0.11429063975811005, 0.11502491682767868, -0.09857552498579025, -0.04074689745903015, -0.013927421532571316, 0.11616934835910797, -0.020080244168639183, 0.12340829521417618, -0.0014887121506035328, 0.0150887630879879, 0.0177235696464777, -0.07468875497579575, 0.07207435369491577, -0.12025957554578781, -0.0912904217839241, -0.05351400375366211, -0.07569762319326401, -0.0033226499799638987, 0.07045912742614746, 0.04809064418077469, -0.0947011336684227, -0.1291433572769165, 0.06663145124912262, 0.12598319351673126, -0.08842989057302475, 0.03050239011645317, -0.07347715646028519, 0.054553281515836716, -0.0676078051328659, 0.00018819503020495176, -0.16973690688610077, -0.10636269301176071, 0.008018498308956623, -0.08094712346792221, -0.0011440522503107786, -0.0418904647231102, 0.06250390410423279, 0.06900820136070251, -0.046467769891023636, -0.06957484781742096, -0.11876513063907623, -0.030909791588783264, -0.06747561693191528, -0.18190672993659973, -0.12300582230091095, -0.00860509742051363, 0.14488576352596283, -0.21783936023712158, 0.02638181485235691, 0.05068724974989891, 0.12504328787326813, 0.018965791910886765, -0.043944861739873886, -0.04092588275671005, 0.0648934468626976, -0.04503335431218147, -0.07872182130813599, 0.02667820081114769, 0.015182365663349628, -0.09637386351823807, 0.009344998747110367, -0.1052842065691948, 0.14023391902446747, 0.1180935949087143, -0.07966407388448715, -0.06538543850183487, 0.006070537492632866, -0.09792284667491913, -0.05386875197291374, -0.030995698645710945, -0.032483797520399094, 0.08963043987751007, 0.008430098183453083, 0.1280103176832199, -0.0794861689209938, -0.04380084201693535, 0.04748089984059334, -0.00566138094291091, -0.03962826728820801, 0.10880885273218155, 0.07884712517261505, -0.04612269252538681, 0.12530392408370972, 0.1007290631532669, -0.09354472905397415, 0.12837280333042145, -0.06644374132156372, -0.11995889246463776, -0.01889818161725998, 0.027850696817040443, 0.040834907442331314, 0.14611487090587616, -0.1422574371099472, -0.0040361215360462666, 0.026862138882279396, 0.029130512848496437, 0.045658037066459656, -0.22553998231887817, -0.006898994091898203, 0.04146239534020424, -0.06319280713796616, -0.03993004932999611, -0.012366233393549919, -0.0026374177541583776, 0.08979635685682297, 0.01645069383084774, -0.015755822882056236, 0.002377212280407548, -0.010628602467477322, -0.07461771368980408, 0.21404749155044556, -0.07919196784496307, -0.16425372660160065, -0.1722627431154251, -0.014636827632784843, -0.04156949743628502, -0.006213418208062649, 0.04609450697898865, -0.10794185847043991, -0.04104642570018768, -0.049242179840803146, 0.048661790788173676, -0.08050765097141266, 0.022966578602790833, 0.003680447582155466, 0.014263039454817772, 0.12287141382694244, -0.11131175607442856, 0.02128293551504612, 0.009369557723402977, -0.08812399953603745, 0.025588229298591614, 0.023012913763523102, 0.12453678995370865, 0.14510361850261688, 0.0057045090943574905, 0.03618764877319336, -0.02213645726442337, 0.19202308356761932, -0.09640591591596603, -0.04658253490924835, 0.12756729125976562, 0.03616848215460777, 0.05777387693524361, 0.07884544879198074, 0.059002585709095, -0.08357565850019455, 0.018181074410676956, 0.06851914525032043, -0.0376247838139534, -0.21294212341308594, -0.027651876211166382, -0.07534707337617874, 0.0021576895378530025, 0.12534523010253906, 0.03715764358639717, 0.013128199614584446, 0.06636623293161392, -0.04744119569659233, 0.05648012459278107, -0.04059857875108719, 0.09442775696516037, 0.09012863785028458, 0.04138428717851639, 0.11147508770227432, -0.04911521449685097, -0.024183617904782295, 0.03982711210846901, -0.0018301602685824037, 0.25383755564689636, -0.018087109550833702, 0.13587233424186707, 0.06838881224393845, 0.17210188508033752, 0.009683225303888321, 0.07310111820697784, 0.011020844802260399, -0.025468673557043076, 0.02442338690161705, -0.059867072850465775, -0.018336070701479912, 0.035696420818567276, 0.034266188740730286, 0.05981152504682541, -0.16390255093574524, -0.036969494074583054, 0.02511689066886902, 0.31999924778938293, 0.07710560411214828, -0.35181793570518494, -0.11979970335960388, 0.010555265471339226, -0.038936663419008255, -0.06463996320962906, 0.008615849539637566, 0.11297860741615295, -0.10127051174640656, 0.06382137537002563, -0.0835137590765953, 0.11457084119319916, -0.0316397100687027, -0.015657346695661545, 0.10769560933113098, 0.08689165860414505, 0.011222217231988907, 0.08385250717401505, -0.24114322662353516, 0.2806897461414337, -0.01786050945520401, 0.06560180336236954, -0.03197818249464035, 0.029516072943806648, 0.04670148715376854, 0.02634974755346775, 0.04525017365813255, -0.00573374517261982, -0.07716517150402069, -0.22155572474002838, -0.061025988310575485, 0.032903362065553665, 0.12937094271183014, -0.06185419484972954, 0.13411647081375122, -0.029081284999847412, -0.0019655274227261543, 0.05684760957956314, -0.002811771584674716, -0.11077655106782913, -0.0960615947842598, 0.022009070962667465, -0.001323463162407279, 0.08880281448364258, -0.11826222389936447, -0.10933715105056763, -0.06858546286821365, 0.16091765463352203, -0.06521964818239212, -0.02393385022878647, -0.13042256236076355, 0.12072170525789261, 0.15584276616573334, -0.07304803282022476, 0.06560743600130081, -0.010071953758597374, 0.16156189143657684, 0.040462784469127655, -0.04977627471089363, 0.09039891511201859, -0.08590976893901825, -0.190002903342247, -0.031005749478936195, 0.11683903634548187, 0.015531417913734913, 0.046437788754701614, -0.0290150735527277, 0.01611865684390068, -0.02943028323352337, -0.09674038738012314, 0.04526417329907417, 0.004037197679281235, 0.021526504307985306, 0.08044920116662979, -0.03159822151064873, 0.049812234938144684, -0.03798813000321388, -0.031185142695903778, 0.119715616106987, 0.2569681406021118, -0.07942303270101547, -0.06624361872673035, 0.018549639731645584, -0.05216311663389206, -0.15605342388153076, 0.0524141751229763, 0.1222665086388588, 0.028830120339989662, -0.011451976373791695, -0.22030536830425262, 0.07022725045681, 0.1186317428946495, -0.030270755290985107, 0.1406230926513672, -0.3230743408203125, -0.11716735363006592, 0.07946405559778214, 0.12188618630170822, -0.023986808955669403, -0.18103289604187012, -0.05085982382297516, -0.023188704624772072, -0.1520540714263916, 0.10575477033853531, -0.01722533255815506, 0.12369327247142792, -0.023973992094397545, 0.06259862333536148, 0.011770072393119335, -0.05518357828259468, 0.19033357501029968, -0.025860672816634178, 0.069059357047081, -0.020874707028269768, 0.056889794766902924, 0.114003024995327, -0.07748059928417206, 0.022512728348374367, -0.0426073782145977, 0.040869858115911484, -0.13971127569675446, -0.011685382574796677, -0.09571602940559387, 0.026587648317217827, -0.04382115975022316, -0.018494226038455963, -0.051665693521499634, 0.049842528998851776, 0.04205557703971863, 0.011530157178640366, 0.1725502461194992, -0.0006440396537072957, 0.15345995128154755, 0.05769141763448715, 0.03699124604463577, 0.003552250796929002, -0.14123030006885529, -0.01995878294110298, -0.004605618771165609, 0.0864790827035904, -0.1700584590435028, 0.025174986571073532, 0.14616794884204865, 0.053760841488838196, 0.13614287972450256, 0.08299034088850021, -0.07668276876211166, 0.03956926614046097, 0.09117048978805542, -0.07249771803617477, -0.1339653879404068, -0.032636962831020355, 0.06581245362758636, -0.1591469645500183, 0.04562933370471001, 0.07771844416856766, -0.08363422751426697, 0.007870770990848541, 0.0041042775847017765, 0.013060230761766434, -0.07212276756763458, 0.196356862783432, 0.05713522061705589, 0.08445090800523758, -0.09000607579946518, 0.09024245291948318, 0.05170330032706261, -0.16657625138759613, -0.010905388742685318, 0.07161309570074081, -0.0308124590665102, -0.011327342130243778, 0.016021287068724632, 0.03937825933098793, -0.07461495697498322, -0.06747151166200638, -0.11489741504192352, -0.1452304571866989, 0.09090851992368698, 0.1059226468205452, 0.0608685165643692, 0.046006496995687485, -0.018948858603835106, 0.0636981874704361, -0.12024018913507462, 0.07967095822095871, 0.06845775991678238, 0.10169318318367004, -0.17143569886684418, 0.15483178198337555, 0.026313627138733864, 0.05293087288737297, 0.00281797768548131, -0.018146444112062454, -0.09722787141799927, 0.02390967309474945, -0.14088304340839386, -0.0288416538387537, -0.04476657882332802, -0.004382228013128042, 0.0029223919846117496, -0.04225378856062889, -0.08597515523433685, 0.0360192246735096, -0.1209331825375557, -0.0761801078915596, 0.00001615367000340484, 0.06634689122438431, -0.10695575177669525, -0.013436750508844852, 0.05410817265510559, -0.12290269136428833, 0.061525605618953705, 0.046268485486507416, 0.04410538822412491, 0.027196718379855156, -0.06456977128982544, 0.003180883824825287, 0.05608857795596123, 0.003691386664286256, 0.032615598291158676, -0.10425639152526855, 0.016380375251173973, -0.020944368094205856, 0.04018086567521095, -0.011064641177654266, 0.014153540134429932, -0.14402776956558228, -0.04878974333405495, -0.028858525678515434, -0.0446290522813797, -0.05182904005050659, 0.06071409583091736, 0.04474393650889397, 0.06079618260264397, 0.14401119947433472, -0.0631428211927414, 0.018536940217018127, -0.24617265164852142, 0.00541585823521018, -0.031188156455755234, -0.08507164567708969, -0.032603949308395386, -0.02480522356927395, 0.07196751981973648, -0.06513389945030212, 0.09752000123262405, -0.040247008204460144, 0.07417568564414978, 0.04138389974832535, -0.04860595613718033, 0.030644753947854042, 0.03854044899344444, 0.2623461186885834, 0.04166773706674576, -0.007693972438573837, 0.0888843983411789, 0.02635948732495308, 0.07967735081911087, 0.12065111100673676, 0.17166249454021454, 0.13013367354869843, -0.026801522821187973, 0.1209130510687828, 0.05839519202709198, -0.09472770243883133, -0.14091971516609192, 0.04387437924742699, -0.025589367374777794, 0.14605864882469177, -0.016739651560783386, 0.14151029288768768, 0.13774016499519348, -0.1903744339942932, 0.05052231252193451, -0.018117466941475868, -0.07319138944149017, -0.07787319272756577, -0.045686159282922745, -0.06832798570394516, -0.19106727838516235, 0.018988395109772682, -0.12513677775859833, 0.046770691871643066, 0.12277507781982422, 0.018377184867858887, 0.008437377400696278, 0.16618625819683075, 0.05102601647377014, -0.0070413448847830296, 0.09468211978673935, 0.0034023120533674955, -0.00792971346527338, -0.034127574414014816, -0.08881371468305588, 0.05447514355182648, -0.020900754258036613, 0.05898071825504303, -0.04536856338381767, -0.08054275065660477, 0.08957044780254364, 0.02523987367749214, -0.09995557367801666, 0.009203432127833366, 0.005419191904366016, 0.06766384094953537, 0.04661959037184715, 0.0017376819159835577, 0.01171090453863144, -0.031435512006282806, 0.21928918361663818, -0.10999941825866699, -0.037069715559482574, -0.12451609969139099, 0.25397971272468567, 0.014834382571280003, -0.03731505572795868, 0.027815289795398712, -0.08663319796323776, -0.03433064743876457, 0.15826372802257538, 0.10453828424215317, -0.014945601113140583, -0.037661198526620865, 0.007428981829434633, -0.02132045291364193, -0.04452776908874512, 0.11326782405376434, 0.11456800997257233, 0.09582876414060593, -0.08430836349725723, -0.044297102838754654, -0.05517372488975525, -0.03622778132557869, -0.011462370865046978, 0.054330840706825256, 0.01965707540512085, -0.01252467930316925, -0.048560719937086105, 0.09082026034593582, -0.05734604597091675, -0.04379291087388992, 0.09642694145441055, -0.15529568493366241, -0.17223717272281647, -0.038349736481904984, 0.05894595757126808, 0.012312747538089752, 0.058776892721652985, -0.03656376525759697, -0.02591106668114662, 0.11242061108350754, -0.003763156943023205, -0.06787796318531036, -0.15551671385765076, 0.08415581285953522, -0.05642198771238327, 0.23382030427455902, -0.04632209241390228, 0.010224158875644207, 0.12482109665870667, 0.04258327558636665, -0.10481370240449905, 0.014652026817202568, 0.05614795535802841, -0.12041818350553513, 0.02348008006811142, 0.17500339448451996, -0.039843954145908356, 0.09883657842874527, 0.020117199048399925, -0.14759084582328796, -0.0007826197543181479, -0.056226205080747604, -0.015660446137189865, -0.06662321835756302, -0.04019008204340935, -0.06149474158883095, 0.11539095640182495, 0.23617641627788544, -0.063785120844841, -0.03789869323372841, -0.0709281712770462, 0.02855399064719677, 0.06350447237491608, 0.11876462399959564, -0.025705639272928238, -0.24617476761341095, 0.012651091441512108, -0.0019440021133050323, -0.008303973823785782, -0.229838564991951, -0.08896633982658386, 0.06805598735809326, -0.0694287046790123, -0.034286268055438995, 0.11131711304187775, 0.09142138808965683, 0.04694420471787453, -0.061775729060173035, -0.10200117528438568, -0.06973980367183685, 0.18613389134407043, -0.1557258516550064, -0.08185620605945587 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # ViTGPT2_VW This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.0771 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - distributed_type: multi-GPU - num_devices: 2 - total_train_batch_size: 4 - total_eval_batch_size: 4 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 0.1256 | 0.03 | 1000 | 0.0928 | | 0.0947 | 0.07 | 2000 | 0.0897 | | 0.0889 | 0.1 | 3000 | 0.0859 | | 0.0888 | 0.14 | 4000 | 0.0842 | | 0.0866 | 0.17 | 5000 | 0.0831 | | 0.0852 | 0.2 | 6000 | 0.0819 | | 0.0833 | 0.24 | 7000 | 0.0810 | | 0.0835 | 0.27 | 8000 | 0.0802 | | 0.081 | 0.31 | 9000 | 0.0796 | | 0.0803 | 0.34 | 10000 | 0.0789 | | 0.0814 | 0.38 | 11000 | 0.0785 | | 0.0799 | 0.41 | 12000 | 0.0780 | | 0.0786 | 0.44 | 13000 | 0.0776 | | 0.0796 | 0.48 | 14000 | 0.0771 | ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.2+cu113 - Datasets 1.18.3 - Tokenizers 0.11.0
{"tags": ["generated_from_trainer"], "model-index": [{"name": "ViTGPT2_VW", "results": []}]}
null
gagan3012/ViTGPT2_VW
[ "transformers", "pytorch", "vision-encoder-decoder", "generated_from_trainer", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #vision-encoder-decoder #generated_from_trainer #endpoints_compatible #region-us
ViTGPT2\_VW =========== This model is a fine-tuned version of [](URL on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 0.0771 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 2 * eval\_batch\_size: 2 * seed: 42 * distributed\_type: multi-GPU * num\_devices: 2 * total\_train\_batch\_size: 4 * total\_eval\_batch\_size: 4 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 1.0 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.16.2 * Pytorch 1.10.2+cu113 * Datasets 1.18.3 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 2\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 2\n* total\\_train\\_batch\\_size: 4\n* total\\_eval\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.2+cu113\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #vision-encoder-decoder #generated_from_trainer #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 2\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 2\n* total\\_train\\_batch\\_size: 4\n* total\\_eval\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.2+cu113\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ 36, 162, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #vision-encoder-decoder #generated_from_trainer #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 2\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 2\n* total\\_train\\_batch\\_size: 4\n* total\\_eval\\_batch\\_size: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.2+cu113\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ -0.12251439690589905, 0.09916622936725616, -0.0023411728907376528, 0.0996495932340622, 0.15530121326446533, 0.04797770455479622, 0.11069605499505997, 0.12950323522090912, -0.09264809638261795, 0.07918142527341843, 0.1236235648393631, 0.08457641303539276, 0.054743699729442596, 0.15180200338363647, -0.02558768354356289, -0.2665317952632904, 0.015926633030176163, 0.02164847031235695, -0.08022303134202957, 0.13337568938732147, 0.07656338065862656, -0.12636731564998627, 0.06755586713552475, -0.016855362802743912, -0.1886378675699234, -0.029028555378317833, -0.027584649622440338, -0.02077995054423809, 0.12801487743854523, 0.034569934010505676, 0.12153388559818268, 0.004845323972404003, 0.08867481350898743, -0.21997101604938507, -0.0028380178846418858, 0.06184002012014389, 0.009768957272171974, 0.06622921675443649, 0.07333870232105255, 0.004192512482404709, 0.07668593525886536, -0.12425185739994049, 0.053991977125406265, 0.007713899482041597, -0.144813671708107, -0.24969957768917084, -0.08738578855991364, 0.03356972336769104, 0.08208952099084854, 0.0981917753815651, -0.019710231572389603, 0.092440165579319, -0.07845088839530945, 0.0810246616601944, 0.24001534283161163, -0.2375209480524063, -0.0721760243177414, 0.012455152347683907, 0.002027589362114668, 0.0853595957159996, -0.1046769842505455, -0.0410466343164444, 0.04370402172207832, 0.05568133294582367, 0.09844411164522171, 0.00270130205899477, -0.037446968257427216, -0.0017249686643481255, -0.15290436148643494, -0.09016755223274231, 0.13715647161006927, 0.05117296054959297, -0.02461208775639534, -0.046376969665288925, -0.062409013509750366, -0.22858285903930664, -0.028454704210162163, 0.026911726221442223, 0.03193457052111626, -0.0474746897816658, -0.0898624137043953, 0.03356451541185379, -0.09688498079776764, -0.062155384570360184, -0.02096451446413994, 0.12195891886949539, 0.05137094482779503, 0.009512179531157017, -0.0034453398548066616, 0.11610212922096252, 0.014048948884010315, -0.1579311043024063, 0.0016831314424052835, 0.01935691572725773, -0.06562421470880508, -0.026793621480464935, -0.05785099044442177, 0.010563008487224579, -0.014395195990800858, 0.12395401298999786, -0.04061754420399666, 0.05747894570231438, 0.01892448589205742, 0.02913963608443737, -0.074625663459301, 0.1869034767150879, -0.08220592886209488, -0.002035437850281596, -0.03308126702904701, 0.08590841293334961, 0.0010399175807833672, -0.016454268246889114, -0.08323006331920624, 0.004668539389967918, 0.1329573690891266, 0.030018480494618416, -0.038451697677373886, 0.04234856739640236, -0.05452471226453781, -0.031753942370414734, 0.010507766157388687, -0.1006346195936203, 0.028056826442480087, 0.008380151353776455, -0.06625020503997803, -0.0023487426806241274, 0.0045588198117911816, -0.007532342802733183, -0.019430063664913177, 0.10619892925024033, -0.09306883066892624, 0.016023602336645126, -0.07923875749111176, -0.11163847148418427, 0.029440460726618767, -0.04927508533000946, 0.008544973097741604, -0.07565393298864365, -0.12785863876342773, -0.02737891487777233, 0.04539630562067032, -0.05184626206755638, -0.061350900679826736, -0.04988035932183266, -0.08526763319969177, 0.02653481811285019, -0.018526460975408554, 0.13708829879760742, -0.05757829546928406, 0.11835326254367828, 0.047176674008369446, 0.05834534019231796, -0.00326257455162704, 0.05206954851746559, -0.06933204084634781, 0.05574469640851021, -0.17228107154369354, 0.08347124606370926, -0.049212269484996796, 0.03461861237883568, -0.10272589325904846, -0.1275627315044403, 0.021963810548186302, -0.03014044649899006, 0.10274989157915115, 0.10740923136472702, -0.19421853125095367, -0.06982764601707458, 0.17640581727027893, -0.0882268026471138, -0.11636863648891449, 0.13724321126937866, -0.04367471858859062, -0.04933122545480728, 0.034524064511060715, 0.15409918129444122, 0.08980213105678558, -0.07958129048347473, -0.02592073753476143, -0.03920191153883934, 0.10243228077888489, -0.030741218477487564, 0.1159747987985611, 0.019065259024500847, 0.0399378202855587, 0.0043558659963309765, -0.068541020154953, 0.07259555906057358, -0.11806141585111618, -0.09506692737340927, -0.029188551008701324, -0.07637850940227509, 0.027943795546889305, 0.0674055889248848, 0.04170048609375954, -0.09314791113138199, -0.09133046865463257, 0.02409449592232704, 0.09667002409696579, -0.09606015682220459, 0.028989041224122047, -0.04982319846749306, 0.07223822176456451, -0.0943087488412857, -0.02068561501801014, -0.18401964008808136, -0.08532948046922684, 0.022154858335852623, -0.009941521100699902, -0.007845339365303516, -0.004781965631991625, 0.0629601702094078, 0.08283894509077072, -0.05564277619123459, -0.05650509148836136, -0.07144461572170258, -0.016760654747486115, -0.10370149463415146, -0.19313479959964752, -0.07936453819274902, -0.013551843352615833, 0.15633487701416016, -0.20977677404880524, 0.005029740277677774, 0.018067017197608948, 0.10776377469301224, 0.00033329619327560067, -0.043535590171813965, -0.04020014405250549, 0.08129677176475525, -0.0314587764441967, -0.07539103180170059, 0.045050233602523804, -0.005901559256017208, -0.08209545165300369, -0.011291708797216415, -0.12032880634069443, 0.11215943843126297, 0.1152489185333252, -0.05539291724562645, -0.08426502346992493, 0.001884737517684698, -0.07801076769828796, -0.04682786762714386, -0.03780081495642662, 0.0023053965996950865, 0.15892678499221802, 0.014904404990375042, 0.1278979778289795, -0.0655861347913742, -0.038035858422517776, 0.04173824191093445, -0.0019053168362006545, -0.013711366802453995, 0.13069719076156616, 0.07887811213731766, -0.0930061787366867, 0.14458823204040527, 0.10653918236494064, -0.04912004992365837, 0.09884874522686005, -0.05747147649526596, -0.11078812181949615, -0.04820779711008072, 0.0012452565133571625, 0.021748045459389687, 0.13444679975509644, -0.1056208685040474, -0.001632557949051261, 0.023403886705636978, 0.03636133298277855, 0.028619710355997086, -0.2102440595626831, -0.03264869377017021, 0.043820880353450775, -0.05080858990550041, -0.0042770542204380035, -0.03812588006258011, 0.01758374460041523, 0.10280124098062515, 0.020942991599440575, -0.05008058249950409, 0.011251760646700859, -0.020063500851392746, -0.08627726137638092, 0.2132318913936615, -0.07129190862178802, -0.16253243386745453, -0.126630499958992, -0.06765244156122208, -0.026568571105599403, 0.0018656979082152247, 0.04292136803269386, -0.08514895290136337, -0.026079656556248665, -0.0635705441236496, 0.02529112435877323, -0.045438144356012344, 0.041797786951065063, 0.019751116633415222, -0.016625406220555305, 0.07217756658792496, -0.08208706229925156, 0.0026203589513897896, -0.018168095499277115, -0.04151434451341629, 0.0555286668241024, 0.022271249443292618, 0.11958900094032288, 0.14793913066387177, 0.003921166528016329, 0.03204047679901123, -0.0234503373503685, 0.24524739384651184, -0.08078975975513458, -0.05776062607765198, 0.08172056823968887, -0.019661979749798775, 0.06122834235429764, 0.1405048817396164, 0.04742082208395004, -0.09522053599357605, 0.0008151039364747703, 0.033325329422950745, -0.03254346922039986, -0.20043756067752838, -0.050354961305856705, -0.06857074797153473, -0.010464079678058624, 0.12341759353876114, 0.012208066880702972, -0.03433416038751602, 0.05017200484871864, -0.0017064353451132774, 0.06505525857210159, -0.04894271865487099, 0.07219775021076202, 0.0948951467871666, 0.03928837552666664, 0.10480908304452896, -0.029514500871300697, -0.0509830117225647, 0.03969232738018036, -0.010591624304652214, 0.2535244822502136, -0.039520129561424255, 0.153542622923851, 0.04603062942624092, 0.17335332930088043, 0.006045271176844835, 0.06594441831111908, 0.006786929909139872, -0.011592849157750607, -0.0059206304140388966, -0.042678821831941605, -0.055261850357055664, 0.01567978225648403, 0.018416263163089752, 0.03525933623313904, -0.14720414578914642, -0.018210191279649734, 0.038279302418231964, 0.27505841851234436, 0.10245899111032486, -0.34736087918281555, -0.09062184393405914, 0.006756848189979792, -0.012744961306452751, -0.02713584341108799, -0.012397919781506062, 0.13518399000167847, -0.0945989266037941, 0.04019727557897568, -0.07008739560842514, 0.07972627878189087, -0.08860565721988678, 0.014386115595698357, 0.08260740339756012, 0.052340857684612274, 0.007716692518442869, 0.06960485130548477, -0.25393787026405334, 0.29858633875846863, -0.006922123488038778, 0.03835838660597801, -0.060476113110780716, -0.003481233259662986, 0.008498718962073326, -0.0031942385248839855, 0.06058713048696518, 0.005927641410380602, -0.04669129103422165, -0.2243155986070633, -0.09594940394163132, 0.0346953347325325, 0.11725807189941406, -0.055593255907297134, 0.13616733253002167, -0.02070593647658825, 0.0034043879713863134, 0.050677359104156494, 0.030131401494145393, -0.025299474596977234, -0.09967003017663956, 0.028354963287711143, -0.012156790122389793, 0.005058609880506992, -0.07370374351739883, -0.12289289385080338, -0.07947050034999847, 0.16745217144489288, -0.016640111804008484, -0.02974952757358551, -0.12065959721803665, 0.12754061818122864, 0.14466722309589386, -0.09451183676719666, 0.04013130068778992, 0.02052207663655281, 0.11322608590126038, 0.029026392847299576, -0.0561615414917469, 0.08528977632522583, -0.05194781720638275, -0.1796826869249344, -0.05040069669485092, 0.12717458605766296, 0.034222546964883804, 0.06368899345397949, -0.034986015409231186, 0.030183428898453712, -0.012109042145311832, -0.0822860598564148, 0.03899242356419563, 0.011482811532914639, 0.07219237834215164, 0.07431124150753021, -0.012350298464298248, 0.024368172511458397, -0.06215687841176987, -0.021152477711439133, 0.1692902147769928, 0.2845449447631836, -0.0865652933716774, -0.018973907455801964, 0.02089780755341053, -0.05078372359275818, -0.15845762193202972, 0.02293221652507782, 0.10091263800859451, 0.023484185338020325, 0.01445908285677433, -0.17115330696105957, 0.061809979379177094, 0.10769974440336227, -0.00667753629386425, 0.0898711308836937, -0.33874258399009705, -0.1158987283706665, 0.0663943961262703, 0.13290560245513916, 0.0484163723886013, -0.15641586482524872, -0.039777692407369614, -0.003972921054810286, -0.16459332406520844, 0.10280197858810425, 0.004431595094501972, 0.13584399223327637, -0.028181148692965508, 0.07067664712667465, 0.016364196315407753, -0.0680905133485794, 0.16178829967975616, 0.01113792136311531, 0.06251289695501328, -0.04454108700156212, -0.016755424439907074, 0.06497697532176971, -0.06561014801263809, -0.0022434864658862352, -0.020630599930882454, 0.03933992609381676, -0.12846525013446808, -0.021059034392237663, -0.09574318677186966, -0.021896967664361, -0.038145117461681366, -0.04359295219182968, -0.03784118965268135, 0.05656373128294945, 0.08585984259843826, -0.009472888894379139, 0.1102442592382431, 0.02513759210705757, 0.10829844325780869, 0.04981643706560135, 0.03671178221702576, -0.005350038409233093, -0.10938464850187302, -0.02818959206342697, -0.0033562721218913794, 0.03705417737364769, -0.1199483722448349, 0.02703663520514965, 0.16914790868759155, 0.032411377876996994, 0.14764411747455597, 0.08166136592626572, -0.05329586938023567, 0.01905968226492405, 0.057728491723537445, -0.13895583152770996, -0.12454518675804138, -0.002573329024016857, -0.028029263019561768, -0.15583322942256927, 0.011548416689038277, 0.09630101174116135, -0.06138850003480911, -0.013749903999269009, -0.016178550198674202, 0.026847288012504578, -0.0511900931596756, 0.21403998136520386, 0.034757085144519806, 0.07333226501941681, -0.11050691455602646, 0.0690675675868988, 0.07170256972312927, -0.12892648577690125, 0.016821760684251785, 0.07770714908838272, -0.0655631348490715, -0.016456173732876778, 0.07289235293865204, 0.11148987710475922, -0.047932013869285583, -0.021229716017842293, -0.11619560420513153, -0.12347889691591263, 0.09948161244392395, 0.12727341055870056, 0.0833607167005539, 0.05132799223065376, -0.021218042820692062, 0.008670692332088947, -0.12592367827892303, 0.09098827093839645, 0.06458643823862076, 0.07852527499198914, -0.1371850073337555, 0.15633675456047058, -0.006804528180509806, 0.060290779918432236, -0.007312323432415724, 0.013517936691641808, -0.11051958799362183, 0.021582290530204773, -0.1240251213312149, 0.0010072875302284956, -0.031800925731658936, -0.004604925401508808, -0.008527391590178013, -0.0626339241862297, -0.05999903380870819, 0.01361649576574564, -0.10487506538629532, -0.056766100227832794, -0.0005986005999147892, 0.04162417724728584, -0.11894498765468597, -0.04140691086649895, 0.03128756582736969, -0.10116709768772125, 0.08548048883676529, 0.054230980575084686, 0.039342962205410004, 0.024587923660874367, -0.07540673017501831, -0.015487810596823692, 0.051080018281936646, 0.00440580677241087, 0.06569569557905197, -0.13350576162338257, -0.0030168548692017794, -0.013845820911228657, 0.02903401292860508, 0.02528490126132965, 0.07235245406627655, -0.13021941483020782, -0.007910158485174179, -0.027357520535588264, -0.05095900222659111, -0.0560540072619915, 0.044574297964572906, 0.05670572444796562, 0.04889317601919174, 0.14947360754013062, -0.07071121782064438, 0.06795241683721542, -0.2386290431022644, -0.026812899857759476, -0.023646635934710503, -0.08218798041343689, -0.0408148430287838, -0.042884018272161484, 0.08175414800643921, -0.05936846137046814, 0.11658347398042679, -0.031170589849352837, 0.06840861588716507, 0.02763024903833866, -0.016657864674925804, 0.023285560309886932, 0.047225888818502426, 0.2078157365322113, 0.028885243460536003, -0.05720077455043793, 0.08402791619300842, 0.04276612028479576, 0.09099084138870239, 0.17627952992916107, 0.1904752403497696, 0.13146694004535675, 0.03577229008078575, 0.07278656959533691, 0.04340258240699768, -0.09062743932008743, -0.145943284034729, 0.027566038072109222, -0.05469575151801109, 0.10592775046825409, -0.003944588825106621, 0.19495128095149994, 0.06116345897316933, -0.1987631767988205, 0.05164531245827675, -0.0511578693985939, -0.09798498451709747, -0.06807927042245865, -0.044199176132678986, -0.08173774182796478, -0.1211671233177185, 0.005496093071997166, -0.12335243076086044, 0.02847905270755291, 0.16461162269115448, 0.02395053766667843, 0.006786149926483631, 0.14807938039302826, 0.07497218251228333, 0.017725501209497452, 0.05363410711288452, 0.030301718041300774, 0.001891792519018054, -0.04262503609061241, -0.06142115220427513, 0.030101066455245018, -0.012430055066943169, 0.06723019480705261, -0.05098653584718704, -0.04031943529844284, 0.06298601627349854, 0.0012585645308718085, -0.11484116315841675, 0.010529277846217155, -0.0017541467677801847, 0.07551030069589615, 0.06150161847472191, 0.014976373873651028, 0.016482112929224968, -0.031175563111901283, 0.20068305730819702, -0.07377105206251144, -0.065543532371521, -0.11134965717792511, 0.23902764916419983, 0.009266585111618042, -0.029022561386227608, 0.04074721410870552, -0.08412875980138779, -0.002340464387089014, 0.18019452691078186, 0.15617120265960693, -0.07383023947477341, -0.022887548431754112, 0.03054540790617466, -0.007885991595685482, -0.003840807592496276, 0.10131926834583282, 0.12360979616641998, 0.07083261013031006, -0.11223826557397842, -0.04557124897837639, -0.06589030474424362, -0.020365292206406593, -0.000007159727374528302, 0.04039699584245682, 0.03100774623453617, 0.021190423518419266, -0.05888185277581215, 0.06629963219165802, -0.06045539304614067, -0.07358189672231674, 0.08493821322917938, -0.23001039028167725, -0.18936313688755035, -0.018810976296663284, 0.07784600555896759, 0.006133981980383396, 0.05955879017710686, -0.029688281938433647, -0.008437755517661572, 0.08207231014966965, -0.010017127729952335, -0.07215309888124466, -0.09921939671039581, 0.08198085427284241, -0.0731777474284172, 0.19260567426681519, -0.03856812044978142, 0.05857310816645622, 0.12872810661792755, 0.06884189695119858, -0.10543415695428848, 0.02719254419207573, 0.06378983706235886, -0.07012084126472473, 0.020299531519412994, 0.1565457135438919, -0.05064477398991585, 0.07182825356721878, 0.04977436363697052, -0.1251337081193924, 0.006649251561611891, -0.03384627029299736, -0.002139213727787137, -0.033681489527225494, -0.06383845955133438, -0.051540546119213104, 0.1460696905851364, 0.22710198163986206, -0.05187398940324783, -0.00639924593269825, -0.06730285286903381, -0.009763404726982117, 0.05007696524262428, 0.1267322450876236, -0.04237567260861397, -0.25369298458099365, 0.008844586089253426, 0.036650802940130234, 0.005355671513825655, -0.22475765645503998, -0.10182204097509384, 0.050533972680568695, -0.07531111687421799, -0.06760463118553162, 0.13151362538337708, 0.0419570729136467, 0.06435772776603699, -0.0690048560500145, -0.07093166559934616, -0.0738760232925415, 0.1638697236776352, -0.15159335732460022, -0.09718494862318039 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # ViTGPT2_vizwiz This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.0719 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - distributed_type: multi-GPU - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 0.1207 | 0.07 | 1000 | 0.0906 | | 0.0916 | 0.14 | 2000 | 0.0861 | | 0.0879 | 0.2 | 3000 | 0.0840 | | 0.0856 | 0.27 | 4000 | 0.0822 | | 0.0834 | 0.34 | 5000 | 0.0806 | | 0.0817 | 0.41 | 6000 | 0.0795 | | 0.0812 | 0.48 | 7000 | 0.0785 | | 0.0808 | 0.55 | 8000 | 0.0779 | | 0.0796 | 0.61 | 9000 | 0.0771 | | 0.0786 | 0.68 | 10000 | 0.0767 | | 0.0774 | 0.75 | 11000 | 0.0762 | | 0.0772 | 0.82 | 12000 | 0.0758 | | 0.0756 | 0.89 | 13000 | 0.0754 | | 0.0759 | 0.96 | 14000 | 0.0750 | | 0.0756 | 1.02 | 15000 | 0.0748 | | 0.0726 | 1.09 | 16000 | 0.0745 | | 0.0727 | 1.16 | 17000 | 0.0745 | | 0.0715 | 1.23 | 18000 | 0.0742 | | 0.0726 | 1.3 | 19000 | 0.0741 | | 0.072 | 1.37 | 20000 | 0.0738 | | 0.0723 | 1.43 | 21000 | 0.0735 | | 0.0715 | 1.5 | 22000 | 0.0734 | | 0.0724 | 1.57 | 23000 | 0.0732 | | 0.0723 | 1.64 | 24000 | 0.0730 | | 0.0718 | 1.71 | 25000 | 0.0729 | | 0.07 | 1.78 | 26000 | 0.0728 | | 0.0702 | 1.84 | 27000 | 0.0726 | | 0.0704 | 1.91 | 28000 | 0.0725 | | 0.0703 | 1.98 | 29000 | 0.0725 | | 0.0686 | 2.05 | 30000 | 0.0726 | | 0.0687 | 2.12 | 31000 | 0.0726 | | 0.0688 | 2.19 | 32000 | 0.0724 | | 0.0677 | 2.25 | 33000 | 0.0724 | | 0.0665 | 2.32 | 34000 | 0.0725 | | 0.0684 | 2.39 | 35000 | 0.0723 | | 0.0678 | 2.46 | 36000 | 0.0722 | | 0.0686 | 2.53 | 37000 | 0.0722 | | 0.067 | 2.59 | 38000 | 0.0721 | | 0.0669 | 2.66 | 39000 | 0.0721 | | 0.0673 | 2.73 | 40000 | 0.0721 | | 0.0673 | 2.8 | 41000 | 0.0720 | | 0.0662 | 2.87 | 42000 | 0.0720 | | 0.0681 | 2.94 | 43000 | 0.0719 | ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2.dev0 - Tokenizers 0.11.0
{"tags": ["generated_from_trainer", "image-to-text"], "model-index": [{"name": "ViTGPT2_vizwiz", "results": []}]}
image-to-text
gagan3012/ViTGPT2_vizwiz
[ "transformers", "pytorch", "vision-encoder-decoder", "generated_from_trainer", "image-to-text", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #vision-encoder-decoder #generated_from_trainer #image-to-text #endpoints_compatible #has_space #region-us
ViTGPT2\_vizwiz =============== This model is a fine-tuned version of [](URL on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 0.0719 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * distributed\_type: multi-GPU * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.17.0.dev0 * Pytorch 1.10.2+cu102 * Datasets 1.18.2.dev0 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* distributed\\_type: multi-GPU\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.2.dev0\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #vision-encoder-decoder #generated_from_trainer #image-to-text #endpoints_compatible #has_space #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* distributed\\_type: multi-GPU\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.2.dev0\n* Tokenizers 0.11.0" ]
[ 46, 124, 4, 39 ]
[ "passage: TAGS\n#transformers #pytorch #vision-encoder-decoder #generated_from_trainer #image-to-text #endpoints_compatible #has_space #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* distributed\\_type: multi-GPU\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.2.dev0\n* Tokenizers 0.11.0" ]
[ -0.12494277209043503, 0.026310095563530922, -0.0009689886355772614, 0.10820595175027847, 0.1920292228460312, 0.03716304525732994, 0.09353043884038925, 0.10481458157300949, -0.11686161905527115, 0.04860696569085121, 0.10306460410356522, 0.13101664185523987, 0.03381987288594246, 0.13447269797325134, -0.029573552310466766, -0.2893730700016022, 0.02438468299806118, 0.05888247862458229, -0.07009861618280411, 0.12907572090625763, 0.07646871358156204, -0.15750013291835785, 0.08660975843667984, -0.007995836436748505, -0.23372767865657806, -0.024045394733548164, -0.005872034002095461, -0.04287375509738922, 0.1368333399295807, 0.03223785012960434, 0.10862207412719727, -0.007195751648396254, 0.08838310092687607, -0.17553330957889557, 0.011251447722315788, 0.09190226346254349, 0.012272344902157784, 0.06367696821689606, 0.06585489958524704, 0.016385581344366074, 0.12371515482664108, -0.1274397075176239, 0.04613540694117546, -0.0015890809008851647, -0.13630105555057526, -0.24998095631599426, -0.04975935071706772, -0.006752932444214821, 0.05978160351514816, 0.10259619355201721, -0.01269228383898735, 0.10967864096164703, -0.05912031978368759, 0.10478987544775009, 0.20208972692489624, -0.2225361317396164, -0.07215879112482071, 0.009188788011670113, 0.011799626983702183, 0.098630391061306, -0.09785860776901245, 0.017644444480538368, 0.02697659470140934, 0.063432477414608, 0.12950336933135986, -0.020302681252360344, -0.1132156103849411, 0.001758650178089738, -0.15380972623825073, -0.06038239225745201, 0.0971546620130539, 0.03890667483210564, -0.021940002217888832, -0.052255112677812576, -0.09007855504751205, -0.15770705044269562, -0.0560046024620533, 0.015144958160817623, 0.03991109877824783, -0.043266016989946365, -0.10313191264867783, -0.01409975253045559, -0.09854064881801605, -0.0717478096485138, -0.0662909671664238, 0.16269180178642273, 0.03549676388502121, 0.04085657373070717, -0.014667258597910404, 0.125075563788414, -0.013291366398334503, -0.12481465190649033, -0.0006565190269611776, 0.003724014153704047, -0.04500347375869751, -0.006440374534577131, -0.08660344779491425, -0.046203263103961945, -0.03758497163653374, 0.06425976008176804, -0.07193738967180252, 0.05051734298467636, 0.03525017201900482, 0.04055788740515709, -0.08114422112703323, 0.19120211899280548, -0.10283641517162323, 0.009895858354866505, -0.0075689260847866535, 0.05309683457016945, 0.013299359939992428, -0.027249135076999664, -0.08940628170967102, -0.01712607406079769, 0.11108393222093582, -0.03198826685547829, -0.06742790341377258, 0.07866543531417847, -0.047665923833847046, -0.03425346687436104, 0.032662488520145416, -0.0740903913974762, 0.04926172271370888, -0.0041880481876432896, -0.06610839813947678, 0.011327264830470085, 0.04689616337418556, -0.0016545167891308665, -0.03838874399662018, 0.10115062445402145, -0.0736088752746582, 0.05938361957669258, -0.12750989198684692, -0.1304485946893692, 0.0024454155936837196, -0.039286889135837555, 0.011022382415831089, -0.09564878791570663, -0.15216396749019623, -0.00934879295527935, 0.06845085322856903, -0.02913275919854641, -0.007112058345228434, -0.02643844299018383, -0.07730454951524734, 0.022836942225694656, -0.02418392151594162, 0.15478405356407166, -0.06081683933734894, 0.12489151954650879, 0.04552875831723213, 0.07997436821460724, -0.0739700123667717, 0.055795036256313324, -0.05936901271343231, 0.018706930801272392, -0.21740451455116272, 0.11540571600198746, -0.03390725329518318, 0.035241175442934036, -0.05680476874113083, -0.13967637717723846, 0.014729863964021206, 0.006616436410695314, 0.11009979248046875, 0.0870525985956192, -0.1723758727312088, -0.05739610269665718, 0.14611537754535675, -0.08177734911441803, -0.10834261029958725, 0.11818798631429672, -0.056421130895614624, -0.0233916062861681, 0.0695064440369606, 0.15757805109024048, 0.06361459940671921, -0.10027693957090378, -0.006015402730554342, -0.018812263384461403, 0.035724736750125885, -0.059395283460617065, 0.07459376007318497, 0.06992288678884506, 0.03494961932301521, 0.024938397109508514, -0.0562785379588604, 0.09035681933164597, -0.11440732330083847, -0.0825885683298111, -0.05102337524294853, -0.07017860561609268, 0.0016954757738858461, 0.07946452498435974, 0.06045306846499443, -0.09458564221858978, -0.09503639489412308, 0.05827046185731888, 0.07270193845033646, -0.10419029742479324, 0.04488953948020935, -0.06237737834453583, 0.04474601894617081, -0.0940198302268982, -0.006990599445998669, -0.20032371580600739, -0.03317277505993843, -0.007763129658997059, 0.013306687586009502, -0.007228757254779339, -0.008168946951627731, 0.05943601205945015, 0.08227444440126419, -0.05979059636592865, -0.03505714610219002, -0.05354256182909012, -0.01613842509686947, -0.11536405980587006, -0.15672002732753754, -0.04837972670793533, -0.016555799171328545, 0.11343293637037277, -0.18924084305763245, 0.007887161336839199, 0.046243030577898026, 0.0733683854341507, 0.00753651512786746, -0.02605326846241951, -0.04450288414955139, 0.0922357439994812, -0.018943654373288155, -0.08277372270822525, 0.07643694430589676, 0.015076125971972942, -0.0505678616464138, -0.008272085338830948, -0.1308189332485199, 0.15032193064689636, 0.13893915712833405, -0.16673219203948975, -0.05300362780690193, 0.022802263498306274, -0.05438200384378433, -0.03135185316205025, -0.04173818975687027, 0.015971725806593895, 0.1804656684398651, -0.009169220924377441, 0.13420502841472626, -0.06430370360612869, -0.0029171507339924574, 0.035183873027563095, 0.007372788619250059, 0.013173487037420273, 0.09029845148324966, 0.08634712547063828, -0.06869018822908401, 0.11738228797912598, 0.10040809959173203, -0.07135140895843506, 0.12788322567939758, -0.022582249715924263, -0.08901116997003555, -0.010329988785088062, -0.036009933799505234, -0.013230212032794952, 0.14245033264160156, -0.145452082157135, -0.014035350643098354, 0.027970274910330772, 0.026361480355262756, 0.043166980147361755, -0.24408337473869324, -0.011804581619799137, 0.028990816324949265, -0.04116680845618248, 0.0032877440098673105, -0.01931406743824482, 0.011808083392679691, 0.10007275640964508, 0.0010814720299094915, -0.03899671137332916, 0.028437741100788116, -0.003686359850689769, -0.06067486107349396, 0.19573061168193817, -0.07632347196340561, -0.20595096051692963, -0.09949758648872375, -0.06355458498001099, -0.047416623681783676, 0.015551173128187656, 0.04815440624952316, -0.06779970228672028, -0.04397592693567276, -0.059504859149456024, 0.031396761536598206, -0.0017623575404286385, 0.030156081542372704, 0.01583773083984852, -0.029297547414898872, 0.06866811960935593, -0.09300976991653442, -0.004318946041166782, -0.03459545224905014, -0.0604424849152565, 0.09125927835702896, 0.058023903518915176, 0.13239561021327972, 0.14596252143383026, -0.02434677444398403, 0.03098357655107975, -0.015715615823864937, 0.26492780447006226, -0.11006037145853043, -0.029901430010795593, 0.12075851112604141, 0.02647547796368599, 0.048693928867578506, 0.09627705067396164, 0.048484738916158676, -0.10943186283111572, 0.002123554004356265, 0.02399301528930664, -0.04630998894572258, -0.14084842801094055, -0.062737837433815, -0.0711822658777237, -0.0680437907576561, 0.08282872289419174, 0.01760270819067955, -0.025456627830863, 0.07531734555959702, 0.03477165102958679, 0.07861422002315521, -0.04158366471529007, 0.058004796504974365, 0.14618359506130219, 0.036224983632564545, 0.12798412144184113, -0.04832374304533005, -0.06534592062234879, 0.04817400127649307, -0.010551751591265202, 0.23931406438350677, -0.07403358072042465, 0.0847642570734024, 0.04225512966513634, 0.14006493985652924, 0.01300507877022028, 0.07612443715333939, 0.0031276443041861057, -0.03129332512617111, -0.0242694653570652, -0.03377678990364075, -0.05859339237213135, 0.0004375968419481069, -0.04168551787734032, 0.01159663312137127, -0.10735882073640823, -0.019836416468024254, 0.05811905115842819, 0.26269808411598206, 0.06775696575641632, -0.36543819308280945, -0.0909804180264473, 0.004286071751266718, 0.005540751852095127, -0.06393671035766602, -0.0018116463907063007, 0.16010677814483643, -0.07758641242980957, 0.05546458810567856, -0.09614953398704529, 0.09380605071783066, -0.071356400847435, 0.03734482452273369, 0.0734395906329155, 0.09999437630176544, 0.012334572151303291, 0.06356927007436752, -0.28855642676353455, 0.23753365874290466, 0.002031225711107254, 0.07756450772285461, -0.07778055965900421, -0.016220560297369957, 0.0303009282797575, 0.08746316283941269, 0.027672747150063515, -0.0035705426707863808, -0.018327416852116585, -0.1993456929922104, -0.04074220731854439, 0.053437843918800354, 0.11101116985082626, 0.03185243159532547, 0.10447364300489426, -0.014144626446068287, 0.003144934307783842, 0.052792586386203766, 0.009551948867738247, -0.052631791681051254, -0.104215107858181, 0.00943074095994234, 0.03435754403471947, 0.01644926518201828, -0.057576797902584076, -0.11243001371622086, -0.10290441662073135, 0.11694817990064621, 0.049250274896621704, -0.024649566039443016, -0.12477722018957138, 0.12488248944282532, 0.0852169319987297, -0.07096148282289505, 0.04631819576025009, 0.01302863098680973, 0.0874657928943634, 0.045747626572847366, -0.0705053061246872, 0.11146284639835358, -0.04401928558945656, -0.13199491798877716, -0.027970127761363983, 0.05169728025794029, 0.02628309652209282, 0.03273272514343262, -0.03459317609667778, 0.024677354842424393, -0.04376649111509323, -0.09953989088535309, 0.03284342959523201, -0.005320059601217508, 0.07654353976249695, 0.08665888011455536, -0.010100184008479118, 0.022299904376268387, -0.0359540730714798, 0.0003420534485485405, 0.15340971946716309, 0.19341091811656952, -0.07852136343717575, -0.03537321835756302, 0.022980201989412308, -0.04156188666820526, -0.2180757373571396, 0.07270293682813644, 0.06160740926861763, 0.030841127038002014, 0.0027248403057456017, -0.16900238394737244, 0.08052514493465424, 0.08129680901765823, -0.00743471086025238, 0.1411304622888565, -0.2956026792526245, -0.1206708699464798, 0.05381237715482712, 0.1828778237104416, 0.10826846957206726, -0.14834515750408173, 0.016559582203626633, -0.03065202385187149, -0.15638089179992676, 0.12118184566497803, -0.04821179434657097, 0.1537473350763321, -0.02321532741189003, 0.0848446860909462, 0.004245354328304529, -0.06699661910533905, 0.1420827955007553, -0.0109712490811944, 0.10872060060501099, -0.05700921267271042, 0.03173993527889252, 0.11749666184186935, -0.048953667283058167, -0.0010137754725292325, -0.015523805283010006, 0.044734347611665726, -0.08718498796224594, -0.024149464443325996, -0.10165191441774368, -0.01931987889111042, -0.00013865633809473366, -0.049401119351387024, -0.04418979957699776, 0.041733767837285995, 0.05773337557911873, -0.007695488631725311, 0.10359921306371689, 0.000034746630262816325, 0.05360478535294533, 0.05000731721520424, 0.03124937042593956, -0.093571737408638, -0.08499875664710999, -0.024018464609980583, 0.005384268704801798, 0.07297676056623459, -0.14990296959877014, 0.03500878065824509, 0.14456425607204437, 0.007803905755281448, 0.13451644778251648, 0.09200853109359741, -0.023483533412218094, 0.04355725646018982, 0.06435055285692215, -0.14040043950080872, -0.13440978527069092, -0.003973798826336861, -0.048162370920181274, -0.06180749461054802, 0.04833446815609932, 0.10123687237501144, -0.06680183112621307, 0.011380610056221485, -0.030109677463769913, 0.0031508116517215967, -0.07130920141935349, 0.194069966673851, 0.03560537472367287, 0.03559881076216698, -0.11794916540384293, 0.06921915709972382, 0.031918469816446304, -0.1346484273672104, 0.0030816730577498674, 0.07232267409563065, -0.07949928194284439, -0.03926688805222511, 0.06806682795286179, 0.13434162735939026, -0.047788310796022415, -0.045578476041555405, -0.10464970022439957, -0.11674274504184723, 0.0905347689986229, 0.09653746336698532, 0.0944940373301506, 0.015028820373117924, -0.058031272143125534, 0.02048913575708866, -0.13851948082447052, 0.055947329849004745, 0.03467453643679619, 0.07293420284986496, -0.14698627591133118, 0.16265583038330078, 0.016319308429956436, 0.07166329771280289, -0.04310596361756325, 0.0012942575849592686, -0.08732854574918747, 0.031237145885825157, -0.15967817604541779, -0.01954449526965618, -0.02366436831653118, 0.0070799002423882484, -0.01628868281841278, -0.06056416407227516, -0.07869582623243332, 0.01564396545290947, -0.11502885818481445, -0.04113486409187317, 0.017293963581323624, 0.01413331925868988, -0.09607386589050293, -0.054426826536655426, 0.010783434845507145, -0.07639428973197937, 0.0773816853761673, 0.057211775332689285, -0.011503185145556927, 0.028793124482035637, -0.12130961567163467, -0.04298662766814232, 0.08623222261667252, 0.011295684613287449, 0.05344116687774658, -0.05055122822523117, 0.01858999766409397, 0.00991449598222971, 0.08085218071937561, 0.010581012815237045, 0.09835116565227509, -0.10628022253513336, 0.008814655244350433, -0.06361236423254013, -0.046919357031583786, -0.05691614747047424, 0.06818153709173203, 0.046034447848796844, 0.06605338305234909, 0.1476273089647293, -0.07827258110046387, 0.01576950028538704, -0.21783505380153656, -0.016975615173578262, -0.013147291727364063, -0.11696670204401016, -0.03975831717252731, -0.06245347857475281, 0.0951748788356781, -0.03646048530936241, 0.15617512166500092, 0.03637285158038139, 0.058413583785295486, 0.014373058453202248, 0.0062142289243638515, -0.012918289750814438, 0.033142250031232834, 0.21174414455890656, 0.02442595176398754, -0.03104740008711815, 0.10843613743782043, 0.08587794005870819, 0.13501156866550446, 0.2147967517375946, 0.15606440603733063, 0.1572536677122116, 0.019537441432476044, 0.11308766156435013, 0.009029948152601719, -0.06353544443845749, -0.15270549058914185, 0.03303837031126022, -0.08989130705595016, 0.11749924719333649, -0.06126093491911888, 0.1398305743932724, 0.04183609038591385, -0.15828078985214233, 0.0604419969022274, -0.03906172886490822, -0.09721647202968597, -0.0583515428006649, -0.03259577974677086, -0.08201627433300018, -0.1609075963497162, 0.009579875506460667, -0.10479558259248734, 0.032175932079553604, 0.17536166310310364, 0.0170004703104496, -0.02163800038397312, 0.1846541464328766, 0.05652373284101486, 0.016518572345376015, 0.0857788547873497, 0.00550204748287797, 0.003575144335627556, -0.04041239619255066, -0.06697878986597061, 0.03098268248140812, 0.0009939565788954496, 0.08139928430318832, -0.03850659728050232, -0.07107330858707428, 0.06758321076631546, -0.010966107249259949, -0.09283819049596786, 0.004240270704030991, 0.034354403614997864, 0.06149957701563835, 0.05309000238776207, 0.005047667771577835, 0.0023296980652958155, -0.02450558729469776, 0.21443094313144684, -0.07869744300842285, -0.08506827056407928, -0.10458604991436005, 0.18355773389339447, 0.013570541515946388, -0.0064864372834563255, 0.030572067946195602, -0.07132534682750702, -0.028701581060886383, 0.20020191371440887, 0.16541528701782227, -0.14987188577651978, -0.0243880245834589, 0.007366749923676252, -0.011237573809921741, -0.020255252718925476, 0.14623388648033142, 0.12659204006195068, 0.06796491891145706, -0.10758452117443085, -0.058492932468652725, -0.056008122861385345, -0.018392415717244148, -0.021885523572564125, 0.0067279464565217495, 0.03317447751760483, 0.011550875380635262, -0.06806915998458862, 0.04647505283355713, -0.03641003370285034, -0.06614915281534195, 0.10206381976604462, -0.20046143233776093, -0.15897595882415771, -0.002950820606201887, 0.06579803675413132, -0.00959545187652111, 0.05452612042427063, -0.047984104603528976, 0.019050640985369682, 0.07456211000680923, -0.023506516590714455, -0.06630677729845047, -0.05903437361121178, 0.07463918626308441, -0.11865345388650894, 0.17816130816936493, -0.04781348630785942, 0.05071273073554039, 0.12613855302333832, 0.06673602014780045, -0.0861847996711731, 0.07374857366085052, -0.0036660332698374987, -0.08765081316232681, 0.02044857107102871, 0.12842106819152832, -0.05533406883478165, 0.004442690871655941, 0.03227713704109192, -0.1276206374168396, 0.019670607522130013, -0.05583181604743004, -0.01443337183445692, -0.029398856684565544, -0.09960617125034332, -0.06908657401800156, 0.12018339335918427, 0.19773904979228973, -0.01141440961509943, 0.021641086786985397, -0.08368663489818573, 0.0039240517653524876, 0.0342940092086792, 0.10045864433050156, -0.06451576948165894, -0.23320266604423523, 0.0013555187033489347, 0.054819148033857346, -0.02525850385427475, -0.21291527152061462, -0.09086013585329056, 0.02636619471013546, -0.08004269003868103, -0.06443962454795837, 0.12193279713392258, 0.049625564366579056, 0.03813020884990692, -0.06664661318063736, -0.1048392802476883, -0.07168464362621307, 0.16480225324630737, -0.14239005744457245, -0.08785706758499146 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-tiny-finetuned-ner This model is a fine-tuned version of [prajjwal1/bert-tiny](https://huggingface.co/prajjwal1/bert-tiny) on the conll2003 dataset. It achieves the following results on the evaluation set: - Loss: 0.1689 - Precision: 0.8083 - Recall: 0.8274 - F1: 0.8177 - Accuracy: 0.9598 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.0355 | 1.0 | 878 | 0.1692 | 0.8072 | 0.8248 | 0.8159 | 0.9594 | | 0.0411 | 2.0 | 1756 | 0.1678 | 0.8101 | 0.8277 | 0.8188 | 0.9600 | | 0.0386 | 3.0 | 2634 | 0.1697 | 0.8103 | 0.8269 | 0.8186 | 0.9599 | | 0.0373 | 4.0 | 3512 | 0.1694 | 0.8106 | 0.8263 | 0.8183 | 0.9600 | | 0.0383 | 5.0 | 4390 | 0.1689 | 0.8083 | 0.8274 | 0.8177 | 0.9598 | ### Framework versions - Transformers 4.10.0 - Pytorch 1.9.0+cu102 - Datasets 1.11.0 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["conll2003"], "metrics": ["precision", "recall", "f1", "accuracy"], "model-index": [{"name": "bert-tiny-finetuned-ner", "results": [{"task": {"type": "token-classification", "name": "Token Classification"}, "dataset": {"name": "conll2003", "type": "conll2003", "args": "conll2003"}, "metrics": [{"type": "precision", "value": 0.8083060109289617, "name": "Precision"}, {"type": "recall", "value": 0.8273856136033113, "name": "Recall"}, {"type": "f1", "value": 0.8177345348001547, "name": "F1"}, {"type": "accuracy", "value": 0.9597597979252387, "name": "Accuracy"}]}]}]}
token-classification
gagan3012/bert-tiny-finetuned-ner
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "generated_from_trainer", "dataset:conll2003", "model-index", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #bert #token-classification #generated_from_trainer #dataset-conll2003 #model-index #autotrain_compatible #endpoints_compatible #has_space #region-us
bert-tiny-finetuned-ner ======================= This model is a fine-tuned version of prajjwal1/bert-tiny on the conll2003 dataset. It achieves the following results on the evaluation set: * Loss: 0.1689 * Precision: 0.8083 * Recall: 0.8274 * F1: 0.8177 * Accuracy: 0.9598 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 5 ### Training results ### Framework versions * Transformers 4.10.0 * Pytorch 1.9.0+cu102 * Datasets 1.11.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.10.0\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #token-classification #generated_from_trainer #dataset-conll2003 #model-index #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.10.0\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
[ 63, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #token-classification #generated_from_trainer #dataset-conll2003 #model-index #autotrain_compatible #endpoints_compatible #has_space #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.10.0\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
[ -0.10851065069437027, 0.09310617297887802, -0.001412251265719533, 0.117517851293087, 0.16889190673828125, 0.04091661050915718, 0.10526427626609802, 0.11664805561304092, -0.10379039496183395, 0.040877580642700195, 0.136621356010437, 0.14917369186878204, -0.0036566026974469423, 0.11660538613796234, -0.05544554442167282, -0.26446211338043213, -0.007292426656931639, 0.056602802127599716, -0.09542742371559143, 0.12092223763465881, 0.09102808684110641, -0.15461155772209167, 0.0974874198436737, 0.004473096691071987, -0.22943924367427826, 0.01544885616749525, 0.02598424442112446, -0.05801746994256973, 0.13800238072872162, 0.02597144991159439, 0.16156379878520966, 0.00011968720355071127, 0.09269753098487854, -0.14746850728988647, 0.009431312792003155, 0.056484367698431015, 0.0089796744287014, 0.09894203394651413, 0.054574429988861084, -0.006927283946424723, 0.11282449215650558, -0.09998921304941177, 0.06032496690750122, 0.005306342151015997, -0.1315913200378418, -0.20095469057559967, -0.07051599025726318, 0.02006491646170616, 0.055883705615997314, 0.09076684713363647, -0.012328884564340115, 0.15436610579490662, -0.0887133926153183, 0.0935465469956398, 0.21535035967826843, -0.28615811467170715, -0.0806771069765091, 0.06940582394599915, 0.0259799025952816, 0.060594592243433, -0.11363501101732254, -0.013319210149347782, 0.06349534541368484, 0.04337430000305176, 0.1389639675617218, -0.04290880262851715, -0.10392977297306061, 0.03176910802721977, -0.15028856694698334, -0.010631218552589417, 0.14588582515716553, 0.035179734230041504, -0.017812855541706085, -0.031040480360388756, -0.062007494270801544, -0.16443437337875366, -0.028956154361367226, -0.029097484424710274, 0.04213835299015045, -0.04936312139034271, -0.10484380275011063, -0.007494967896491289, -0.11323652416467667, -0.06268981099128723, -0.08021875470876694, 0.16529430449008942, 0.02624213509261608, 0.00761394901201129, -0.027380380779504776, 0.1016533151268959, -0.017381779849529266, -0.11964582651853561, 0.0280102901160717, 0.026443565264344215, -0.032129187136888504, -0.07457926869392395, -0.07340732216835022, -0.07647908478975296, -0.001775235403329134, 0.10346588492393494, -0.052893225103616714, 0.0395694226026535, 0.0682888776063919, 0.033519137650728226, -0.06882523000240326, 0.20981058478355408, -0.06895705312490463, -0.03117731399834156, -0.015932640060782433, 0.04862901195883751, -0.008576195687055588, -0.008385146036744118, -0.1191345825791359, -0.009353508241474628, 0.11945652216672897, -0.011324266903102398, -0.0746125802397728, 0.0844932273030281, -0.0485830195248127, -0.023878376930952072, -0.011892174370586872, -0.08385151624679565, 0.0525968074798584, -0.010531365871429443, -0.08497463911771774, -0.01680188998579979, 0.015650829300284386, 0.014346067793667316, 0.008207064121961594, 0.11654838919639587, -0.11436958611011505, 0.028614552691578865, -0.10094019770622253, -0.13494138419628143, 0.0032479276414960623, -0.09664890170097351, 0.03525485843420029, -0.10108695179224014, -0.1314050257205963, -0.020305931568145752, 0.05479036271572113, -0.0336579903960228, -0.03943055123090744, -0.04882042482495308, -0.08235012739896774, 0.01565122976899147, -0.00010606464638840407, 0.1383436620235443, -0.05632685497403145, 0.08953659236431122, 0.036301545798778534, 0.08115644007921219, -0.04910840466618538, 0.049775972962379456, -0.08672000467777252, 0.018116675317287445, -0.17943130433559418, 0.03996224328875542, -0.0485551692545414, 0.06076303496956825, -0.08485246449708939, -0.10324999690055847, 0.0035113489720970392, 0.013423006050288677, 0.07494992017745972, 0.08537232875823975, -0.18106265366077423, -0.0778985545039177, 0.13424761593341827, -0.049388282001018524, -0.1047324389219284, 0.10894458740949631, -0.07865729928016663, 0.05680156499147415, 0.057712309062480927, 0.1635415405035019, 0.06306229531764984, -0.07929389923810959, -0.01018932182341814, 0.0076987664215266705, 0.04540882632136345, -0.05108916386961937, 0.052612196654081345, 0.027660464867949486, 0.026229513809084892, 0.013320306316018105, -0.014941699802875519, 0.04700924828648567, -0.1062072366476059, -0.08656259626150131, -0.03343743458390236, -0.10889794677495956, 0.06644990295171738, 0.07213430106639862, 0.08953361213207245, -0.10060448199510574, -0.08897903561592102, 0.10295234620571136, 0.08244771510362625, -0.058211106806993484, 0.018135840073227882, -0.06484300643205643, 0.06506700813770294, -0.07622118294239044, -0.04068775847554207, -0.1844286024570465, -0.051980942487716675, -0.00580257922410965, 0.04141847789287567, 0.018057214096188545, 0.06476056575775146, 0.08755466341972351, 0.05926943197846413, -0.054530512541532516, -0.01753699593245983, -0.015738220885396004, 0.013595975935459137, -0.1508117914199829, -0.1945943683385849, -0.05042348429560661, -0.030002376064658165, 0.14073923230171204, -0.2349630743265152, 0.0238609928637743, 0.0014234468108043075, 0.09267852455377579, 0.02852676250040531, -0.010645754635334015, -0.03998879715800285, 0.07759372144937515, -0.05026913061738014, -0.05713275447487831, 0.05104757100343704, -0.004468596540391445, -0.08562622219324112, -0.06779976934194565, -0.12706656754016876, 0.17051726579666138, 0.13545961678028107, -0.1371810883283615, -0.10626784712076187, 0.003821637248620391, -0.053187061101198196, -0.02253914624452591, -0.045284856110811234, 0.024966081604361534, 0.16709579527378082, -0.020064057782292366, 0.14535397291183472, -0.054906729608774185, -0.048386383801698685, 0.019467389211058617, -0.03435048460960388, 0.00032899933285079896, 0.11921361833810806, 0.1127033606171608, -0.09596514701843262, 0.14695420861244202, 0.14545200765132904, -0.11215837299823761, 0.13567650318145752, -0.016399677842855453, -0.07629000395536423, -0.0328228659927845, -0.04494311660528183, -0.0015426726313307881, 0.1183319166302681, -0.12934404611587524, -0.029338102787733078, 0.019230887293815613, 0.016479061916470528, 0.023851383477449417, -0.21821120381355286, -0.042487453669309616, 0.04637179151177406, -0.008244953118264675, 0.012990890070796013, -0.012190168723464012, 0.011893801391124725, 0.10983454436063766, 0.011917205527424812, -0.08798178285360336, 0.028826141729950905, 0.008985728025436401, -0.06475771963596344, 0.20923781394958496, -0.0753529742360115, -0.1280842423439026, -0.1015564426779747, -0.06450572609901428, -0.04013390466570854, 0.020122580230236053, 0.03544406220316887, -0.09060198068618774, -0.025546886026859283, -0.06383027881383896, 0.01814015954732895, -0.02274281717836857, 0.044969331473112106, -0.005637366324663162, -0.0099213607609272, 0.0657699927687645, -0.10145746171474457, -0.004282607231289148, -0.06418060511350632, -0.07145627588033676, 0.048005517572164536, 0.035001084208488464, 0.1235254779458046, 0.15934886038303375, -0.035980336368083954, 0.012504682876169682, -0.028378136456012726, 0.25308528542518616, -0.07765321433544159, -0.020163899287581444, 0.0976046472787857, -0.013021591119468212, 0.031405746936798096, 0.11228388547897339, 0.071604885160923, -0.0857938677072525, 0.0041284565813839436, 0.043214213103055954, -0.037301961332559586, -0.18920184671878815, -0.047349169850349426, -0.0535295195877552, -0.030123502016067505, 0.0989004522562027, 0.014095211401581764, 0.04519230127334595, 0.07759270817041397, 0.05891833081841469, 0.09347666054964066, -0.06816263496875763, 0.05298265814781189, 0.09704805165529251, 0.041959501802921295, 0.13072903454303741, -0.03330061212182045, -0.08189912885427475, 0.03549032658338547, -0.00979532115161419, 0.20903562009334564, 0.0006430840003304183, 0.11205164343118668, 0.03816533461213112, 0.1661272644996643, -0.0026234977412968874, 0.0728212371468544, 0.000014426103007281199, -0.05738401040434837, -0.01299789547920227, -0.03490901738405228, -0.028267133980989456, 0.03555767610669136, -0.019433706998825073, 0.05050061270594597, -0.12679272890090942, 0.0008955094963312149, 0.055844925343990326, 0.20463833212852478, 0.0660671666264534, -0.34572601318359375, -0.09356946498155594, 0.004018361214548349, -0.02313896268606186, -0.02556699328124523, 0.014287619851529598, 0.13186295330524445, -0.07574447244405746, 0.012767969630658627, -0.06382741779088974, 0.08059363067150116, -0.05765148624777794, 0.04398515820503235, 0.0720900371670723, 0.11026260256767273, -0.009164826944470406, 0.07056176662445068, -0.26604244112968445, 0.28319159150123596, 0.01779603399336338, 0.05672617256641388, -0.06916454434394836, -0.013118606060743332, 0.026902830228209496, 0.07014062255620956, 0.053424935787916183, -0.00879045482724905, -0.052769605070352554, -0.22425968945026398, -0.033399879932403564, 0.023352455347776413, 0.09798038750886917, -0.03075386956334114, 0.10865376889705658, -0.026616597548127174, -0.0013898767065256834, 0.08839830011129379, 0.006131375208497047, -0.04071977362036705, -0.08996504545211792, -0.010538396425545216, 0.03511824086308479, -0.06097088381648064, -0.056906502693891525, -0.11672823876142502, -0.13794197142124176, 0.15848992764949799, 0.008645790629088879, -0.009795091114938259, -0.12465869635343552, 0.10938813537359238, 0.08891958743333817, -0.08112511783838272, 0.04030446335673332, 0.016929566860198975, 0.07624054700136185, 0.03986888751387596, -0.056383147835731506, 0.12144214659929276, -0.05915285646915436, -0.14367230236530304, -0.06367211043834686, 0.07915802299976349, 0.04143132269382477, 0.06552194058895111, -0.008939765393733978, 0.03394607827067375, -0.040326979011297226, -0.07563050836324692, 0.0338282585144043, -0.020686965435743332, 0.0798838883638382, 0.018643025308847427, -0.045471422374248505, 0.028433218598365784, -0.04109795391559601, -0.004374103154987097, 0.1962665468454361, 0.2246885597705841, -0.11161912232637405, -0.022659121081233025, 0.0031525674276053905, -0.06311734020709991, -0.20021916925907135, 0.08843480795621872, 0.05768256261944771, 0.021652117371559143, 0.040061332285404205, -0.16262021660804749, 0.13535867631435394, 0.09482738375663757, -0.002449475694447756, 0.09196596592664719, -0.2953207790851593, -0.12606818974018097, 0.12445911020040512, 0.13885250687599182, 0.109819196164608, -0.13430707156658173, -0.006151608657091856, -0.010064540430903435, -0.1015922799706459, 0.11086509376764297, -0.09152062982320786, 0.12304336577653885, -0.00784317683428526, 0.09877722710371017, 0.017336245626211166, -0.061996325850486755, 0.11940629035234451, 0.02170788310468197, 0.10680950433015823, -0.05100943148136139, -0.08049743622541428, 0.040836259722709656, -0.04796139523386955, 0.0008523669093847275, -0.04443255439400673, 0.02214697003364563, -0.10004464536905289, -0.01924801617860794, -0.07847951352596283, 0.04561753198504448, -0.029372116550803185, -0.07256504893302917, -0.04527580738067627, 0.03763801231980324, 0.04276851564645767, -0.01620507799088955, 0.1334780752658844, 0.013993973843753338, 0.15610621869564056, 0.10290861129760742, 0.07547213137149811, -0.06850628554821014, -0.05249202623963356, 0.015000305138528347, -0.01031446922570467, 0.0644153356552124, -0.13619323074817657, 0.031030401587486267, 0.1452377885580063, 0.0212516151368618, 0.13653132319450378, 0.09224054962396622, -0.0216926671564579, 0.00668678293004632, 0.0647168755531311, -0.1637469381093979, -0.07307752221822739, -0.007644400000572205, -0.08440171182155609, -0.11946803331375122, 0.057799506932497025, 0.10898009687662125, -0.07527831196784973, -0.0033702473156154156, -0.013350804336369038, -0.016227148473262787, -0.06421148777008057, 0.20100320875644684, 0.0790751725435257, 0.04208708181977272, -0.08713502436876297, 0.04415683075785637, 0.05364285036921501, -0.06859400123357773, 0.007891546003520489, 0.05145861580967903, -0.0686916634440422, -0.04228265956044197, 0.06721599400043488, 0.18874968588352203, -0.07138743996620178, -0.02400326356291771, -0.13441334664821625, -0.10144300013780594, 0.06760632991790771, 0.1807299703359604, 0.11744200438261032, 0.013952280394732952, -0.0548926405608654, 0.02120281010866165, -0.13203410804271698, 0.08389996737241745, 0.04872473329305649, 0.08285129815340042, -0.1602906733751297, 0.20538505911827087, -0.013176160864531994, 0.041904669255018234, -0.033649761229753494, 0.0257680993527174, -0.12414045631885529, 0.002219865331426263, -0.11620473861694336, -0.03736773878335953, -0.02806941606104374, 0.0029757919255644083, 0.006212544161826372, -0.07349809259176254, -0.07016507536172867, -0.002824201248586178, -0.12093567103147507, -0.016811329871416092, 0.041418809443712234, 0.04829418286681175, -0.10606560111045837, -0.04032998904585838, 0.018081294372677803, -0.058711323887109756, 0.06184021383523941, 0.03454452380537987, 0.029751930385828018, 0.04993724450469017, -0.12651094794273376, 0.002691007684916258, 0.06586013734340668, 0.00841064564883709, 0.09805874526500702, -0.07070374488830566, -0.005903712008148432, -0.018082594498991966, 0.08508369326591492, 0.018465561792254448, 0.0666104257106781, -0.11549877375364304, 0.0030119759030640125, -0.03699091449379921, -0.07871933281421661, -0.0681939497590065, 0.03846713900566101, 0.0904303714632988, 0.012586350552737713, 0.18220041692256927, -0.07308758050203323, 0.03236876428127289, -0.2128032147884369, 0.00012081336899427697, -0.018621094524860382, -0.11531507223844528, -0.09405719488859177, -0.06495710462331772, 0.078412726521492, -0.061857663094997406, 0.14052553474903107, 0.04606918618083, 0.04240649566054344, 0.030084265395998955, -0.015366950072348118, 0.015202861279249191, 0.037955392152071, 0.1941741704940796, 0.04197048395872116, -0.04476552456617355, 0.053848203271627426, 0.07384210079908371, 0.12408031523227692, 0.1353912651538849, 0.19538621604442596, 0.13258200883865356, -0.03948548808693886, 0.10097026079893112, 0.026780717074871063, -0.07104583084583282, -0.15095685422420502, 0.04440682381391525, -0.06823071837425232, 0.0911441445350647, -0.0359705351293087, 0.18335647881031036, 0.062314584851264954, -0.16754232347011566, 0.031042004004120827, -0.06681594252586365, -0.09486961364746094, -0.10295119136571884, -0.03572171553969383, -0.09019481390714645, -0.13061656057834625, -0.005620657466351986, -0.11658543348312378, 0.010044384747743607, 0.11703633517026901, 0.011764981783926487, -0.014762050472199917, 0.17885376513004303, 0.012319452129304409, 0.04430896043777466, 0.051909804344177246, 0.017371589317917824, -0.012667772360146046, -0.11601801216602325, -0.06146613508462906, -0.03062458708882332, -0.006087614689022303, 0.028701702132821083, -0.0793815404176712, -0.05382435396313667, 0.03259314224123955, -0.010217485949397087, -0.1045135110616684, 0.0030541361775249243, 0.01782124489545822, 0.052547160536050797, 0.0003089373931288719, 0.0035863681696355343, 0.015091609209775925, -0.027289796620607376, 0.2240452617406845, -0.09070254862308502, -0.0454566553235054, -0.1131172627210617, 0.24914687871932983, 0.029682893306016922, 0.008655634708702564, 0.0218137726187706, -0.08065266162157059, 0.004847451578825712, 0.24735882878303528, 0.21906177699565887, -0.11259223520755768, 0.0032925314735621214, 0.009483627043664455, -0.00793035514652729, -0.032472215592861176, 0.09703393280506134, 0.10416484624147415, 0.03658410534262657, -0.09909473359584808, -0.05616239830851555, -0.05268486589193344, -0.0118641909211874, -0.01924712397158146, 0.05512215197086334, 0.06511228531599045, 0.019566873088479042, -0.05319736897945404, 0.04197051376104355, -0.0602317675948143, -0.11225342750549316, 0.07361909002065659, -0.22829513251781464, -0.16444607079029083, -0.010141493752598763, 0.10112632811069489, -0.005788969341665506, 0.08474399149417877, -0.035320308059453964, -0.01310842577368021, 0.049923188984394073, -0.01663096994161606, -0.08685854077339172, -0.07746492326259613, 0.10538898408412933, -0.09351781755685806, 0.20075778663158417, -0.054789479821920395, 0.053909800946712494, 0.1354697346687317, 0.05348153039813042, -0.06822726130485535, 0.043371282517910004, 0.05355537310242653, -0.08439458906650543, 0.024852506816387177, 0.09726933389902115, -0.043493665754795074, 0.081654392182827, 0.04647965729236603, -0.16956031322479248, 0.02888507768511772, -0.08607993274927139, -0.046778518706560135, -0.0403146855533123, -0.05392315611243248, -0.05213787406682968, 0.13957753777503967, 0.2248205840587616, -0.02181561104953289, 0.004091895185410976, -0.06317225098609924, 0.0108333770185709, 0.06580552458763123, 0.0474945567548275, -0.08285938948392868, -0.2336096465587616, 0.012158900499343872, 0.03350745141506195, -0.02997819148004055, -0.2161354124546051, -0.08639595657587051, 0.008078946731984615, -0.06754406541585922, -0.07080955803394318, 0.08755079656839371, 0.09193979948759079, 0.05341792106628418, -0.06592284888029099, -0.0375513918697834, -0.07513395696878433, 0.15002623200416565, -0.1424294263124466, -0.09099958091974258 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-ner This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the conll2003 dataset. It achieves the following results on the evaluation set: - Loss: 0.0614 - Precision: 0.9274 - Recall: 0.9363 - F1: 0.9319 - Accuracy: 0.9840 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.2403 | 1.0 | 878 | 0.0701 | 0.9101 | 0.9202 | 0.9151 | 0.9805 | | 0.0508 | 2.0 | 1756 | 0.0600 | 0.9220 | 0.9350 | 0.9285 | 0.9833 | | 0.0301 | 3.0 | 2634 | 0.0614 | 0.9274 | 0.9363 | 0.9319 | 0.9840 | ### Framework versions - Transformers 4.10.2 - Pytorch 1.9.0+cu102 - Datasets 1.12.0 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["conll2003"], "metrics": ["precision", "recall", "f1", "accuracy"], "model-index": [{"name": "distilbert-base-uncased-finetuned-ner", "results": [{"task": {"type": "token-classification", "name": "Token Classification"}, "dataset": {"name": "conll2003", "type": "conll2003", "args": "conll2003"}, "metrics": [{"type": "precision", "value": 0.9274238227146815, "name": "Precision"}, {"type": "recall", "value": 0.9363463474661595, "name": "Recall"}, {"type": "f1", "value": 0.9318637274549098, "name": "F1"}, {"type": "accuracy", "value": 0.9839865283492462, "name": "Accuracy"}]}]}]}
token-classification
gagan3012/distilbert-base-uncased-finetuned-ner
[ "transformers", "pytorch", "tensorboard", "distilbert", "token-classification", "generated_from_trainer", "dataset:conll2003", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #token-classification #generated_from_trainer #dataset-conll2003 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-ner ===================================== This model is a fine-tuned version of distilbert-base-uncased on the conll2003 dataset. It achieves the following results on the evaluation set: * Loss: 0.0614 * Precision: 0.9274 * Recall: 0.9363 * F1: 0.9319 * Accuracy: 0.9840 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3 ### Training results ### Framework versions * Transformers 4.10.2 * Pytorch 1.9.0+cu102 * Datasets 1.12.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3", "### Training results", "### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #token-classification #generated_from_trainer #dataset-conll2003 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3", "### Training results", "### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.0\n* Tokenizers 0.10.3" ]
[ 69, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #token-classification #generated_from_trainer #dataset-conll2003 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.0\n* Tokenizers 0.10.3" ]
[ -0.10817645490169525, 0.11320975422859192, -0.002463343320414424, 0.1325080841779709, 0.1525668501853943, 0.030717479065060616, 0.12243320047855377, 0.11324058473110199, -0.09156515449285507, 0.025485731661319733, 0.13229374587535858, 0.16207194328308105, 0.015222606249153614, 0.11622914671897888, -0.05286842957139015, -0.24673697352409363, -0.003482274943962693, 0.04362938925623894, -0.04988817870616913, 0.13168871402740479, 0.09821856766939163, -0.13306692242622375, 0.09448322653770447, 0.01561434380710125, -0.1954789161682129, -0.007503822445869446, 0.003234093077480793, -0.05218910798430443, 0.1421356499195099, 0.015798868611454964, 0.12576164305210114, -0.011038488708436489, 0.09547081589698792, -0.17366495728492737, 0.00519853038713336, 0.04452250152826309, 0.012629679404199123, 0.0965389832854271, 0.041965991258621216, 0.01288489531725645, 0.09556606411933899, -0.06101891025900841, 0.06007314845919609, 0.011879269033670425, -0.11892972141504288, -0.20520564913749695, -0.09283290803432465, 0.04748104512691498, 0.08332131057977676, 0.09796521067619324, 0.004302435088902712, 0.1391502022743225, -0.08987855166196823, 0.08536813408136368, 0.2053663432598114, -0.2851874530315399, -0.06802612543106079, 0.04824196919798851, 0.010730392299592495, 0.04269130527973175, -0.09855502843856812, -0.0438108891248703, 0.04361143335700035, 0.050126537680625916, 0.1312054693698883, -0.027818024158477783, -0.11129193007946014, 0.012868739664554596, -0.1425374299287796, -0.042615797370672226, 0.16715098917484283, 0.04990904778242111, -0.03604314848780632, -0.041193317621946335, -0.06359817087650299, -0.164606511592865, -0.0288506131619215, -0.013921602629125118, 0.04442150890827179, -0.026521621271967888, -0.0562010258436203, 0.0006074340781196952, -0.10039065033197403, -0.06885430961847305, -0.08101525902748108, 0.1350230872631073, 0.035472165793180466, 0.015971506014466286, -0.021399779245257378, 0.11426833271980286, 0.0004156090726610273, -0.12140396237373352, 0.02053394354879856, 0.021590907126665115, 0.004075511358678341, -0.04597953334450722, -0.05174160748720169, -0.04296202212572098, 0.007668188773095608, 0.14446380734443665, -0.03264591470360756, 0.03329899162054062, 0.0537223219871521, 0.04418426379561424, -0.08855423331260681, 0.18396814167499542, -0.042952101677656174, -0.03383726254105568, 0.008014483377337456, 0.054429519921541214, 0.02557644620537758, -0.004421702120453119, -0.12368816137313843, 0.009974930435419083, 0.0978240892291069, 0.00805945135653019, -0.0656750500202179, 0.06479966640472412, -0.06243341416120529, -0.028439724817872047, 0.019054798409342766, -0.08543899655342102, 0.028429865837097168, -0.010553550906479359, -0.08018594980239868, -0.02219662256538868, 0.018496351316571236, 0.02436514012515545, -0.003702285699546337, 0.11057643592357635, -0.09640969336032867, 0.019585998728871346, -0.09028411656618118, -0.09864632040262222, 0.016174878925085068, -0.10933615267276764, 0.031578194350004196, -0.09535062313079834, -0.1961366981267929, -0.0036772103048861027, 0.06431546062231064, -0.0231940820813179, -0.07015454024076462, -0.0454469658434391, -0.06731566041707993, 0.009105165489017963, -0.009637190029025078, 0.12045573443174362, -0.06551243364810944, 0.0895988941192627, 0.020354699343442917, 0.05777054652571678, -0.053436312824487686, 0.05229693278670311, -0.10850567370653152, 0.02443447895348072, -0.15330395102500916, 0.03093011863529682, -0.04774477705359459, 0.062403757125139236, -0.08881653100252151, -0.10016635060310364, 0.01641538180410862, -0.01878579519689083, 0.06655380129814148, 0.08686811476945877, -0.18543602526187897, -0.06253620982170105, 0.13510993123054504, -0.06166541948914528, -0.12211087346076965, 0.12273122370243073, -0.06652545183897018, 0.04114757478237152, 0.05760481581091881, 0.151268869638443, 0.06827706098556519, -0.07533963769674301, 0.0043062870390713215, 0.010957700200378895, 0.051408860832452774, -0.06191996857523918, 0.07675977796316147, 0.005168503616005182, 0.015956176444888115, 0.0292753204703331, -0.03666900470852852, 0.05517169088125229, -0.08859530091285706, -0.1008458063006401, -0.04041777178645134, -0.09800264984369278, 0.04665899649262428, 0.0626826137304306, 0.06456942111253738, -0.08851597458124161, -0.07604947686195374, 0.052120909094810486, 0.09142126888036728, -0.044396091252565384, 0.020580459386110306, -0.06487052142620087, 0.0796298161149025, -0.04605669155716896, -0.03179631754755974, -0.17417313158512115, -0.033221110701560974, 0.01266591064631939, 0.0014970781048759818, 0.014374371618032455, 0.025865375995635986, 0.06246580556035042, 0.07420652359724045, -0.04267241433262825, -0.01950160041451454, -0.037407711148262024, 0.005499699618667364, -0.13156111538410187, -0.19282323122024536, -0.04342617839574814, -0.01942722126841545, 0.15563999116420746, -0.20186986029148102, 0.034146133810281754, -0.025317687541246414, 0.08778396993875504, 0.016044404357671738, -0.015395388938486576, -0.04330144077539444, 0.0685446634888649, -0.049985047429800034, -0.054318640381097794, 0.0652972161769867, 0.012191365472972393, -0.09109502285718918, -0.06641801446676254, -0.08756392449140549, 0.1621599644422531, 0.12625901401042938, -0.10113069415092468, -0.0736972913146019, -0.01609187014400959, -0.06495258212089539, -0.03419007360935211, -0.0506848506629467, 0.03021325170993805, 0.17371228337287903, -0.005152401514351368, 0.14251960813999176, -0.07014402747154236, -0.04465324804186821, 0.020611366257071495, -0.03425287827849388, 0.01945994608104229, 0.1139461100101471, 0.13531753420829773, -0.08171374350786209, 0.15208883583545685, 0.1549399048089981, -0.09335020929574966, 0.1156088262796402, -0.03903908282518387, -0.06310130655765533, -0.02674637921154499, -0.028943706303834915, -0.00767382001504302, 0.11595501750707626, -0.13964858651161194, 0.008336789906024933, 0.03642470762133598, 0.021943621337413788, 0.009799150750041008, -0.219939723610878, -0.04100724309682846, 0.03671945631504059, -0.0339110791683197, -0.006095096468925476, -0.01094911154359579, 0.006276230327785015, 0.09935463964939117, 0.0047756945714354515, -0.10570273548364639, 0.04776104539632797, 0.008964328095316887, -0.07248826324939728, 0.2040858119726181, -0.08690149337053299, -0.14122022688388824, -0.12400861084461212, -0.08494829386472702, -0.05893726646900177, 0.01116553321480751, 0.0529136024415493, -0.07159104198217392, -0.03631899505853653, -0.0723022073507309, 0.0022558560594916344, 0.0009501348831690848, 0.029797973111271858, 0.016634253785014153, -0.008368164300918579, 0.06798326969146729, -0.10520795732736588, -0.012365929782390594, -0.05160174146294594, -0.048995353281497955, 0.03591780737042427, 0.042148180305957794, 0.11395005881786346, 0.14934197068214417, -0.012972189113497734, 0.007610892411321402, -0.02072436548769474, 0.2539505660533905, -0.05908921733498573, -0.02042444981634617, 0.13773727416992188, -0.01955288276076317, 0.053054459393024445, 0.12119343876838684, 0.07568088173866272, -0.08348016440868378, -0.0023108357563614845, 0.031455181539058685, -0.03918725624680519, -0.21109984815120697, -0.053314097225666046, -0.05534370243549347, -0.007145646493881941, 0.09790794551372528, 0.02383042499423027, 0.038233764469623566, 0.08166204392910004, 0.03835240751504898, 0.09565010666847229, -0.052014078944921494, 0.06341741979122162, 0.12053665518760681, 0.04688894748687744, 0.12359534949064255, -0.03213420882821083, -0.06087375804781914, 0.0471905879676342, 0.005095073953270912, 0.2220466136932373, 0.012302805669605732, 0.12432891875505447, 0.06067948788404465, 0.18183033168315887, -0.010069633834064007, 0.0759282186627388, -0.00884784385561943, -0.03117067739367485, -0.02179272286593914, -0.037693291902542114, -0.040242258459329605, 0.027858445420861244, -0.05666995048522949, 0.07338742911815643, -0.10428783297538757, 0.02284305728971958, 0.051644206047058105, 0.2555640935897827, 0.038054946810007095, -0.34092211723327637, -0.09854152053594589, 0.00109822116792202, -0.03457309305667877, -0.024386251345276833, 0.02938893437385559, 0.0815875455737114, -0.0963740348815918, 0.022230898961424828, -0.06638722121715546, 0.09087297320365906, -0.051630064845085144, 0.041319768875837326, 0.08099466562271118, 0.0915423035621643, 0.013591844588518143, 0.08538056910037994, -0.27146899700164795, 0.2695838510990143, 0.0020335684530436993, 0.059616442769765854, -0.07820868492126465, 0.007642426528036594, 0.03506791591644287, 0.06423255056142807, 0.07269669324159622, -0.005352470558136702, -0.02157154679298401, -0.19570811092853546, -0.06258122622966766, 0.021919243037700653, 0.05955266207456589, -0.04061947762966156, 0.08851595968008041, -0.031296901404857635, 0.008749022148549557, 0.06803934276103973, 0.0062082731164991856, -0.048033151775598526, -0.10001352429389954, -0.005992088466882706, 0.037084951996803284, -0.045628871768713, -0.06241396814584732, -0.10849673300981522, -0.12127772718667984, 0.1436549872159958, -0.03217245265841484, -0.037102799862623215, -0.10771254450082779, 0.07576044648885727, 0.08138390630483627, -0.08232399821281433, 0.051328133791685104, -0.005709149409085512, 0.07703285664319992, 0.03309512883424759, -0.058879245072603226, 0.0984581857919693, -0.08060692250728607, -0.1688452810049057, -0.07240866124629974, 0.10352367162704468, 0.03739522024989128, 0.06483674049377441, -0.005450844764709473, 0.01844576932489872, -0.049744948744773865, -0.08899864554405212, 0.02327617071568966, 0.0023729519452899694, 0.08808863162994385, 0.01538530271500349, -0.04917232692241669, 0.025563480332493782, -0.05283571034669876, -0.03200049325823784, 0.18853305280208588, 0.23288851976394653, -0.10219664126634598, 0.019824828952550888, 0.028952129185199738, -0.06224953010678291, -0.17716413736343384, 0.02657325752079487, 0.05439988523721695, 0.004062270745635033, 0.040895476937294006, -0.17603285610675812, 0.14484943449497223, 0.11823732405900955, -0.018414992839097977, 0.10456965118646622, -0.3253220319747925, -0.11916499584913254, 0.13174839317798615, 0.13228487968444824, 0.1055133044719696, -0.12217845022678375, -0.020775066688656807, -0.018731476739048958, -0.1456051766872406, 0.11531335115432739, -0.07345319539308548, 0.11319600045681, -0.03303464874625206, 0.09853953868150711, 0.0030358540825545788, -0.05728635564446449, 0.12637527287006378, 0.034807153046131134, 0.09825865924358368, -0.0573401004076004, -0.04132261499762535, 0.03259928897023201, -0.04245501011610031, 0.02306911163032055, -0.08006823062896729, 0.038028109818696976, -0.1071585938334465, -0.0195018257945776, -0.06471694260835648, 0.04148260876536369, -0.03478411212563515, -0.07371383905410767, -0.04367890581488609, 0.02920873649418354, 0.0542772002518177, -0.011224198155105114, 0.1301102340221405, 0.049098215997219086, 0.1319703310728073, 0.09840372949838638, 0.06539186090230942, -0.07598528265953064, -0.0864650085568428, -0.029904864728450775, -0.016162382438778877, 0.05857016146183014, -0.1186198815703392, 0.02580508589744568, 0.14418600499629974, 0.023877494037151337, 0.13740704953670502, 0.08120670169591904, -0.016980841755867004, 0.006630040239542723, 0.05120522528886795, -0.17020870745182037, -0.06955239176750183, -0.00207059015519917, -0.03581907972693443, -0.11978787183761597, 0.05151791498064995, 0.09553749114274979, -0.07083508372306824, -0.007862416096031666, -0.00329012144356966, 0.013019414618611336, -0.05001607909798622, 0.19012168049812317, 0.05593651905655861, 0.04586886614561081, -0.10153039544820786, 0.07285546511411667, 0.054516613483428955, -0.052903033792972565, -0.00386789720505476, 0.047017667442560196, -0.08960752189159393, -0.041729580610990524, 0.05017653852701187, 0.16984573006629944, -0.07022494822740555, -0.04390936344861984, -0.13073822855949402, -0.11442448198795319, 0.08023715019226074, 0.13468000292778015, 0.11675281077623367, 0.016358956694602966, -0.06769761443138123, 0.000990710104815662, -0.1090991348028183, 0.09828060865402222, 0.047513313591480255, 0.07371482998132706, -0.15733586251735687, 0.13783763349056244, 0.004347299225628376, 0.04014428332448006, -0.01569807529449463, 0.02853160910308361, -0.0922224372625351, 0.007776952814310789, -0.11451610922813416, -0.020412571728229523, -0.037293434143066406, 0.013678283430635929, -0.004293015226721764, -0.05913880467414856, -0.05720795691013336, 0.01445155218243599, -0.10744114220142365, -0.019912000745534897, 0.03903994336724281, 0.06187763437628746, -0.11213311553001404, -0.03753271326422691, 0.029462359845638275, -0.06141684204339981, 0.0757351815700531, 0.04615128040313721, 0.025802932679653168, 0.042793937027454376, -0.11831625550985336, 0.011860872618854046, 0.06629679352045059, 0.029679937288165092, 0.07761236280202866, -0.10005507618188858, -0.013756772503256798, -0.002232906175777316, 0.037544623017311096, 0.014099475927650928, 0.0775010734796524, -0.1385369747877121, -0.010671062394976616, -0.010623650625348091, -0.07799592614173889, -0.06507866829633713, 0.017960112541913986, 0.10501421242952347, 0.017397819086909294, 0.2117377072572708, -0.060620490461587906, 0.04301229119300842, -0.20622839033603668, 0.0028291158378124237, -0.009644659236073494, -0.10717295855283737, -0.13312563300132751, -0.06061554327607155, 0.05153422802686691, -0.057664208114147186, 0.15196025371551514, 0.02733137086033821, 0.026553161442279816, 0.022098751738667488, 0.0024877474643290043, 0.02094321697950363, 0.009928583167493343, 0.19421900808811188, 0.04273122921586037, -0.03530365973711014, 0.057670336216688156, 0.03872951120138168, 0.10562846064567566, 0.10216277837753296, 0.18794463574886322, 0.13751569390296936, -0.0006151287234388292, 0.08835700154304504, 0.037604618817567825, -0.06515274196863174, -0.1754055619239807, 0.03262369707226753, -0.03752991184592247, 0.10625074803829193, -0.015471761114895344, 0.2279808521270752, 0.05541122704744339, -0.16879211366176605, 0.0331403985619545, -0.05034196004271507, -0.08079177886247635, -0.10232765227556229, -0.0639391541481018, -0.07805169373750687, -0.1259469985961914, -0.0004644935834221542, -0.11075980961322784, 0.008620215579867363, 0.12805825471878052, 0.00445753987878561, -0.02505398541688919, 0.14523586630821228, 0.0017235492123290896, 0.038010526448488235, 0.03869561105966568, 0.01353471539914608, -0.03502430394291878, -0.1110701784491539, -0.07435449212789536, -0.024660782888531685, -0.014861605130136013, 0.037946898490190506, -0.07131223380565643, -0.036500174552202225, 0.02696118876338005, -0.010986940935254097, -0.09129567444324493, 0.007318954914808273, 0.004903777968138456, 0.05119321867823601, 0.03402009233832359, 0.006444526370614767, 0.03641402721405029, -0.008416485972702503, 0.19665338099002838, -0.0748487114906311, -0.06379985064268112, -0.10773223638534546, 0.2278040498495102, 0.028930043801665306, -0.021108688786625862, 0.04188592731952667, -0.06633566319942474, 0.0050066313706338406, 0.23243562877178192, 0.19834978878498077, -0.09656862169504166, -0.01400521956384182, 0.009468605741858482, -0.014175452291965485, -0.034356601536273956, 0.09437727928161621, 0.13050082325935364, 0.04275999218225479, -0.08997633308172226, -0.0398663729429245, -0.07035182416439056, -0.012330877594649792, -0.033772196620702744, 0.058181677013635635, 0.03981747478246689, 0.005713402759283781, -0.04382667317986488, 0.047551676630973816, -0.06796412169933319, -0.09200788289308548, 0.061600495129823685, -0.199005588889122, -0.17041145265102386, -0.009912470355629921, 0.09833572804927826, 0.004328522831201553, 0.05889531224966049, -0.03365052863955498, -0.0014563101576641202, 0.08455890417098999, -0.019923580810427666, -0.09293641149997711, -0.08293168246746063, 0.10648664087057114, -0.0718170553445816, 0.23169483244419098, -0.04480297118425369, 0.07209361344575882, 0.12203482538461685, 0.06703657656908035, -0.0822090283036232, 0.056391771882772446, 0.05491861328482628, -0.05440649762749672, 0.02168349176645279, 0.0693649873137474, -0.02778610773384571, 0.08277266472578049, 0.044878680258989334, -0.13061197102069855, 0.010545195080339909, -0.043900374323129654, -0.0553804486989975, -0.04542292654514313, -0.033240627497434616, -0.055771492421627045, 0.13874585926532745, 0.20725072920322418, -0.03533142805099487, -0.014447090215981007, -0.06954600661993027, 0.023784196004271507, 0.06087842211127281, 0.00841572880744934, -0.061394672840833664, -0.21567288041114807, 0.016550946980714798, 0.04382448270916939, -0.019952859729528427, -0.2107541561126709, -0.1027153804898262, 0.001159833511337638, -0.07493405789136887, -0.08686104416847229, 0.0711950734257698, 0.08035538345575333, 0.05000495910644531, -0.05983276665210724, -0.025278707966208458, -0.08488844335079193, 0.13487741351127625, -0.1352304220199585, -0.08917593955993652 ]
null
null
transformers
# keytotext ![keytotext (1)](https://user-images.githubusercontent.com/49101362/116334480-f5e57a00-a7dd-11eb-987c-186477f94b6e.png) Idea is to build a model which will take keywords as inputs and generate sentences as outputs. ### Keytotext is powered by Huggingface 🤗 [![pypi Version](https://img.shields.io/pypi/v/keytotext.svg?style=flat-square&logo=pypi&logoColor=white)](https://pypi.org/project/keytotext/) [![Downloads](https://static.pepy.tech/personalized-badge/keytotext?period=total&units=none&left_color=grey&right_color=orange&left_text=Pip%20Downloads)](https://pepy.tech/project/keytotext) [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/gagan3012/keytotext/blob/master/Examples/K2T.ipynb) [![Streamlit App](https://static.streamlit.io/badges/streamlit_badge_black_white.svg)](https://share.streamlit.io/gagan3012/keytotext/UI/app.py) ## Model: Keytotext is based on the Amazing T5 Model: - `k2t`: [Model](https://huggingface.co/gagan3012/k2t) - `k2t-tiny`: [Model](https://huggingface.co/gagan3012/k2t-tiny) - `k2t-base`: [Model](https://huggingface.co/gagan3012/k2t-base) Training Notebooks can be found in the [`Training Notebooks`](https://github.com/gagan3012/keytotext/tree/master/Training%20Notebooks) Folder ## Usage: Example usage: [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/gagan3012/keytotext/blob/master/Examples/K2T.ipynb) Example Notebooks can be found in the [`Notebooks`](https://github.com/gagan3012/keytotext/tree/master/Examples) Folder ``` pip install keytotext ``` ![carbon (3)](https://user-images.githubusercontent.com/49101362/116220679-90e64180-a755-11eb-9246-82d93d924a6c.png) ## UI: UI: [![Streamlit App](https://static.streamlit.io/badges/streamlit_badge_black_white.svg)](https://share.streamlit.io/gagan3012/keytotext/UI/app.py) ``` pip install streamlit-tags ``` This uses a custom streamlit component built by me: [GitHub](https://github.com/gagan3012/streamlit-tags) ![image](https://user-images.githubusercontent.com/49101362/116162205-fc042980-a6fd-11eb-892e-8f6902f193f4.png)
{"language": "en", "license": "mit", "tags": ["keytotext", "k2t-base", "Keywords to Sentences"], "datasets": ["WebNLG", "Dart"], "metrics": ["NLG"], "thumbnail": "Keywords to Sentences"}
text2text-generation
gagan3012/k2t-base
[ "transformers", "pytorch", "t5", "text2text-generation", "keytotext", "k2t-base", "Keywords to Sentences", "en", "dataset:WebNLG", "dataset:Dart", "license:mit", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #t5 #text2text-generation #keytotext #k2t-base #Keywords to Sentences #en #dataset-WebNLG #dataset-Dart #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# keytotext !keytotext (1) Idea is to build a model which will take keywords as inputs and generate sentences as outputs. ### Keytotext is powered by Huggingface ![pypi Version](URL ![Downloads](URL ![Open In Colab](URL ![Streamlit App](URL ## Model: Keytotext is based on the Amazing T5 Model: - 'k2t': Model - 'k2t-tiny': Model - 'k2t-base': Model Training Notebooks can be found in the 'Training Notebooks' Folder ## Usage: Example usage: ![Open In Colab](URL Example Notebooks can be found in the 'Notebooks' Folder !carbon (3) ## UI: UI: ![Streamlit App](URL This uses a custom streamlit component built by me: GitHub !image
[ "# keytotext\n\n!keytotext (1)\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.", "### Keytotext is powered by Huggingface \n\n![pypi Version](URL\n![Downloads](URL\n![Open In Colab](URL\n![Streamlit App](URL", "## Model:\n\nKeytotext is based on the Amazing T5 Model: \n\n- 'k2t': Model\n- 'k2t-tiny': Model\n- 'k2t-base': Model\n\nTraining Notebooks can be found in the 'Training Notebooks' Folder", "## Usage:\n\nExample usage: ![Open In Colab](URL\n\nExample Notebooks can be found in the 'Notebooks' Folder\n\n\n\n!carbon (3)", "## UI:\n\nUI: ![Streamlit App](URL\n\n\nThis uses a custom streamlit component built by me: GitHub\n\n!image" ]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #keytotext #k2t-base #Keywords to Sentences #en #dataset-WebNLG #dataset-Dart #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# keytotext\n\n!keytotext (1)\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.", "### Keytotext is powered by Huggingface \n\n![pypi Version](URL\n![Downloads](URL\n![Open In Colab](URL\n![Streamlit App](URL", "## Model:\n\nKeytotext is based on the Amazing T5 Model: \n\n- 'k2t': Model\n- 'k2t-tiny': Model\n- 'k2t-base': Model\n\nTraining Notebooks can be found in the 'Training Notebooks' Folder", "## Usage:\n\nExample usage: ![Open In Colab](URL\n\nExample Notebooks can be found in the 'Notebooks' Folder\n\n\n\n!carbon (3)", "## UI:\n\nUI: ![Streamlit App](URL\n\n\nThis uses a custom streamlit component built by me: GitHub\n\n!image" ]
[ 85, 31, 44, 59, 36, 30 ]
[ "passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #keytotext #k2t-base #Keywords to Sentences #en #dataset-WebNLG #dataset-Dart #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# keytotext\n\n!keytotext (1)\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.### Keytotext is powered by Huggingface \n\n![pypi Version](URL\n![Downloads](URL\n![Open In Colab](URL\n![Streamlit App](URL## Model:\n\nKeytotext is based on the Amazing T5 Model: \n\n- 'k2t': Model\n- 'k2t-tiny': Model\n- 'k2t-base': Model\n\nTraining Notebooks can be found in the 'Training Notebooks' Folder## Usage:\n\nExample usage: ![Open In Colab](URL\n\nExample Notebooks can be found in the 'Notebooks' Folder\n\n\n\n!carbon (3)## UI:\n\nUI: ![Streamlit App](URL\n\n\nThis uses a custom streamlit component built by me: GitHub\n\n!image" ]
[ 0.022545116022229195, 0.15081577003002167, -0.006902134511619806, 0.052540652453899384, 0.09278816729784012, 0.01311257854104042, 0.17454983294010162, 0.12590664625167847, -0.018588101491332054, -0.013497188687324524, 0.08154649287462234, 0.1299619823694229, 0.05814589187502861, 0.19820158183574677, 0.014237146824598312, -0.23642423748970032, 0.03850706294178963, -0.008892025798559189, -0.0028976411558687687, 0.10865090787410736, 0.08588062971830368, -0.043456047773361206, 0.0833170935511589, 0.05551088601350784, -0.10083552449941635, 0.022495029494166374, -0.028258133679628372, -0.10732574760913849, 0.04806932434439659, 0.0617348775267601, 0.06288667023181915, 0.031502339988946915, -0.013700815849006176, -0.2369968146085739, 0.02069343999028206, 0.07490342110395432, 0.024024486541748047, 0.0483853854238987, 0.04558282345533371, 0.010756809264421463, 0.14172306656837463, 0.03496374562382698, 0.05292564257979393, 0.02259654738008976, -0.09434576332569122, -0.16997012495994568, -0.09155035018920898, 0.10970201343297958, 0.059319786727428436, 0.055495914071798325, -0.002940942533314228, 0.10594742745161057, 0.005288696847856045, 0.08593574166297913, 0.2135380357503891, -0.14506693184375763, -0.03953763470053673, -0.03880137950181961, 0.07200276851654053, 0.02832774817943573, -0.01873260922729969, 0.001543957507237792, -0.015178302302956581, -0.0016625849530100822, 0.13335269689559937, -0.07336549460887909, -0.14837856590747833, -0.019431648775935173, -0.07766515761613846, -0.03649429976940155, 0.15613505244255066, -0.01871725544333458, -0.011934855952858925, -0.16198115050792694, -0.02509462833404541, 0.06331460177898407, -0.0313296914100647, -0.002948393113911152, 0.00861147791147232, 0.011877329088747501, 0.041616544127464294, -0.17549116909503937, -0.11535188555717468, -0.020367801189422607, 0.06806381791830063, 0.02671070396900177, 0.012718943879008293, 0.0278181005269289, -0.04817581549286842, 0.13763687014579773, 0.03296466916799545, -0.14253519475460052, -0.009899149648845196, -0.06671847403049469, 0.026587683707475662, -0.00635969964787364, 0.06668263673782349, -0.09051209688186646, 0.0983467698097229, -0.0031992364674806595, 0.08098486810922623, 0.04385959729552269, -0.01606752537190914, 0.04241253063082695, 0.05995640903711319, 0.23326949775218964, -0.04028060659766197, -0.04890187829732895, 0.055764339864254, 0.04258882254362106, -0.0064017316326498985, -0.033427830785512924, -0.04353144392371178, 0.0143967280164361, -0.0029190191999077797, 0.04709590598940849, 0.08708037436008453, 0.10391177982091904, -0.01082056574523449, -0.03959058225154877, 0.09987955540418625, -0.06850359588861465, 0.03594982996582985, 0.054528042674064636, -0.026954423636198044, 0.05372387915849686, 0.11217240244150162, 0.006138554308563471, -0.1153751090168953, -0.023054102435708046, -0.029029246419668198, 0.002474575536325574, -0.10103631019592285, -0.11247534304857254, 0.02126908116042614, -0.07819097489118576, -0.031256452202796936, -0.07014752924442291, -0.09908804297447205, -0.03927423059940338, 0.08399124443531036, 0.016604825854301453, -0.0462651401758194, -0.051712214946746826, -0.014559201896190643, 0.011480668559670448, 0.03386947140097618, -0.10297947376966476, -0.013330486603081226, 0.06566943973302841, -0.1034608855843544, 0.0316072516143322, -0.051753487437963486, 0.05409645661711693, -0.11838389933109283, 0.024119319394230843, -0.2373759001493454, 0.09775228798389435, -0.052119579166173935, 0.10353872179985046, -0.10909638553857803, -0.010242831893265247, 0.1010846421122551, 0.002252631587907672, 0.014695393852889538, 0.09322601556777954, -0.09447291493415833, -0.0068136719055473804, 0.12370065599679947, -0.09127791970968246, -0.11193704605102539, 0.052230190485715866, -0.052090078592300415, 0.2100726217031479, 0.09385054558515549, 0.22132442891597748, 0.19504114985466003, -0.08513114601373672, -0.017375340685248375, 0.038304343819618225, -0.025688594207167625, 0.024546239525079727, 0.025284962728619576, -0.054299965500831604, 0.026378704234957695, 0.031893279403448105, -0.008042980916798115, 0.026990875601768494, 0.04419061914086342, -0.06442292034626007, 0.008282242342829704, -0.03106115572154522, -0.04110864922404289, -0.05812711641192436, -0.007419821340590715, -0.021690785884857178, -0.05642687901854515, 0.14573487639427185, 0.033548060804605484, -0.12408580631017685, -0.003229083027690649, -0.033544301986694336, 0.012125040404498577, 0.004788061138242483, 0.03342171758413315, -0.03396947681903839, -0.06551282852888107, 0.05194422975182533, -0.08027449250221252, 0.059400007128715515, -0.02685382403433323, 0.031658511608839035, 0.05185358226299286, 0.03880618140101433, -0.029133597388863564, -0.028964539989829063, 0.02665497176349163, -0.04319031164050102, -0.09239407628774643, 0.01786033995449543, -0.007288882043212652, 0.013264508917927742, -0.15513934195041656, 0.052053485065698624, 0.06586689502000809, 0.03875255212187767, 0.07139819115400314, -0.010071815922856331, 0.06611894816160202, -0.03595002740621567, 0.00997342448681593, -0.04509425908327103, 0.023828050121665, 0.06030678004026413, -0.027367735281586647, 0.09705083817243576, -0.14619654417037964, -0.011807648465037346, 0.06746806204319, -0.0060943434946238995, -0.030387766659259796, -0.038696739822626114, -0.023616356775164604, -0.018923776224255562, 0.007469849195331335, -0.003423800226300955, 0.15449516475200653, 0.06983716040849686, 0.10733337700366974, -0.07772118598222733, -0.07480408996343613, 0.00022269572946242988, -0.12292400747537613, 0.02712484449148178, 0.06992083787918091, 0.03127242252230644, -0.15398655831813812, 0.07997414469718933, 0.058878641575574875, -0.008707177825272083, 0.23312702775001526, -0.013233670964837074, -0.031163357198238373, -0.0333353728055954, 0.0293995700776577, 0.009693898260593414, 0.05086406320333481, -0.03329385071992874, 0.04813429340720177, 0.046700041741132736, -0.04243609681725502, 0.03536639362573624, -0.1381666511297226, 0.024658123031258583, 0.042210645973682404, -0.03003903292119503, -0.0500653013586998, 0.022211870178580284, 0.03489299863576889, 0.018621405586600304, 0.01694958098232746, 0.10951731353998184, 0.002329146722331643, -0.013438333757221699, -0.1166003942489624, 0.1691775918006897, -0.14203445613384247, -0.24584287405014038, -0.14314725995063782, -0.04016396403312683, 0.045231763273477554, -0.02181398496031761, 0.09163672477006912, -0.0044869547709822655, -0.014104612171649933, -0.09365416318178177, 0.08245370537042618, 0.04880950227379799, -0.11717446893453598, -0.08624111860990524, 0.05527671426534653, 0.001971637597307563, -0.1391879916191101, -0.030612492933869362, 0.07353970408439636, -0.07788117229938507, 0.028543872758746147, 0.0055036903358995914, 0.060905009508132935, 0.057555705308914185, -0.055338241159915924, 0.008062232285737991, -0.04124325141310692, 0.21081294119358063, -0.06260419636964798, 0.08578063547611237, 0.16405144333839417, -0.11542203277349472, 0.12109792232513428, 0.16720785200595856, 0.03794170543551445, 0.012375577352941036, 0.013895658776164055, -0.020891128107905388, -0.041222911328077316, -0.19690722227096558, 0.016830001026391983, -0.024180054664611816, 0.09396608918905258, 0.06122767925262451, 0.013967623934149742, 0.042371977120637894, 0.11335936933755875, -0.03126807510852814, 0.06436755508184433, 0.005893285386264324, 0.08173245936632156, 0.15976378321647644, -0.0006327145965769887, 0.021205520257353783, -0.10305735468864441, -0.027322763577103615, 0.058553218841552734, 0.09706276655197144, 0.11801059544086456, -0.04685762897133827, 0.1640271693468094, 0.09327437728643417, 0.05009303241968155, 0.017427388578653336, 0.07338934391736984, -0.06959163397550583, 0.06889096647500992, -0.016539746895432472, -0.12590165436267853, 0.022483497858047485, 0.0620657354593277, -0.026553254574537277, -0.08403993397951126, 0.0395926833152771, 0.010387920774519444, 0.09778935462236404, 0.24687451124191284, 0.0006333679775707424, -0.15127496421337128, 0.009416666813194752, 0.033955369144678116, 0.0046923719346523285, -0.1151227280497551, -0.02451666258275509, 0.07868783921003342, -0.17725777626037598, 0.009056966751813889, -0.0024235451128333807, 0.09950704872608185, -0.19137617945671082, -0.007148639298975468, 0.10451571643352509, 0.06115587055683136, -0.017391838133335114, 0.10509774088859558, -0.11340613663196564, 0.10442392528057098, 0.00634601479396224, 0.06980947405099869, -0.057221099734306335, 0.05097886174917221, 0.01223108358681202, -0.012519906274974346, 0.04914087429642677, 0.015447711572051048, 0.029171383008360863, -0.09805883467197418, -0.09036890417337418, 0.05772661417722702, -0.010836447589099407, -0.10961100459098816, 0.07063097506761551, -0.03856273740530014, 0.001754601951688528, -0.02989710308611393, -0.1420813947916031, -0.11948671191930771, -0.19851593673229218, 0.04168296977877617, -0.037011872977018356, 0.04207144305109978, -0.07887893170118332, 0.024311117827892303, 0.0662786066532135, 0.16778382658958435, 0.005767800845205784, -0.1548028290271759, -0.0953119546175003, -0.1000278890132904, 0.043845728039741516, -0.09962601959705353, 0.03395194560289383, -0.060768429189920425, 0.048506319522857666, -0.03590540960431099, -0.11836900562047958, 0.08789975941181183, -0.05272824317216873, -0.08148285746574402, -0.02572224661707878, 0.05252411216497421, -0.0034648794680833817, -0.015731437131762505, 0.014506221748888493, -0.004055740311741829, -0.0413513146340847, -0.11186568439006805, -0.044849641621112823, 0.10995081812143326, -0.04101148620247841, -0.01443497370928526, -0.05848793312907219, -0.11416122317314148, -0.08386364579200745, 0.0017410257132723927, 0.18850106000900269, 0.09899953752756119, -0.09413772821426392, 0.10445785522460938, 0.15878364443778992, -0.12411423027515411, -0.2125559002161026, -0.02615324594080448, 0.050072748214006424, -0.03733695298433304, 0.0241223331540823, -0.2663562595844269, 0.08189708739519119, 0.013726986944675446, -0.0198329109698534, 0.11520818620920181, -0.2331390678882599, -0.0920267179608345, 0.06239045411348343, 0.09630648046731949, -0.13612554967403412, -0.11028426140546799, -0.06213045120239258, -0.029889123514294624, -0.14992234110832214, 0.14835499227046967, -0.03763756528496742, 0.07475124299526215, 0.01900438964366913, 0.08237672597169876, 0.05174664407968521, -0.041067954152822495, 0.034860458225011826, -0.15678520500659943, 0.023700503632426262, -0.09086356312036514, -0.12081509083509445, 0.03404313325881958, -0.06940251588821411, 0.14375105500221252, -0.11732757091522217, 0.05240410938858986, -0.06658269464969635, -0.008412039838731289, -0.039364226162433624, 0.06966441869735718, -0.06270581483840942, -0.09007927030324936, -0.07965081185102463, -0.027090048417448997, 0.06389220803976059, -0.006064510438591242, -0.02969827689230442, -0.030135642737150192, -0.0015581499319523573, 0.15345820784568787, 0.01308775320649147, 0.17146699130535126, -0.11153212934732437, -0.04921746253967285, -0.026206690818071365, 0.0517355240881443, -0.17757664620876312, 0.02398303709924221, 0.06589053571224213, -0.012259835377335548, 0.13104109466075897, -0.018619615584611893, -0.127187579870224, 0.030060410499572754, 0.07694786787033081, -0.16647084057331085, -0.09988244622945786, -0.06037774309515953, 0.07045237720012665, -0.10180282592773438, -0.04057376831769943, 0.12742362916469574, -0.07817928493022919, -0.033585358411073685, -0.020569264888763428, 0.02497117966413498, -0.004393950570374727, 0.058804843574762344, 0.03157716616988182, 0.02963932231068611, -0.05010799691081047, 0.038766007870435715, 0.06991427391767502, 0.04731541872024536, 0.047500040382146835, 0.1359023153781891, -0.058502353727817535, -0.034616388380527496, 0.10732892900705338, 0.11324290931224823, 0.052162665873765945, -0.03213081136345863, -0.026527702808380127, -0.04812341555953026, -0.009151863865554333, 0.14459145069122314, 0.014677193947136402, 0.03961319103837013, -0.011617427691817284, 0.00306242355145514, -0.011362100020051003, 0.1424752175807953, -0.01663299649953842, 0.005010548979043961, -0.1069493517279625, 0.031898315995931625, 0.024049753323197365, 0.03824908286333084, 0.015312561765313148, -0.04658392071723938, -0.10333685576915741, -0.0466628223657608, -0.03849329799413681, 0.058767326176166534, -0.056691285222768784, 0.028813915327191353, 0.006322459317743778, 0.03625589236617088, 0.009655868634581566, 0.01873527094721794, -0.08054285496473312, -0.032363370060920715, -0.03752721846103668, 0.14416876435279846, -0.14532098174095154, -0.019870828837156296, 0.08076098561286926, -0.0757378339767456, 0.10201238095760345, -0.02063811756670475, -0.07879634201526642, 0.0036226478405296803, -0.10891518741846085, -0.0015249920543283224, 0.03142784163355827, 0.03261999413371086, 0.02543170563876629, -0.03519392013549805, 0.007022773381322622, -0.05133458599448204, 0.0021537013817578554, -0.029568996280431747, 0.11521641910076141, -0.12275374680757523, 0.02536102943122387, 0.041603971272706985, -0.07794715464115143, -0.031408507376909256, 0.04838091880083084, 0.0533154234290123, -0.013550236821174622, 0.13298669457435608, -0.041605137288570404, 0.07914933562278748, -0.11627219617366791, -0.015955850481987, 0.06613870710134506, -0.06735352426767349, -0.051519978791475296, -0.060331057757139206, 0.006248754914849997, -0.023229455575346947, 0.07770801335573196, 0.022113433107733727, -0.07849334180355072, -0.014932543970644474, -0.0399477556347847, 0.010670029558241367, 0.052514612674713135, 0.1133427694439888, -0.05499953031539917, -0.021637769415974617, -0.056107230484485626, -0.02923954650759697, -0.012793832458555698, -0.06692278385162354, 0.011365634389221668, 0.05717932805418968, 0.05268361046910286, 0.034060679376125336, 0.09420473873615265, 0.054380908608436584, -0.10789783298969269, -0.019370777532458305, 0.060083020478487015, 0.15351547300815582, -0.12039913237094879, 0.14092323184013367, 0.1545504778623581, -0.07345551997423172, 0.04808354377746582, 0.02904207445681095, -0.06829848140478134, -0.0912906676530838, -0.18982748687267303, -0.07103186845779419, -0.06342338770627975, -0.01583869569003582, -0.09735507518053055, 0.08734620362520218, -0.021711207926273346, -0.0030966748017817736, -0.056778568774461746, 0.15299223363399506, 0.06715944409370422, -0.07305724173784256, 0.07273288816213608, -0.019266260787844658, -0.016892388463020325, -0.04033241048455238, 0.049911998212337494, 0.016073761507868767, 0.06443750113248825, 0.04469245299696922, 0.09546102583408356, -0.04650736227631569, 0.007124239578843117, -0.1181122437119484, -0.08709001541137695, 0.004467806778848171, 0.022059323266148567, -0.016438014805316925, 0.05959261581301689, 0.05501271411776543, -0.05528569221496582, -0.00878556165844202, 0.1613909900188446, -0.060061000287532806, -0.117015540599823, -0.1487704962491989, 0.0650118738412857, -0.04942469298839569, 0.002726155100390315, -0.002290961565449834, -0.047152329236269, -0.049946390092372894, 0.2485342174768448, 0.20890015363693237, 0.011623305268585682, 0.014789954759180546, -0.030082322657108307, 0.015544656664133072, -0.02350914664566517, 0.15224429965019226, -0.0017789971316233277, 0.1511124223470688, -0.020765243098139763, 0.06685739755630493, -0.051698099821805954, -0.08942621201276779, -0.09956152737140656, 0.02824302203953266, 0.04044751822948456, 0.00812632404267788, -0.03271397948265076, 0.13004207611083984, -0.08000196516513824, -0.12879298627376556, -0.12010882049798965, -0.04281949624419212, -0.07736004889011383, 0.015960192307829857, 0.12471622973680496, -0.02698645368218422, 0.056134942919015884, 0.009203892201185226, -0.048284322023391724, 0.10872267931699753, 0.007059378083795309, -0.09114867448806763, 0.027696164324879646, 0.053300488740205765, -0.22131067514419556, 0.21900667250156403, 0.011493288911879063, 0.008702498860657215, 0.08237596601247787, -0.018939584493637085, -0.09386390447616577, 0.053713470697402954, 0.04702640324831009, -0.07483606040477753, -0.03664999082684517, 0.020574428141117096, -0.025836877524852753, -0.009551682509481907, 0.050284553319215775, -0.020690063014626503, 0.07640552520751953, -0.08696035295724869, -0.0018195667071267962, -0.08172968029975891, -0.018251944333314896, -0.08522951602935791, 0.0771171897649765, 0.16696636378765106, -0.024955620989203453, 0.007793879602104425, -0.039207007735967636, -0.01348145492374897, 0.045798804610967636, -0.0670144185423851, -0.002168738516047597, -0.05937298387289047, -0.05703030154109001, 0.05195050686597824, 0.1024816706776619, -0.07863403111696243, -0.002256637206301093, -0.03309889882802963, 0.02197904698550701, -0.07546041905879974, 0.1087217628955841, 0.12903115153312683, 0.0018792079063132405, 0.0075655654072761536, -0.11606748402118683, -0.06169462576508522, 0.10237009823322296, -0.07062531262636185, -0.0747082382440567 ]
null
null
transformers
# keytotext ![keytotext (1)](https://user-images.githubusercontent.com/49101362/116334480-f5e57a00-a7dd-11eb-987c-186477f94b6e.png) Idea is to build a model which will take keywords as inputs and generate sentences as outputs. ### Keytotext is powered by Huggingface 🤗 [![pypi Version](https://img.shields.io/pypi/v/keytotext.svg?style=flat-square&logo=pypi&logoColor=white)](https://pypi.org/project/keytotext/) [![Downloads](https://static.pepy.tech/personalized-badge/keytotext?period=total&units=none&left_color=grey&right_color=orange&left_text=Pip%20Downloads)](https://pepy.tech/project/keytotext) [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/gagan3012/keytotext/blob/master/Examples/K2T.ipynb) [![Streamlit App](https://static.streamlit.io/badges/streamlit_badge_black_white.svg)](https://share.streamlit.io/gagan3012/keytotext/UI/app.py) ## Model: Keytotext is based on the Amazing T5 Model: - `k2t`: [Model](https://huggingface.co/gagan3012/k2t) - `k2t-tiny`: [Model](https://huggingface.co/gagan3012/k2t-tiny) - `k2t-base`: [Model](https://huggingface.co/gagan3012/k2t-base) Training Notebooks can be found in the [`Training Notebooks`](https://github.com/gagan3012/keytotext/tree/master/Training%20Notebooks) Folder ## Usage: Example usage: [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/gagan3012/keytotext/blob/master/Examples/K2T.ipynb) Example Notebooks can be found in the [`Notebooks`](https://github.com/gagan3012/keytotext/tree/master/Examples) Folder ``` pip install keytotext ``` ![carbon (3)](https://user-images.githubusercontent.com/49101362/116220679-90e64180-a755-11eb-9246-82d93d924a6c.png) ## UI: UI: [![Streamlit App](https://static.streamlit.io/badges/streamlit_badge_black_white.svg)](https://share.streamlit.io/gagan3012/keytotext/UI/app.py) ``` pip install streamlit-tags ``` This uses a custom streamlit component built by me: [GitHub](https://github.com/gagan3012/streamlit-tags) ![image](https://user-images.githubusercontent.com/49101362/116162205-fc042980-a6fd-11eb-892e-8f6902f193f4.png)
{"language": "en", "license": "mit", "tags": ["keytotext", "k2t", "Keywords to Sentences"], "datasets": ["common_gen"], "metrics": ["NLG"], "thumbnail": "Keywords to Sentences"}
text2text-generation
gagan3012/k2t-new
[ "transformers", "pytorch", "jax", "t5", "text2text-generation", "keytotext", "k2t", "Keywords to Sentences", "en", "dataset:common_gen", "license:mit", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #t5 #text2text-generation #keytotext #k2t #Keywords to Sentences #en #dataset-common_gen #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# keytotext !keytotext (1) Idea is to build a model which will take keywords as inputs and generate sentences as outputs. ### Keytotext is powered by Huggingface ![pypi Version](URL ![Downloads](URL ![Open In Colab](URL ![Streamlit App](URL ## Model: Keytotext is based on the Amazing T5 Model: - 'k2t': Model - 'k2t-tiny': Model - 'k2t-base': Model Training Notebooks can be found in the 'Training Notebooks' Folder ## Usage: Example usage: ![Open In Colab](URL Example Notebooks can be found in the 'Notebooks' Folder !carbon (3) ## UI: UI: ![Streamlit App](URL This uses a custom streamlit component built by me: GitHub !image
[ "# keytotext\n\n!keytotext (1)\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.", "### Keytotext is powered by Huggingface \n\n![pypi Version](URL\n![Downloads](URL\n![Open In Colab](URL\n![Streamlit App](URL", "## Model:\n\nKeytotext is based on the Amazing T5 Model: \n\n- 'k2t': Model\n- 'k2t-tiny': Model\n- 'k2t-base': Model\n\nTraining Notebooks can be found in the 'Training Notebooks' Folder", "## Usage:\n\nExample usage: ![Open In Colab](URL\n\nExample Notebooks can be found in the 'Notebooks' Folder\n\n\n\n!carbon (3)", "## UI:\n\nUI: ![Streamlit App](URL\n\n\nThis uses a custom streamlit component built by me: GitHub\n\n!image" ]
[ "TAGS\n#transformers #pytorch #jax #t5 #text2text-generation #keytotext #k2t #Keywords to Sentences #en #dataset-common_gen #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# keytotext\n\n!keytotext (1)\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.", "### Keytotext is powered by Huggingface \n\n![pypi Version](URL\n![Downloads](URL\n![Open In Colab](URL\n![Streamlit App](URL", "## Model:\n\nKeytotext is based on the Amazing T5 Model: \n\n- 'k2t': Model\n- 'k2t-tiny': Model\n- 'k2t-base': Model\n\nTraining Notebooks can be found in the 'Training Notebooks' Folder", "## Usage:\n\nExample usage: ![Open In Colab](URL\n\nExample Notebooks can be found in the 'Notebooks' Folder\n\n\n\n!carbon (3)", "## UI:\n\nUI: ![Streamlit App](URL\n\n\nThis uses a custom streamlit component built by me: GitHub\n\n!image" ]
[ 81, 31, 44, 59, 36, 30 ]
[ "passage: TAGS\n#transformers #pytorch #jax #t5 #text2text-generation #keytotext #k2t #Keywords to Sentences #en #dataset-common_gen #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# keytotext\n\n!keytotext (1)\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.### Keytotext is powered by Huggingface \n\n![pypi Version](URL\n![Downloads](URL\n![Open In Colab](URL\n![Streamlit App](URL## Model:\n\nKeytotext is based on the Amazing T5 Model: \n\n- 'k2t': Model\n- 'k2t-tiny': Model\n- 'k2t-base': Model\n\nTraining Notebooks can be found in the 'Training Notebooks' Folder## Usage:\n\nExample usage: ![Open In Colab](URL\n\nExample Notebooks can be found in the 'Notebooks' Folder\n\n\n\n!carbon (3)## UI:\n\nUI: ![Streamlit App](URL\n\n\nThis uses a custom streamlit component built by me: GitHub\n\n!image" ]
[ 0.006397270131856203, 0.12119467556476593, -0.00590087566524744, 0.04985640197992325, 0.09001443535089493, 0.01826499216258526, 0.1998293101787567, 0.10257604718208313, 0.008477992378175259, -0.027575716376304626, 0.0788995698094368, 0.1310560554265976, 0.05929138883948326, 0.22064153850078583, 0.012605082243680954, -0.20708215236663818, 0.04966270178556442, -0.008622536435723305, -0.014222823083400726, 0.10715412348508835, 0.0931643471121788, -0.049494173377752304, 0.10004328191280365, 0.06921125203371048, -0.11669337004423141, 0.00906766764819622, -0.024502474814653397, -0.11332830041646957, 0.06877505779266357, 0.06913188099861145, 0.05654539167881012, 0.02772047370672226, -0.02661272883415222, -0.23392127454280853, 0.020741425454616547, 0.07818910479545593, 0.02551918849349022, 0.04941052943468094, 0.03388575464487076, 0.03285922110080719, 0.1386808156967163, 0.05016682296991348, 0.048793673515319824, 0.02455788664519787, -0.10560416430234909, -0.16413560509681702, -0.07794831693172455, 0.08170226216316223, 0.05034711956977844, 0.039424747228622437, 0.0019127666018903255, 0.12292821705341339, 0.019750770181417465, 0.07254661619663239, 0.23101098835468292, -0.12051776051521301, -0.028109634295105934, -0.026147253811359406, 0.09886861592531204, 0.0375593937933445, -0.006282700691372156, 0.009463390335440636, 0.003206290304660797, -0.020984098315238953, 0.14630815386772156, -0.07979520410299301, -0.12680670619010925, -0.01755974441766739, -0.06514742970466614, -0.03300822526216507, 0.17011509835720062, -0.017218654975295067, -0.02111414633691311, -0.17360606789588928, -0.04085926711559296, 0.05712735280394554, -0.037224799394607544, -0.011049901135265827, 0.01054622046649456, 0.01842445693910122, 0.053971562534570694, -0.19773757457733154, -0.10891011357307434, -0.034868139773607254, 0.10386069118976593, 0.044292252510786057, 0.00883780512958765, 0.03725901246070862, -0.057260215282440186, 0.1406261920928955, 0.012279875576496124, -0.13020870089530945, -0.02239752933382988, -0.09441494941711426, 0.028508830815553665, -0.0065095736645162106, 0.07561057806015015, -0.09472326934337616, 0.09785943478345871, -0.012141058221459389, 0.09619349986314774, 0.04964089393615723, -0.0006919896695762873, 0.043570246547460556, 0.06861558556556702, 0.20238849520683289, -0.03214366361498833, -0.05008570849895477, 0.05151636525988579, 0.04918406158685684, -0.032772962003946304, -0.05138697475194931, -0.038355473428964615, 0.005719910841435194, -0.005974899046123028, 0.03233739361166954, 0.06884697824716568, 0.10625043511390686, -0.0058809309266507626, -0.06347157806158066, 0.09628187119960785, -0.05664493143558502, 0.0388018861413002, 0.05532247945666313, -0.0112197520211339, 0.040083013474941254, 0.10052290558815002, 0.021047798916697502, -0.09130235761404037, -0.04612040892243385, -0.04048054665327072, 0.0002508911129552871, -0.12218309193849564, -0.1264169067144394, 0.012094970792531967, -0.1270582675933838, -0.031205253675580025, -0.07185633480548859, -0.10153736919164658, -0.033226121217012405, 0.09136775135993958, 0.04064224660396576, -0.04906972497701645, -0.04099789261817932, -0.017355795949697495, -0.004108247347176075, 0.03994401544332504, -0.11166982352733612, -0.0194099061191082, 0.060713671147823334, -0.10859790444374084, 0.03270038589835167, -0.07208371162414551, 0.0474274680018425, -0.11903944611549377, 0.02647705003619194, -0.20255440473556519, 0.0920279249548912, -0.04093341529369354, 0.10696131736040115, -0.0922727882862091, -0.0013390538515523076, 0.10083820670843124, 0.004175614099949598, 0.004281109664589167, 0.09237373620271683, -0.10040923207998276, -0.004592945799231529, 0.13461433351039886, -0.09256008267402649, -0.12207933515310287, 0.043564047664403915, -0.0697411298751831, 0.23423349857330322, 0.09615597128868103, 0.19714458286762238, 0.22026757895946503, -0.0993659570813179, 0.0010870774276554585, 0.03086230717599392, -0.023027321323752403, 0.02885289117693901, 0.015929043292999268, -0.050072554498910904, -0.0014917460503056645, 0.03415055200457573, -0.018775640055537224, 0.02834319695830345, 0.04753059148788452, -0.06110581383109093, 0.0085652656853199, -0.041449807584285736, -0.03182658925652504, -0.07802807539701462, -0.021159972995519638, -0.016907047480344772, -0.02238926663994789, 0.1561933010816574, 0.014825302176177502, -0.13179029524326324, -0.014679827727377415, -0.042634159326553345, 0.008523396216332912, 0.008939862251281738, 0.03636502847075462, -0.02867487072944641, -0.05787556990981102, 0.06687703728675842, -0.0799633339047432, 0.05137545242905617, 0.012002786621451378, 0.03423691540956497, 0.06587030738592148, 0.05105215311050415, -0.019528275355696678, -0.04235578328371048, 0.029138920828700066, -0.03624945133924484, -0.10525015741586685, 0.021780826151371002, -0.017863450571894646, 0.019872242584824562, -0.13585211336612701, 0.05918414518237114, 0.06500757485628128, 0.013119290582835674, 0.09840483963489532, -0.018373878672719002, 0.06120520830154419, -0.05095146223902702, 0.010872626677155495, -0.046587713062763214, 0.016339460387825966, 0.06775509566068649, -0.02376544289290905, 0.10574133694171906, -0.17385272681713104, 0.008338596671819687, 0.05869927257299423, -0.019404597580432892, -0.025666458532214165, -0.04722633585333824, -0.01818147301673889, -0.00959508866071701, -0.001998155377805233, -0.013993090018630028, 0.15213759243488312, 0.05710633099079132, 0.10310550779104233, -0.08802515268325806, -0.06924742460250854, 0.014607079327106476, -0.12826119363307953, 0.03750409930944443, 0.03019028529524803, 0.04287910461425781, -0.14439532160758972, 0.0774555578827858, 0.06592988967895508, 0.006204277742654085, 0.26038050651550293, -0.005987087264657021, -0.020050764083862305, -0.03339599072933197, 0.0382959246635437, -0.000704508856870234, 0.07774100452661514, -0.0658390000462532, 0.04545092210173607, 0.045758072286844254, -0.0370832122862339, 0.029205268248915672, -0.14168401062488556, 0.030596818774938583, 0.05280774086713791, -0.027322370558977127, -0.025330541655421257, 0.02397061511874199, 0.04127941653132439, 0.018317386507987976, 0.02425987459719181, 0.10590039938688278, 0.007451916579157114, -0.012213419191539288, -0.12809865176677704, 0.17400498688220978, -0.14575693011283875, -0.2214692384004593, -0.13214324414730072, -0.04820789769291878, 0.05396675318479538, -0.005884500686079264, 0.08030573278665543, -0.0016321009024977684, -0.013654260896146297, -0.09745324403047562, 0.08261644095182419, 0.047936249524354935, -0.11570192128419876, -0.08898986130952835, 0.06162247434258461, -0.0021001079585403204, -0.13846038281917572, -0.03681021183729172, 0.08044838160276413, -0.0653359666466713, 0.0289872158318758, 0.02040022425353527, 0.07006373256444931, 0.05694552883505821, -0.07218831032514572, 0.00967677403241396, -0.02915598452091217, 0.23768974840641022, -0.052825022488832474, 0.07428918033838272, 0.15125465393066406, -0.12298260629177094, 0.12043972313404083, 0.17997422814369202, 0.04017086327075958, 0.009699232876300812, 0.015031575225293636, -0.021363964304327965, -0.05291076377034187, -0.16851423680782318, 0.03416527435183525, -0.04592995345592499, 0.09993699193000793, 0.05033775418996811, 0.0006955497665330768, 0.06985212117433548, 0.12489651143550873, -0.0317973867058754, 0.10069173574447632, 0.01625707931816578, 0.08612319082021713, 0.1474475860595703, 0.003061983734369278, 0.012508764863014221, -0.09541555494070053, -0.03495185449719429, 0.050920307636260986, 0.09629859775304794, 0.15810884535312653, -0.03267643228173256, 0.17389658093452454, 0.11118455976247787, 0.03260675445199013, 0.016545668244361877, 0.07909045368432999, -0.07267772406339645, 0.07375752925872803, -0.028074610978364944, -0.12453492730855942, 0.018258297815918922, 0.08910203725099564, -0.04963058978319168, -0.09692999720573425, 0.06162383407354355, 0.02408589981496334, 0.09645988047122955, 0.22092531621456146, 0.0102114612236619, -0.1527230590581894, 0.016107140108942986, 0.03786129876971245, 0.018040476366877556, -0.13899554312229156, -0.02034766413271427, 0.08825046569108963, -0.18373189866542816, 0.016279617324471474, -0.03190998733043671, 0.10100369900465012, -0.1880820393562317, -0.0006596698076464236, 0.11167159676551819, 0.053554851561784744, -0.02288399450480938, 0.09542245417833328, -0.1417216956615448, 0.10862380266189575, 0.009309278801083565, 0.06515426188707352, -0.06857169419527054, 0.05543091520667076, 0.011161920614540577, -0.023723410442471504, 0.022103343158960342, 0.015735307708382607, 0.02509714476764202, -0.10309664160013199, -0.0642174482345581, 0.07118342071771622, -0.006615477614104748, -0.11428491771221161, 0.0696503221988678, -0.04296273738145828, 0.018362130969762802, -0.03637552633881569, -0.124579057097435, -0.1552421748638153, -0.1993626207113266, 0.05362161248922348, -0.03838283568620682, 0.04745079204440117, -0.08196760714054108, 0.04678402468562126, 0.08220906555652618, 0.16381987929344177, 0.012234571389853954, -0.16022656857967377, -0.08764772862195969, -0.11591674387454987, 0.04215402528643608, -0.09013472497463226, 0.047305937856435776, -0.06728208065032959, 0.05099615082144737, -0.0553465262055397, -0.11673151701688766, 0.10226179659366608, -0.05087314918637276, -0.08839969336986542, -0.03653038293123245, 0.0221414677798748, 0.00922833289951086, -0.027334434911608696, 0.020028049126267433, -0.0029872639570385218, -0.04682982340455055, -0.1192813441157341, -0.046532340347766876, 0.12508288025856018, -0.08120070397853851, -0.024195818230509758, -0.049691881984472275, -0.1328153759241104, -0.07178358733654022, 0.011016800999641418, 0.1828717440366745, 0.12523797154426575, -0.08559996634721756, 0.10591965168714523, 0.16338428854942322, -0.12225178629159927, -0.18638348579406738, -0.0068282573483884335, 0.03738740459084511, -0.03020087443292141, 0.026890475302934647, -0.27494657039642334, 0.04789354279637337, -0.008778178133070469, -0.019336747005581856, 0.13768011331558228, -0.22268150746822357, -0.08734731376171112, 0.0610230416059494, 0.10820960253477097, -0.1251632422208786, -0.1010999009013176, -0.06172490492463112, -0.027707353234291077, -0.1924157291650772, 0.17615477740764618, -0.040408674627542496, 0.06898581236600876, 0.021089766174554825, 0.07100752741098404, 0.0631515309214592, -0.046996746212244034, 0.03731486201286316, -0.16057200729846954, 0.03731607645750046, -0.09086383879184723, -0.12494693696498871, 0.04646284133195877, -0.06974950432777405, 0.14511947333812714, -0.14033183455467224, 0.05599023774266243, -0.047732703387737274, -0.019723117351531982, -0.03876183182001114, 0.07849995046854019, -0.05946517735719681, -0.1164768785238266, -0.08666805922985077, -0.04063170775771141, 0.04784954711794853, 0.0009260172373615205, -0.05496755614876747, -0.02314520627260208, 0.004757160320878029, 0.13353092968463898, 0.022509466856718063, 0.138665571808815, -0.08481486141681671, -0.03258923441171646, -0.009819099679589272, 0.06900790333747864, -0.18951524794101715, 0.02897205390036106, 0.05110177397727966, -0.0050001428462564945, 0.1403309851884842, -0.018387161195278168, -0.14364929497241974, 0.042836885899305344, 0.07557272911071777, -0.173936665058136, -0.09523696452379227, -0.06382003426551819, 0.08386265486478806, -0.09628656506538391, -0.03739813715219498, 0.12111350148916245, -0.08348014205694199, -0.031776703894138336, -0.03433509171009064, 0.017749959602952003, -0.0007487880066037178, 0.03718329593539238, 0.025121517479419708, 0.0355016253888607, -0.05139429494738579, 0.004403641913086176, 0.04315340518951416, 0.046485599130392075, 0.03613803908228874, 0.10725808888673782, -0.06379794329404831, -0.029298316687345505, 0.11946853250265121, 0.1268007904291153, 0.05531023442745209, -0.03233453631401062, -0.026001250371336937, -0.0539093054831028, -0.015576512552797794, 0.12361860275268555, 0.010767494328320026, 0.04683016240596771, -0.0025885282084345818, 0.0020893034525215626, 0.0072632236406207085, 0.1409498155117035, -0.018711963668465614, 0.00711481599137187, -0.10889719426631927, 0.02021137811243534, 0.020530760288238525, 0.0398346409201622, 0.020300570875406265, -0.04106844216585159, -0.08168765902519226, -0.05276931822299957, -0.03862809017300606, 0.053864240646362305, -0.06823062896728516, 0.026056353002786636, 0.004964424297213554, 0.05668537691235542, 0.01563034951686859, 0.01664869487285614, -0.09525235742330551, -0.03440261259675026, -0.03559383749961853, 0.14418290555477142, -0.14135588705539703, -0.008675787597894669, 0.09781572222709656, -0.08732236921787262, 0.11464694142341614, -0.03297751396894455, -0.09114479273557663, -0.0030078489799052477, -0.09819291532039642, 0.010054919868707657, 0.05026966333389282, 0.03447281941771507, 0.025119060650467873, -0.018256429582834244, 0.015087001025676727, -0.064223513007164, 0.02130219340324402, -0.0406096950173378, 0.1171492338180542, -0.12074559181928635, 0.04027499258518219, 0.04767632484436035, -0.08235785365104675, -0.026213116943836212, 0.04531453922390938, 0.05775214359164238, -0.017567552626132965, 0.11693090200424194, -0.02297193370759487, 0.06954316794872284, -0.10742456465959549, -0.022608160972595215, 0.07415220886468887, -0.06387952715158463, -0.06449215859174728, -0.07441972941160202, -0.008318808861076832, -0.022702915593981743, 0.10362212359905243, 0.02446352317929268, -0.08809523284435272, -0.02735498547554016, -0.038903623819351196, 0.008655883371829987, 0.0575067475438118, 0.13404618203639984, -0.058960530906915665, -0.01170092262327671, -0.0660456195473671, -0.020349716767668724, -0.008306472562253475, -0.07279200106859207, 0.015671368688344955, 0.05017375946044922, 0.03410590812563896, 0.034185513854026794, 0.09461900591850281, 0.03760331869125366, -0.0910060927271843, -0.005792669951915741, 0.07329554110765457, 0.17203636467456818, -0.13045090436935425, 0.14601485431194305, 0.15847888588905334, -0.06532561033964157, 0.045181240886449814, 0.04248795285820961, -0.07425811886787415, -0.09667836874723434, -0.1584727019071579, -0.07056698203086853, -0.07364056259393692, -0.018811648711562157, -0.09476049989461899, 0.08753763884305954, -0.04786815866827965, -0.01705213263630867, -0.0536799319088459, 0.1601305902004242, 0.07390434294939041, -0.09232184290885925, 0.06942444294691086, -0.012245004996657372, -0.023764124140143394, -0.04938637837767601, 0.030763452872633934, 0.015352501533925533, 0.05541202053427696, 0.04227210953831673, 0.08834075182676315, -0.055831119418144226, 0.012330743484199047, -0.11662497371435165, -0.08884448558092117, -0.0008801512885838747, 0.005009234882891178, -0.023375308141112328, 0.03505311906337738, 0.040228087455034256, -0.058494508266448975, -0.0068830568343400955, 0.14811238646507263, -0.05639500170946121, -0.10991320013999939, -0.15787599980831146, 0.01329825259745121, -0.06424568593502045, 0.0072470358572900295, 0.0010161587269976735, -0.03956611454486847, -0.05217769369482994, 0.2347848117351532, 0.22433121502399445, 0.006577456835657358, 0.0186973474919796, -0.034553080797195435, 0.012560910545289516, -0.02248571068048477, 0.1524742692708969, -0.04919350519776344, 0.15303586423397064, -0.012000972405076027, 0.07658783346414566, -0.050322845578193665, -0.09078732132911682, -0.09269822388887405, 0.02059575356543064, 0.050671759992837906, 0.0037633441388607025, -0.03152357414364815, 0.13050514459609985, -0.06514988839626312, -0.11814668774604797, -0.1066751629114151, -0.024741774424910545, -0.05866260081529617, 0.01510835625231266, 0.10372033715248108, -0.03311283886432648, 0.05558241158723831, -0.005994274280965328, -0.051878422498703, 0.12508155405521393, 0.010022918693721294, -0.11376495659351349, 0.01910240575671196, 0.05730541795492172, -0.20283037424087524, 0.22015702724456787, 0.01160853449255228, -0.0019177410285919905, 0.07527907937765121, -0.024538878351449966, -0.09229271113872528, 0.03972962498664856, 0.051930513232946396, -0.06366978585720062, -0.05701316148042679, -0.009701867587864399, -0.02300127223134041, -0.01924893818795681, 0.05440395697951317, -0.032926104962825775, 0.07832535356283188, -0.09658661484718323, -0.006534182000905275, -0.09215812385082245, -0.02157982625067234, -0.09115232527256012, 0.0697273388504982, 0.14890868961811066, -0.01868288777768612, -0.004538996145129204, -0.030332662165164948, 0.006867953576147556, 0.0640227422118187, -0.07766830176115036, -0.025120889768004417, -0.06451641023159027, -0.07271898537874222, 0.07487855106592178, 0.10218413919210434, -0.054678164422512054, 0.0011143723968416452, -0.03022201545536518, 0.038342271000146866, -0.06249852478504181, 0.10087922215461731, 0.14120851457118988, -0.007158010266721249, 0.01334556844085455, -0.12261556088924408, -0.06548746675252914, 0.1026679128408432, -0.06890170276165009, -0.08426952362060547 ]
null
null
transformers
<h1 align="center">keytotext</h1> [![pypi Version](https://img.shields.io/pypi/v/keytotext.svg?logo=pypi&logoColor=white)](https://pypi.org/project/keytotext/) [![Downloads](https://static.pepy.tech/personalized-badge/keytotext?period=total&units=none&left_color=grey&right_color=orange&left_text=Pip%20Downloads)](https://pepy.tech/project/keytotext) [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/gagan3012/keytotext/blob/master/notebooks/K2T.ipynb) [![Streamlit App](https://static.streamlit.io/badges/streamlit_badge_black_white.svg)](https://share.streamlit.io/gagan3012/keytotext/UI/app.py) [![API Call](https://img.shields.io/badge/-FastAPI-red?logo=fastapi&labelColor=white)](https://github.com/gagan3012/keytotext#api) [![Docker Call](https://img.shields.io/badge/-Docker%20Image-blue?logo=docker&labelColor=white)](https://hub.docker.com/r/gagan30/keytotext) [![HuggingFace](https://img.shields.io/badge/%F0%9F%A4%97-Models%20on%20Hub-yellow)](https://huggingface.co/models?filter=keytotext) [![Documentation Status](https://readthedocs.org/projects/keytotext/badge/?version=latest)](https://keytotext.readthedocs.io/en/latest/?badge=latest) [![Code style: black](https://img.shields.io/badge/code%20style-black-000000.svg)](https://github.com/psf/black) ![keytotext](https://socialify.git.ci/gagan3012/keytotext/image?description=1&forks=1&language=1&owner=1&stargazers=1&theme=Light) Idea is to build a model which will take keywords as inputs and generate sentences as outputs. Potential use case can include: - Marketing - Search Engine Optimization - Topic generation etc. - Fine tuning of topic modeling models
{"language": "en", "license": "MIT", "tags": ["keytotext", "k2t", "Keywords to Sentences"], "datasets": ["WebNLG", "Dart"], "metrics": ["NLG"], "thumbnail": "Keywords to Sentences"}
text2text-generation
gagan3012/k2t-test
[ "transformers", "pytorch", "t5", "text2text-generation", "keytotext", "k2t", "Keywords to Sentences", "en", "dataset:WebNLG", "dataset:Dart", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #t5 #text2text-generation #keytotext #k2t #Keywords to Sentences #en #dataset-WebNLG #dataset-Dart #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
<h1 align="center">keytotext</h1> ![pypi Version](URL ![Downloads](URL ![Open In Colab](URL ![Streamlit App](URL ![API Call](URL ![Docker Call](URL ![HuggingFace](URL ![Documentation Status](URL ![Code style: black](URL !keytotext Idea is to build a model which will take keywords as inputs and generate sentences as outputs. Potential use case can include: - Marketing - Search Engine Optimization - Topic generation etc. - Fine tuning of topic modeling models
[]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #keytotext #k2t #Keywords to Sentences #en #dataset-WebNLG #dataset-Dart #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 78 ]
[ "passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #keytotext #k2t #Keywords to Sentences #en #dataset-WebNLG #dataset-Dart #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.03903217613697052, 0.07688179612159729, -0.0061773881316185, 0.007946956902742386, 0.10117848217487335, 0.02843649871647358, 0.09411219507455826, 0.15868787467479706, 0.006724410690367222, -0.057724714279174805, 0.1404922604560852, 0.2025938332080841, 0.017940107733011246, 0.13757465779781342, -0.06942672282457352, -0.2564336359500885, 0.03815453499555588, 0.037285901606082916, -0.019459573552012444, 0.14391419291496277, 0.08486589789390564, -0.05302097275853157, 0.04979008436203003, -0.019126372411847115, -0.1216764822602272, 0.054594650864601135, 0.030159002169966698, -0.13637526333332062, 0.10469778627157211, 0.018528535962104797, 0.08310289680957794, 0.05006900057196617, -0.05860939249396324, -0.2161225974559784, 0.025030845776200294, 0.008722852915525436, -0.045329101383686066, 0.03354315459728241, 0.03338639438152313, -0.046357449144124985, 0.05640164762735367, -0.06243095174431801, -0.017651746049523354, 0.04220695421099663, -0.09832804650068283, -0.04187982156872749, -0.05007176101207733, 0.043057892471551895, 0.08319853991270065, 0.0737672746181488, -0.040723055601119995, 0.15581513941287994, -0.07629773765802383, 0.14664098620414734, 0.1368003636598587, -0.3269534111022949, 0.012729325331747532, -0.0035922133829444647, 0.004724225029349327, 0.06208911910653114, -0.02152496948838234, 0.06762268394231796, 0.012842950411140919, 0.017713449895381927, 0.04426150768995285, -0.09849993884563446, -0.24517373740673065, 0.04561421275138855, -0.07974745333194733, -0.020331554114818573, 0.289760559797287, -0.02673063613474369, 0.08189614117145538, -0.06428142637014389, -0.08825667947530746, 0.03182201087474823, -0.023756423965096474, -0.030087575316429138, -0.05645490437746048, 0.04441112279891968, -0.010185645893216133, -0.08027484267950058, -0.15373766422271729, 0.005433336831629276, -0.20648396015167236, 0.10115378350019455, 0.024363122880458832, 0.035511452704668045, -0.18737928569316864, 0.05462529510259628, 0.03907693922519684, -0.12055029720067978, 0.04595165699720383, -0.053805362433195114, 0.018455632030963898, -0.0044195144437253475, -0.03764479234814644, -0.08237682282924652, 0.15702591836452484, -0.021671166643500328, 0.003712681820616126, 0.003861523699015379, -0.08378283679485321, 0.05623997002840042, 0.062132399529218674, 0.07123706489801407, -0.02089962735772133, -0.07825490087270737, 0.05290557071566582, -0.1396183967590332, 0.009325655177235603, -0.04026453197002411, -0.09688538312911987, -0.06370905786752701, 0.10109986364841461, 0.08713588118553162, 0.08178628981113434, 0.10759539902210236, -0.025620993226766586, -0.017821846529841423, 0.012486998923122883, -0.07545977830886841, -0.027385571971535683, 0.0017266568029299378, -0.014242880046367645, 0.10150834172964096, 0.0165895763784647, 0.07835089415311813, -0.134853333234787, 0.019631991162896156, -0.07318528741598129, -0.0271466001868248, 0.03464220091700554, -0.08904528617858887, 0.05325685441493988, -0.0927010253071785, 0.0202367901802063, -0.1247885525226593, -0.17206141352653503, 0.01621382124722004, -0.03229447826743126, 0.0012832974316552281, -0.0709434524178505, -0.08504429459571838, -0.04151918739080429, 0.05391307175159454, -0.0692572221159935, 0.008550930768251419, -0.04580128937959671, 0.11258784681558609, -0.07574059069156647, 0.089997299015522, -0.11348620057106018, 0.042653366923332214, -0.15919475257396698, -0.07423004508018494, -0.053964316844940186, 0.10881427675485611, -0.014113583602011204, 0.12761420011520386, -0.08176441490650177, -0.007769408170133829, -0.03169988840818405, 0.034524861723184586, -0.045046985149383545, 0.24183803796768188, -0.12320122867822647, -0.10078071057796478, 0.2652226984500885, -0.039097171276807785, -0.16684569418430328, 0.07894851267337799, 0.0009547838126309216, 0.14258721470832825, 0.15320105850696564, 0.2657999098300934, 0.07426482439041138, -0.02078670635819435, 0.00763301644474268, 0.11893578618764877, -0.0869615450501442, -0.035517383366823196, 0.01223339419811964, -0.018862366676330566, -0.013180753216147423, 0.054628338664770126, 0.12370645999908447, 0.052965909242630005, -0.027910005301237106, -0.05075078457593918, -0.018453463912010193, -0.018193617463111877, 0.08231326192617416, -0.02094157226383686, 0.07350154966115952, -0.07670266926288605, -0.028671003878116608, 0.03127618506550789, 0.006361823063343763, -0.019136076793074608, 0.04919913783669472, -0.03604045882821083, -0.027698056772351265, -0.0022763845045119524, 0.052324894815683365, -0.1082569807767868, -0.12414075434207916, -0.012394402176141739, 0.18391548097133636, 0.03078749030828476, 0.0533563457429409, 0.030673427507281303, -0.07504592090845108, -0.01635466329753399, -0.0037628004793077707, 0.1724305897951126, 0.03408222645521164, -0.08487121015787125, -0.11941272765398026, 0.1011456772685051, -0.032992489635944366, -0.015726841986179352, 0.04496084898710251, -0.004146311432123184, 0.1094067245721817, 0.10359770059585571, -0.01346167828887701, 0.06498674303293228, 0.020176568999886513, -0.005791624076664448, -0.03126189485192299, 0.02556530199944973, 0.06685677170753479, -0.006117179058492184, -0.12800662219524384, 0.20311257243156433, -0.07922592759132385, 0.1098286360502243, 0.17013971507549286, -0.1630544364452362, 0.052374500781297684, -0.05337982624769211, -0.03546663746237755, -0.02642180025577545, 0.052608367055654526, -0.02748814783990383, 0.1092694103717804, 0.04293446242809296, 0.1276484578847885, -0.055808719247579575, -0.08623116463422775, -0.0014845937257632613, -0.07218868285417557, -0.03463004156947136, 0.07184917479753494, -0.05464067682623863, -0.2767590880393982, 0.151351660490036, 0.14284931123256683, 0.07120442390441895, 0.218695268034935, -0.020985165610909462, -0.021876879036426544, 0.03600084409117699, -0.02205980382859707, -0.0722251608967781, -0.06470290571451187, -0.13558830320835114, -0.0033426089212298393, 0.07773612439632416, 0.021976018324494362, 0.07477480918169022, -0.10188990086317062, -0.04983449727296829, -0.03033437579870224, -0.024645287543535233, -0.0009964666096493602, 0.09942738711833954, 0.06909600645303726, 0.13495886325836182, 0.01318643894046545, 0.04001353681087494, 0.08726658672094345, -0.011574335396289825, -0.12594421207904816, 0.183069109916687, -0.20184381306171417, -0.3897164463996887, -0.06920620054006577, -0.08202673494815826, -0.0433579757809639, -0.023706885054707527, 0.134678915143013, -0.158100888133049, 0.007603765465319157, 0.0012842394644394517, 0.04123978316783905, -0.06566179543733597, 0.0188005268573761, -0.06776919960975647, 0.02648961916565895, -0.09561283141374588, -0.0856005847454071, -0.05406622216105461, -0.013013672083616257, -0.069023497402668, 0.1326999068260193, -0.1190083771944046, 0.0657150074839592, 0.15551768243312836, 0.01996716484427452, 0.04895695298910141, -0.08226631581783295, 0.16687576472759247, -0.06638176739215851, 0.06961362808942795, 0.13507559895515442, -0.028373628854751587, 0.08798438310623169, 0.15129351615905762, -0.013650152832269669, -0.0232254546135664, 0.049437958747148514, 0.018354935571551323, -0.026459643617272377, -0.26186463236808777, -0.12826767563819885, -0.07437894493341446, 0.08185781538486481, -0.0017615138785913587, 0.04023335501551628, 0.14391960203647614, 0.06444190442562103, -0.027538392692804337, -0.020490268245339394, 0.022095942869782448, 0.06953368335962296, 0.25473225116729736, -0.02774951606988907, 0.1169879361987114, -0.06025315448641777, -0.11819146573543549, 0.08268919587135315, 0.11264616996049881, 0.050551220774650574, 0.01648484170436859, 0.10656298696994781, 0.05184439942240715, 0.11337745189666748, 0.04947604611515999, 0.07878274470567703, 0.03159871697425842, 0.0007107985438778996, -0.015850944444537163, -0.05846184492111206, -0.01383562758564949, 0.05181414261460304, 0.034767843782901764, -0.10243432223796844, -0.03443082794547081, -0.01150580495595932, 0.12185569107532501, 0.13546621799468994, 0.08351317793130875, -0.12537677586078644, 0.005218727979809046, 0.07765185832977295, -0.044701624661684036, -0.10390593111515045, 0.08798053115606308, -0.005960504990071058, -0.13519296050071716, 0.11843277513980865, -0.032555289566516876, 0.0880533754825592, -0.04732156917452812, 0.06625839322805405, -0.052118733525276184, -0.09939749538898468, 0.002937457524240017, 0.13495902717113495, -0.3500785231590271, 0.18479159474372864, 0.020522842183709145, -0.03739010915160179, -0.11735451966524124, -0.024229316040873528, -0.020654641091823578, 0.018029769882559776, 0.15043169260025024, 0.0040740673430264, 0.022580603137612343, -0.023493951186537743, -0.055756766349077225, 0.054868098348379135, 0.08672568202018738, -0.03158225491642952, 0.02804802730679512, -0.047701746225357056, -0.008736777119338512, 0.001647863071411848, -0.03301027789711952, -0.0661146268248558, -0.13900674879550934, 0.05676543340086937, 0.08871498703956604, -0.018053695559501648, 0.011931195855140686, -0.030276063829660416, -0.023456571623682976, 0.2034592479467392, -0.06985244899988174, -0.13113707304000854, -0.12003286927938461, -0.034163400530815125, 0.059256117790937424, -0.10904631018638611, 0.007828399538993835, -0.06547124683856964, 0.016343815252184868, -0.08073829859495163, -0.19135794043540955, 0.10531875491142273, -0.07745100557804108, -0.033804330974817276, -0.04372392222285271, 0.1793138086795807, -0.01073919702321291, 0.015988832339644432, 0.03128870576620102, 0.010686389170587063, -0.06278391927480698, -0.06007697433233261, 0.041904956102371216, -0.029453665018081665, 0.035968679934740067, 0.02807362750172615, -0.06198997050523758, -0.02893623523414135, -0.10285694152116776, -0.04338344931602478, 0.3439633548259735, 0.13939659297466278, -0.06957757472991943, 0.21206419169902802, 0.1378963142633438, -0.10091788321733475, -0.31464821100234985, -0.09116147458553314, -0.06621747463941574, -0.05524633452296257, -0.014028315432369709, -0.17787262797355652, 0.09584298729896545, -0.042744219303131104, -0.024806752800941467, 0.03977423906326294, -0.2556311786174774, -0.082899309694767, 0.16976392269134521, 0.01596090942621231, 0.2834286689758301, -0.13093051314353943, -0.08922819793224335, -0.03937128558754921, -0.12698093056678772, 0.19979402422904968, -0.08418729156255722, 0.06153452768921852, -0.004666736349463463, 0.1776805818080902, 0.04617675021290779, -0.011990601196885109, 0.012206112034618855, 0.003995159175246954, -0.02480645291507244, -0.06926066428422928, -0.09844497591257095, 0.08762900531291962, -0.008778671734035015, 0.020381171256303787, -0.06595955789089203, 0.05049559473991394, -0.0949656069278717, -0.0006382031133398414, -0.12513716518878937, 0.07758493721485138, -0.007955621927976608, -0.09927115589380264, -0.019221661612391472, -0.04650452733039856, 0.06919635832309723, -0.020818738266825676, 0.15716107189655304, -0.03872097283601761, 0.11444004625082016, 0.10615856200456619, 0.09696070849895477, -0.0019096651813015342, 0.06675493717193604, -0.062093157321214676, -0.09167221933603287, 0.027807479724287987, -0.13657724857330322, 0.05464359000325203, 0.11414086073637009, -0.018889576196670532, 0.042606230825185776, 0.06232642009854317, -0.030024895444512367, -0.03841933608055115, 0.10966654866933823, -0.24939408898353577, 0.008710821159183979, -0.09975350648164749, -0.07740998268127441, 0.03193715959787369, 0.056537844240665436, 0.20132313668727875, 0.019023699685931206, -0.06170305982232094, -0.020352233201265335, 0.05099378526210785, -0.04595917835831642, 0.08228398114442825, 0.03132554888725281, 0.029461823403835297, -0.13575321435928345, 0.10601908713579178, 0.015725204721093178, -0.04993676394224167, 0.08034761995077133, 0.17141593992710114, -0.11595481634140015, -0.08235012739896774, 0.018206652253866196, 0.09943213313817978, -0.09247054904699326, -0.02033846080303192, -0.05460567772388458, -0.09290233254432678, 0.046055182814598083, 0.2177998423576355, 0.020063476637005806, 0.1118638664484024, -0.04517297074198723, -0.04891025274991989, -0.047661323100328445, 0.10537919402122498, 0.0014809079002588987, -0.04026184603571892, -0.10279271751642227, 0.05047539249062538, -0.04830300062894821, 0.17378509044647217, -0.06399358063936234, -0.0592978410422802, -0.10721044987440109, 0.023348625749349594, -0.06541787832975388, -0.029305675998330116, -0.0824657455086708, -0.03429308906197548, -0.021074894815683365, -0.031610239297151566, -0.03873627260327339, -0.03024277836084366, -0.0764424055814743, 0.02725931815803051, -0.016917074099183083, 0.07398903369903564, -0.11142293363809586, -0.05420057475566864, 0.07131079584360123, 0.00017522448615636677, 0.13711436092853546, 0.08519066870212555, -0.10575205832719803, 0.1608172357082367, -0.12382415682077408, -0.056644801050424576, 0.11797840893268585, 0.0074418759904801846, 0.050511684268713, 0.03570277988910675, -0.018321089446544647, 0.048318617045879364, 0.04591407999396324, 0.048932358622550964, 0.03741343691945076, -0.0999658852815628, -0.02282877452671528, -0.057729750871658325, -0.1148778423666954, -0.08515831083059311, -0.027044016867876053, 0.10125566273927689, -0.01629621535539627, 0.10339433699846268, -0.08726240694522858, 0.10801132768392563, -0.08006537705659866, 0.013444874435663223, 0.006591364275664091, -0.15751560032367706, -0.08619159460067749, -0.06383692473173141, 0.036279238760471344, -0.029388047754764557, 0.17760130763053894, -0.019755007699131966, -0.005182204302400351, 0.030871329829096794, -0.016395168378949165, -0.01168188638985157, 0.024826066568493843, 0.2648567259311676, 0.04261944815516472, -0.10843217372894287, -0.0639122724533081, 0.06574015319347382, -0.01914919540286064, 0.023890914395451546, 0.1350153535604477, -0.003828831482678652, 0.10972201079130173, 0.08876129984855652, 0.005746243055909872, 0.03159472718834877, -0.03779241442680359, -0.1040029525756836, 0.001482845051214099, 0.04779302701354027, -0.010397776961326599, 0.05456288903951645, 0.1640816479921341, -0.07143199443817139, 0.0039795925840735435, -0.01002996601164341, -0.07117529958486557, -0.17887158691883087, -0.16366270184516907, -0.10530594736337662, -0.060769401490688324, 0.019913695752620697, -0.161081925034523, 0.06848270446062088, -0.021497240290045738, 0.04664139449596405, -0.07486061006784439, 0.08085466921329498, 0.10443606972694397, -0.11811155080795288, 0.10797092318534851, -0.02451854571700096, 0.03894788771867752, 0.011804025620222092, 0.030835168436169624, -0.050016775727272034, 0.06286969780921936, -0.010726607404649258, 0.05615726858377457, -0.07746545970439911, 0.018269198015332222, -0.18184438347816467, -0.12726320326328278, -0.030422940850257874, 0.03916741535067558, 0.0003359842812642455, 0.07872729003429413, 0.07855517417192459, -0.045163124799728394, 0.04764260724186897, 0.20738162100315094, -0.07398447394371033, -0.1045374646782875, -0.023633794859051704, 0.1414511501789093, 0.0523374080657959, 0.07870625704526901, 0.026478974148631096, -0.031166523694992065, -0.09295572340488434, 0.2916955053806305, 0.2669696509838104, -0.031840093433856964, 0.024861881509423256, -0.01131228357553482, 0.030655555427074432, 0.09602819383144379, 0.13856270909309387, 0.07178739458322525, 0.2547357678413391, -0.06891409307718277, 0.009557278826832771, -0.029259970411658287, -0.056100137531757355, -0.07058566808700562, 0.06937406957149506, 0.06757080554962158, -0.04952109605073929, -0.027773842215538025, 0.13015484809875488, -0.22022801637649536, 0.06527146697044373, -0.11274199187755585, -0.17195047438144684, -0.06769340485334396, -0.003549414686858654, 0.11553569883108139, 0.025867212563753128, 0.06929931044578552, 0.018015000969171524, -0.047917440533638, 0.05909794569015503, 0.03272463381290436, -0.20910649001598358, 0.044403325766325, 0.05030481889843941, -0.18334515392780304, 0.008507284335792065, -0.02192201465368271, 0.12356370687484741, 0.09216096252202988, 0.01715623028576374, -0.04188229516148567, 0.046647462993860245, 0.019858231768012047, 0.004706459119915962, 0.024247458204627037, 0.03574637696146965, 0.04167858511209488, -0.06116395443677902, 0.13154110312461853, -0.12377110868692398, 0.059437453746795654, -0.02207612246274948, -0.01617778278887272, -0.051059331744909286, 0.038034092634916306, -0.06564220786094666, 0.048679251223802567, 0.06022518500685692, -0.0346447192132473, 0.04353766143321991, -0.07530952990055084, -0.07014861702919006, -0.006826259661465883, -0.008663292042911053, -0.0499395914375782, -0.08878384530544281, -0.090488500893116, 0.10712049156427383, 0.025276314467191696, -0.2150489240884781, 0.02117997035384178, -0.11495615541934967, 0.024585455656051636, -0.17837341129779816, 0.13729168474674225, 0.06493893265724182, -0.004544015508145094, -0.0009455821709707379, -0.07390233874320984, 0.03382629528641701, 0.10528970509767532, -0.1486518681049347, -0.06735222786664963 ]
null
null
transformers
#keytotext [![pypi Version](https://img.shields.io/pypi/v/keytotext.svg?logo=pypi&logoColor=white)](https://pypi.org/project/keytotext/) [![Downloads](https://static.pepy.tech/personalized-badge/keytotext?period=total&units=none&left_color=grey&right_color=orange&left_text=Pip%20Downloads)](https://pepy.tech/project/keytotext) [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/gagan3012/keytotext/blob/master/notebooks/K2T.ipynb) [![Streamlit App](https://static.streamlit.io/badges/streamlit_badge_black_white.svg)](https://share.streamlit.io/gagan3012/keytotext/UI/app.py) [![API Call](https://img.shields.io/badge/-FastAPI-red?logo=fastapi&labelColor=white)](https://github.com/gagan3012/keytotext#api) [![Docker Call](https://img.shields.io/badge/-Docker%20Image-blue?logo=docker&labelColor=white)](https://hub.docker.com/r/gagan30/keytotext) [![HuggingFace](https://img.shields.io/badge/%F0%9F%A4%97-Models%20on%20Hub-yellow)](https://huggingface.co/models?filter=keytotext) [![Documentation Status](https://readthedocs.org/projects/keytotext/badge/?version=latest)](https://keytotext.readthedocs.io/en/latest/?badge=latest) [![Code style: black](https://img.shields.io/badge/code%20style-black-000000.svg)](https://github.com/psf/black) ![keytotext](https://socialify.git.ci/gagan3012/keytotext/image?description=1&forks=1&language=1&owner=1&stargazers=1&theme=Light) Idea is to build a model which will take keywords as inputs and generate sentences as outputs. Potential use case can include: - Marketing - Search Engine Optimization - Topic generation etc. - Fine tuning of topic modeling models
{"language": "en", "license": "MIT", "tags": ["keytotext", "k2t", "Keywords to Sentences"], "datasets": ["WebNLG", "Dart"], "metrics": ["NLG"], "thumbnail": "Keywords to Sentences"}
text2text-generation
gagan3012/k2t-test3
[ "transformers", "pytorch", "t5", "text2text-generation", "keytotext", "k2t", "Keywords to Sentences", "en", "dataset:WebNLG", "dataset:Dart", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #t5 #text2text-generation #keytotext #k2t #Keywords to Sentences #en #dataset-WebNLG #dataset-Dart #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
#keytotext ![pypi Version](URL ![Downloads](URL ![Open In Colab](URL ![Streamlit App](URL ![API Call](URL ![Docker Call](URL ![HuggingFace](URL ![Documentation Status](URL ![Code style: black](URL !keytotext Idea is to build a model which will take keywords as inputs and generate sentences as outputs. Potential use case can include: - Marketing - Search Engine Optimization - Topic generation etc. - Fine tuning of topic modeling models
[]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #keytotext #k2t #Keywords to Sentences #en #dataset-WebNLG #dataset-Dart #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 78 ]
[ "passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #keytotext #k2t #Keywords to Sentences #en #dataset-WebNLG #dataset-Dart #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.03903217613697052, 0.07688179612159729, -0.0061773881316185, 0.007946956902742386, 0.10117848217487335, 0.02843649871647358, 0.09411219507455826, 0.15868787467479706, 0.006724410690367222, -0.057724714279174805, 0.1404922604560852, 0.2025938332080841, 0.017940107733011246, 0.13757465779781342, -0.06942672282457352, -0.2564336359500885, 0.03815453499555588, 0.037285901606082916, -0.019459573552012444, 0.14391419291496277, 0.08486589789390564, -0.05302097275853157, 0.04979008436203003, -0.019126372411847115, -0.1216764822602272, 0.054594650864601135, 0.030159002169966698, -0.13637526333332062, 0.10469778627157211, 0.018528535962104797, 0.08310289680957794, 0.05006900057196617, -0.05860939249396324, -0.2161225974559784, 0.025030845776200294, 0.008722852915525436, -0.045329101383686066, 0.03354315459728241, 0.03338639438152313, -0.046357449144124985, 0.05640164762735367, -0.06243095174431801, -0.017651746049523354, 0.04220695421099663, -0.09832804650068283, -0.04187982156872749, -0.05007176101207733, 0.043057892471551895, 0.08319853991270065, 0.0737672746181488, -0.040723055601119995, 0.15581513941287994, -0.07629773765802383, 0.14664098620414734, 0.1368003636598587, -0.3269534111022949, 0.012729325331747532, -0.0035922133829444647, 0.004724225029349327, 0.06208911910653114, -0.02152496948838234, 0.06762268394231796, 0.012842950411140919, 0.017713449895381927, 0.04426150768995285, -0.09849993884563446, -0.24517373740673065, 0.04561421275138855, -0.07974745333194733, -0.020331554114818573, 0.289760559797287, -0.02673063613474369, 0.08189614117145538, -0.06428142637014389, -0.08825667947530746, 0.03182201087474823, -0.023756423965096474, -0.030087575316429138, -0.05645490437746048, 0.04441112279891968, -0.010185645893216133, -0.08027484267950058, -0.15373766422271729, 0.005433336831629276, -0.20648396015167236, 0.10115378350019455, 0.024363122880458832, 0.035511452704668045, -0.18737928569316864, 0.05462529510259628, 0.03907693922519684, -0.12055029720067978, 0.04595165699720383, -0.053805362433195114, 0.018455632030963898, -0.0044195144437253475, -0.03764479234814644, -0.08237682282924652, 0.15702591836452484, -0.021671166643500328, 0.003712681820616126, 0.003861523699015379, -0.08378283679485321, 0.05623997002840042, 0.062132399529218674, 0.07123706489801407, -0.02089962735772133, -0.07825490087270737, 0.05290557071566582, -0.1396183967590332, 0.009325655177235603, -0.04026453197002411, -0.09688538312911987, -0.06370905786752701, 0.10109986364841461, 0.08713588118553162, 0.08178628981113434, 0.10759539902210236, -0.025620993226766586, -0.017821846529841423, 0.012486998923122883, -0.07545977830886841, -0.027385571971535683, 0.0017266568029299378, -0.014242880046367645, 0.10150834172964096, 0.0165895763784647, 0.07835089415311813, -0.134853333234787, 0.019631991162896156, -0.07318528741598129, -0.0271466001868248, 0.03464220091700554, -0.08904528617858887, 0.05325685441493988, -0.0927010253071785, 0.0202367901802063, -0.1247885525226593, -0.17206141352653503, 0.01621382124722004, -0.03229447826743126, 0.0012832974316552281, -0.0709434524178505, -0.08504429459571838, -0.04151918739080429, 0.05391307175159454, -0.0692572221159935, 0.008550930768251419, -0.04580128937959671, 0.11258784681558609, -0.07574059069156647, 0.089997299015522, -0.11348620057106018, 0.042653366923332214, -0.15919475257396698, -0.07423004508018494, -0.053964316844940186, 0.10881427675485611, -0.014113583602011204, 0.12761420011520386, -0.08176441490650177, -0.007769408170133829, -0.03169988840818405, 0.034524861723184586, -0.045046985149383545, 0.24183803796768188, -0.12320122867822647, -0.10078071057796478, 0.2652226984500885, -0.039097171276807785, -0.16684569418430328, 0.07894851267337799, 0.0009547838126309216, 0.14258721470832825, 0.15320105850696564, 0.2657999098300934, 0.07426482439041138, -0.02078670635819435, 0.00763301644474268, 0.11893578618764877, -0.0869615450501442, -0.035517383366823196, 0.01223339419811964, -0.018862366676330566, -0.013180753216147423, 0.054628338664770126, 0.12370645999908447, 0.052965909242630005, -0.027910005301237106, -0.05075078457593918, -0.018453463912010193, -0.018193617463111877, 0.08231326192617416, -0.02094157226383686, 0.07350154966115952, -0.07670266926288605, -0.028671003878116608, 0.03127618506550789, 0.006361823063343763, -0.019136076793074608, 0.04919913783669472, -0.03604045882821083, -0.027698056772351265, -0.0022763845045119524, 0.052324894815683365, -0.1082569807767868, -0.12414075434207916, -0.012394402176141739, 0.18391548097133636, 0.03078749030828476, 0.0533563457429409, 0.030673427507281303, -0.07504592090845108, -0.01635466329753399, -0.0037628004793077707, 0.1724305897951126, 0.03408222645521164, -0.08487121015787125, -0.11941272765398026, 0.1011456772685051, -0.032992489635944366, -0.015726841986179352, 0.04496084898710251, -0.004146311432123184, 0.1094067245721817, 0.10359770059585571, -0.01346167828887701, 0.06498674303293228, 0.020176568999886513, -0.005791624076664448, -0.03126189485192299, 0.02556530199944973, 0.06685677170753479, -0.006117179058492184, -0.12800662219524384, 0.20311257243156433, -0.07922592759132385, 0.1098286360502243, 0.17013971507549286, -0.1630544364452362, 0.052374500781297684, -0.05337982624769211, -0.03546663746237755, -0.02642180025577545, 0.052608367055654526, -0.02748814783990383, 0.1092694103717804, 0.04293446242809296, 0.1276484578847885, -0.055808719247579575, -0.08623116463422775, -0.0014845937257632613, -0.07218868285417557, -0.03463004156947136, 0.07184917479753494, -0.05464067682623863, -0.2767590880393982, 0.151351660490036, 0.14284931123256683, 0.07120442390441895, 0.218695268034935, -0.020985165610909462, -0.021876879036426544, 0.03600084409117699, -0.02205980382859707, -0.0722251608967781, -0.06470290571451187, -0.13558830320835114, -0.0033426089212298393, 0.07773612439632416, 0.021976018324494362, 0.07477480918169022, -0.10188990086317062, -0.04983449727296829, -0.03033437579870224, -0.024645287543535233, -0.0009964666096493602, 0.09942738711833954, 0.06909600645303726, 0.13495886325836182, 0.01318643894046545, 0.04001353681087494, 0.08726658672094345, -0.011574335396289825, -0.12594421207904816, 0.183069109916687, -0.20184381306171417, -0.3897164463996887, -0.06920620054006577, -0.08202673494815826, -0.0433579757809639, -0.023706885054707527, 0.134678915143013, -0.158100888133049, 0.007603765465319157, 0.0012842394644394517, 0.04123978316783905, -0.06566179543733597, 0.0188005268573761, -0.06776919960975647, 0.02648961916565895, -0.09561283141374588, -0.0856005847454071, -0.05406622216105461, -0.013013672083616257, -0.069023497402668, 0.1326999068260193, -0.1190083771944046, 0.0657150074839592, 0.15551768243312836, 0.01996716484427452, 0.04895695298910141, -0.08226631581783295, 0.16687576472759247, -0.06638176739215851, 0.06961362808942795, 0.13507559895515442, -0.028373628854751587, 0.08798438310623169, 0.15129351615905762, -0.013650152832269669, -0.0232254546135664, 0.049437958747148514, 0.018354935571551323, -0.026459643617272377, -0.26186463236808777, -0.12826767563819885, -0.07437894493341446, 0.08185781538486481, -0.0017615138785913587, 0.04023335501551628, 0.14391960203647614, 0.06444190442562103, -0.027538392692804337, -0.020490268245339394, 0.022095942869782448, 0.06953368335962296, 0.25473225116729736, -0.02774951606988907, 0.1169879361987114, -0.06025315448641777, -0.11819146573543549, 0.08268919587135315, 0.11264616996049881, 0.050551220774650574, 0.01648484170436859, 0.10656298696994781, 0.05184439942240715, 0.11337745189666748, 0.04947604611515999, 0.07878274470567703, 0.03159871697425842, 0.0007107985438778996, -0.015850944444537163, -0.05846184492111206, -0.01383562758564949, 0.05181414261460304, 0.034767843782901764, -0.10243432223796844, -0.03443082794547081, -0.01150580495595932, 0.12185569107532501, 0.13546621799468994, 0.08351317793130875, -0.12537677586078644, 0.005218727979809046, 0.07765185832977295, -0.044701624661684036, -0.10390593111515045, 0.08798053115606308, -0.005960504990071058, -0.13519296050071716, 0.11843277513980865, -0.032555289566516876, 0.0880533754825592, -0.04732156917452812, 0.06625839322805405, -0.052118733525276184, -0.09939749538898468, 0.002937457524240017, 0.13495902717113495, -0.3500785231590271, 0.18479159474372864, 0.020522842183709145, -0.03739010915160179, -0.11735451966524124, -0.024229316040873528, -0.020654641091823578, 0.018029769882559776, 0.15043169260025024, 0.0040740673430264, 0.022580603137612343, -0.023493951186537743, -0.055756766349077225, 0.054868098348379135, 0.08672568202018738, -0.03158225491642952, 0.02804802730679512, -0.047701746225357056, -0.008736777119338512, 0.001647863071411848, -0.03301027789711952, -0.0661146268248558, -0.13900674879550934, 0.05676543340086937, 0.08871498703956604, -0.018053695559501648, 0.011931195855140686, -0.030276063829660416, -0.023456571623682976, 0.2034592479467392, -0.06985244899988174, -0.13113707304000854, -0.12003286927938461, -0.034163400530815125, 0.059256117790937424, -0.10904631018638611, 0.007828399538993835, -0.06547124683856964, 0.016343815252184868, -0.08073829859495163, -0.19135794043540955, 0.10531875491142273, -0.07745100557804108, -0.033804330974817276, -0.04372392222285271, 0.1793138086795807, -0.01073919702321291, 0.015988832339644432, 0.03128870576620102, 0.010686389170587063, -0.06278391927480698, -0.06007697433233261, 0.041904956102371216, -0.029453665018081665, 0.035968679934740067, 0.02807362750172615, -0.06198997050523758, -0.02893623523414135, -0.10285694152116776, -0.04338344931602478, 0.3439633548259735, 0.13939659297466278, -0.06957757472991943, 0.21206419169902802, 0.1378963142633438, -0.10091788321733475, -0.31464821100234985, -0.09116147458553314, -0.06621747463941574, -0.05524633452296257, -0.014028315432369709, -0.17787262797355652, 0.09584298729896545, -0.042744219303131104, -0.024806752800941467, 0.03977423906326294, -0.2556311786174774, -0.082899309694767, 0.16976392269134521, 0.01596090942621231, 0.2834286689758301, -0.13093051314353943, -0.08922819793224335, -0.03937128558754921, -0.12698093056678772, 0.19979402422904968, -0.08418729156255722, 0.06153452768921852, -0.004666736349463463, 0.1776805818080902, 0.04617675021290779, -0.011990601196885109, 0.012206112034618855, 0.003995159175246954, -0.02480645291507244, -0.06926066428422928, -0.09844497591257095, 0.08762900531291962, -0.008778671734035015, 0.020381171256303787, -0.06595955789089203, 0.05049559473991394, -0.0949656069278717, -0.0006382031133398414, -0.12513716518878937, 0.07758493721485138, -0.007955621927976608, -0.09927115589380264, -0.019221661612391472, -0.04650452733039856, 0.06919635832309723, -0.020818738266825676, 0.15716107189655304, -0.03872097283601761, 0.11444004625082016, 0.10615856200456619, 0.09696070849895477, -0.0019096651813015342, 0.06675493717193604, -0.062093157321214676, -0.09167221933603287, 0.027807479724287987, -0.13657724857330322, 0.05464359000325203, 0.11414086073637009, -0.018889576196670532, 0.042606230825185776, 0.06232642009854317, -0.030024895444512367, -0.03841933608055115, 0.10966654866933823, -0.24939408898353577, 0.008710821159183979, -0.09975350648164749, -0.07740998268127441, 0.03193715959787369, 0.056537844240665436, 0.20132313668727875, 0.019023699685931206, -0.06170305982232094, -0.020352233201265335, 0.05099378526210785, -0.04595917835831642, 0.08228398114442825, 0.03132554888725281, 0.029461823403835297, -0.13575321435928345, 0.10601908713579178, 0.015725204721093178, -0.04993676394224167, 0.08034761995077133, 0.17141593992710114, -0.11595481634140015, -0.08235012739896774, 0.018206652253866196, 0.09943213313817978, -0.09247054904699326, -0.02033846080303192, -0.05460567772388458, -0.09290233254432678, 0.046055182814598083, 0.2177998423576355, 0.020063476637005806, 0.1118638664484024, -0.04517297074198723, -0.04891025274991989, -0.047661323100328445, 0.10537919402122498, 0.0014809079002588987, -0.04026184603571892, -0.10279271751642227, 0.05047539249062538, -0.04830300062894821, 0.17378509044647217, -0.06399358063936234, -0.0592978410422802, -0.10721044987440109, 0.023348625749349594, -0.06541787832975388, -0.029305675998330116, -0.0824657455086708, -0.03429308906197548, -0.021074894815683365, -0.031610239297151566, -0.03873627260327339, -0.03024277836084366, -0.0764424055814743, 0.02725931815803051, -0.016917074099183083, 0.07398903369903564, -0.11142293363809586, -0.05420057475566864, 0.07131079584360123, 0.00017522448615636677, 0.13711436092853546, 0.08519066870212555, -0.10575205832719803, 0.1608172357082367, -0.12382415682077408, -0.056644801050424576, 0.11797840893268585, 0.0074418759904801846, 0.050511684268713, 0.03570277988910675, -0.018321089446544647, 0.048318617045879364, 0.04591407999396324, 0.048932358622550964, 0.03741343691945076, -0.0999658852815628, -0.02282877452671528, -0.057729750871658325, -0.1148778423666954, -0.08515831083059311, -0.027044016867876053, 0.10125566273927689, -0.01629621535539627, 0.10339433699846268, -0.08726240694522858, 0.10801132768392563, -0.08006537705659866, 0.013444874435663223, 0.006591364275664091, -0.15751560032367706, -0.08619159460067749, -0.06383692473173141, 0.036279238760471344, -0.029388047754764557, 0.17760130763053894, -0.019755007699131966, -0.005182204302400351, 0.030871329829096794, -0.016395168378949165, -0.01168188638985157, 0.024826066568493843, 0.2648567259311676, 0.04261944815516472, -0.10843217372894287, -0.0639122724533081, 0.06574015319347382, -0.01914919540286064, 0.023890914395451546, 0.1350153535604477, -0.003828831482678652, 0.10972201079130173, 0.08876129984855652, 0.005746243055909872, 0.03159472718834877, -0.03779241442680359, -0.1040029525756836, 0.001482845051214099, 0.04779302701354027, -0.010397776961326599, 0.05456288903951645, 0.1640816479921341, -0.07143199443817139, 0.0039795925840735435, -0.01002996601164341, -0.07117529958486557, -0.17887158691883087, -0.16366270184516907, -0.10530594736337662, -0.060769401490688324, 0.019913695752620697, -0.161081925034523, 0.06848270446062088, -0.021497240290045738, 0.04664139449596405, -0.07486061006784439, 0.08085466921329498, 0.10443606972694397, -0.11811155080795288, 0.10797092318534851, -0.02451854571700096, 0.03894788771867752, 0.011804025620222092, 0.030835168436169624, -0.050016775727272034, 0.06286969780921936, -0.010726607404649258, 0.05615726858377457, -0.07746545970439911, 0.018269198015332222, -0.18184438347816467, -0.12726320326328278, -0.030422940850257874, 0.03916741535067558, 0.0003359842812642455, 0.07872729003429413, 0.07855517417192459, -0.045163124799728394, 0.04764260724186897, 0.20738162100315094, -0.07398447394371033, -0.1045374646782875, -0.023633794859051704, 0.1414511501789093, 0.0523374080657959, 0.07870625704526901, 0.026478974148631096, -0.031166523694992065, -0.09295572340488434, 0.2916955053806305, 0.2669696509838104, -0.031840093433856964, 0.024861881509423256, -0.01131228357553482, 0.030655555427074432, 0.09602819383144379, 0.13856270909309387, 0.07178739458322525, 0.2547357678413391, -0.06891409307718277, 0.009557278826832771, -0.029259970411658287, -0.056100137531757355, -0.07058566808700562, 0.06937406957149506, 0.06757080554962158, -0.04952109605073929, -0.027773842215538025, 0.13015484809875488, -0.22022801637649536, 0.06527146697044373, -0.11274199187755585, -0.17195047438144684, -0.06769340485334396, -0.003549414686858654, 0.11553569883108139, 0.025867212563753128, 0.06929931044578552, 0.018015000969171524, -0.047917440533638, 0.05909794569015503, 0.03272463381290436, -0.20910649001598358, 0.044403325766325, 0.05030481889843941, -0.18334515392780304, 0.008507284335792065, -0.02192201465368271, 0.12356370687484741, 0.09216096252202988, 0.01715623028576374, -0.04188229516148567, 0.046647462993860245, 0.019858231768012047, 0.004706459119915962, 0.024247458204627037, 0.03574637696146965, 0.04167858511209488, -0.06116395443677902, 0.13154110312461853, -0.12377110868692398, 0.059437453746795654, -0.02207612246274948, -0.01617778278887272, -0.051059331744909286, 0.038034092634916306, -0.06564220786094666, 0.048679251223802567, 0.06022518500685692, -0.0346447192132473, 0.04353766143321991, -0.07530952990055084, -0.07014861702919006, -0.006826259661465883, -0.008663292042911053, -0.0499395914375782, -0.08878384530544281, -0.090488500893116, 0.10712049156427383, 0.025276314467191696, -0.2150489240884781, 0.02117997035384178, -0.11495615541934967, 0.024585455656051636, -0.17837341129779816, 0.13729168474674225, 0.06493893265724182, -0.004544015508145094, -0.0009455821709707379, -0.07390233874320984, 0.03382629528641701, 0.10528970509767532, -0.1486518681049347, -0.06735222786664963 ]
null
null
transformers
# keytotext ![keytotext (1)](https://user-images.githubusercontent.com/49101362/116334480-f5e57a00-a7dd-11eb-987c-186477f94b6e.png) Idea is to build a model which will take keywords as inputs and generate sentences as outputs. ### Keytotext is powered by Huggingface 🤗 [![pypi Version](https://img.shields.io/pypi/v/keytotext.svg?style=flat-square&logo=pypi&logoColor=white)](https://pypi.org/project/keytotext/) [![Downloads](https://static.pepy.tech/personalized-badge/keytotext?period=total&units=none&left_color=grey&right_color=orange&left_text=Pip%20Downloads)](https://pepy.tech/project/keytotext) [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/gagan3012/keytotext/blob/master/Examples/K2T.ipynb) [![Streamlit App](https://static.streamlit.io/badges/streamlit_badge_black_white.svg)](https://share.streamlit.io/gagan3012/keytotext/UI/app.py) ## Model: Keytotext is based on the Amazing T5 Model: - `k2t`: [Model](https://huggingface.co/gagan3012/k2t) - `k2t-tiny`: [Model](https://huggingface.co/gagan3012/k2t-tiny) - `k2t-base`: [Model](https://huggingface.co/gagan3012/k2t-base) Training Notebooks can be found in the [`Training Notebooks`](https://github.com/gagan3012/keytotext/tree/master/Training%20Notebooks) Folder ## Usage: Example usage: [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/gagan3012/keytotext/blob/master/Examples/K2T.ipynb) Example Notebooks can be found in the [`Notebooks`](https://github.com/gagan3012/keytotext/tree/master/Examples) Folder ``` pip install keytotext ``` ![carbon (3)](https://user-images.githubusercontent.com/49101362/116220679-90e64180-a755-11eb-9246-82d93d924a6c.png) ## UI: UI: [![Streamlit App](https://static.streamlit.io/badges/streamlit_badge_black_white.svg)](https://share.streamlit.io/gagan3012/keytotext/UI/app.py) ``` pip install streamlit-tags ``` This uses a custom streamlit component built by me: [GitHub](https://github.com/gagan3012/streamlit-tags) ![image](https://user-images.githubusercontent.com/49101362/116162205-fc042980-a6fd-11eb-892e-8f6902f193f4.png)
{"language": "en", "license": "mit", "tags": ["keytotext", "k2t-tiny", "Keywords to Sentences"], "datasets": ["WebNLG", "Dart"], "metrics": ["NLG"], "thumbnail": "Keywords to Sentences"}
text2text-generation
gagan3012/k2t-tiny
[ "transformers", "pytorch", "t5", "text2text-generation", "keytotext", "k2t-tiny", "Keywords to Sentences", "en", "dataset:WebNLG", "dataset:Dart", "license:mit", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #t5 #text2text-generation #keytotext #k2t-tiny #Keywords to Sentences #en #dataset-WebNLG #dataset-Dart #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# keytotext !keytotext (1) Idea is to build a model which will take keywords as inputs and generate sentences as outputs. ### Keytotext is powered by Huggingface ![pypi Version](URL ![Downloads](URL ![Open In Colab](URL ![Streamlit App](URL ## Model: Keytotext is based on the Amazing T5 Model: - 'k2t': Model - 'k2t-tiny': Model - 'k2t-base': Model Training Notebooks can be found in the 'Training Notebooks' Folder ## Usage: Example usage: ![Open In Colab](URL Example Notebooks can be found in the 'Notebooks' Folder !carbon (3) ## UI: UI: ![Streamlit App](URL This uses a custom streamlit component built by me: GitHub !image
[ "# keytotext\n\n!keytotext (1)\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.", "### Keytotext is powered by Huggingface \n\n![pypi Version](URL\n![Downloads](URL\n![Open In Colab](URL\n![Streamlit App](URL", "## Model:\n\nKeytotext is based on the Amazing T5 Model: \n\n- 'k2t': Model\n- 'k2t-tiny': Model\n- 'k2t-base': Model\n\nTraining Notebooks can be found in the 'Training Notebooks' Folder", "## Usage:\n\nExample usage: ![Open In Colab](URL\n\nExample Notebooks can be found in the 'Notebooks' Folder\n\n\n\n!carbon (3)", "## UI:\n\nUI: ![Streamlit App](URL\n\n\nThis uses a custom streamlit component built by me: GitHub\n\n!image" ]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #keytotext #k2t-tiny #Keywords to Sentences #en #dataset-WebNLG #dataset-Dart #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# keytotext\n\n!keytotext (1)\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.", "### Keytotext is powered by Huggingface \n\n![pypi Version](URL\n![Downloads](URL\n![Open In Colab](URL\n![Streamlit App](URL", "## Model:\n\nKeytotext is based on the Amazing T5 Model: \n\n- 'k2t': Model\n- 'k2t-tiny': Model\n- 'k2t-base': Model\n\nTraining Notebooks can be found in the 'Training Notebooks' Folder", "## Usage:\n\nExample usage: ![Open In Colab](URL\n\nExample Notebooks can be found in the 'Notebooks' Folder\n\n\n\n!carbon (3)", "## UI:\n\nUI: ![Streamlit App](URL\n\n\nThis uses a custom streamlit component built by me: GitHub\n\n!image" ]
[ 85, 31, 44, 59, 36, 30 ]
[ "passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #keytotext #k2t-tiny #Keywords to Sentences #en #dataset-WebNLG #dataset-Dart #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# keytotext\n\n!keytotext (1)\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.### Keytotext is powered by Huggingface \n\n![pypi Version](URL\n![Downloads](URL\n![Open In Colab](URL\n![Streamlit App](URL## Model:\n\nKeytotext is based on the Amazing T5 Model: \n\n- 'k2t': Model\n- 'k2t-tiny': Model\n- 'k2t-base': Model\n\nTraining Notebooks can be found in the 'Training Notebooks' Folder## Usage:\n\nExample usage: ![Open In Colab](URL\n\nExample Notebooks can be found in the 'Notebooks' Folder\n\n\n\n!carbon (3)## UI:\n\nUI: ![Streamlit App](URL\n\n\nThis uses a custom streamlit component built by me: GitHub\n\n!image" ]
[ 0.024995075538754463, 0.14030449092388153, -0.00688745966181159, 0.05371125787496567, 0.09252085536718369, 0.011265386827290058, 0.17432823777198792, 0.1297273337841034, -0.02333986386656761, -0.010075799189507961, 0.08039990812540054, 0.12820541858673096, 0.058661989867687225, 0.19613204896450043, 0.012298434972763062, -0.23589542508125305, 0.03694259002804756, -0.007861614227294922, -0.0015794665087014437, 0.10738363116979599, 0.08568063378334045, -0.04710492491722107, 0.08283750712871552, 0.05674067884683609, -0.10011783987283707, 0.020098431035876274, -0.027024751529097557, -0.1080639660358429, 0.04501970484852791, 0.061884623020887375, 0.06174857169389725, 0.030151894316077232, -0.013569816015660763, -0.23539257049560547, 0.01983245089650154, 0.07639086991548538, 0.028337780386209488, 0.04778263717889786, 0.048111531883478165, 0.016737984493374825, 0.14135541021823883, 0.02783457189798355, 0.050733111798763275, 0.023331215605139732, -0.09024925529956818, -0.17152327299118042, -0.09124598652124405, 0.10769323259592056, 0.05629228800535202, 0.05544678121805191, -0.0019970491994172335, 0.1136937215924263, -0.001226095831952989, 0.08358126878738403, 0.21705745160579681, -0.1397777646780014, -0.04068900644779205, -0.03797862306237221, 0.07277955859899521, 0.027733780443668365, -0.02027874067425728, 0.0020451312884688377, -0.015332418493926525, 0.00033808738226071, 0.1312500536441803, -0.07180513441562653, -0.14757685363292694, -0.021014844998717308, -0.07561668753623962, -0.03761134669184685, 0.15363186597824097, -0.020054690539836884, -0.011228138580918312, -0.169596865773201, -0.02356056310236454, 0.06053680554032326, -0.03615835681557655, -0.002490988001227379, 0.01076037809252739, 0.01065027341246605, 0.04052159562706947, -0.1717325747013092, -0.11735369265079498, -0.01813943311572075, 0.06888317316770554, 0.0232700128108263, 0.010938100516796112, 0.025154268369078636, -0.054037466645240784, 0.1315704584121704, 0.033614397048950195, -0.139610156416893, -0.010425319895148277, -0.06688583642244339, 0.035928260535001755, -0.006205300334841013, 0.0655200332403183, -0.0892176702618599, 0.101591095328331, -0.004818984307348728, 0.08163543045520782, 0.04232843965291977, -0.013047019951045513, 0.041932910680770874, 0.05688682198524475, 0.23163431882858276, -0.03632815554738045, -0.0457066148519516, 0.0586722195148468, 0.04556727409362793, -0.002144832629710436, -0.03253912180662155, -0.04361628368496895, 0.015914209187030792, -0.00441478006541729, 0.04869350790977478, 0.09074681252241135, 0.10236667841672897, -0.010241974145174026, -0.037559133023023605, 0.09941548109054565, -0.06977292895317078, 0.037259168922901154, 0.0540984645485878, -0.023715848103165627, 0.059634625911712646, 0.10693136602640152, 0.002064355416223407, -0.11836988478899002, -0.019567903131246567, -0.028937654569745064, 0.0015560077736154199, -0.0983179584145546, -0.11314541846513748, 0.021165743470191956, -0.07746593654155731, -0.029063653200864792, -0.07221050560474396, -0.08986565470695496, -0.0376027375459671, 0.08036952465772629, 0.016572941094636917, -0.04944324493408203, -0.047852735966444016, -0.019545378163456917, 0.013820881024003029, 0.03495611250400543, -0.10439212620258331, -0.01345013827085495, 0.06653877347707748, -0.10364475101232529, 0.032698217779397964, -0.048123545944690704, 0.053627438843250275, -0.11984430998563766, 0.022258717566728592, -0.23762722313404083, 0.09764107316732407, -0.04997044801712036, 0.10432974249124527, -0.10805120319128036, -0.012822129763662815, 0.10111555457115173, 0.0029059078078716993, 0.012652317993342876, 0.09276165068149567, -0.09379729628562927, -0.008533555082976818, 0.12301164865493774, -0.08939606696367264, -0.10744664818048477, 0.053236864507198334, -0.04869657754898071, 0.20688855648040771, 0.09427877515554428, 0.22767096757888794, 0.1959385871887207, -0.08367648720741272, -0.021181650459766388, 0.0414641797542572, -0.026042822748422623, 0.022187331691384315, 0.02338891662657261, -0.0547923780977726, 0.03111194260418415, 0.03133793920278549, -0.007468126714229584, 0.027101224288344383, 0.044605035334825516, -0.06286831945180893, 0.0072025759145617485, -0.03001459501683712, -0.04274512827396393, -0.05418230593204498, -0.006383947096765041, -0.025855643674731255, -0.05588292330503464, 0.14454308152198792, 0.03468687832355499, -0.12386874854564667, -0.0030052009969949722, -0.03681667521595955, 0.0095125213265419, 0.004086702596396208, 0.03379896655678749, -0.0313272699713707, -0.0623643696308136, 0.05532481148838997, -0.08447068929672241, 0.05694752559065819, -0.02445484697818756, 0.03358250856399536, 0.04840325936675072, 0.039110664278268814, -0.02607872150838375, -0.02864030748605728, 0.025604255497455597, -0.0454937107861042, -0.09026117622852325, 0.017043760046362877, -0.005110852420330048, 0.007806000299751759, -0.1556251049041748, 0.04973781853914261, 0.0695522353053093, 0.03620342165231705, 0.07226837426424026, -0.00970686599612236, 0.06609852612018585, -0.03748936578631401, 0.00871951226145029, -0.0463426411151886, 0.02686132863163948, 0.060644738376140594, -0.02107195183634758, 0.09356561303138733, -0.14310133457183838, -0.014653549529612064, 0.06960168480873108, -0.004292188212275505, -0.028972366824746132, -0.03430827707052231, -0.023994747549295425, -0.017181729897856712, 0.005894950125366449, -0.005103824660181999, 0.15037894248962402, 0.06791122257709503, 0.10749312490224838, -0.08033598214387894, -0.07713249325752258, 0.002223394112661481, -0.12140020728111267, 0.027172399684786797, 0.07302220165729523, 0.031447380781173706, -0.14981307089328766, 0.07986278831958771, 0.055649060755968094, -0.005707486066967249, 0.2301097810268402, -0.014875931665301323, -0.03197643160820007, -0.028906861320137978, 0.02887301705777645, 0.016388533636927605, 0.04604732245206833, -0.028350675478577614, 0.050029460340738297, 0.04890911281108856, -0.04173046723008156, 0.03493441268801689, -0.13820695877075195, 0.02380211651325226, 0.04499790817499161, -0.029277844354510307, -0.04621884226799011, 0.0226488895714283, 0.0368768647313118, 0.01915420964360237, 0.017963962629437447, 0.11160967499017715, 0.0015846412861719728, -0.012469417415559292, -0.11443059146404266, 0.16639703512191772, -0.13787004351615906, -0.2510277032852173, -0.1435827761888504, -0.03719417750835419, 0.04887444153428078, -0.020948093384504318, 0.09444747120141983, -0.010006133466959, -0.016402512788772583, -0.0949258804321289, 0.08114266395568848, 0.05110377073287964, -0.11469253152608871, -0.08276130259037018, 0.06017355993390083, 0.0031711161136627197, -0.14071202278137207, -0.02994425594806671, 0.07336137443780899, -0.07513703405857086, 0.02906428650021553, 0.006128497421741486, 0.05873383209109306, 0.05683797225356102, -0.055179398506879807, 0.007499208208173513, -0.039867252111434937, 0.2108268439769745, -0.061787575483322144, 0.08674777299165726, 0.16877318918704987, -0.10944821685552597, 0.12033623456954956, 0.16460946202278137, 0.03782214969396591, 0.011952975764870644, 0.012491719797253609, -0.02053709700703621, -0.040984466671943665, -0.1948627531528473, 0.014648974873125553, -0.02470399998128414, 0.09925607591867447, 0.0636381134390831, 0.01497957669198513, 0.038701605051755905, 0.11261492222547531, -0.0338454507291317, 0.06924333423376083, 0.0032070144079625607, 0.0814962312579155, 0.1643826812505722, 0.0034779796842485666, 0.020596878603100777, -0.10533605515956879, -0.02933509647846222, 0.061456069350242615, 0.08794375509023666, 0.1142175942659378, -0.048174258321523666, 0.1688806265592575, 0.09050484001636505, 0.04551348835229874, 0.02083335630595684, 0.0746377483010292, -0.07060838490724564, 0.06643718481063843, -0.016097573563456535, -0.1269092708826065, 0.022002290934324265, 0.0583825521171093, -0.026364872232079506, -0.08344146609306335, 0.03601553291082382, 0.014193377457559109, 0.09740503132343292, 0.24947267770767212, 0.0019145484548062086, -0.149860218167305, 0.0055871568620204926, 0.032867807894945145, 0.002049689879640937, -0.113380067050457, -0.02331884205341339, 0.08590183407068253, -0.17602871358394623, 0.005818719509989023, -0.0014576609246432781, 0.0989941880106926, -0.19248543679714203, -0.003316704649478197, 0.10395124554634094, 0.06316983699798584, -0.014354513958096504, 0.10439068078994751, -0.11109405010938644, 0.09743006527423859, 0.0071489145047962666, 0.07277993112802505, -0.05867486074566841, 0.05102211609482765, 0.012042777612805367, -0.01681806705892086, 0.049814894795417786, 0.015716996043920517, 0.03259989246726036, -0.10192497074604034, -0.09167812019586563, 0.05752389878034592, -0.008040985092520714, -0.10525090247392654, 0.07105924189090729, -0.044375840574502945, 0.0011613416718319058, -0.02827291376888752, -0.15093295276165009, -0.11699921637773514, -0.19974887371063232, 0.04309593886137009, -0.03145726025104523, 0.029705215245485306, -0.07991917431354523, 0.02349832095205784, 0.06639472395181656, 0.1689148247241974, 0.007068308535963297, -0.15462934970855713, -0.09356410801410675, -0.10066873580217361, 0.045134373009204865, -0.09922536462545395, 0.03539860621094704, -0.06124498322606087, 0.05051109939813614, -0.034565556794404984, -0.11857955157756805, 0.08569219708442688, -0.055327318608760834, -0.08602004498243332, -0.023118451237678528, 0.05210811644792557, -0.007624072954058647, -0.014451192691922188, 0.015202612616121769, -0.004178833682090044, -0.040827784687280655, -0.11222334951162338, -0.04568849503993988, 0.11096150428056717, -0.04349927976727486, -0.014702445827424526, -0.05664420500397682, -0.11352772265672684, -0.08039477467536926, 0.0006884374306537211, 0.19067911803722382, 0.10307908803224564, -0.0958952009677887, 0.10408475250005722, 0.15647323429584503, -0.11996016651391983, -0.2171185463666916, -0.026848984882235527, 0.04824924096465111, -0.04027858003973961, 0.024906374514102936, -0.2615738809108734, 0.08596266061067581, 0.013591689057648182, -0.018340030685067177, 0.11762043833732605, -0.2417357712984085, -0.09058918803930283, 0.057625629007816315, 0.0931064561009407, -0.1376917064189911, -0.11142109334468842, -0.06353124976158142, -0.03174322098493576, -0.14642155170440674, 0.14831919968128204, -0.03629058972001076, 0.07754133641719818, 0.016388971358537674, 0.07810371369123459, 0.05350594222545624, -0.040558282285928726, 0.03458939865231514, -0.16060391068458557, 0.021894583478569984, -0.09199047088623047, -0.1246395856142044, 0.034243326634168625, -0.06829296052455902, 0.13982440531253815, -0.11894640326499939, 0.05156601592898369, -0.06486575305461884, -0.008677111007273197, -0.04010757431387901, 0.07085780799388885, -0.060880403965711594, -0.08985516428947449, -0.08519288152456284, -0.02909078262746334, 0.0657423734664917, -0.004646745510399342, -0.03257936239242554, -0.030250458046793938, -0.005706427618861198, 0.14894850552082062, 0.017018843442201614, 0.16744406521320343, -0.11421959847211838, -0.04732290282845497, -0.024918116629123688, 0.0487566813826561, -0.17661286890506744, 0.02477722056210041, 0.06380817294120789, -0.01215664017945528, 0.12485647201538086, -0.018924569711089134, -0.12311534583568573, 0.028548166155815125, 0.07638831436634064, -0.1624039262533188, -0.10081293433904648, -0.05961509793996811, 0.07379063963890076, -0.10291638970375061, -0.04201006144285202, 0.12379484623670578, -0.07828745990991592, -0.03315090015530586, -0.021111030131578445, 0.025777174159884453, -0.0010654849465936422, 0.05893583968281746, 0.030722051858901978, 0.030809545889496803, -0.051285792142152786, 0.03775995224714279, 0.07207036018371582, 0.04278387874364853, 0.046475380659103394, 0.1424095630645752, -0.05833388492465019, -0.03330834209918976, 0.11601122468709946, 0.10781192779541016, 0.05517483130097389, -0.029436107724905014, -0.026596970856189728, -0.05091795325279236, -0.007562849670648575, 0.13677987456321716, 0.015907656401395798, 0.0360708087682724, -0.011623965576291084, 0.0029856995679438114, -0.017767958343029022, 0.14234668016433716, -0.018568402156233788, 0.0055197980254888535, -0.10848535597324371, 0.03542444109916687, 0.022682908922433853, 0.03842630237340927, 0.016335943713784218, -0.04176370054483414, -0.10396482795476913, -0.04798395559191704, -0.034115277230739594, 0.06573821604251862, -0.05604863166809082, 0.028306081891059875, 0.006032607983797789, 0.040795087814331055, 0.007119992282241583, 0.01803656667470932, -0.08085919171571732, -0.035212356597185135, -0.03961751237511635, 0.1468965858221054, -0.13952910900115967, -0.017506299540400505, 0.07950784265995026, -0.07586473226547241, 0.10129406303167343, -0.025392016395926476, -0.07987003028392792, 0.0014988231705501676, -0.1064298003911972, -0.0072630541399121284, 0.03142418712377548, 0.03297394514083862, 0.027351731434464455, -0.03233439475297928, 0.008069573901593685, -0.04995797201991081, 0.0035353489220142365, -0.02912072464823723, 0.11755695939064026, -0.12421239167451859, 0.02736419253051281, 0.03772677481174469, -0.07819084078073502, -0.032433487474918365, 0.0449192188680172, 0.054926201701164246, -0.01773894764482975, 0.13462409377098083, -0.04258014261722565, 0.07930361479520798, -0.11771156638860703, -0.01524745486676693, 0.06309600919485092, -0.07034765183925629, -0.050994813442230225, -0.055825669318437576, 0.008596843108534813, -0.02199074625968933, 0.0741553008556366, 0.023777376860380173, -0.08082462847232819, -0.01230253092944622, -0.04456142336130142, 0.008751473389565945, 0.05204923823475838, 0.1169717013835907, -0.05228330194950104, -0.023981144651770592, -0.054632339626550674, -0.03436471149325371, -0.00836858805269003, -0.0635390505194664, 0.018316812813282013, 0.0637097954750061, 0.04964745417237282, 0.03718689829111099, 0.08868524432182312, 0.05001123994588852, -0.11291985958814621, -0.01610209420323372, 0.060130346566438675, 0.15483127534389496, -0.12063876539468765, 0.14226558804512024, 0.15535281598567963, -0.07219333946704865, 0.04679069668054581, 0.021300723776221275, -0.06634090095758438, -0.09045311063528061, -0.18765735626220703, -0.07296416163444519, -0.061890069395303726, -0.01919015310704708, -0.09579436480998993, 0.08710591495037079, -0.020360056310892105, -0.004484465345740318, -0.059302859008312225, 0.15330356359481812, 0.059149935841560364, -0.07101581990718842, 0.07184372842311859, -0.020942028611898422, -0.019514063373208046, -0.039126716554164886, 0.046257272362709045, 0.013727505691349506, 0.0644061341881752, 0.046856168657541275, 0.09585399925708771, -0.04645054042339325, 0.006356291007250547, -0.12016595155000687, -0.08873769640922546, 0.006450107786804438, 0.021090375259518623, -0.014215377159416676, 0.055957455188035965, 0.05502203851938248, -0.05656920745968819, -0.00995885394513607, 0.1593572497367859, -0.060573577880859375, -0.12035110592842102, -0.14424405992031097, 0.061720144003629684, -0.05160142853856087, 0.002509242622181773, -0.0025318986736238003, -0.04859745875000954, -0.05394479259848595, 0.25779780745506287, 0.20950859785079956, 0.013542374595999718, 0.015420363284647465, -0.031024854630231857, 0.014986719936132431, -0.02616109512746334, 0.156338170170784, -0.0001880061608972028, 0.15049150586128235, -0.020607395097613335, 0.06740039587020874, -0.05099349096417427, -0.0868765264749527, -0.10692528635263443, 0.02908043935894966, 0.042772579938173294, 0.010193194262683392, -0.03153599798679352, 0.1278180330991745, -0.07608497887849808, -0.12145509570837021, -0.1163143515586853, -0.043327510356903076, -0.07670149207115173, 0.01804991438984871, 0.12396638840436935, -0.02913801744580269, 0.05774857848882675, 0.008823325857520103, -0.04833270236849785, 0.10012924671173096, 0.004694374278187752, -0.09403476119041443, 0.03185480460524559, 0.054119858890771866, -0.2229204773902893, 0.22270438075065613, 0.012450028210878372, 0.006130959838628769, 0.08371143043041229, -0.01946672797203064, -0.09922774136066437, 0.05232040584087372, 0.04832000285387039, -0.07687369734048843, -0.0325583852827549, 0.023545756936073303, -0.028822751715779305, -0.010750970803201199, 0.04936787113547325, -0.024803631007671356, 0.07584589719772339, -0.08026382327079773, 0.004706664010882378, -0.08076269179582596, -0.021631568670272827, -0.08529434353113174, 0.07597936689853668, 0.16719844937324524, -0.022540321573615074, 0.0031857448630034924, -0.03834319859743118, -0.012019718065857887, 0.047733571380376816, -0.06563003361225128, -0.003621784271672368, -0.062123969197273254, -0.05734916403889656, 0.0489530973136425, 0.10569150000810623, -0.07805590331554413, -0.00024635778390802443, -0.03408396244049072, 0.021529333665966988, -0.08025958389043808, 0.10917623341083527, 0.12972094118595123, -0.00016000917821656913, 0.007076601963490248, -0.1120702475309372, -0.06294792145490646, 0.10333576798439026, -0.06960733234882355, -0.07543110847473145 ]
null
null
transformers
# keytotext ![keytotext (1)](https://user-images.githubusercontent.com/49101362/116334480-f5e57a00-a7dd-11eb-987c-186477f94b6e.png) Idea is to build a model which will take keywords as inputs and generate sentences as outputs. ### Keytotext is powered by Huggingface 🤗 [![pypi Version](https://img.shields.io/pypi/v/keytotext.svg?style=flat-square&logo=pypi&logoColor=white)](https://pypi.org/project/keytotext/) [![Downloads](https://static.pepy.tech/personalized-badge/keytotext?period=total&units=none&left_color=grey&right_color=orange&left_text=Pip%20Downloads)](https://pepy.tech/project/keytotext) [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/gagan3012/keytotext/blob/master/Examples/K2T.ipynb) [![Streamlit App](https://static.streamlit.io/badges/streamlit_badge_black_white.svg)](https://share.streamlit.io/gagan3012/keytotext/UI/app.py) ## Model: Keytotext is based on the Amazing T5 Model: - `k2t`: [Model](https://huggingface.co/gagan3012/k2t) - `k2t-tiny`: [Model](https://huggingface.co/gagan3012/k2t-tiny) - `k2t-base`: [Model](https://huggingface.co/gagan3012/k2t-base) Training Notebooks can be found in the [`Training Notebooks`](https://github.com/gagan3012/keytotext/tree/master/Training%20Notebooks) Folder ## Usage: Example usage: [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/gagan3012/keytotext/blob/master/Examples/K2T.ipynb) Example Notebooks can be found in the [`Notebooks`](https://github.com/gagan3012/keytotext/tree/master/Examples) Folder ``` pip install keytotext ``` ![carbon (3)](https://user-images.githubusercontent.com/49101362/116220679-90e64180-a755-11eb-9246-82d93d924a6c.png) ## UI: UI: [![Streamlit App](https://static.streamlit.io/badges/streamlit_badge_black_white.svg)](https://share.streamlit.io/gagan3012/keytotext/UI/app.py) ``` pip install streamlit-tags ``` This uses a custom streamlit component built by me: [GitHub](https://github.com/gagan3012/streamlit-tags) ![image](https://user-images.githubusercontent.com/49101362/116162205-fc042980-a6fd-11eb-892e-8f6902f193f4.png)
{"language": "en", "license": "mit", "tags": ["keytotext", "k2t", "Keywords to Sentences"], "datasets": ["WebNLG", "Dart"], "metrics": ["NLG"], "thumbnail": "Keywords to Sentences"}
text2text-generation
gagan3012/k2t
[ "transformers", "pytorch", "t5", "text2text-generation", "keytotext", "k2t", "Keywords to Sentences", "en", "dataset:WebNLG", "dataset:Dart", "license:mit", "autotrain_compatible", "endpoints_compatible", "has_space", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #t5 #text2text-generation #keytotext #k2t #Keywords to Sentences #en #dataset-WebNLG #dataset-Dart #license-mit #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us
# keytotext !keytotext (1) Idea is to build a model which will take keywords as inputs and generate sentences as outputs. ### Keytotext is powered by Huggingface ![pypi Version](URL ![Downloads](URL ![Open In Colab](URL ![Streamlit App](URL ## Model: Keytotext is based on the Amazing T5 Model: - 'k2t': Model - 'k2t-tiny': Model - 'k2t-base': Model Training Notebooks can be found in the 'Training Notebooks' Folder ## Usage: Example usage: ![Open In Colab](URL Example Notebooks can be found in the 'Notebooks' Folder !carbon (3) ## UI: UI: ![Streamlit App](URL This uses a custom streamlit component built by me: GitHub !image
[ "# keytotext\n\n!keytotext (1)\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.", "### Keytotext is powered by Huggingface \n\n![pypi Version](URL\n![Downloads](URL\n![Open In Colab](URL\n![Streamlit App](URL", "## Model:\n\nKeytotext is based on the Amazing T5 Model: \n\n- 'k2t': Model\n- 'k2t-tiny': Model\n- 'k2t-base': Model\n\nTraining Notebooks can be found in the 'Training Notebooks' Folder", "## Usage:\n\nExample usage: ![Open In Colab](URL\n\nExample Notebooks can be found in the 'Notebooks' Folder\n\n\n\n!carbon (3)", "## UI:\n\nUI: ![Streamlit App](URL\n\n\nThis uses a custom streamlit component built by me: GitHub\n\n!image" ]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #keytotext #k2t #Keywords to Sentences #en #dataset-WebNLG #dataset-Dart #license-mit #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n", "# keytotext\n\n!keytotext (1)\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.", "### Keytotext is powered by Huggingface \n\n![pypi Version](URL\n![Downloads](URL\n![Open In Colab](URL\n![Streamlit App](URL", "## Model:\n\nKeytotext is based on the Amazing T5 Model: \n\n- 'k2t': Model\n- 'k2t-tiny': Model\n- 'k2t-base': Model\n\nTraining Notebooks can be found in the 'Training Notebooks' Folder", "## Usage:\n\nExample usage: ![Open In Colab](URL\n\nExample Notebooks can be found in the 'Notebooks' Folder\n\n\n\n!carbon (3)", "## UI:\n\nUI: ![Streamlit App](URL\n\n\nThis uses a custom streamlit component built by me: GitHub\n\n!image" ]
[ 87, 31, 44, 59, 36, 30 ]
[ "passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #keytotext #k2t #Keywords to Sentences #en #dataset-WebNLG #dataset-Dart #license-mit #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n# keytotext\n\n!keytotext (1)\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.### Keytotext is powered by Huggingface \n\n![pypi Version](URL\n![Downloads](URL\n![Open In Colab](URL\n![Streamlit App](URL## Model:\n\nKeytotext is based on the Amazing T5 Model: \n\n- 'k2t': Model\n- 'k2t-tiny': Model\n- 'k2t-base': Model\n\nTraining Notebooks can be found in the 'Training Notebooks' Folder## Usage:\n\nExample usage: ![Open In Colab](URL\n\nExample Notebooks can be found in the 'Notebooks' Folder\n\n\n\n!carbon (3)## UI:\n\nUI: ![Streamlit App](URL\n\n\nThis uses a custom streamlit component built by me: GitHub\n\n!image" ]
[ 0.022291600704193115, 0.14366869628429413, -0.006325175985693932, 0.06701385974884033, 0.10081034898757935, 0.01401167269796133, 0.16786429286003113, 0.12444747984409332, -0.03176899254322052, -0.02147694118320942, 0.07114984095096588, 0.14324456453323364, 0.05764023959636688, 0.19172796607017517, 0.021508794277906418, -0.24650834500789642, 0.03117373213171959, -0.0019516743486747146, -0.0003842453006654978, 0.10903267562389374, 0.08014189451932907, -0.04874717816710472, 0.08721155673265457, 0.06654877960681915, -0.11160929501056671, 0.022266613319516182, -0.03257383406162262, -0.11623802036046982, 0.05732593685388565, 0.053252048790454865, 0.06102795526385307, 0.03544801473617554, -0.016754258424043655, -0.2049153447151184, 0.02186303213238716, 0.07407193630933762, 0.024834858253598213, 0.04469124600291252, 0.043097738176584244, -0.0037474718410521746, 0.136617973446846, 0.03926548734307289, 0.055095646530389786, 0.012672059237957, -0.09386029094457626, -0.18562249839305878, -0.08821094036102295, 0.09402373433113098, 0.07889734953641891, 0.0587194599211216, 0.0004711266083177179, 0.09722766280174255, 0.005322757642716169, 0.09101824462413788, 0.2264639139175415, -0.1531449258327484, -0.04592042416334152, -0.0255826935172081, 0.06044980138540268, 0.02574883960187435, -0.02093898504972458, 0.003179749706760049, -0.022116756066679955, -0.006470547057688236, 0.14773260056972504, -0.07081420719623566, -0.1671750396490097, -0.023040512576699257, -0.0786372646689415, -0.03242846950888634, 0.15074913203716278, -0.023515714332461357, -0.017292115837335587, -0.16399165987968445, -0.01569240726530552, 0.05099000781774521, -0.026432646438479424, 0.006698263343423605, 0.01758384332060814, 0.0016074756858870387, 0.025148313492536545, -0.16858038306236267, -0.11666915565729141, -0.015425346791744232, 0.06636879593133926, 0.022136086598038673, 0.018735267221927643, 0.027497831732034683, -0.04963213577866554, 0.15660496056079865, 0.02251764014363289, -0.13521061837673187, -0.005411303602159023, -0.06772743165493011, 0.0314314067363739, -0.005028547719120979, 0.06309635937213898, -0.13313539326190948, 0.07679663598537445, -0.019015146419405937, 0.08324501663446426, 0.037293337285518646, -0.022931640967726707, 0.05086406692862511, 0.04613116383552551, 0.24017247557640076, -0.043900493532419205, -0.054154928773641586, 0.039795227348804474, 0.0433650016784668, -0.007173685356974602, -0.029933463782072067, -0.050929661840200424, -0.0003343747230246663, -0.0050294408574700356, 0.037742942571640015, 0.06949231773614883, 0.11333812028169632, -0.011742147617042065, -0.03307401016354561, 0.0851207822561264, -0.056141823530197144, 0.028778938576579094, 0.037251345813274384, -0.040622226893901825, 0.06358257681131363, 0.12812203168869019, 0.0035930010490119457, -0.10818260908126831, 0.0019075130112469196, -0.02591337077319622, 0.008749696426093578, -0.10599445551633835, -0.1145147904753685, 0.015331320464611053, -0.09108850359916687, -0.02146848663687706, -0.07769231498241425, -0.09391078352928162, -0.02493879944086075, 0.09145358204841614, 0.022591937333345413, -0.033519528806209564, -0.0517214871942997, -0.020511439070105553, 0.007029294036328793, 0.03541480377316475, -0.08678470551967621, -0.010807734914124012, 0.06253182888031006, -0.10476846992969513, 0.05489591509103775, -0.05873965844511986, 0.06064549833536148, -0.111722432076931, 0.01832789182662964, -0.2595916986465454, 0.10240332037210464, -0.046625155955553055, 0.09979960322380066, -0.1048014909029007, -0.021635746583342552, 0.10789041966199875, -0.006418358068913221, 0.016216693446040154, 0.0826629251241684, -0.09342444688081741, -0.006628822535276413, 0.13674098253250122, -0.09133647382259369, -0.11041688174009323, 0.04304220899939537, -0.05678139254450798, 0.2006136029958725, 0.09497205913066864, 0.22624404728412628, 0.17996057868003845, -0.09805156290531158, -0.02583463117480278, 0.03765176609158516, -0.04118731617927551, 0.02862188220024109, 0.02815132588148117, -0.03106788732111454, 0.03088035248219967, 0.03702470660209656, -0.021712107583880424, 0.0441720187664032, 0.03994910418987274, -0.057694289833307266, 0.007038130424916744, -0.03423401340842247, -0.02485601417720318, -0.05915321037173271, -0.005312096327543259, -0.013517370447516441, -0.06816482543945312, 0.1428651064634323, 0.03400332108139992, -0.13007549941539764, -0.006204906385391951, -0.03338337317109108, 0.022275619208812714, 0.002882251050323248, 0.026355329900979996, -0.034866128116846085, -0.05342342332005501, 0.043593283742666245, -0.08427619189023972, 0.05610998347401619, -0.021145576611161232, 0.028601281344890594, 0.06707987934350967, 0.03275243192911148, -0.03152012079954147, -0.04562782868742943, 0.023925352841615677, -0.03615910932421684, -0.07753700762987137, 0.01892245188355446, -0.004184740595519543, 0.024379834532737732, -0.1526339203119278, 0.05125151947140694, 0.07051065564155579, 0.027225246652960777, 0.07976266741752625, -0.008640522137284279, 0.0630321279168129, -0.03888492286205292, 0.004523749928921461, -0.041818201541900635, 0.024259407073259354, 0.0603303499519825, -0.014961488544940948, 0.10419974476099014, -0.13785317540168762, -0.016522852703928947, 0.07818256318569183, -0.00831534806638956, -0.0393846370279789, -0.04035515710711479, -0.0237126424908638, -0.021566595882177353, 0.007161118555814028, 0.0036948604974895716, 0.16652996838092804, 0.0632794126868248, 0.12500087916851044, -0.0781930685043335, -0.06321500986814499, 0.010223624296486378, -0.11795705556869507, 0.03196929022669792, 0.06303264945745468, 0.012175354175269604, -0.14183203876018524, 0.07110696285963058, 0.07237213850021362, -0.011589422821998596, 0.2233506292104721, -0.005310917738825083, -0.031204810366034508, -0.03758886829018593, 0.021047445014119148, 0.01207607239484787, 0.058114681392908096, -0.04501604288816452, 0.05061586573719978, 0.04741273075342178, -0.04946472868323326, 0.028376534581184387, -0.14829722046852112, 0.01825484074652195, 0.045887794345617294, -0.029032081365585327, -0.03673198074102402, 0.03883068636059761, 0.035948269069194794, 0.01774686947464943, 0.008014461025595665, 0.10563728958368301, 0.0007870895205996931, -0.013599592261016369, -0.09527229517698288, 0.1715124547481537, -0.14248284697532654, -0.2640177607536316, -0.1337573379278183, -0.049634773284196854, 0.04970448091626167, -0.02613730914890766, 0.09834820032119751, 0.005817966070026159, -0.014406946487724781, -0.09852556139230728, 0.09179670363664627, 0.04757792875170708, -0.11369309574365616, -0.09110303968191147, 0.05361846461892128, -0.007290625479072332, -0.1351536363363266, -0.024078425019979477, 0.07374080270528793, -0.09452120214700699, 0.049373794347047806, 0.009547783993184566, 0.057962071150541306, 0.04585440456867218, -0.07239130139350891, 0.009646887890994549, -0.04161710664629936, 0.20343755185604095, -0.06943441182374954, 0.0887872651219368, 0.15774372220039368, -0.11135411262512207, 0.12375571578741074, 0.14574465155601501, 0.026733551174402237, 0.009804380126297474, 0.01611669361591339, -0.029776476323604584, -0.05056294426321983, -0.19419588148593903, 0.01712428405880928, -0.014435926452279091, 0.10016937553882599, 0.06529238075017929, 0.010844217613339424, 0.04515435919165611, 0.12759335339069366, -0.010478314943611622, 0.06832465529441833, 0.007017099764198065, 0.07723091542720795, 0.15042707324028015, -0.005354158580303192, 0.01788293570280075, -0.10679128021001816, -0.023915566504001617, 0.060601815581321716, 0.11086463183164597, 0.14108970761299133, -0.0627383142709732, 0.12899227440357208, 0.09511376917362213, 0.04787454381585121, 0.018842676654458046, 0.07619525492191315, -0.07034188508987427, 0.05892598628997803, -0.02271013706922531, -0.12623631954193115, 0.03782932832837105, 0.05622543394565582, -0.03109796531498432, -0.07588090002536774, 0.026421846821904182, 0.017972616478800774, 0.08838188648223877, 0.2634253203868866, -0.0030089537613093853, -0.1702772080898285, 0.01470138132572174, 0.020088398829102516, 0.011440594680607319, -0.1108887568116188, -0.019554467871785164, 0.09419567883014679, -0.17903730273246765, 0.016568677499890327, -0.018758367747068405, 0.08955051004886627, -0.19868963956832886, -0.007367934565991163, 0.12207800894975662, 0.04687223583459854, -0.008233273401856422, 0.09864534437656403, -0.11121541261672974, 0.1011006087064743, -0.0005996341351419687, 0.06812907010316849, -0.04944818094372749, 0.039724916219711304, 0.016833141446113586, 0.002995863789692521, 0.05946020036935806, 0.01474169921129942, 0.044102996587753296, -0.09639355540275574, -0.09982302784919739, 0.04497263580560684, -0.013477115891873837, -0.10157785564661026, 0.07170688360929489, -0.03413049131631851, -0.0020265814382582903, -0.030925294384360313, -0.11335325986146927, -0.10725348442792892, -0.20028558373451233, 0.03753950074315071, -0.03460736200213432, 0.04508151113986969, -0.07205129414796829, 0.016662707552313805, 0.05284937098622322, 0.14154060184955597, 0.021425476297736168, -0.15467314422130585, -0.08464102447032928, -0.09871716797351837, 0.02981640212237835, -0.09849030524492264, 0.04364549368619919, -0.063319131731987, 0.04961768537759781, -0.0316227525472641, -0.12253692001104355, 0.08533959835767746, -0.05064263939857483, -0.06622791290283203, -0.026015108451247215, 0.033192139118909836, -0.007648590952157974, -0.018805822357535362, 0.01540655642747879, -0.007026237901300192, -0.0405685193836689, -0.1059722751379013, -0.04136212542653084, 0.11150071024894714, -0.028907975181937218, -0.0032488014549016953, -0.05925319716334343, -0.12595057487487793, -0.07608219236135483, 0.007821016944944859, 0.1978665441274643, 0.09303677827119827, -0.10798085480928421, 0.09821770340204239, 0.13585765659809113, -0.11793147027492523, -0.22485274076461792, -0.005555274896323681, 0.06061425805091858, -0.031438566744327545, 0.04621633514761925, -0.2659594714641571, 0.07626710832118988, 0.02028582990169525, -0.016895592212677002, 0.12504954636096954, -0.27003738284111023, -0.09036025404930115, 0.05144832283258438, 0.11496017128229141, -0.1365801990032196, -0.1084994226694107, -0.041438013315200806, -0.018267778679728508, -0.160466730594635, 0.15364302694797516, -0.04130431264638901, 0.08556616306304932, 0.010351655073463917, 0.0912030041217804, 0.05447736755013466, -0.0434492789208889, 0.02691716141998768, -0.15037305653095245, 0.0386657789349556, -0.09442616254091263, -0.12027308344841003, 0.053743135184049606, -0.07134487479925156, 0.1524481177330017, -0.11205929517745972, 0.05548923835158348, -0.06217586249113083, 0.0010114590404555202, -0.041603267192840576, 0.0685969665646553, -0.05800243839621544, -0.08985070884227753, -0.08873479068279266, -0.019780419766902924, 0.06633738428354263, -0.00003809652116615325, -0.03962576761841774, -0.013390647247433662, -0.012786230072379112, 0.15001243352890015, -0.0022533447481691837, 0.18079206347465515, -0.13412003219127655, -0.05533527582883835, -0.017913924530148506, 0.055516310036182404, -0.18348443508148193, 0.019409285858273506, 0.06929946690797806, -0.01287901122123003, 0.1312919557094574, -0.016928967088460922, -0.12046392261981964, 0.051304906606674194, 0.08204927295446396, -0.17523643374443054, -0.10938964784145355, -0.07551673799753189, 0.07307753711938858, -0.06997094303369522, -0.0332120917737484, 0.12168412655591965, -0.09102246910333633, -0.02713819034397602, -0.020715678110718727, 0.01546412892639637, -0.005667718127369881, 0.04641745239496231, 0.019669385626912117, 0.017224455252289772, -0.04662904143333435, 0.052093375474214554, 0.06622157990932465, 0.022909531369805336, 0.038980137556791306, 0.12061891704797745, -0.06496439129114151, -0.035229798406362534, 0.08982527256011963, 0.10309699177742004, 0.04345525801181793, -0.024460840970277786, -0.013316557742655277, -0.04089318960905075, -0.017512330785393715, 0.1314600110054016, 0.01692531816661358, 0.03602007403969765, -0.011274605058133602, 0.0013606243301182985, -0.0189929511398077, 0.1420082449913025, -0.028510136529803276, 0.018022853881120682, -0.11257326602935791, 0.031398095190525055, 0.031579915434122086, 0.042640671133995056, 0.012751255184412003, -0.05933380872011185, -0.08848222345113754, -0.04543519765138626, -0.04235587269067764, 0.06171838566660881, -0.042476460337638855, 0.022008463740348816, -0.001242202240973711, 0.04371897131204605, 0.013092502020299435, 0.024380329996347427, -0.07980169355869293, -0.023630011826753616, -0.03414617478847504, 0.14455363154411316, -0.133024662733078, -0.03114376775920391, 0.07476817071437836, -0.0706813856959343, 0.10496052354574203, -0.023707399144768715, -0.08455859124660492, 0.0009556979057379067, -0.10049677640199661, -0.006849248427897692, 0.035546135157346725, 0.03051786869764328, 0.03225121647119522, -0.03184839338064194, 0.008111679926514626, -0.06335002183914185, 0.0008534779190085828, -0.035616129636764526, 0.1078813374042511, -0.12411364912986755, 0.01322268322110176, 0.046610720455646515, -0.06847646832466125, -0.03594456613063812, 0.05470053106546402, 0.05865354835987091, -0.0053869495168328285, 0.13665546476840973, -0.033184971660375595, 0.07715751975774765, -0.10486968606710434, -0.022375427186489105, 0.06580139696598053, -0.07556906342506409, -0.04939121752977371, -0.061731643974781036, -0.005423069931566715, -0.036646630614995956, 0.07221332937479019, 0.044364381581544876, -0.08961714804172516, -0.022772548720240593, -0.04769522324204445, -0.007779452484101057, 0.05214359611272812, 0.10056804120540619, -0.061270665377378464, -0.018267260864377022, -0.053916387259960175, -0.01887614093720913, -0.006847048178315163, -0.05361723154783249, 0.007879534736275673, 0.05124152451753616, 0.04248322919011116, 0.03555990383028984, 0.09247110784053802, 0.07595916837453842, -0.15202799439430237, -0.029506562277674675, 0.057336099445819855, 0.17070706188678741, -0.12822532653808594, 0.15195585787296295, 0.14618901908397675, -0.08433417975902557, 0.05819129943847656, 0.03873870149254799, -0.06927244365215302, -0.07580643147230148, -0.19167301058769226, -0.06737460941076279, -0.06934864819049835, -0.010144409723579884, -0.09576407074928284, 0.09555552154779434, -0.014714212156832218, 0.006536961533129215, -0.05912042036652565, 0.1662202626466751, 0.05568646267056465, -0.07766444981098175, 0.09424235671758652, -0.02032095566391945, -0.02397722192108631, -0.052487194538116455, 0.05849381163716316, 0.012514138594269753, 0.06848955899477005, 0.03969571366906166, 0.09559446573257446, -0.04757385700941086, -0.004006420262157917, -0.11357604712247849, -0.07875996828079224, 0.007513747550547123, 0.026904847472906113, -0.007549542002379894, 0.03203290328383446, 0.05310723930597305, -0.05720044672489166, -0.019035054370760918, 0.17290467023849487, -0.05395381152629852, -0.1011146530508995, -0.139743372797966, 0.04740719124674797, -0.04494895413517952, -0.00016984993999358267, -0.0011121132411062717, -0.03909524530172348, -0.056515976786613464, 0.24327188730239868, 0.19613303244113922, 0.005420766770839691, 0.013016389682888985, -0.030942926183342934, 0.013269256800413132, -0.032697781920433044, 0.166756734251976, 0.004828507546335459, 0.14895407855510712, -0.020487463101744652, 0.0745854452252388, -0.060398709028959274, -0.08374275267124176, -0.09041357040405273, 0.018285011872649193, 0.03045808896422386, 0.007479547988623381, -0.039042796939611435, 0.13101907074451447, -0.07804228365421295, -0.11673708260059357, -0.11888165771961212, -0.04980773106217384, -0.09353215247392654, 0.017900483682751656, 0.12897923588752747, -0.029072033241391182, 0.058150749653577805, 0.01151998620480299, -0.04891311749815941, 0.10428396612405777, 0.004433372989296913, -0.07675319910049438, 0.02718404121696949, 0.05336865037679672, -0.23298557102680206, 0.22006963193416595, 0.009867630898952484, 0.0015614696312695742, 0.07337167859077454, -0.011923013255000114, -0.09060639142990112, 0.044264256954193115, 0.03621489927172661, -0.07249027490615845, -0.039365608245134354, 0.03192152455449104, -0.02398524060845375, -0.012148725800216198, 0.038881439715623856, -0.007162688300013542, 0.08388779312372208, -0.08420084416866302, 0.0019397392170503736, -0.0752178281545639, -0.025403227657079697, -0.0756336897611618, 0.08540317416191101, 0.18041542172431946, -0.021247990429401398, 0.012652318924665451, -0.04194134101271629, -0.025743909180164337, 0.0446651428937912, -0.06889607757329941, 0.0011761229252442718, -0.044039759784936905, -0.05562314763665199, 0.05410892888903618, 0.10255105048418045, -0.07336302846670151, -0.00486784428358078, -0.0321677140891552, 0.012736781500279903, -0.0737491026520729, 0.11329631507396698, 0.14341764152050018, 0.004668378736823797, 0.00638585677370429, -0.09423459321260452, -0.06734520941972733, 0.1079636886715889, -0.07703813165426254, -0.08550699800252914 ]
null
null
transformers
# keytotext Idea is to build a model which will take keywords as inputs and generate sentences as outputs. ### Model: Two Models have been built: - Using T5-base size = 850 MB can be found here: https://huggingface.co/gagan3012/keytotext - Using T5-small size = 230 MB can be found here: https://huggingface.co/gagan3012/keytotext-small #### Usage: ```python from transformers import AutoTokenizer, AutoModelWithLMHead tokenizer = AutoTokenizer.from_pretrained("gagan3012/keytotext-small") model = AutoModelWithLMHead.from_pretrained("gagan3012/keytotext-small") ``` ### Demo: [![Streamlit App](https://static.streamlit.io/badges/streamlit_badge_black_white.svg)](https://share.streamlit.io/gagan3012/keytotext/app.py) https://share.streamlit.io/gagan3012/keytotext/app.py ![image](https://user-images.githubusercontent.com/49101362/110660053-3b20fe80-81d4-11eb-9275-ba402134e8d9.png) ### Example: ['India', 'Wedding'] -> We are celebrating today in New Delhi with three wedding anniversary parties.
{}
text2text-generation
gagan3012/keytotext-small
[ "transformers", "pytorch", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# keytotext Idea is to build a model which will take keywords as inputs and generate sentences as outputs. ### Model: Two Models have been built: - Using T5-base size = 850 MB can be found here: URL - Using T5-small size = 230 MB can be found here: URL #### Usage: ### Demo: ![Streamlit App](URL URL !image ### Example: ['India', 'Wedding'] -> We are celebrating today in New Delhi with three wedding anniversary parties.
[ "# keytotext\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.", "### Model:\n\nTwo Models have been built: \n\n- Using T5-base size = 850 MB can be found here: URL\n- Using T5-small size = 230 MB can be found here: URL", "#### Usage:", "### Demo:\n\n![Streamlit App](URL\n\nURL\n\n!image", "### Example: \n\n['India', 'Wedding'] -> We are celebrating today in New Delhi with three wedding anniversary parties." ]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# keytotext\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.", "### Model:\n\nTwo Models have been built: \n\n- Using T5-base size = 850 MB can be found here: URL\n- Using T5-small size = 230 MB can be found here: URL", "#### Usage:", "### Demo:\n\n![Streamlit App](URL\n\nURL\n\n!image", "### Example: \n\n['India', 'Wedding'] -> We are celebrating today in New Delhi with three wedding anniversary parties." ]
[ 48, 26, 44, 5, 15, 30 ]
[ "passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# keytotext\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.### Model:\n\nTwo Models have been built: \n\n- Using T5-base size = 850 MB can be found here: URL\n- Using T5-small size = 230 MB can be found here: URL#### Usage:### Demo:\n\n![Streamlit App](URL\n\nURL\n\n!image### Example: \n\n['India', 'Wedding'] -> We are celebrating today in New Delhi with three wedding anniversary parties." ]
[ 0.01082890946418047, -0.034107934683561325, -0.0005711333360522985, 0.09137126803398132, 0.09778378158807755, -0.032812681049108505, 0.08776213973760605, 0.11602818965911865, 0.006385213695466518, -0.008260245434939861, 0.14921413362026215, 0.08520209044218063, -0.030830487608909607, 0.22713811695575714, -0.06687697023153305, -0.26192936301231384, -0.009469025768339634, 0.08418028056621552, -0.11059223860502243, 0.17225269973278046, 0.13310764729976654, -0.06893830001354218, 0.11066621541976929, -0.06370043754577637, -0.1936110109090805, 0.08047802001237869, -0.09941473603248596, -0.05954384803771973, 0.010702691972255707, -0.02426263503730297, 0.007665828336030245, 0.10478097945451736, -0.045252662152051926, -0.11358386278152466, 0.008020811714231968, -0.06130457669496536, 0.013469340279698372, -0.023161228746175766, 0.010063536465168, 0.08530083298683167, 0.24626579880714417, 0.051679227501153946, -0.06096404418349266, 0.046582095324993134, -0.0844264104962349, 0.07734347134828568, -0.01195458136498928, 0.15905609726905823, -0.060297004878520966, 0.07750838249921799, -0.01194777712225914, 0.12797144055366516, -0.055454690009355545, 0.06808626651763916, 0.08607526123523712, -0.16248375177383423, -0.05077774077653885, 0.046166181564331055, 0.11639919131994247, 0.0819336324930191, 0.050231341272592545, -0.006985163316130638, 0.05483411252498627, 0.056084368377923965, 0.04870054870843887, -0.05686640366911888, -0.04970265179872513, 0.03140411898493767, -0.10291090607643127, -0.06456564366817474, 0.25128331780433655, 0.026791173964738846, 0.03524200618267059, -0.06336678564548492, -0.09618581086397171, -0.00751235568895936, -0.03748608008027077, -0.058540914207696915, -0.01852208375930786, 0.043952278792858124, -0.08920054882764816, -0.09926385432481766, -0.13421474397182465, 0.10065236687660217, 0.06457102298736572, -0.02508234605193138, 0.0095314746722579, 0.04855702072381973, -0.23754504323005676, 0.0010253081563860178, -0.04232592135667801, -0.0850696936249733, 0.072418712079525, -0.08365172147750854, 0.18084797263145447, 0.029127798974514008, 0.03298729285597801, -0.13558855652809143, 0.13195902109146118, -0.1791558861732483, 0.022598247975111008, -0.04350108280777931, 0.037057556211948395, 0.05769864469766617, 0.08358026295900345, 0.08664364367723465, -0.061484936624765396, -0.10632236301898956, -0.0004131099267397076, -0.03217538446187973, -0.005216586869210005, 0.04540923237800598, -0.06438393145799637, -0.029827114194631577, -0.031078403815627098, 0.003792481031268835, -0.053158506751060486, 0.12600138783454895, 0.053477317094802856, -0.031378913670778275, 0.10840144753456116, 0.0229225754737854, -0.03820040822029114, -0.06562305986881256, 0.026471981778740883, 0.12079627066850662, -0.008507238700985909, 0.025695955380797386, -0.08120325952768326, -0.09847112745046616, -0.0492849163711071, -0.0010852674022316933, -0.0467209592461586, -0.05664613097906113, 0.005415990483015776, -0.13924212753772736, -0.04492224380373955, -0.11201518774032593, -0.08631544560194016, 0.026455221697688103, 0.03205759823322296, -0.019227147102355957, -0.0727638229727745, -0.10971380770206451, -0.0054255081340670586, -0.03390522673726082, -0.03642328828573227, -0.06967028230428696, 0.004682891070842743, 0.12489855289459229, 0.004681999329477549, 0.12024988979101181, -0.13066507875919342, 0.09001724421977997, -0.15911242365837097, 0.012262502685189247, -0.04682888835668564, 0.10985518991947174, 0.15057310461997986, 0.14727963507175446, -0.03775244206190109, -0.011348102241754532, -0.0292346328496933, 0.04650068283081055, 0.0003586002276279032, 0.162837952375412, -0.10813798010349274, -0.02147473581135273, 0.14282040297985077, -0.08814603835344315, -0.20663724839687347, 0.10161472111940384, 0.002019454725086689, 0.3336494266986847, 0.06755223125219345, 0.11011698096990585, -0.02787983976304531, 0.07910285145044327, 0.01719873957335949, -0.034751441329717636, -0.1408502757549286, -0.13960057497024536, -0.01408513356000185, 0.040430668741464615, -0.01010325737297535, 0.029896285384893417, 0.12608955800533295, 0.02256832830607891, -0.10154379904270172, -0.03509034961462021, -0.07313129305839539, -0.13099196553230286, 0.010110507719218731, 0.012092423625290394, 0.08568774163722992, -0.08959967643022537, -0.023648688569664955, 0.038288455456495285, -0.0040451353415846825, -0.08116248995065689, 0.03437834978103638, -0.09904711693525314, 0.012004530988633633, -0.04181871563196182, -0.015329147689044476, -0.09998992830514908, 0.14460249245166779, 0.04123696684837341, 0.22133797407150269, 0.040396783500909805, 0.16990846395492554, 0.050432804971933365, -0.06778179854154587, 0.018741097301244736, 0.06313913315534592, 0.11753237992525101, -0.014537551440298557, -0.10829723626375198, 0.060312509536743164, 0.03340299427509308, -0.04432058706879616, 0.009100891649723053, -0.046639423817396164, -0.009466694667935371, 0.04673588648438454, 0.12184130400419235, 0.04918409883975983, 0.02329730987548828, 0.020249776542186737, 0.01636873371899128, -0.08046601712703705, -0.006382588762789965, 0.07246751338243484, 0.018777957186102867, -0.043158408254384995, 0.19181153178215027, -0.03547145798802376, 0.07444905489683151, 0.05931861698627472, -0.17301501333713531, 0.0030696962494403124, -0.046710483729839325, -0.09051419049501419, 0.014617757871747017, 0.10024750232696533, 0.02805197983980179, 0.03598613291978836, -0.015306541696190834, 0.14059355854988098, -0.013478974811732769, -0.076138436794281, 0.06023531034588814, -0.04790833219885826, -0.06376554071903229, 0.06527985632419586, 0.06574670970439911, -0.29548147320747375, 0.10827288031578064, 0.12074505537748337, 0.00957865733653307, 0.14550530910491943, 0.007877187803387642, 0.01693057455122471, 0.019128749147057533, 0.08058905601501465, -0.08335660398006439, -0.05682707205414772, -0.21483953297138214, -0.026705604046583176, 0.005149407312273979, -0.10471135377883911, 0.07249483466148376, 0.002407593885436654, -0.00727874506264925, -0.0016304254531860352, 0.011142352595925331, -0.05241917818784714, 0.09179190546274185, -0.00890041422098875, 0.089470274746418, 0.032205548137426376, 0.060992613434791565, 0.13223595917224884, 0.05640945956110954, -0.09224635362625122, 0.07981666922569275, -0.014230815693736076, -0.3524332344532013, -0.07776977121829987, -0.08503005653619766, 0.07276752591133118, 0.012777956202626228, 0.12844207882881165, -0.13303765654563904, -0.05206049978733063, -0.04631683975458145, 0.07893726229667664, 0.013055853545665741, -0.02959877997636795, -0.10094918310642242, 0.014459332451224327, -0.004496203735470772, -0.09311854839324951, -0.03411954641342163, 0.0270443893969059, 0.0018547001527622342, 0.15166360139846802, -0.12054844200611115, 0.049127478152513504, 0.08113346248865128, -0.14585493505001068, 0.06868340075016022, -0.009453179314732552, 0.09957946836948395, -0.03837045282125473, 0.08997402340173721, 0.16009151935577393, 0.00477764243260026, 0.09466411918401718, 0.17649416625499725, -0.01705397292971611, -0.02724151499569416, 0.04842684790492058, -0.014683657325804234, -0.0724385529756546, -0.1120031327009201, -0.0633048266172409, -0.05177382007241249, -0.010144074447453022, 0.012341433204710484, 0.0744282528758049, 0.04312784597277641, 0.17019076645374298, -0.06594827026128769, 0.0535443089902401, -0.1419581174850464, 0.019538072869181633, 0.14619489014148712, -0.03163960576057434, -0.00014829292194917798, -0.10701519250869751, -0.16118071973323822, 0.07572086155414581, -0.007469928357750177, 0.08079590648412704, 0.015894122421741486, -0.020450709387660027, 0.11724889278411865, 0.031011095270514488, 0.07966947555541992, 0.026727134361863136, -0.10332746058702469, 0.01266884058713913, -0.024949971586465836, -0.05412982031702995, 0.0364188551902771, 0.10117433965206146, 0.062476176768541336, -0.10735323280096054, -0.0021743429824709892, -0.02861718460917473, 0.03462967276573181, 0.1889367550611496, 0.12161342799663544, -0.21514299511909485, 0.01069377176463604, 0.03774520382285118, -0.08331664651632309, -0.09440625458955765, 0.06247501075267792, 0.11140041053295135, -0.016055120155215263, 0.058712828904390335, -0.02375781536102295, 0.08859636634588242, -0.029445840045809746, 0.0798061341047287, 0.05252521485090256, -0.03940609097480774, 0.03527453914284706, 0.06961151212453842, -0.3550260365009308, 0.22299548983573914, -0.008314825594425201, -0.09938815981149673, -0.09260500222444534, -0.00986372958868742, -0.07000842690467834, -0.00972070824354887, 0.010848643258213997, -0.026651455089449883, 0.22760434448719025, -0.030160436406731606, 0.02015610970556736, 0.047253984957933426, 0.04566110298037529, -0.018895452842116356, 0.08732179552316666, -0.044939033687114716, -0.012262066826224327, -0.01834316924214363, -0.0705653503537178, -0.15315288305282593, -0.05645991116762161, -0.016330603510141373, 0.062425579875707626, -0.02991853468120098, 0.01837010122835636, 0.008866565302014351, 0.034005098044872284, 0.12693819403648376, 0.13864143192768097, -0.15174263715744019, -0.04214521124958992, 0.010241634212434292, 0.08456294238567352, -0.03329229727387428, 0.06539548188447952, -0.04408327117562294, 0.001481515821069479, -0.025787660852074623, -0.10165808349847794, 0.07005500048398972, 0.00147929263766855, -0.02250482700765133, -0.023888520896434784, 0.040588025003671646, -0.011061779223382473, -0.041392140090465546, 0.006963638123124838, -0.019512835890054703, -0.06584134697914124, -0.030729644000530243, -0.037231847643852234, -0.08309119194746017, -0.07507694512605667, -0.02378462627530098, -0.027835121378302574, -0.0075820148922502995, -0.028732402250170708, -0.06188717111945152, 0.18773213028907776, 0.09702329337596893, -0.019745297729969025, 0.10128287225961685, 0.13449178636074066, 0.017844216898083687, -0.23566998541355133, -0.11628154665231705, -0.06093135103583336, 0.011361788958311081, -0.010657204315066338, -0.05168123170733452, 0.006400249432772398, -0.03359615057706833, 0.004349325783550739, 0.014682142063975334, -0.1889932006597519, -0.01888631284236908, 0.056496039032936096, 0.08818715810775757, 0.22621215879917145, -0.12237068265676498, -0.06723935902118683, -0.06163504719734192, -0.1131981834769249, 0.08168534934520721, 0.04052843898534775, 0.001059756032191217, -0.04927881807088852, 0.18566225469112396, 0.054081182926893234, -0.021107252687215805, -0.10586195439100266, -0.07040673494338989, -0.03490530699491501, -0.15280291438102722, -0.17713411152362823, 0.16253027319908142, -0.020188990980386734, 0.1643989533185959, -0.1630563884973526, 0.03656895086169243, -0.08099699020385742, -0.030154459178447723, -0.05900393798947334, -0.02484537661075592, -0.030173130333423615, -0.10029400140047073, -0.09191545099020004, -0.06870589405298233, 0.0015137521550059319, 0.009730175137519836, 0.01002420298755169, -0.04965392127633095, 0.03194653242826462, 0.10216201841831207, 0.12728580832481384, -0.020709404721856117, 0.041814785450696945, -0.07069884240627289, -0.04454528167843819, 0.04345665127038956, -0.24261952936649323, 0.0014690157258883119, 0.051265642046928406, -0.02148653008043766, 0.08646336197853088, -0.0025550085119903088, 0.00031842352473177016, 0.08825352787971497, 0.19088855385780334, -0.08338882774114609, -0.018312163650989532, -0.1171351745724678, -0.020013459026813507, -0.011992980726063251, 0.013240945525467396, 0.10634678602218628, -0.08281367272138596, -0.014296418987214565, -0.05053159222006798, 0.03317386284470558, -0.08075668662786484, -0.03546195477247238, 0.06615728884935379, 0.014058179222047329, -0.04470339044928551, -0.0703742504119873, 0.06636954843997955, -0.07147262990474701, -0.007939723320305347, 0.24594871699810028, -0.07360000908374786, -0.08462967723608017, 0.12977291643619537, 0.06473758816719055, -0.13670013844966888, 0.015391889028251171, -0.05158970132470131, -0.03609005734324455, 0.04845979064702988, 0.19375014305114746, -0.03772180527448654, 0.09962382912635803, -0.09065043181180954, 0.029278263449668884, -0.11792987585067749, 0.03989604115486145, 0.051565006375312805, 0.029495617374777794, -0.1911115199327469, 0.014230059459805489, -0.0564739927649498, 0.09884310513734818, -0.0712791383266449, -0.06167111173272133, -0.027686195448040962, -0.010395675897598267, -0.12182820588350296, 0.013789637945592403, -0.07935506850481033, -0.0717480257153511, 0.014924961142241955, -0.02740117348730564, -0.03712143003940582, -0.016848929226398468, -0.058863017708063126, 0.035020530223846436, 0.040038902312517166, 0.10164206475019455, -0.05147524178028107, -0.02871425822377205, 0.021904824301600456, -0.02144690230488777, 0.1303510069847107, -0.011808107607066631, -0.18214863538742065, -0.002467659767717123, -0.0899997428059578, -0.040547315031290054, 0.12251103669404984, 0.01757558435201645, 0.06671656668186188, 0.19423891603946686, -0.0067499903962016106, 0.04359569400548935, 0.08808828145265579, 0.03774895891547203, 0.12239863723516464, -0.059625912457704544, 0.07157192379236221, -0.07309483736753464, -0.08835454285144806, -0.04235561564564705, 0.04349001869559288, 0.07585523277521133, 0.015659116208553314, 0.10519582033157349, -0.08903355896472931, 0.021632174029946327, -0.09773826599121094, 0.04435759782791138, 0.040601734071969986, -0.06728749722242355, -0.1624782383441925, -0.04461563378572464, -0.00011523235298227519, -0.004251922946423292, 0.11695127934217453, 0.0542314276099205, -0.05698946490883827, 0.02839941903948784, -0.025824343785643578, 0.024219801649451256, -0.01859203539788723, 0.29201072454452515, 0.017094820737838745, -0.03712693974375725, -0.1643575131893158, 0.023092474788427353, -0.011540628969669342, -0.12300024926662445, 0.08605404943227768, 0.07721847295761108, -0.009752578102052212, 0.05248025804758072, 0.12826813757419586, 0.16888652741909027, -0.03014942817389965, -0.04240608960390091, 0.025163548067212105, 0.07696802914142609, -0.11240799725055695, 0.1052575558423996, 0.19255663454532623, 0.03731829673051834, -0.051706861704587936, 0.009307246655225754, 0.0027445561718195677, 0.023734409362077713, -0.07189902663230896, -0.1175154522061348, -0.16289347410202026, 0.019688380882143974, -0.0691247209906578, 0.06713484227657318, 0.02611451968550682, 0.035348664969205856, -0.10550089180469513, 0.018729371950030327, 0.12874434888362885, -0.008022873662412167, 0.0944441482424736, -0.017568625509738922, -0.01706542819738388, -0.0005294526927173138, 0.027341529726982117, -0.10668651759624481, 0.12619401514530182, 0.004730683285742998, 0.07958899438381195, -0.06628315150737762, -0.04174678027629852, -0.11339453607797623, -0.09367872029542923, -0.04577641934156418, 0.05476958304643631, 0.019463473930954933, 0.019578181207180023, 0.009242166765034199, -0.07726223766803741, -0.044836800545454025, 0.010594986379146576, -0.010891340672969818, -0.12927383184432983, -0.013084380887448788, 0.17935657501220703, -0.007919969968497753, 0.03360243886709213, 0.01391671970486641, -0.012732738628983498, -0.13551419973373413, 0.2385387271642685, 0.23441046476364136, -0.027859840542078018, -0.014673241414129734, -0.019813017919659615, 0.03475429117679596, -0.049412429332733154, 0.19776000082492828, 0.0028357000555843115, 0.2602749764919281, -0.06271476298570633, -0.04589061066508293, -0.0779002234339714, -0.009737488813698292, -0.06773922592401505, 0.017782721668481827, 0.041555337607860565, -0.04930737242102623, -0.013523943722248077, 0.1079082265496254, -0.1977352648973465, 0.10427799075841904, -0.15153776109218597, -0.08611235022544861, 0.003022788790985942, 0.0006565612275153399, -0.04241715371608734, 0.014161449857056141, 0.05282851681113243, -0.00480167381465435, -0.019914433360099792, 0.006723520811647177, 0.0302690751850605, -0.2192714661359787, 0.10469471663236618, 0.08205097913742065, -0.23269595205783844, 0.028347091749310493, -0.019223365932703018, 0.05988841503858566, 0.053471386432647705, -0.016955379396677017, -0.03164546564221382, 0.07445292174816132, 0.009085859172046185, 0.11564026772975922, -0.04280761256814003, -0.015395250171422958, 0.005340672098100185, -0.05629056319594383, 0.06171998009085655, -0.1284298151731491, -0.0038267469499260187, 0.021795347332954407, 0.09658944606781006, -0.1410149782896042, 0.040930673480033875, -0.002251217607408762, 0.08053545653820038, 0.05671023204922676, -0.036111511290073395, -0.010603683069348335, -0.029856812208890915, 0.018657758831977844, -0.028602106496691704, -0.037007641047239304, -0.015919150784611702, -0.10215046256780624, -0.10934906452894211, -0.03756320849061012, -0.022666219621896744, -0.2746734023094177, -0.00997922196984291, -0.08015692234039307, 0.08928319811820984, -0.09636471420526505, 0.08627842366695404, 0.22401681542396545, 0.04238072410225868, 0.028648605570197105, 0.06490515172481537, -0.06347111612558365, 0.1325930804014206, -0.08154750615358353, -0.06537696719169617 ]
null
null
transformers
# keytotext Idea is to build a model which will take keywords as inputs and generate sentences as outputs. ### Model: Two Models have been built: - Using T5-base size = 850 MB can be found here: https://huggingface.co/gagan3012/keytotext - Using T5-small size = 230 MB can be found here: https://huggingface.co/gagan3012/keytotext-small #### Usage: ```python from transformers import AutoTokenizer, AutoModelWithLMHead tokenizer = AutoTokenizer.from_pretrained("gagan3012/keytotext-small") model = AutoModelWithLMHead.from_pretrained("gagan3012/keytotext-small") ``` ### Demo: [![Streamlit App](https://static.streamlit.io/badges/streamlit_badge_black_white.svg)](https://share.streamlit.io/gagan3012/keytotext/app.py) https://share.streamlit.io/gagan3012/keytotext/app.py ![image](https://user-images.githubusercontent.com/49101362/110660053-3b20fe80-81d4-11eb-9275-ba402134e8d9.png) ### Example: ['India', 'Wedding'] -> We are celebrating today in New Delhi with three wedding anniversary parties.
{}
text2text-generation
gagan3012/keytotext
[ "transformers", "pytorch", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# keytotext Idea is to build a model which will take keywords as inputs and generate sentences as outputs. ### Model: Two Models have been built: - Using T5-base size = 850 MB can be found here: URL - Using T5-small size = 230 MB can be found here: URL #### Usage: ### Demo: ![Streamlit App](URL URL !image ### Example: ['India', 'Wedding'] -> We are celebrating today in New Delhi with three wedding anniversary parties.
[ "# keytotext\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.", "### Model:\n\nTwo Models have been built: \n\n- Using T5-base size = 850 MB can be found here: URL\n- Using T5-small size = 230 MB can be found here: URL", "#### Usage:", "### Demo:\n\n![Streamlit App](URL\n\nURL\n\n!image", "### Example: \n\n['India', 'Wedding'] -> We are celebrating today in New Delhi with three wedding anniversary parties." ]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# keytotext\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.", "### Model:\n\nTwo Models have been built: \n\n- Using T5-base size = 850 MB can be found here: URL\n- Using T5-small size = 230 MB can be found here: URL", "#### Usage:", "### Demo:\n\n![Streamlit App](URL\n\nURL\n\n!image", "### Example: \n\n['India', 'Wedding'] -> We are celebrating today in New Delhi with three wedding anniversary parties." ]
[ 48, 26, 44, 5, 15, 30 ]
[ "passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# keytotext\n\nIdea is to build a model which will take keywords as inputs and generate sentences as outputs.### Model:\n\nTwo Models have been built: \n\n- Using T5-base size = 850 MB can be found here: URL\n- Using T5-small size = 230 MB can be found here: URL#### Usage:### Demo:\n\n![Streamlit App](URL\n\nURL\n\n!image### Example: \n\n['India', 'Wedding'] -> We are celebrating today in New Delhi with three wedding anniversary parties." ]
[ 0.01082890946418047, -0.034107934683561325, -0.0005711333360522985, 0.09137126803398132, 0.09778378158807755, -0.032812681049108505, 0.08776213973760605, 0.11602818965911865, 0.006385213695466518, -0.008260245434939861, 0.14921413362026215, 0.08520209044218063, -0.030830487608909607, 0.22713811695575714, -0.06687697023153305, -0.26192936301231384, -0.009469025768339634, 0.08418028056621552, -0.11059223860502243, 0.17225269973278046, 0.13310764729976654, -0.06893830001354218, 0.11066621541976929, -0.06370043754577637, -0.1936110109090805, 0.08047802001237869, -0.09941473603248596, -0.05954384803771973, 0.010702691972255707, -0.02426263503730297, 0.007665828336030245, 0.10478097945451736, -0.045252662152051926, -0.11358386278152466, 0.008020811714231968, -0.06130457669496536, 0.013469340279698372, -0.023161228746175766, 0.010063536465168, 0.08530083298683167, 0.24626579880714417, 0.051679227501153946, -0.06096404418349266, 0.046582095324993134, -0.0844264104962349, 0.07734347134828568, -0.01195458136498928, 0.15905609726905823, -0.060297004878520966, 0.07750838249921799, -0.01194777712225914, 0.12797144055366516, -0.055454690009355545, 0.06808626651763916, 0.08607526123523712, -0.16248375177383423, -0.05077774077653885, 0.046166181564331055, 0.11639919131994247, 0.0819336324930191, 0.050231341272592545, -0.006985163316130638, 0.05483411252498627, 0.056084368377923965, 0.04870054870843887, -0.05686640366911888, -0.04970265179872513, 0.03140411898493767, -0.10291090607643127, -0.06456564366817474, 0.25128331780433655, 0.026791173964738846, 0.03524200618267059, -0.06336678564548492, -0.09618581086397171, -0.00751235568895936, -0.03748608008027077, -0.058540914207696915, -0.01852208375930786, 0.043952278792858124, -0.08920054882764816, -0.09926385432481766, -0.13421474397182465, 0.10065236687660217, 0.06457102298736572, -0.02508234605193138, 0.0095314746722579, 0.04855702072381973, -0.23754504323005676, 0.0010253081563860178, -0.04232592135667801, -0.0850696936249733, 0.072418712079525, -0.08365172147750854, 0.18084797263145447, 0.029127798974514008, 0.03298729285597801, -0.13558855652809143, 0.13195902109146118, -0.1791558861732483, 0.022598247975111008, -0.04350108280777931, 0.037057556211948395, 0.05769864469766617, 0.08358026295900345, 0.08664364367723465, -0.061484936624765396, -0.10632236301898956, -0.0004131099267397076, -0.03217538446187973, -0.005216586869210005, 0.04540923237800598, -0.06438393145799637, -0.029827114194631577, -0.031078403815627098, 0.003792481031268835, -0.053158506751060486, 0.12600138783454895, 0.053477317094802856, -0.031378913670778275, 0.10840144753456116, 0.0229225754737854, -0.03820040822029114, -0.06562305986881256, 0.026471981778740883, 0.12079627066850662, -0.008507238700985909, 0.025695955380797386, -0.08120325952768326, -0.09847112745046616, -0.0492849163711071, -0.0010852674022316933, -0.0467209592461586, -0.05664613097906113, 0.005415990483015776, -0.13924212753772736, -0.04492224380373955, -0.11201518774032593, -0.08631544560194016, 0.026455221697688103, 0.03205759823322296, -0.019227147102355957, -0.0727638229727745, -0.10971380770206451, -0.0054255081340670586, -0.03390522673726082, -0.03642328828573227, -0.06967028230428696, 0.004682891070842743, 0.12489855289459229, 0.004681999329477549, 0.12024988979101181, -0.13066507875919342, 0.09001724421977997, -0.15911242365837097, 0.012262502685189247, -0.04682888835668564, 0.10985518991947174, 0.15057310461997986, 0.14727963507175446, -0.03775244206190109, -0.011348102241754532, -0.0292346328496933, 0.04650068283081055, 0.0003586002276279032, 0.162837952375412, -0.10813798010349274, -0.02147473581135273, 0.14282040297985077, -0.08814603835344315, -0.20663724839687347, 0.10161472111940384, 0.002019454725086689, 0.3336494266986847, 0.06755223125219345, 0.11011698096990585, -0.02787983976304531, 0.07910285145044327, 0.01719873957335949, -0.034751441329717636, -0.1408502757549286, -0.13960057497024536, -0.01408513356000185, 0.040430668741464615, -0.01010325737297535, 0.029896285384893417, 0.12608955800533295, 0.02256832830607891, -0.10154379904270172, -0.03509034961462021, -0.07313129305839539, -0.13099196553230286, 0.010110507719218731, 0.012092423625290394, 0.08568774163722992, -0.08959967643022537, -0.023648688569664955, 0.038288455456495285, -0.0040451353415846825, -0.08116248995065689, 0.03437834978103638, -0.09904711693525314, 0.012004530988633633, -0.04181871563196182, -0.015329147689044476, -0.09998992830514908, 0.14460249245166779, 0.04123696684837341, 0.22133797407150269, 0.040396783500909805, 0.16990846395492554, 0.050432804971933365, -0.06778179854154587, 0.018741097301244736, 0.06313913315534592, 0.11753237992525101, -0.014537551440298557, -0.10829723626375198, 0.060312509536743164, 0.03340299427509308, -0.04432058706879616, 0.009100891649723053, -0.046639423817396164, -0.009466694667935371, 0.04673588648438454, 0.12184130400419235, 0.04918409883975983, 0.02329730987548828, 0.020249776542186737, 0.01636873371899128, -0.08046601712703705, -0.006382588762789965, 0.07246751338243484, 0.018777957186102867, -0.043158408254384995, 0.19181153178215027, -0.03547145798802376, 0.07444905489683151, 0.05931861698627472, -0.17301501333713531, 0.0030696962494403124, -0.046710483729839325, -0.09051419049501419, 0.014617757871747017, 0.10024750232696533, 0.02805197983980179, 0.03598613291978836, -0.015306541696190834, 0.14059355854988098, -0.013478974811732769, -0.076138436794281, 0.06023531034588814, -0.04790833219885826, -0.06376554071903229, 0.06527985632419586, 0.06574670970439911, -0.29548147320747375, 0.10827288031578064, 0.12074505537748337, 0.00957865733653307, 0.14550530910491943, 0.007877187803387642, 0.01693057455122471, 0.019128749147057533, 0.08058905601501465, -0.08335660398006439, -0.05682707205414772, -0.21483953297138214, -0.026705604046583176, 0.005149407312273979, -0.10471135377883911, 0.07249483466148376, 0.002407593885436654, -0.00727874506264925, -0.0016304254531860352, 0.011142352595925331, -0.05241917818784714, 0.09179190546274185, -0.00890041422098875, 0.089470274746418, 0.032205548137426376, 0.060992613434791565, 0.13223595917224884, 0.05640945956110954, -0.09224635362625122, 0.07981666922569275, -0.014230815693736076, -0.3524332344532013, -0.07776977121829987, -0.08503005653619766, 0.07276752591133118, 0.012777956202626228, 0.12844207882881165, -0.13303765654563904, -0.05206049978733063, -0.04631683975458145, 0.07893726229667664, 0.013055853545665741, -0.02959877997636795, -0.10094918310642242, 0.014459332451224327, -0.004496203735470772, -0.09311854839324951, -0.03411954641342163, 0.0270443893969059, 0.0018547001527622342, 0.15166360139846802, -0.12054844200611115, 0.049127478152513504, 0.08113346248865128, -0.14585493505001068, 0.06868340075016022, -0.009453179314732552, 0.09957946836948395, -0.03837045282125473, 0.08997402340173721, 0.16009151935577393, 0.00477764243260026, 0.09466411918401718, 0.17649416625499725, -0.01705397292971611, -0.02724151499569416, 0.04842684790492058, -0.014683657325804234, -0.0724385529756546, -0.1120031327009201, -0.0633048266172409, -0.05177382007241249, -0.010144074447453022, 0.012341433204710484, 0.0744282528758049, 0.04312784597277641, 0.17019076645374298, -0.06594827026128769, 0.0535443089902401, -0.1419581174850464, 0.019538072869181633, 0.14619489014148712, -0.03163960576057434, -0.00014829292194917798, -0.10701519250869751, -0.16118071973323822, 0.07572086155414581, -0.007469928357750177, 0.08079590648412704, 0.015894122421741486, -0.020450709387660027, 0.11724889278411865, 0.031011095270514488, 0.07966947555541992, 0.026727134361863136, -0.10332746058702469, 0.01266884058713913, -0.024949971586465836, -0.05412982031702995, 0.0364188551902771, 0.10117433965206146, 0.062476176768541336, -0.10735323280096054, -0.0021743429824709892, -0.02861718460917473, 0.03462967276573181, 0.1889367550611496, 0.12161342799663544, -0.21514299511909485, 0.01069377176463604, 0.03774520382285118, -0.08331664651632309, -0.09440625458955765, 0.06247501075267792, 0.11140041053295135, -0.016055120155215263, 0.058712828904390335, -0.02375781536102295, 0.08859636634588242, -0.029445840045809746, 0.0798061341047287, 0.05252521485090256, -0.03940609097480774, 0.03527453914284706, 0.06961151212453842, -0.3550260365009308, 0.22299548983573914, -0.008314825594425201, -0.09938815981149673, -0.09260500222444534, -0.00986372958868742, -0.07000842690467834, -0.00972070824354887, 0.010848643258213997, -0.026651455089449883, 0.22760434448719025, -0.030160436406731606, 0.02015610970556736, 0.047253984957933426, 0.04566110298037529, -0.018895452842116356, 0.08732179552316666, -0.044939033687114716, -0.012262066826224327, -0.01834316924214363, -0.0705653503537178, -0.15315288305282593, -0.05645991116762161, -0.016330603510141373, 0.062425579875707626, -0.02991853468120098, 0.01837010122835636, 0.008866565302014351, 0.034005098044872284, 0.12693819403648376, 0.13864143192768097, -0.15174263715744019, -0.04214521124958992, 0.010241634212434292, 0.08456294238567352, -0.03329229727387428, 0.06539548188447952, -0.04408327117562294, 0.001481515821069479, -0.025787660852074623, -0.10165808349847794, 0.07005500048398972, 0.00147929263766855, -0.02250482700765133, -0.023888520896434784, 0.040588025003671646, -0.011061779223382473, -0.041392140090465546, 0.006963638123124838, -0.019512835890054703, -0.06584134697914124, -0.030729644000530243, -0.037231847643852234, -0.08309119194746017, -0.07507694512605667, -0.02378462627530098, -0.027835121378302574, -0.0075820148922502995, -0.028732402250170708, -0.06188717111945152, 0.18773213028907776, 0.09702329337596893, -0.019745297729969025, 0.10128287225961685, 0.13449178636074066, 0.017844216898083687, -0.23566998541355133, -0.11628154665231705, -0.06093135103583336, 0.011361788958311081, -0.010657204315066338, -0.05168123170733452, 0.006400249432772398, -0.03359615057706833, 0.004349325783550739, 0.014682142063975334, -0.1889932006597519, -0.01888631284236908, 0.056496039032936096, 0.08818715810775757, 0.22621215879917145, -0.12237068265676498, -0.06723935902118683, -0.06163504719734192, -0.1131981834769249, 0.08168534934520721, 0.04052843898534775, 0.001059756032191217, -0.04927881807088852, 0.18566225469112396, 0.054081182926893234, -0.021107252687215805, -0.10586195439100266, -0.07040673494338989, -0.03490530699491501, -0.15280291438102722, -0.17713411152362823, 0.16253027319908142, -0.020188990980386734, 0.1643989533185959, -0.1630563884973526, 0.03656895086169243, -0.08099699020385742, -0.030154459178447723, -0.05900393798947334, -0.02484537661075592, -0.030173130333423615, -0.10029400140047073, -0.09191545099020004, -0.06870589405298233, 0.0015137521550059319, 0.009730175137519836, 0.01002420298755169, -0.04965392127633095, 0.03194653242826462, 0.10216201841831207, 0.12728580832481384, -0.020709404721856117, 0.041814785450696945, -0.07069884240627289, -0.04454528167843819, 0.04345665127038956, -0.24261952936649323, 0.0014690157258883119, 0.051265642046928406, -0.02148653008043766, 0.08646336197853088, -0.0025550085119903088, 0.00031842352473177016, 0.08825352787971497, 0.19088855385780334, -0.08338882774114609, -0.018312163650989532, -0.1171351745724678, -0.020013459026813507, -0.011992980726063251, 0.013240945525467396, 0.10634678602218628, -0.08281367272138596, -0.014296418987214565, -0.05053159222006798, 0.03317386284470558, -0.08075668662786484, -0.03546195477247238, 0.06615728884935379, 0.014058179222047329, -0.04470339044928551, -0.0703742504119873, 0.06636954843997955, -0.07147262990474701, -0.007939723320305347, 0.24594871699810028, -0.07360000908374786, -0.08462967723608017, 0.12977291643619537, 0.06473758816719055, -0.13670013844966888, 0.015391889028251171, -0.05158970132470131, -0.03609005734324455, 0.04845979064702988, 0.19375014305114746, -0.03772180527448654, 0.09962382912635803, -0.09065043181180954, 0.029278263449668884, -0.11792987585067749, 0.03989604115486145, 0.051565006375312805, 0.029495617374777794, -0.1911115199327469, 0.014230059459805489, -0.0564739927649498, 0.09884310513734818, -0.0712791383266449, -0.06167111173272133, -0.027686195448040962, -0.010395675897598267, -0.12182820588350296, 0.013789637945592403, -0.07935506850481033, -0.0717480257153511, 0.014924961142241955, -0.02740117348730564, -0.03712143003940582, -0.016848929226398468, -0.058863017708063126, 0.035020530223846436, 0.040038902312517166, 0.10164206475019455, -0.05147524178028107, -0.02871425822377205, 0.021904824301600456, -0.02144690230488777, 0.1303510069847107, -0.011808107607066631, -0.18214863538742065, -0.002467659767717123, -0.0899997428059578, -0.040547315031290054, 0.12251103669404984, 0.01757558435201645, 0.06671656668186188, 0.19423891603946686, -0.0067499903962016106, 0.04359569400548935, 0.08808828145265579, 0.03774895891547203, 0.12239863723516464, -0.059625912457704544, 0.07157192379236221, -0.07309483736753464, -0.08835454285144806, -0.04235561564564705, 0.04349001869559288, 0.07585523277521133, 0.015659116208553314, 0.10519582033157349, -0.08903355896472931, 0.021632174029946327, -0.09773826599121094, 0.04435759782791138, 0.040601734071969986, -0.06728749722242355, -0.1624782383441925, -0.04461563378572464, -0.00011523235298227519, -0.004251922946423292, 0.11695127934217453, 0.0542314276099205, -0.05698946490883827, 0.02839941903948784, -0.025824343785643578, 0.024219801649451256, -0.01859203539788723, 0.29201072454452515, 0.017094820737838745, -0.03712693974375725, -0.1643575131893158, 0.023092474788427353, -0.011540628969669342, -0.12300024926662445, 0.08605404943227768, 0.07721847295761108, -0.009752578102052212, 0.05248025804758072, 0.12826813757419586, 0.16888652741909027, -0.03014942817389965, -0.04240608960390091, 0.025163548067212105, 0.07696802914142609, -0.11240799725055695, 0.1052575558423996, 0.19255663454532623, 0.03731829673051834, -0.051706861704587936, 0.009307246655225754, 0.0027445561718195677, 0.023734409362077713, -0.07189902663230896, -0.1175154522061348, -0.16289347410202026, 0.019688380882143974, -0.0691247209906578, 0.06713484227657318, 0.02611451968550682, 0.035348664969205856, -0.10550089180469513, 0.018729371950030327, 0.12874434888362885, -0.008022873662412167, 0.0944441482424736, -0.017568625509738922, -0.01706542819738388, -0.0005294526927173138, 0.027341529726982117, -0.10668651759624481, 0.12619401514530182, 0.004730683285742998, 0.07958899438381195, -0.06628315150737762, -0.04174678027629852, -0.11339453607797623, -0.09367872029542923, -0.04577641934156418, 0.05476958304643631, 0.019463473930954933, 0.019578181207180023, 0.009242166765034199, -0.07726223766803741, -0.044836800545454025, 0.010594986379146576, -0.010891340672969818, -0.12927383184432983, -0.013084380887448788, 0.17935657501220703, -0.007919969968497753, 0.03360243886709213, 0.01391671970486641, -0.012732738628983498, -0.13551419973373413, 0.2385387271642685, 0.23441046476364136, -0.027859840542078018, -0.014673241414129734, -0.019813017919659615, 0.03475429117679596, -0.049412429332733154, 0.19776000082492828, 0.0028357000555843115, 0.2602749764919281, -0.06271476298570633, -0.04589061066508293, -0.0779002234339714, -0.009737488813698292, -0.06773922592401505, 0.017782721668481827, 0.041555337607860565, -0.04930737242102623, -0.013523943722248077, 0.1079082265496254, -0.1977352648973465, 0.10427799075841904, -0.15153776109218597, -0.08611235022544861, 0.003022788790985942, 0.0006565612275153399, -0.04241715371608734, 0.014161449857056141, 0.05282851681113243, -0.00480167381465435, -0.019914433360099792, 0.006723520811647177, 0.0302690751850605, -0.2192714661359787, 0.10469471663236618, 0.08205097913742065, -0.23269595205783844, 0.028347091749310493, -0.019223365932703018, 0.05988841503858566, 0.053471386432647705, -0.016955379396677017, -0.03164546564221382, 0.07445292174816132, 0.009085859172046185, 0.11564026772975922, -0.04280761256814003, -0.015395250171422958, 0.005340672098100185, -0.05629056319594383, 0.06171998009085655, -0.1284298151731491, -0.0038267469499260187, 0.021795347332954407, 0.09658944606781006, -0.1410149782896042, 0.040930673480033875, -0.002251217607408762, 0.08053545653820038, 0.05671023204922676, -0.036111511290073395, -0.010603683069348335, -0.029856812208890915, 0.018657758831977844, -0.028602106496691704, -0.037007641047239304, -0.015919150784611702, -0.10215046256780624, -0.10934906452894211, -0.03756320849061012, -0.022666219621896744, -0.2746734023094177, -0.00997922196984291, -0.08015692234039307, 0.08928319811820984, -0.09636471420526505, 0.08627842366695404, 0.22401681542396545, 0.04238072410225868, 0.028648605570197105, 0.06490515172481537, -0.06347111612558365, 0.1325930804014206, -0.08154750615358353, -0.06537696719169617 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # model This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 3.6250 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 2 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.9.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "model", "results": []}]}
text-generation
gagan3012/model
[ "transformers", "pytorch", "tensorboard", "gpt2", "text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# model This model is a fine-tuned version of distilgpt2 on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 3.6250 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 2 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.9.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
[ "# model\n\nThis model is a fine-tuned version of distilgpt2 on an unknown dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 3.6250", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.9.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# model\n\nThis model is a fine-tuned version of distilgpt2 on an unknown dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 3.6250", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.9.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ 66, 43, 6, 12, 8, 3, 90, 4, 37 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# model\n\nThis model is a fine-tuned version of distilgpt2 on an unknown dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 3.6250## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.9.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ -0.07980087399482727, 0.14180368185043335, -0.0025029443204402924, 0.09985678642988205, 0.15788549184799194, 0.03454642742872238, 0.09689822047948837, 0.13622651994228363, -0.11403767764568329, 0.06251244992017746, 0.07859344780445099, 0.07814902067184448, 0.05085495114326477, 0.1174091249704361, -0.0259455144405365, -0.23991964757442474, 0.008535833097994328, -0.00026586902095004916, -0.044098030775785446, 0.09922844916582108, 0.08454442024230957, -0.09284605830907822, 0.0758858323097229, 0.002261000918224454, -0.17486967146396637, 0.026403846219182014, -0.02267080545425415, -0.04141160473227501, 0.08808533102273941, 0.013599776662886143, 0.09475315362215042, -0.008737158961594105, 0.12139137834310532, -0.20063449442386627, -0.010017418302595615, 0.0848366841673851, 0.03602277487516403, 0.07258696109056473, 0.0642918050289154, 0.005532122682780027, 0.14776648581027985, -0.14801861345767975, 0.08057919144630432, 0.025365222245454788, -0.06467054039239883, -0.11178971827030182, -0.08194712549448013, 0.06902363896369934, 0.08092283457517624, 0.11347732692956924, 0.014214404858648777, 0.13778093457221985, -0.08036335557699203, 0.07747480273246765, 0.1999237984418869, -0.24630403518676758, -0.055018484592437744, 0.06835037469863892, 0.04537949711084366, 0.07097543030977249, -0.09098397195339203, -0.010556047782301903, 0.04102195426821709, 0.04026085510849953, 0.08493669331073761, -0.02011759765446186, -0.10855141282081604, 0.006193559151142836, -0.14051055908203125, -0.03376327082514763, 0.17576974630355835, 0.02649731934070587, -0.030229777097702026, -0.08443796634674072, -0.06176334619522095, -0.10419738292694092, -0.004027712624520063, -0.029637131839990616, 0.039796795696020126, -0.038403160870075226, -0.034936949610710144, -0.0703086331486702, -0.07472225278615952, -0.07423581182956696, -0.012602390721440315, 0.07495209574699402, 0.05396682396531105, 0.02348587103188038, -0.04137212038040161, 0.1137976348400116, -0.02380365878343582, -0.09866422414779663, -0.015324418433010578, -0.004835702478885651, -0.06234116852283478, -0.0475592203438282, -0.038949064910411835, -0.06487593054771423, -0.0008073485223576427, 0.09836344420909882, -0.07956823706626892, 0.0704944059252739, 0.018983814865350723, 0.018302632495760918, -0.024716153740882874, 0.14865249395370483, -0.045385781675577164, -0.021219663321971893, 0.0072090052999556065, 0.08653036504983902, 0.021692495793104172, -0.008058581501245499, -0.09691483527421951, -0.01694321632385254, 0.09855970740318298, 0.06752963364124298, -0.03553489223122597, 0.045777615159749985, -0.029616324231028557, -0.02357380837202072, 0.0055772410705685616, -0.13221289217472076, 0.04677606001496315, -0.011686512269079685, -0.1025756374001503, 0.008028700016438961, 0.053321801126003265, -0.02289918065071106, -0.06507285684347153, 0.07447660714387894, -0.08426637947559357, 0.018122268840670586, -0.10378669947385788, -0.08598703145980835, 0.019730469211935997, -0.0743710994720459, -0.016037428751587868, -0.06764615327119827, -0.2363254278898239, -0.03333490714430809, 0.043183572590351105, -0.0552205890417099, -0.03891230374574661, -0.061387479305267334, -0.06545154005289078, 0.00422430457547307, -0.007843772880733013, 0.1024712398648262, -0.04699287191033363, 0.07228078693151474, 0.003748774528503418, 0.031162161380052567, 0.012322415597736835, 0.04317225515842438, -0.08209158480167389, 0.018073903396725655, -0.11996070295572281, 0.07987435907125473, -0.07309383153915405, 0.024752788245677948, -0.10005329549312592, -0.11372607201337814, 0.005181761458516121, -0.015014786273241043, 0.053816575556993484, 0.12537983059883118, -0.17805762588977814, -0.028665300458669662, 0.14671960473060608, -0.06829787790775299, -0.047800462692976, 0.09113272279500961, -0.042502131313085556, 0.02391091175377369, 0.06881065666675568, 0.16481941938400269, 0.09771402925252914, -0.11776653677225113, -0.02157984860241413, 0.019210096448659897, 0.029233748093247414, -0.016358299180865288, 0.034980494529008865, -0.011698317714035511, 0.02286207675933838, 0.019898120313882828, -0.060303423553705215, 0.006642854772508144, -0.08496677875518799, -0.08720213174819946, -0.05978325381875038, -0.09291254729032516, 0.029817065224051476, 0.03964313864707947, 0.06283123791217804, -0.05939749628305435, -0.11190507560968399, 0.13986697793006897, 0.12924683094024658, -0.07051918655633926, 0.017801642417907715, -0.0588277131319046, 0.0418495275080204, -0.04658420756459236, -0.014562013559043407, -0.19751982390880585, -0.10263489186763763, 0.04096745699644089, -0.08907622843980789, 0.04761289805173874, 0.021621916443109512, 0.06525813043117523, 0.07177424430847168, -0.03261466324329376, -0.01042852085083723, -0.06471218913793564, -0.007324239704757929, -0.09840787947177887, -0.1814502328634262, -0.0384938046336174, -0.011926479637622833, 0.12110558152198792, -0.24210242927074432, 0.023315515369176865, -0.03482704237103462, 0.1090107336640358, 0.004929742775857449, -0.05848567187786102, 0.0016268015606328845, 0.05033765360713005, -0.0202578604221344, -0.09488902240991592, 0.04794824868440628, -0.002446049591526389, -0.0716247409582138, -0.0778624415397644, -0.14069737493991852, 0.02170441299676895, 0.08261238783597946, 0.015733733773231506, -0.09161486476659775, 0.02080574817955494, -0.05966043099761009, -0.05376139283180237, -0.07674483954906464, 0.0015685721300542355, 0.1950031965970993, -0.0010746610350906849, 0.13408948481082916, -0.05327008292078972, -0.0631343275308609, 0.00019723486911971122, 0.002032988704741001, -0.006569425575435162, 0.08012761175632477, 0.1086740791797638, -0.06885747611522675, 0.09587769210338593, 0.04844691976904869, -0.09073473513126373, 0.14382165670394897, -0.037180446088314056, -0.08469231426715851, -0.014017139561474323, 0.016485322266817093, -0.008425931446254253, 0.10609696060419083, -0.10930570960044861, -0.0020734297577291727, 0.02270486019551754, 0.03384429216384888, 0.05552666634321213, -0.18600070476531982, -0.0023269604425877333, 0.019475840032100677, -0.0531899519264698, -0.006706753745675087, -0.025632020086050034, 0.023009609431028366, 0.08647231757640839, 0.019461147487163544, -0.01654851995408535, 0.029462222009897232, -0.005117153283208609, -0.09284129738807678, 0.17968684434890747, -0.12185678631067276, -0.17024752497673035, -0.13330279290676117, 0.058032698929309845, -0.07694363594055176, -0.022849339991807938, 0.017786171287298203, -0.07800427079200745, -0.05618568882346153, -0.09476511180400848, 0.0013953434536233544, -0.07011561095714569, -0.01186030637472868, 0.03260866552591324, 0.009777111932635307, 0.0762602761387825, -0.13933975994586945, 0.004067768808454275, -0.020132051780819893, -0.11410260200500488, -0.007117049768567085, 0.042749371379613876, 0.10940667241811752, 0.13836491107940674, -0.01478655356913805, 0.026999354362487793, -0.02743053250014782, 0.2044120579957962, -0.059578560292720795, -0.0028218410443514585, 0.11195462197065353, 0.031031304970383644, 0.060521647334098816, 0.0957857221364975, 0.03085455670952797, -0.09080690145492554, 0.02077248878777027, 0.06294660270214081, -0.023653961718082428, -0.23397794365882874, -0.05199849233031273, -0.03909169137477875, -0.04909005016088486, 0.08307670801877975, 0.06031053140759468, 0.036741696298122406, 0.035374585539102554, 0.01207709964364767, 0.07814044505357742, -0.03122614696621895, 0.0836004838347435, 0.10488393157720566, 0.04472887143492699, 0.09031295031309128, -0.04228988289833069, -0.015736274421215057, 0.07314008474349976, -0.002111671958118677, 0.2947677969932556, -0.034317344427108765, 0.10006655752658844, 0.04438478499650955, 0.11965629458427429, -0.023322027176618576, 0.040589336305856705, 0.005885569844394922, 0.00946816522628069, 0.012899440713226795, -0.06347716599702835, -0.028507543727755547, 0.01711675524711609, -0.03492818400263786, 0.05511461943387985, -0.10338747501373291, 0.028222301974892616, 0.03504578769207001, 0.22853797674179077, 0.021690454334020615, -0.30233633518218994, -0.07757797837257385, 0.010813961736857891, -0.027314087375998497, -0.05901104211807251, 0.0113318907096982, 0.11051045358181, -0.1282687485218048, 0.050009217113256454, -0.05256422236561775, 0.09278390556573868, -0.07083594053983688, 0.00402840506285429, 0.03899049013853073, 0.13410741090774536, 0.0048629059456288815, 0.10323353856801987, -0.23032157123088837, 0.20106728374958038, 0.0219076257199049, 0.13024349510669708, -0.059673141688108444, 0.03262534737586975, 0.01461114827543497, 0.08456964045763016, 0.09841717034578323, -0.0007358089205808938, -0.023038482293486595, -0.1537684202194214, -0.07855153828859329, 0.033788032829761505, 0.10217670351266861, 0.01238889992237091, 0.07937301695346832, -0.04374543949961662, 0.006103213876485825, 0.05244719609618187, -0.10529806464910507, -0.16306239366531372, -0.12595517933368683, 0.024800041690468788, 0.010818960145115852, -0.05710747465491295, -0.06177528575062752, -0.09591109305620193, -0.03156401962041855, 0.19748353958129883, -0.017946790903806686, -0.06361090391874313, -0.1259988397359848, 0.06975758820772171, 0.11735022068023682, -0.06282650679349899, 0.01513333898037672, 0.013452990911900997, 0.09591245651245117, 0.04739761725068092, -0.1012219786643982, 0.044929273426532745, -0.0703231543302536, -0.14179548621177673, -0.0413224957883358, 0.10166554152965546, 0.06785427778959274, 0.053178202360868454, -0.017141537740826607, 0.00798086728900671, -0.006541681941598654, -0.10714368522167206, -0.0039571793749928474, 0.12344076484441757, 0.08323145657777786, 0.0693773701786995, -0.08614220470190048, 0.013113359920680523, -0.028255101293325424, -0.015048434026539326, 0.12973478436470032, 0.17519456148147583, -0.08449798822402954, 0.08443987369537354, 0.06583734601736069, -0.10776517540216446, -0.17615462839603424, 0.07072966545820236, 0.0884745717048645, 0.008365496061742306, 0.022269247099757195, -0.19771550595760345, 0.10909710824489594, 0.12296777218580246, -0.015128342434763908, 0.08457513153553009, -0.392463356256485, -0.1189471036195755, 0.058828867971897125, 0.11981672048568726, 0.0456266850233078, -0.15185090899467468, -0.03442953899502754, -0.026323875412344933, -0.14287638664245605, 0.11128058284521103, -0.07389217615127563, 0.11857271194458008, -0.020490774884819984, 0.11064810305833817, 0.022768855094909668, -0.05004409700632095, 0.13488887250423431, 0.04026150330901146, 0.07875259965658188, -0.0585009790956974, 0.02246745675802231, 0.08625327050685883, -0.07336565852165222, 0.08673402667045593, -0.04581617936491966, 0.06184788793325424, -0.14172221720218658, -0.021131787449121475, -0.0684533640742302, 0.0796462669968605, -0.04181629791855812, -0.0582423098385334, -0.056753747165203094, 0.038037654012441635, 0.06432735174894333, -0.020202351734042168, 0.04540933668613434, 0.02969258278608322, 0.0768112912774086, 0.06651481240987778, 0.08523818850517273, -0.014020590111613274, -0.12108826637268066, 0.00296094361692667, -0.00714262668043375, 0.05873354151844978, -0.12731562554836273, 0.012538875453174114, 0.14176970720291138, 0.035150766372680664, 0.14005304872989655, 0.04970572143793106, -0.052830759435892105, 0.007844432257115841, 0.03352578729391098, -0.13024234771728516, -0.14268983900547028, -0.0072279381565749645, -0.07127845287322998, -0.11063893139362335, 0.035361189395189285, 0.08485160768032074, -0.06841246783733368, -0.0022437642328441143, -0.015571481548249722, 0.0303659550845623, -0.02652497962117195, 0.1913663148880005, 0.03225450962781906, 0.049288660287857056, -0.08182075619697571, 0.1311601996421814, 0.07251915335655212, -0.06237510219216347, 0.06102357804775238, 0.09676092118024826, -0.09190724045038223, -0.02187195047736168, 0.07709448784589767, 0.17573539912700653, -0.061832934617996216, -0.052533626556396484, -0.09577620029449463, -0.06880359351634979, 0.051491159945726395, 0.1020653173327446, 0.05056365206837654, -0.005066245794296265, -0.050934772938489914, 0.03048151172697544, -0.16581057012081146, 0.07661203294992447, 0.047794755548238754, 0.0634465217590332, -0.1574997901916504, 0.1361805647611618, 0.02629767544567585, 0.05029548332095146, -0.01854945346713066, 0.0210587028414011, -0.10072726011276245, -0.012859306298196316, -0.11818007379770279, -0.009960565716028214, -0.0297772828489542, 0.00038748231600038707, -0.0066275871358811855, -0.02476891502737999, -0.04797027260065079, 0.0611993633210659, -0.06430288404226303, -0.06654537469148636, 0.01012368779629469, 0.04822836071252823, -0.1367597132921219, -0.0050453501753509045, 0.002656920114532113, -0.0810113251209259, 0.061607688665390015, 0.06477002799510956, 0.02212626114487648, 0.045453622937202454, -0.10545200854539871, -0.009200059808790684, 0.039769820868968964, 0.028354071080684662, 0.06305237859487534, -0.06431393325328827, -0.0012609867844730616, -0.011062968522310257, 0.05936884880065918, 0.01779504492878914, 0.048834990710020065, -0.12568596005439758, -0.00342611619271338, -0.06611186265945435, -0.03994978219270706, -0.06845808774232864, 0.060974400490522385, 0.10980749875307083, 0.03836185112595558, 0.15714557468891144, -0.08276783674955368, 0.03971223905682564, -0.18120521306991577, -0.034925851970911026, 0.004674912896007299, -0.03355924412608147, -0.049081843346357346, -0.023571092635393143, 0.0806943029165268, -0.057628706097602844, 0.13890181481838226, 0.016034187749028206, 0.0823148861527443, 0.02854384109377861, -0.02510933205485344, -0.024289332330226898, -0.004559630528092384, 0.15951214730739594, 0.054935697466135025, -0.02071324922144413, 0.0855848491191864, 0.002954519586637616, 0.07152663171291351, 0.030287738889455795, 0.21849863231182098, 0.11434965580701828, -0.0629487931728363, 0.08128981292247772, 0.050692055374383926, -0.11201343685388565, -0.1767391860485077, 0.08300736546516418, -0.03246476501226425, 0.129512757062912, -0.055256038904190063, 0.18272145092487335, 0.09030471742153168, -0.14837861061096191, 0.05862206220626831, -0.042577628046274185, -0.12446235120296478, -0.11903057247400284, -0.05866466462612152, -0.07121433317661285, -0.11586420238018036, 0.01658891886472702, -0.11749621480703354, 0.04768173769116402, 0.0875038430094719, 0.013582494109869003, -0.011274021118879318, 0.15464843809604645, -0.018764639273285866, 0.01985940709710121, 0.03646746277809143, 0.0025552420411258936, -0.013279836624860764, -0.06251271069049835, -0.04421377182006836, 0.028234055265784264, 0.003266590880230069, 0.08737756311893463, -0.03393721580505371, 0.004839873872697353, 0.03106292337179184, -0.024490075185894966, -0.052465688437223434, 0.016555948182940483, 0.02531040459871292, 0.02782401069998741, 0.04804304242134094, 0.05422702059149742, -0.026101650670170784, -0.04118457809090614, 0.25757473707199097, -0.07717692106962204, -0.09200792014598846, -0.10942771285772324, 0.24143587052822113, 0.044967249035835266, -0.02780100516974926, 0.061910927295684814, -0.12224610894918442, -0.018216297030448914, 0.20421580970287323, 0.16767317056655884, -0.046208515763282776, -0.031473129987716675, -0.017914647236466408, -0.01567118614912033, -0.04245210066437721, 0.12293551117181778, 0.10303229093551636, 0.0925564393401146, -0.051547273993492126, -0.010412691161036491, -0.01636824570596218, -0.017699863761663437, -0.10550069808959961, 0.057872433215379715, 0.029264304786920547, 0.009977679699659348, -0.021874751895666122, 0.05964723229408264, -0.02896023355424404, -0.15508031845092773, 0.025874506682157516, -0.14646083116531372, -0.1595909744501114, -0.012668316252529621, 0.09299079328775406, -0.04835936427116394, 0.05047006160020828, -0.0222356915473938, -0.011535427533090115, 0.1323917657136917, -0.018674949184060097, -0.10395179688930511, -0.09324786812067032, 0.05855054780840874, -0.06986255943775177, 0.22987158596515656, -0.001122010056860745, 0.06720379739999771, 0.10570161044597626, 0.021103311330080032, -0.1291019171476364, 0.061168912798166275, 0.05239250138401985, -0.061369311064481735, 0.034011732786893845, 0.15923616290092468, -0.05218588560819626, 0.06458207219839096, 0.0241700392216444, -0.0902256965637207, -0.026461021974682808, -0.049135979264974594, -0.01786636933684349, -0.07573045790195465, -0.03476784750819206, -0.06768970191478729, 0.15049424767494202, 0.20318585634231567, -0.018864253535866737, 0.01551029086112976, -0.09621891379356384, 0.030168188735842705, 0.04208087921142578, 0.10288789123296738, -0.030089061707258224, -0.19740214943885803, 0.014971788972616196, -0.0009626282844692469, 0.030784819275140762, -0.23960553109645844, -0.08655320107936859, 0.032549455761909485, -0.05582233890891075, -0.06902063637971878, 0.10162349045276642, 0.07158396393060684, 0.042051397264003754, -0.04314007610082626, -0.09099157154560089, -0.05642164498567581, 0.14615091681480408, -0.16220013797283173, -0.06296814233064651 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # pickuplines This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 5.7873 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 2 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 100.0 ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.9.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
{"license": "mit", "tags": ["generated_from_trainer"], "model-index": [{"name": "pickuplines", "results": []}]}
text-generation
gagan3012/pickuplines
[ "transformers", "pytorch", "tensorboard", "gpt2", "text-generation", "generated_from_trainer", "license:mit", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# pickuplines This model is a fine-tuned version of gpt2 on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 5.7873 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 2 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 100.0 ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.9.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
[ "# pickuplines\n\nThis model is a fine-tuned version of gpt2 on an unknown dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 5.7873", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 100.0", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.9.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# pickuplines\n\nThis model is a fine-tuned version of gpt2 on an unknown dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 5.7873", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 100.0", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.9.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ 63, 43, 6, 12, 8, 3, 91, 4, 37 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# pickuplines\n\nThis model is a fine-tuned version of gpt2 on an unknown dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 5.7873## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 100.0### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.9.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ -0.07381478697061539, 0.1252884417772293, -0.0021630642004311085, 0.09240163862705231, 0.16610439121723175, 0.035675570368766785, 0.08921786397695541, 0.12947838008403778, -0.11022061854600906, 0.07958666980266571, 0.09250079095363617, 0.09092976152896881, 0.05290792137384415, 0.13163864612579346, -0.023608975112438202, -0.2507363557815552, 0.0070801242254674435, 0.0012297442881390452, -0.06569917500019073, 0.09247155487537384, 0.07667523622512817, -0.09026961773633957, 0.07093443721532822, -0.006043156608939171, -0.19009172916412354, 0.01868113875389099, -0.01331247203052044, -0.04149207845330238, 0.08092442899942398, 0.017347514629364014, 0.09392992407083511, -0.008374117314815521, 0.11892750859260559, -0.19395501911640167, -0.0064322324469685555, 0.09448989480733871, 0.02564818412065506, 0.06206158176064491, 0.06652071326971054, 0.001558762975037098, 0.1565248668193817, -0.14492188394069672, 0.07813791185617447, 0.024442095309495926, -0.06933668255805969, -0.13489946722984314, -0.07755961269140244, 0.0664670318365097, 0.07541123777627945, 0.10134214162826538, -0.0021319936495274305, 0.1365414559841156, -0.08238154649734497, 0.07583258301019669, 0.2000795155763626, -0.2539159655570984, -0.06752148270606995, 0.10135466605424881, 0.03748398274183273, 0.08215773105621338, -0.09193848073482513, 0.004716963041573763, 0.042570799589157104, 0.03490393981337547, 0.08379855751991272, -0.011950219050049782, -0.08089204877614975, 0.020275713875889778, -0.15118494629859924, -0.022201156243681908, 0.1605713814496994, 0.024316903203725815, -0.02687733806669712, -0.07919564843177795, -0.05558307468891144, -0.10839414596557617, -0.0023686792701482773, -0.037710536271333694, 0.04151686280965805, -0.038131795823574066, -0.06621406972408295, -0.06614184379577637, -0.07009868323802948, -0.07695505023002625, -0.01098029688000679, 0.09314920008182526, 0.05432318150997162, 0.019381051883101463, -0.029083773493766785, 0.11492116004228592, -0.032958243042230606, -0.0914742648601532, -0.021047767251729965, 0.00017403463425580412, -0.07256287336349487, -0.0538344644010067, -0.04760640859603882, -0.062178730964660645, 0.0008815598557703197, 0.10101798921823502, -0.05691804736852646, 0.08708005398511887, 0.017982598394155502, 0.015916841104626656, -0.033228177577257156, 0.14391888678073883, -0.04472563415765762, -0.018351100385189056, -0.005520454607903957, 0.07382121682167053, 0.010690652765333652, -0.01832256093621254, -0.09943974763154984, -0.0036050863564014435, 0.08899209648370743, 0.052193038165569305, -0.045890867710113525, 0.055375173687934875, -0.028837284073233604, -0.023505093529820442, -0.03356136009097099, -0.1224951297044754, 0.04649965837597847, -0.00745974387973547, -0.09478206932544708, 0.01074997615069151, 0.04300851374864578, -0.022709203884005547, -0.05936650559306145, 0.09398694336414337, -0.09676595777273178, 0.0125503521412611, -0.10079407691955566, -0.08763004094362259, 0.01884918473660946, -0.0932827740907669, -0.027720559388399124, -0.06819938868284225, -0.21904337406158447, -0.04101097956299782, 0.048971984535455704, -0.05333229899406433, -0.02754862606525421, -0.0642235279083252, -0.06881773471832275, 0.006309535354375839, 0.001046160003170371, 0.10605132579803467, -0.04433676227927208, 0.07808233052492142, 0.003612736938521266, 0.04276758059859276, 0.02019701711833477, 0.04272184148430824, -0.06973063200712204, 0.017171550542116165, -0.12740767002105713, 0.08452454954385757, -0.08899874240159988, 0.026497192680835724, -0.09637821465730667, -0.11236196756362915, -0.011998273432254791, -0.007991957478225231, 0.049380917102098465, 0.11829136312007904, -0.17280146479606628, -0.05612894520163536, 0.17161908745765686, -0.07123766094446182, -0.03981064632534981, 0.09462238848209381, -0.04261587932705879, 0.020333165302872658, 0.07658030837774277, 0.182987242937088, 0.0844821184873581, -0.11076127737760544, -0.026990974321961403, 0.005252446513622999, 0.03375124931335449, -0.003938889130949974, 0.03657466545701027, -0.020629845559597015, 0.009318222291767597, 0.02032855898141861, -0.07379134744405746, 0.010537373833358288, -0.08902519941329956, -0.08361479640007019, -0.054199256002902985, -0.08895426988601685, 0.03797231987118721, 0.042016614228487015, 0.069635771214962, -0.05958503857254982, -0.11946957558393478, 0.1298792064189911, 0.10909277200698853, -0.0804048627614975, 0.00894977431744337, -0.06555312126874924, 0.04258652403950691, -0.04185556247830391, -0.004379049874842167, -0.18220211565494537, -0.11914901435375214, 0.036113813519477844, -0.08503610640764236, 0.0547264888882637, 0.015142587944865227, 0.06785304099321365, 0.07117687910795212, -0.03597522899508476, 0.006725623272359371, -0.06208318844437599, -0.011613829992711544, -0.10345333814620972, -0.19259445369243622, -0.03562787175178528, -0.011880441568791866, 0.12836599349975586, -0.2600230872631073, 0.02910582721233368, -0.008184715174138546, 0.11897848546504974, 0.010396027006208897, -0.07750099152326584, -0.00916951335966587, 0.05538883060216904, -0.016896860674023628, -0.10115024447441101, 0.047098223119974136, -0.008224650286138058, -0.07150281220674515, -0.08546792715787888, -0.14651937782764435, 0.037089377641677856, 0.0920804888010025, 0.00775856152176857, -0.10470949858427048, 0.01196279563009739, -0.05431186780333519, -0.050371844321489334, -0.08178234100341797, -0.0005308451945893466, 0.18706700205802917, -0.010248957201838493, 0.13116322457790375, -0.04919380322098732, -0.06120102480053902, 0.006963617168366909, 0.008579310029745102, -0.004818309098482132, 0.07956182211637497, 0.126923069357872, -0.07131244242191315, 0.11032801866531372, 0.033805325627326965, -0.10010818392038345, 0.16575771570205688, -0.033165574073791504, -0.08328264951705933, -0.022168751806020737, 0.02115173265337944, 0.0010897343745455146, 0.10370370000600815, -0.14229756593704224, -0.017482731491327286, 0.014366742223501205, 0.0321158841252327, 0.058733630925416946, -0.1848234236240387, -0.009818699210882187, 0.02085108682513237, -0.0529119074344635, -0.0030844174325466156, -0.03043389320373535, 0.007211313582956791, 0.08738166093826294, 0.02459077350795269, 0.004510955419391394, 0.028457866981625557, -0.004000645596534014, -0.09526406973600388, 0.19336606562137604, -0.11616133898496628, -0.1440400779247284, -0.12448178976774216, 0.053909238427877426, -0.0689873918890953, -0.008037038147449493, 0.016332503408193588, -0.1046205461025238, -0.050920892506837845, -0.08381512016057968, 0.015463133342564106, -0.07461205124855042, 0.0021337789949029684, 0.023205086588859558, -0.0008553406805731356, 0.0560242235660553, -0.13660769164562225, 0.0037223100662231445, -0.030657729133963585, -0.12876640260219574, 0.004826158285140991, 0.03415290266275406, 0.10335975140333176, 0.15352720022201538, -0.010732290334999561, 0.03291800245642662, -0.0344104990363121, 0.19908347725868225, -0.07117091864347458, -0.010751515626907349, 0.08186174929141998, 0.03216848894953728, 0.054710082709789276, 0.07504034042358398, 0.03678128495812416, -0.09961194545030594, 0.021876690909266472, 0.0720989853143692, -0.04013662412762642, -0.23753474652767181, -0.047445885837078094, -0.04214844852685928, -0.03961661830544472, 0.08900729566812515, 0.056949201971292496, 0.0417095310986042, 0.05160585418343544, 0.01192847453057766, 0.10267965495586395, -0.033649325370788574, 0.08386603742837906, 0.11106069386005402, 0.040097128599882126, 0.09977526217699051, -0.037579335272312164, -0.023254983127117157, 0.06964804232120514, -0.01757681742310524, 0.3195176124572754, -0.022370196878910065, 0.10081382095813751, 0.043931931257247925, 0.12363812327384949, -0.020886369049549103, 0.030295154079794884, 0.01727229915559292, -0.009279248304665089, 0.018365120515227318, -0.06618980318307877, -0.024368202313780785, 0.026759644970297813, -0.016676394268870354, 0.02948492020368576, -0.0903182402253151, 0.013270385563373566, 0.03498656675219536, 0.2128354012966156, 0.016498813405632973, -0.3040132224559784, -0.07328077405691147, 0.011900672689080238, -0.021195676177740097, -0.04813603311777115, 0.009307332336902618, 0.1260635256767273, -0.1295701265335083, 0.05669058859348297, -0.06059660762548447, 0.08764009177684784, -0.0650152638554573, 0.001600546413101256, 0.04639653488993645, 0.1317891627550125, -0.0074065327644348145, 0.09279964864253998, -0.2225070297718048, 0.20834076404571533, 0.017658760771155357, 0.12706813216209412, -0.061403244733810425, 0.03078387677669525, 0.021497564390301704, 0.061598390340805054, 0.09158258885145187, -0.002144836587831378, -0.03062978945672512, -0.16345663368701935, -0.05845588073134422, 0.04445825517177582, 0.12694492936134338, -0.008561269380152225, 0.08427178859710693, -0.03930345177650452, 0.010780935175716877, 0.0551922507584095, -0.08398638665676117, -0.15353865921497345, -0.12139477580785751, 0.0280105322599411, 0.004175384528934956, -0.049148205667734146, -0.06023123115301132, -0.09910157322883606, -0.04569106549024582, 0.20508287847042084, -0.04589392617344856, -0.058967724442481995, -0.1263529658317566, 0.08508549630641937, 0.09973493963479996, -0.057075534015893936, 0.02087167091667652, 0.025056632235646248, 0.0912461206316948, 0.052170656621456146, -0.09872884303331375, 0.06441928446292877, -0.06575769931077957, -0.1502251923084259, -0.04894119128584862, 0.1011887639760971, 0.0781051442027092, 0.04936707019805908, -0.022105759009718895, 0.01905164122581482, 0.0049736700020730495, -0.1183132529258728, 0.02424827590584755, 0.11236349493265152, 0.061088401824235916, 0.08572868257761002, -0.07302011549472809, 0.015811193734407425, -0.024594832211732864, -0.019196471199393272, 0.12081427872180939, 0.18183284997940063, -0.08137927204370499, 0.07100191712379456, 0.052032794803380966, -0.10685684531927109, -0.1852341890335083, 0.09925908595323563, 0.1009875014424324, 0.002201330615207553, 0.019246781244874, -0.2181486189365387, 0.12701338529586792, 0.13806559145450592, -0.013270151801407337, 0.098170205950737, -0.3704564571380615, -0.12153822183609009, 0.05681236833333969, 0.1247422993183136, 0.05897111818194389, -0.17015455663204193, -0.04585619643330574, -0.018374666571617126, -0.1388999968767166, 0.12268634140491486, -0.10015217959880829, 0.12520454823970795, -0.013407912105321884, 0.10116203874349594, 0.02198551595211029, -0.05028311163187027, 0.13682810962200165, 0.045753784477710724, 0.07552590221166611, -0.05872480943799019, 0.01023939996957779, 0.08126585930585861, -0.055568695068359375, 0.06392836570739746, -0.03907768800854683, 0.05111143738031387, -0.12285071611404419, -0.02939511090517044, -0.07096006721258163, 0.07672876119613647, -0.03456369414925575, -0.06343594193458557, -0.06505206227302551, 0.036156315356492996, 0.05378236621618271, -0.021606402471661568, 0.04921814799308777, 0.021531514823436737, 0.11888717114925385, 0.05631944537162781, 0.07480853796005249, -0.015523917973041534, -0.12225333601236343, -0.002342235529795289, -0.008127160370349884, 0.07033812254667282, -0.13532662391662598, 0.003958570305258036, 0.13957828283309937, 0.04662808030843735, 0.14561305940151215, 0.0658704861998558, -0.05739424377679825, 0.014606769196689129, 0.041923001408576965, -0.12052708864212036, -0.1597444862127304, -0.0068252976052463055, -0.09850500524044037, -0.10143867880105972, 0.04865463078022003, 0.08511711657047272, -0.06727868318557739, 0.007149386219680309, -0.0160045363008976, 0.017586201429367065, -0.027126168832182884, 0.20142018795013428, 0.0335678867995739, 0.0442192517220974, -0.08668269217014313, 0.11741141974925995, 0.06662768125534058, -0.07421810925006866, 0.06348629295825958, 0.08876106888055801, -0.08782031387090683, -0.01894039660692215, 0.06436675786972046, 0.17092889547348022, -0.053000643849372864, -0.03306981921195984, -0.08916796743869781, -0.0783960297703743, 0.0551113486289978, 0.11957500874996185, 0.04910725727677345, 0.0005599515279754996, -0.05211780592799187, 0.057474978268146515, -0.16094112396240234, 0.06096675246953964, 0.039432160556316376, 0.06970687210559845, -0.15562830865383148, 0.16740021109580994, 0.023663679137825966, 0.033072203397750854, -0.02303113229572773, 0.016914118081331253, -0.12336171418428421, -0.01695990189909935, -0.10279903560876846, -0.01580648124217987, -0.0335206501185894, -0.006497511174529791, 0.0043429844081401825, -0.026720689609646797, -0.04890900105237961, 0.053270578384399414, -0.07085474580526352, -0.06278937309980392, 0.012218005955219269, 0.047674428671598434, -0.13123533129692078, 0.007542692590504885, 0.00400262838229537, -0.08024265617132187, 0.06235450878739357, 0.0689961239695549, 0.029175693169236183, 0.053550224751234055, -0.13671621680259705, 0.004188877530395985, 0.04651869833469391, 0.014499984681606293, 0.055205125361680984, -0.047519054263830185, 0.003808635054156184, -0.02981792390346527, 0.07039695978164673, 0.032832056283950806, 0.05930086597800255, -0.12438534200191498, 0.0036823104601353407, -0.0756896585226059, -0.037979912012815475, -0.06749337911605835, 0.05924838408827782, 0.08826345950365067, 0.03548065200448036, 0.14754678308963776, -0.08909733593463898, 0.029291298240423203, -0.18455621600151062, -0.03447047248482704, 0.007563618011772633, -0.036529310047626495, -0.027893073856830597, -0.01737768016755581, 0.08266544342041016, -0.05559685081243515, 0.14216220378875732, 0.017845336347818375, 0.04912903159856796, 0.029424821957945824, -0.020729772746562958, -0.008129974827170372, -0.0019877972081303596, 0.1632683426141739, 0.07158797979354858, -0.02750425785779953, 0.05806940421462059, 0.015314091928303242, 0.07289937883615494, 0.018345322459936142, 0.22447070479393005, 0.101555734872818, -0.09623510390520096, 0.07443676888942719, 0.046440090984106064, -0.12180924415588379, -0.18485383689403534, 0.10118454694747925, -0.045020509511232376, 0.1333584189414978, -0.05220738798379898, 0.18037612736225128, 0.09539539366960526, -0.15520033240318298, 0.05278845876455307, -0.04489561542868614, -0.12050498276948929, -0.12565448880195618, -0.07392915338277817, -0.07233749330043793, -0.13666975498199463, 0.026080509647727013, -0.11625736951828003, 0.05037565529346466, 0.10534091293811798, 0.0251570213586092, 0.0030768841970711946, 0.1628316342830658, -0.028580747544765472, 0.012956253252923489, 0.04590999335050583, 0.0008288621320389211, -0.02521008625626564, -0.06621210277080536, -0.048151157796382904, 0.02421967126429081, -0.011429989710450172, 0.09027262032032013, -0.04119594767689705, -0.012249445542693138, 0.032255105674266815, -0.02317608892917633, -0.04978427290916443, 0.016828373074531555, 0.025982346385717392, 0.01857880689203739, 0.04143919795751572, 0.04044643044471741, -0.036391481757164, -0.04207390919327736, 0.2570926249027252, -0.07633952796459198, -0.06885863095521927, -0.10894515365362167, 0.22783496975898743, 0.04386812821030617, -0.014237673953175545, 0.053993236273527145, -0.12349636107683182, -0.006615626625716686, 0.20043215155601501, 0.16450759768486023, -0.035517364740371704, -0.023433031514286995, -0.002933230483904481, -0.017512725666165352, -0.031434375792741776, 0.12859903275966644, 0.08765941858291626, 0.1185956671833992, -0.0447121188044548, -0.013982155360281467, 0.00006571134144905955, -0.02126006782054901, -0.10234451293945312, 0.04215378686785698, 0.04065919667482376, 0.0072678918950259686, -0.02484547533094883, 0.06397230923175812, -0.03981245681643486, -0.15521839261054993, 0.05551019310951233, -0.15449947118759155, -0.15729297697544098, -0.0198383666574955, 0.07655473798513412, -0.04610922932624817, 0.057001397013664246, -0.017416613176465034, -0.020421599969267845, 0.10972633957862854, -0.013024890795350075, -0.11149772256612778, -0.10234081000089645, 0.059477414935827255, -0.058862827718257904, 0.22680185735225677, -0.013007374480366707, 0.07338626682758331, 0.10640135407447815, 0.028360478579998016, -0.10562340170145035, 0.06148068234324455, 0.04620710015296936, -0.08483518660068512, 0.02367430366575718, 0.15806609392166138, -0.062150850892066956, 0.0760166347026825, 0.030776003375649452, -0.09204231947660446, -0.0172598697245121, -0.03537416458129883, -0.014615640975534916, -0.0820741206407547, -0.033536139875650406, -0.084187813103199, 0.14296096563339233, 0.21840490400791168, -0.01654510758817196, 0.012530521489679813, -0.09377916902303696, 0.027825837954878807, 0.04850367456674576, 0.12141726911067963, -0.03471601754426956, -0.21253912150859833, 0.01625225506722927, -0.00322447856888175, 0.025293339043855667, -0.2518455982208252, -0.08482931554317474, 0.027778953313827515, -0.04412901774048805, -0.0660191997885704, 0.10043949633836746, 0.09078560769557953, 0.04412344470620155, -0.04883434250950813, -0.10626699030399323, -0.05183418467640877, 0.14806856215000153, -0.15708297491073608, -0.06672704219818115 ]
null
null
transformers
# Leetcode using AI :robot: GPT-2 Model for Leetcode Questions in python **Note**: the Answers might not make sense in some cases because of the bias in GPT-2 **Contribtuions:** If you would like to make the model better contributions are welcome Check out [CONTRIBUTIONS.md](https://github.com/gagan3012/project-code-py/blob/master/CONTRIBUTIONS.md) ### 📢 Favour: It would be highly motivating, if you can STAR⭐ this repo if you find it helpful. ## Model Two models have been developed for different use cases and they can be found at https://huggingface.co/gagan3012 The model weights can be found here: [GPT-2](https://huggingface.co/gagan3012/project-code-py) and [DistilGPT-2](https://huggingface.co/gagan3012/project-code-py-small) ### Example usage: ```python from transformers import AutoTokenizer, AutoModelWithLMHead tokenizer = AutoTokenizer.from_pretrained("gagan3012/project-code-py") model = AutoModelWithLMHead.from_pretrained("gagan3012/project-code-py") ``` ## Demo [![Streamlit App](https://static.streamlit.io/badges/streamlit_badge_black_white.svg)](https://share.streamlit.io/gagan3012/project-code-py/app.py) A streamlit webapp has been setup to use the model: https://share.streamlit.io/gagan3012/project-code-py/app.py ![image](https://user-images.githubusercontent.com/49101362/110356042-e69e4780-804a-11eb-94c6-a47fa3cd66b3.png) ## Example results: ### Question: ``` Write a function to delete a node in a singly-linked list. You will not be given access to the head of the list, instead you will be given access to the node to be deleted directly. It is guaranteed that the node to be deleted is not a tail node in the list. ``` ### Answer: ```python """ Write a function to delete a node in a singly-linked list. You will not be given access to the head of the list, instead you will be given access to the node to be deleted directly. It is guaranteed that the node to be deleted is not a tail node in the list. For example, a = 1->2->3 b = 3->1->2 t = ListNode(-1, 1) Note: The lexicographic ordering of the nodes in a tree matters. Do not assign values to nodes in a tree. Example 1: Input: [1,2,3] Output: 1->2->5 Explanation: 1->2->3->3->4, then 1->2->5[2] and then 5->1->3->4. Note: The length of a linked list will be in the range [1, 1000]. Node.val must be a valid LinkedListNode type. Both the length and the value of the nodes in a linked list will be in the range [-1000, 1000]. All nodes are distinct. """ # Definition for singly-linked list. # class ListNode: # def __init__(self, x): # self.val = x # self.next = None class Solution: def deleteNode(self, head: ListNode, val: int) -> None: """ BFS Linked List :param head: ListNode :param val: int :return: ListNode """ if head is not None: return head dummy = ListNode(-1, 1) dummy.next = head dummy.next.val = val dummy.next.next = head dummy.val = "" s1 = Solution() print(s1.deleteNode(head)) print(s1.deleteNode(-1)) print(s1.deleteNode(-1)) ```
{}
text-generation
gagan3012/project-code-py-small
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "autotrain_compatible", "endpoints_compatible", "has_space", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us
# Leetcode using AI :robot: GPT-2 Model for Leetcode Questions in python Note: the Answers might not make sense in some cases because of the bias in GPT-2 Contribtuions: If you would like to make the model better contributions are welcome Check out URL ### Favour: It would be highly motivating, if you can STAR⭐ this repo if you find it helpful. ## Model Two models have been developed for different use cases and they can be found at URL The model weights can be found here: GPT-2 and DistilGPT-2 ### Example usage: ## Demo ![Streamlit App](URL A streamlit webapp has been setup to use the model: URL !image ## Example results: ### Question: ### Answer:
[ "# Leetcode using AI :robot:\nGPT-2 Model for Leetcode Questions in python \n\nNote: the Answers might not make sense in some cases because of the bias in GPT-2\n\nContribtuions: If you would like to make the model better contributions are welcome Check out URL", "### Favour:\n\nIt would be highly motivating, if you can STAR⭐ this repo if you find it helpful.", "## Model\n\nTwo models have been developed for different use cases and they can be found at URL\n\nThe model weights can be found here: GPT-2 and DistilGPT-2", "### Example usage:", "## Demo\n![Streamlit App](URL\n\n\nA streamlit webapp has been setup to use the model: URL\n\n!image", "## Example results:", "### Question:", "### Answer:" ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n", "# Leetcode using AI :robot:\nGPT-2 Model for Leetcode Questions in python \n\nNote: the Answers might not make sense in some cases because of the bias in GPT-2\n\nContribtuions: If you would like to make the model better contributions are welcome Check out URL", "### Favour:\n\nIt would be highly motivating, if you can STAR⭐ this repo if you find it helpful.", "## Model\n\nTwo models have been developed for different use cases and they can be found at URL\n\nThe model weights can be found here: GPT-2 and DistilGPT-2", "### Example usage:", "## Demo\n![Streamlit App](URL\n\n\nA streamlit webapp has been setup to use the model: URL\n\n!image", "## Example results:", "### Question:", "### Answer:" ]
[ 54, 64, 25, 36, 6, 26, 5, 4, 4 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n# Leetcode using AI :robot:\nGPT-2 Model for Leetcode Questions in python \n\nNote: the Answers might not make sense in some cases because of the bias in GPT-2\n\nContribtuions: If you would like to make the model better contributions are welcome Check out URL### Favour:\n\nIt would be highly motivating, if you can STAR⭐ this repo if you find it helpful.## Model\n\nTwo models have been developed for different use cases and they can be found at URL\n\nThe model weights can be found here: GPT-2 and DistilGPT-2### Example usage:## Demo\n![Streamlit App](URL\n\n\nA streamlit webapp has been setup to use the model: URL\n\n!image## Example results:### Question:### Answer:" ]
[ -0.01831289380788803, 0.11950904875993729, -0.0032198207918554544, 0.07458344846963882, 0.06959093362092972, 0.023757020011544228, 0.19028472900390625, 0.13112078607082367, 0.0706179291009903, 0.014789854176342487, 0.14876337349414825, 0.08723016828298569, 0.04902274161577225, 0.1366937756538391, 0.007858491502702236, -0.3096257746219635, 0.04677940160036087, 0.004702850244939327, 0.0007331662927754223, 0.13310031592845917, 0.1381158083677292, -0.09480296820402145, 0.13155293464660645, 0.06612741947174072, -0.10754311829805374, -0.00754258967936039, 0.0014173827366903424, -0.0477561429142952, 0.07384905964136124, 0.04684638977050781, 0.035161539912223816, 0.03926166519522667, 0.07943585515022278, -0.10424615442752838, 0.01137119997292757, 0.026929685845971107, 0.029167253524065018, 0.09803580492734909, 0.019201423972845078, -0.04724644497036934, 0.18998782336711884, 0.056249093264341354, 0.03697758913040161, 0.036918073892593384, -0.10100796818733215, -0.1241474524140358, -0.06612788140773773, 0.05426827445626259, 0.12149685621261597, 0.049968354403972626, -0.058819301426410675, 0.17698583006858826, -0.1119614914059639, 0.07129266113042831, 0.18488140404224396, -0.18061260879039764, -0.054051224142313004, 0.09797936677932739, 0.07426322251558304, 0.015777504071593285, 0.033708252012729645, 0.05702776461839676, 0.0693998858332634, 0.014079472981393337, 0.01819346472620964, -0.04189606010913849, -0.05023352429270744, 0.02125433087348938, -0.14103062450885773, -0.09926300495862961, 0.15425710380077362, 0.06918064504861832, -0.06508957594633102, -0.16047413647174835, -0.06518565118312836, 0.09576868265867233, -0.019747521728277206, -0.06602544337511063, -0.023446818813681602, 0.0026071490719914436, -0.012825348414480686, -0.12633971869945526, -0.08313494920730591, -0.09612447768449783, -0.03549535199999809, 0.1145714819431305, 0.0367124080657959, 0.08114741742610931, -0.10174400359392166, 0.17084190249443054, -0.076485775411129, -0.09168486297130585, -0.006944435648620129, -0.07467985898256302, -0.0008955869707278907, 0.027843935415148735, 0.013222969137132168, -0.018728097900748253, 0.017440909519791603, 0.1097441092133522, 0.133557990193367, -0.011731003411114216, 0.027232427150011063, 0.029564835131168365, 0.048530079424381256, 0.14366154372692108, -0.15275289118289948, -0.041635144501924515, 0.014814377762377262, 0.02847553789615631, 0.005160035099834204, -0.06379912793636322, -0.13208049535751343, -0.032564982771873474, 0.052503801882267, 0.057008255273103714, 0.15404215455055237, 0.0477767176926136, -0.06009260192513466, -0.07529710233211517, 0.08411882072687149, -0.02103036642074585, 0.05799684301018715, 0.007437428925186396, -0.04131093621253967, 0.08413645625114441, 0.024350790306925774, 0.11103609204292297, -0.05494268611073494, -0.10806901752948761, -0.0904783234000206, 0.019337255507707596, -0.06404723227024078, -0.09578244388103485, 0.014725684188306332, -0.036843229085206985, -0.009120495989918709, -0.12704730033874512, -0.17665472626686096, 0.008476898074150085, 0.04096221923828125, -0.05718349665403366, -0.08003930747509003, -0.016145167872309685, -0.019435087218880653, -0.02919965237379074, -0.005707941018044949, 0.04989555850625038, -0.005531011149287224, 0.05751518905162811, -0.027551939710974693, 0.08941439539194107, -0.03308434411883354, 0.027102407068014145, -0.1051919087767601, 0.021182658150792122, -0.19802196323871613, 0.14727367460727692, -0.029829325154423714, 0.040858663618564606, -0.11950724571943283, -0.10360012948513031, 0.07343795895576477, -0.003932136110961437, -0.00984372291713953, 0.18654918670654297, -0.11206547170877457, -0.016826244071125984, 0.08107408136129379, -0.10873256623744965, -0.09731516242027283, 0.13567307591438293, -0.053831011056900024, 0.06744926422834396, 0.09375137835741043, 0.09775828570127487, 0.07005064934492111, -0.1088390126824379, 0.05483501777052879, -0.0019155469490215182, -0.06255849450826645, 0.12347695976495743, 0.11942016333341599, -0.019945388659834862, -0.15352904796600342, 0.02997288852930069, -0.15166746079921722, 0.007193797267973423, -0.0950314849615097, -0.08491512387990952, -0.029194310307502747, -0.029138850048184395, 0.08789137750864029, 0.04290473461151123, 0.03557537868618965, 0.016476519405841827, -0.12253180146217346, -0.07129298895597458, 0.10860872268676758, -0.07984192669391632, -0.023329682648181915, -0.11852233856916428, 0.13137322664260864, -0.1396123766899109, 0.0016217585653066635, -0.13713398575782776, -0.0656389370560646, 0.04544728621840477, -0.02792644500732422, 0.009467520751059055, 0.11408773809671402, 0.048663169145584106, 0.05182568356394768, 0.007811385206878185, 0.0017930876929312944, 0.01751038432121277, -0.008328829891979694, -0.0800812691450119, -0.11173119395971298, -0.06165143847465515, -0.025865335017442703, 0.07227981835603714, -0.0934227854013443, 0.003588012419641018, 0.026206359267234802, 0.14081688225269318, 0.054293882101774216, -0.044811420142650604, 0.044908154755830765, -0.04521309956908226, -0.053996045142412186, -0.02731495536863804, 0.004045251756906509, -0.06328281760215759, -0.10868456214666367, 0.11695855855941772, -0.07727675139904022, 0.03035236895084381, 0.10981105268001556, -0.030330196022987366, -0.03278094530105591, 0.08501166850328445, -0.03802470490336418, 0.025250840932130814, -0.06172006204724312, -0.045842494815588, 0.15903323888778687, 0.019997546449303627, 0.0627874806523323, -0.08723197877407074, -0.01707646995782852, 0.026356060057878494, -0.11369454115629196, 0.012345843017101288, 0.08428173512220383, 0.10912474244832993, -0.057050906121730804, 0.07715130597352982, 0.06284163147211075, 0.05933130159974098, 0.07743821293115616, 0.00992016401141882, -0.08459249138832092, -0.07928436249494553, 0.0175889004021883, -0.02594977244734764, 0.055083975195884705, -0.09078162163496017, 0.017132794484496117, 0.07556940615177155, 0.0163741372525692, 0.02336914837360382, -0.14861854910850525, -0.015300139784812927, 0.00590597465634346, -0.07553844898939133, -0.037946466356515884, 0.030703244730830193, 0.022958682850003242, 0.08986175060272217, -0.010413574054837227, -0.02221348136663437, 0.004062906838953495, -0.0011312896385788918, -0.14564435184001923, 0.1565297544002533, 0.006266733165830374, -0.22296853363513947, -0.09187103807926178, 0.009713927283883095, -0.023319890722632408, 0.006965231616050005, 0.02152933180332184, -0.04737572371959686, -0.044875748455524445, -0.08986871689558029, 0.1183822974562645, 0.03760392591357231, -0.042078714817762375, -0.005632308311760426, -0.012979226186871529, 0.03307637572288513, -0.10599061101675034, -0.043123867362737656, -0.010776514187455177, -0.05557333678007126, 0.05102569982409477, -0.00793323665857315, 0.08563636243343353, 0.10625114291906357, -0.00032998903770931065, 0.029768312349915504, 0.013784751296043396, 0.3244079053401947, -0.0911584198474884, 0.0369490385055542, 0.18089967966079712, -0.029799412935972214, 0.07470865547657013, 0.13748987019062042, 0.029005810618400574, -0.07479708641767502, 0.029851973056793213, -0.0011707688681781292, -0.06525392085313797, -0.11828036606311798, -0.09202557057142258, -0.02685091458261013, 0.011502295732498169, 0.058269668370485306, 0.01020894292742014, -0.025918686762452126, 0.13133201003074646, -0.02582935057580471, 0.06399203091859818, -0.0004957400960847735, 0.07807133346796036, 0.01298727560788393, -0.01383745763450861, 0.10663576424121857, -0.060837939381599426, -0.12990246713161469, 0.12013866752386093, -0.0094899982213974, 0.18455387651920319, -0.016731783747673035, 0.008299713023006916, 0.07881520688533783, 0.07095737010240555, 0.044016070663928986, 0.027309849858283997, 0.011538461782038212, -0.014605834148824215, -0.09022533893585205, -0.0446128286421299, -0.020557435229420662, 0.10579116642475128, 0.1223623976111412, -0.04839971661567688, -0.061497971415519714, 0.0332883857190609, 0.05095285177230835, 0.14479388296604156, 0.16045381128787994, -0.3122703433036804, -0.009265427477657795, 0.09858329594135284, -0.10508754849433899, -0.05089004337787628, 0.015349461697041988, 0.08531555533409119, -0.2326248586177826, 0.01390416081994772, -0.044877730309963226, 0.12071049213409424, -0.13494029641151428, 0.014911926351487637, -0.1013990044593811, -0.0186782106757164, 0.009903156198561192, 0.1147768571972847, -0.18730300664901733, 0.15422308444976807, -0.01911250501871109, 0.001404210226610303, -0.09504339843988419, 0.03511260449886322, 0.02092730440199375, 0.09697376936674118, 0.15242382884025574, 0.021834615617990494, 0.03917688503861427, -0.09371452778577805, -0.12556807696819305, 0.06209259480237961, -0.0015111807733774185, -0.14069968461990356, 0.09171655774116516, -0.03388853371143341, 0.010051247663795948, -0.01593705266714096, -0.023912720382213593, -0.10875500738620758, -0.07416997104883194, 0.07338183373212814, -0.050132058560848236, 0.06931105256080627, -0.06490394473075867, -0.042658183723688126, 0.01879481039941311, 0.1220012679696083, 0.020825695246458054, -0.13404597342014313, -0.09374557435512543, 0.043868791311979294, 0.1414601057767868, -0.09116625785827637, 0.027699284255504608, -0.02195815183222294, 0.08048880100250244, 0.021381454542279243, -0.06652689725160599, 0.07056663930416107, -0.06968580186367035, -0.17823773622512817, -0.017837457358837128, 0.03978341072797775, 0.0067312633618712425, 0.04878396540880203, 0.05035540834069252, 0.030098482966423035, -0.0854545459151268, -0.15740080177783966, 0.0020652697421610355, 0.12583570182323456, -0.01992838829755783, 0.06521911919116974, 0.11896388232707977, -0.03265461325645447, -0.04972037672996521, 0.038666367530822754, 0.1685422658920288, 0.18475903570652008, -0.1006130501627922, 0.07364866137504578, 0.1146516352891922, -0.0197916179895401, -0.19233152270317078, -0.07217424362897873, 0.03455895185470581, 0.031358666718006134, 0.040244948118925095, -0.09819453954696655, 0.06297890096902847, -0.03399910405278206, -0.05522565543651581, -0.017986483871936798, -0.28644677996635437, -0.08296819031238556, 0.09191243350505829, 0.07374449074268341, 0.06139922887086868, -0.11035539954900742, -0.05428389832377434, 0.03606787696480751, -0.27140241861343384, 0.1317131221294403, 0.038766905665397644, 0.040376532822847366, 0.02411186695098877, 0.16868804395198822, 0.05857647955417633, -0.058630917221307755, 0.18758468329906464, -0.047947872430086136, 0.02994425781071186, -0.09284527599811554, -0.11492660641670227, 0.050369590520858765, -0.02885023131966591, 0.19486673176288605, -0.020305976271629333, 0.08458036929368973, -0.1369173675775528, -0.06369104981422424, -0.09972091019153595, 0.03678803890943527, -0.008908746764063835, -0.0932864099740982, -0.0854521170258522, 0.0050461264327168465, -0.014889913611114025, 0.014578672125935555, 0.003989683464169502, -0.013120845891535282, -0.01331683062016964, 0.03791932761669159, 0.1459946632385254, 0.0320991687476635, -0.06229634955525398, -0.046594660729169846, -0.031166136264801025, 0.10109816491603851, -0.22930808365345, 0.05964479595422745, 0.0788034051656723, 0.006973681040108204, 0.14836977422237396, 0.014101861976087093, -0.10672646760940552, 0.09860862046480179, 0.07864200323820114, -0.16402395069599152, -0.26062676310539246, -0.06067324057221413, -0.11731503158807755, -0.06494566053152084, 0.003796936245635152, 0.146030992269516, -0.05573538690805435, -0.06497221440076828, -0.002137592760846019, 0.06218699738383293, -0.009945702739059925, 0.027925437316298485, 0.09745572507381439, 0.028135310858488083, -0.12070854008197784, 0.006091214716434479, 0.032225947827100754, -0.002273811027407646, 0.048591792583465576, -0.008868133649230003, -0.07226262241601944, -0.059243831783533096, -0.13756437599658966, 0.09406045824289322, -0.05682305991649628, -0.0016043902141973376, -0.07895544916391373, -0.051609985530376434, 0.013426462188363075, -0.16411037743091583, 0.05475133657455444, 0.03843051940202713, -0.005886564496904612, 0.010640564374625683, -0.04814746230840683, 0.09670943021774292, 0.10916519165039062, -0.006625303067266941, -0.17076610028743744, 0.0507076159119606, 0.025498539209365845, 0.1048908606171608, -0.05638120695948601, -0.019645772874355316, -0.06360270082950592, -0.026865432038903236, -0.17190669476985931, -0.035431601107120514, -0.1459382176399231, -0.0011061348486691713, 0.016787441447377205, -0.1525813639163971, -0.04562946781516075, 0.02139301598072052, -0.11635465919971466, -0.027958355844020844, -0.0077391378581523895, 0.0691031664609909, -0.12341952323913574, -0.006988669279962778, 0.13320466876029968, -0.050480917096138, 0.14601248502731323, 0.010424603708088398, -0.05544855445623398, 0.012599579989910126, -0.048915497958660126, 0.004332809709012508, 0.02765929140150547, 0.016944075003266335, 0.049342572689056396, -0.12360743433237076, 0.038834307342767715, 0.000662120059132576, 0.08093590289354324, 0.013020569458603859, 0.06410106271505356, -0.11805896461009979, -0.0038361013866961002, 0.030760467052459717, -0.10927009582519531, -0.054563023149967194, 0.06701575219631195, -0.0069759562611579895, 0.07491965591907501, 0.12942050397396088, 0.019305426627397537, 0.11428374797105789, -0.14851176738739014, 0.007033524103462696, -0.0319795198738575, -0.0029673539102077484, -0.05179852619767189, -0.05587449669837952, 0.014027293771505356, -0.05540338531136513, 0.16035403311252594, 0.07456247508525848, 0.029612189158797264, -0.025875931605696678, 0.136451855301857, 0.19397693872451782, 0.018428048118948936, 0.09586082398891449, -0.008922601118683815, 0.006399200297892094, -0.08858747780323029, 0.011571093462407589, 0.010909682139754295, -0.08593513071537018, 0.051661115139722824, 0.050062790513038635, -0.0027633877471089363, 0.054227229207754135, 0.050636522471904755, 0.014928143471479416, -0.19640864431858063, -0.03160027414560318, -0.02635939233005047, 0.09630494564771652, -0.029966287314891815, 0.07727925479412079, 0.14487521350383759, -0.0902261808514595, 0.014232352375984192, 0.06176376715302467, -0.0479385145008564, -0.13811944425106049, -0.2284046858549118, -0.09044378250837326, -0.16672170162200928, 0.04047147557139397, -0.1098480075597763, 0.04898756742477417, 0.06011335179209709, 0.01745699532330036, -0.06658224761486053, 0.0879235789179802, 0.04988009110093117, -0.06443773955106735, 0.03841070830821991, 0.016431113705039024, 0.0062170750461518764, 0.06589597463607788, 0.009380693547427654, 0.005881139542907476, 0.08277296274900436, 0.057611484080553055, 0.0337604284286499, -0.08136166632175446, 0.023625748232007027, -0.09799037128686905, -0.09796681255102158, -0.08199445903301239, -0.016041651368141174, -0.015676548704504967, 0.0730559229850769, 0.03764225170016289, 0.027601854875683784, -0.013576569966971874, 0.22756627202033997, -0.07716436684131622, -0.06492087990045547, -0.24241507053375244, 0.2436901181936264, -0.08986761420965195, 0.05550387501716614, 0.009442746639251709, -0.08001887798309326, -0.026156634092330933, 0.2496470957994461, 0.21894630789756775, -0.06521257013082504, -0.014505277387797832, -0.014814316295087337, -0.014976116828620434, 0.025420671328902245, 0.09290333837270737, -0.053909141570329666, 0.2247876524925232, -0.06671988219022751, 0.07692933082580566, -0.08118098974227905, -0.07806091755628586, -0.030545730143785477, 0.08206500113010406, 0.039877552539110184, -0.028344223275780678, -0.06818082183599472, 0.1431514024734497, -0.04775647819042206, -0.003800211474299431, -0.10057467967271805, -0.07187993824481964, -0.09564530104398727, 0.044499803334474564, 0.006637594662606716, 0.030632875859737396, 0.09203093498945236, -0.04398171603679657, 0.0008315376471728086, 0.1891605406999588, 0.013807080686092377, -0.15658819675445557, 0.019115014001727104, 0.10901311039924622, -0.06515051424503326, 0.18825778365135193, 0.0035999955143779516, 0.08345755934715271, 0.10892047733068466, -0.05323494225740433, -0.14066049456596375, 0.02736116200685501, 0.04342181980609894, 0.01584317721426487, 0.009889237582683563, 0.06472177058458328, 0.024146443232893944, -0.0012248856946825981, 0.09229514002799988, -0.04710526019334793, 0.008427252061665058, -0.04392610117793083, 0.010633488185703754, -0.12104611098766327, 0.01746564358472824, -0.07344898581504822, 0.07539475709199905, 0.13158538937568665, -0.06890976428985596, 0.013023818843066692, -0.03173096105456352, 0.08953497558832169, -0.022106263786554337, 0.030918283388018608, 0.003443476976826787, -0.14915348589420319, -0.03100551664829254, 0.05237823724746704, 0.04499116912484169, -0.14629817008972168, -0.05088253691792488, -0.021450340747833252, -0.024989694356918335, 0.009345301426947117, 0.11694356054067612, 0.16698551177978516, -0.020570658147335052, 0.008157193660736084, -0.07712368667125702, -0.07892249524593353, 0.10112573206424713, -0.0763571634888649, -0.06407050788402557 ]
null
null
transformers
# Leetcode using AI :robot: GPT-2 Model for Leetcode Questions in python **Note**: the Answers might not make sense in some cases because of the bias in GPT-2 **Contribtuions:** If you would like to make the model better contributions are welcome Check out [CONTRIBUTIONS.md](https://github.com/gagan3012/project-code-py/blob/master/CONTRIBUTIONS.md) ### 📢 Favour: It would be highly motivating, if you can STAR⭐ this repo if you find it helpful. ## Model Two models have been developed for different use cases and they can be found at https://huggingface.co/gagan3012 The model weights can be found here: [GPT-2](https://huggingface.co/gagan3012/project-code-py) and [DistilGPT-2](https://huggingface.co/gagan3012/project-code-py-small) ### Example usage: ```python from transformers import AutoTokenizer, AutoModelWithLMHead tokenizer = AutoTokenizer.from_pretrained("gagan3012/project-code-py") model = AutoModelWithLMHead.from_pretrained("gagan3012/project-code-py") ``` ## Demo [![Streamlit App](https://static.streamlit.io/badges/streamlit_badge_black_white.svg)](https://share.streamlit.io/gagan3012/project-code-py/app.py) A streamlit webapp has been setup to use the model: https://share.streamlit.io/gagan3012/project-code-py/app.py ![image](https://user-images.githubusercontent.com/49101362/110356042-e69e4780-804a-11eb-94c6-a47fa3cd66b3.png) ## Example results: ### Question: ``` Write a function to delete a node in a singly-linked list. You will not be given access to the head of the list, instead you will be given access to the node to be deleted directly. It is guaranteed that the node to be deleted is not a tail node in the list. ``` ### Answer: ```python """ Write a function to delete a node in a singly-linked list. You will not be given access to the head of the list, instead you will be given access to the node to be deleted directly. It is guaranteed that the node to be deleted is not a tail node in the list. For example, a = 1->2->3 b = 3->1->2 t = ListNode(-1, 1) Note: The lexicographic ordering of the nodes in a tree matters. Do not assign values to nodes in a tree. Example 1: Input: [1,2,3] Output: 1->2->5 Explanation: 1->2->3->3->4, then 1->2->5[2] and then 5->1->3->4. Note: The length of a linked list will be in the range [1, 1000]. Node.val must be a valid LinkedListNode type. Both the length and the value of the nodes in a linked list will be in the range [-1000, 1000]. All nodes are distinct. """ # Definition for singly-linked list. # class ListNode: # def __init__(self, x): # self.val = x # self.next = None class Solution: def deleteNode(self, head: ListNode, val: int) -> None: """ BFS Linked List :param head: ListNode :param val: int :return: ListNode """ if head is not None: return head dummy = ListNode(-1, 1) dummy.next = head dummy.next.val = val dummy.next.next = head dummy.val = "" s1 = Solution() print(s1.deleteNode(head)) print(s1.deleteNode(-1)) print(s1.deleteNode(-1)) ```
{}
text-generation
gagan3012/project-code-py
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Leetcode using AI :robot: GPT-2 Model for Leetcode Questions in python Note: the Answers might not make sense in some cases because of the bias in GPT-2 Contribtuions: If you would like to make the model better contributions are welcome Check out URL ### Favour: It would be highly motivating, if you can STAR⭐ this repo if you find it helpful. ## Model Two models have been developed for different use cases and they can be found at URL The model weights can be found here: GPT-2 and DistilGPT-2 ### Example usage: ## Demo ![Streamlit App](URL A streamlit webapp has been setup to use the model: URL !image ## Example results: ### Question: ### Answer:
[ "# Leetcode using AI :robot:\nGPT-2 Model for Leetcode Questions in python \n\nNote: the Answers might not make sense in some cases because of the bias in GPT-2\n\nContribtuions: If you would like to make the model better contributions are welcome Check out URL", "### Favour:\n\nIt would be highly motivating, if you can STAR⭐ this repo if you find it helpful.", "## Model\n\nTwo models have been developed for different use cases and they can be found at URL\n\nThe model weights can be found here: GPT-2 and DistilGPT-2", "### Example usage:", "## Demo\n![Streamlit App](URL\n\n\nA streamlit webapp has been setup to use the model: URL\n\n!image", "## Example results:", "### Question:", "### Answer:" ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Leetcode using AI :robot:\nGPT-2 Model for Leetcode Questions in python \n\nNote: the Answers might not make sense in some cases because of the bias in GPT-2\n\nContribtuions: If you would like to make the model better contributions are welcome Check out URL", "### Favour:\n\nIt would be highly motivating, if you can STAR⭐ this repo if you find it helpful.", "## Model\n\nTwo models have been developed for different use cases and they can be found at URL\n\nThe model weights can be found here: GPT-2 and DistilGPT-2", "### Example usage:", "## Demo\n![Streamlit App](URL\n\n\nA streamlit webapp has been setup to use the model: URL\n\n!image", "## Example results:", "### Question:", "### Answer:" ]
[ 50, 64, 25, 36, 6, 26, 5, 4, 4 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Leetcode using AI :robot:\nGPT-2 Model for Leetcode Questions in python \n\nNote: the Answers might not make sense in some cases because of the bias in GPT-2\n\nContribtuions: If you would like to make the model better contributions are welcome Check out URL### Favour:\n\nIt would be highly motivating, if you can STAR⭐ this repo if you find it helpful.## Model\n\nTwo models have been developed for different use cases and they can be found at URL\n\nThe model weights can be found here: GPT-2 and DistilGPT-2### Example usage:## Demo\n![Streamlit App](URL\n\n\nA streamlit webapp has been setup to use the model: URL\n\n!image## Example results:### Question:### Answer:" ]
[ -0.019194507971405983, 0.12044961005449295, -0.00401445385068655, 0.07311122119426727, 0.06294218450784683, 0.04274298623204231, 0.217446431517601, 0.11595702171325684, 0.09536070376634598, 0.0022928747348487377, 0.12468336522579193, 0.07126764208078384, 0.07014438509941101, 0.1494578868150711, 0.022595569491386414, -0.3138583302497864, 0.051876772195100784, 0.017651444301009178, -0.0004495097673498094, 0.12673838436603546, 0.13470374047756195, -0.07687171548604965, 0.1316135972738266, 0.07106712460517883, -0.1011677011847496, -0.020905664190649986, 0.023415518924593925, -0.03996727615594864, 0.06614112108945847, 0.04104682430624962, 0.02974727563560009, 0.050196241587400436, 0.09590881317853928, -0.07745368033647537, 0.01801455020904541, 0.02758140116930008, 0.02743050642311573, 0.10540289431810379, 0.009256914258003235, -0.05346863716840744, 0.1820991039276123, 0.06670358031988144, 0.03940410912036896, 0.03600263223052025, -0.1030910313129425, -0.11513563990592957, -0.0599701814353466, 0.054121389985084534, 0.1214170828461647, 0.0414627268910408, -0.047246403992176056, 0.15924201905727386, -0.09310928732156754, 0.062290776520967484, 0.16464002430438995, -0.18562987446784973, -0.0548706129193306, 0.10365966707468033, 0.04570930078625679, 0.011632093228399754, 0.04817378148436546, 0.06699560582637787, 0.07109297066926956, 0.003980810288339853, 0.03335051238536835, -0.04283907264471054, -0.006481969729065895, -0.008049790747463703, -0.14196564257144928, -0.1003040298819542, 0.15559585392475128, 0.05911967530846596, -0.07804407179355621, -0.12301839143037796, -0.07883960753679276, 0.1135309562087059, -0.009981734678149223, -0.0659651979804039, -0.015966936945915222, 0.011697648093104362, -0.0036705038510262966, -0.15625928342342377, -0.06816801428794861, -0.12028530985116959, -0.05507809668779373, 0.14283645153045654, 0.04490786790847778, 0.09430965036153793, -0.1153780072927475, 0.1802164614200592, -0.09349840134382248, -0.08958036452531815, -0.003239302895963192, -0.06813699007034302, -0.006188873667269945, 0.024951020255684853, 0.028079181909561157, -0.005165465176105499, 0.00937761552631855, 0.131239652633667, 0.13042515516281128, -0.008453424088656902, 0.030168918892741203, 0.017643364146351814, 0.06789883226156235, 0.13031582534313202, -0.16002991795539856, -0.03777404874563217, 0.02170025371015072, 0.016677645966410637, 0.006840296555310488, -0.06536135822534561, -0.12890836596488953, -0.04818456247448921, 0.041219305247068405, 0.056080855429172516, 0.15261851251125336, 0.05390947312116623, -0.06303094327449799, -0.09249410778284073, 0.08739384263753891, -0.018218057230114937, 0.05617962032556534, -0.0002983871381729841, -0.03366784751415253, 0.11724022030830383, 0.011000814847648144, 0.11006636917591095, -0.057740725576877594, -0.11446404457092285, -0.09372980892658234, 0.01946193352341652, -0.059238385409116745, -0.07798200845718384, 0.008069154806435108, -0.044128838926553726, -0.003492968389764428, -0.1215125098824501, -0.20769084990024567, 0.013406451791524887, 0.04432033374905586, -0.07097892463207245, -0.0798858180642128, -0.012718874961137772, -0.005953605752438307, -0.03900953009724617, -0.009680172428488731, 0.03271046280860901, -0.003954572137445211, 0.05054174363613129, -0.005942024290561676, 0.10412599891424179, -0.05002634972333908, 0.02182125300168991, -0.09564134478569031, 0.022425560280680656, -0.20583772659301758, 0.14179587364196777, -0.014940154738724232, 0.026758288964629173, -0.11704938113689423, -0.09429235756397247, 0.07352466136217117, -0.014687348157167435, -0.020303240045905113, 0.17301447689533234, -0.1256110668182373, -0.018870584666728973, 0.11178704351186752, -0.12664151191711426, -0.10310566425323486, 0.1324055790901184, -0.044829726219177246, 0.06511016190052032, 0.10243997722864151, 0.07096860557794571, 0.08467438817024231, -0.12029311060905457, 0.061639994382858276, -0.00824662670493126, -0.08021242916584015, 0.1255287379026413, 0.1372876614332199, -0.006982424296438694, -0.1907627284526825, 0.021629400551319122, -0.18636886775493622, -0.008407370187342167, -0.08869736641645432, -0.08166322857141495, -0.02414705418050289, -0.02639387734234333, 0.1109495609998703, 0.02346285991370678, 0.028100192546844482, 0.027460940182209015, -0.10993432253599167, -0.08634208142757416, 0.12260308116674423, -0.0656479150056839, -0.024145955219864845, -0.12951670587062836, 0.1250154823064804, -0.14313708245754242, 0.009064163081347942, -0.12199640274047852, -0.07357868552207947, 0.04384159669280052, -0.04228402301669121, 0.026967104524374008, 0.11354539543390274, 0.04708174243569374, 0.06580248475074768, -0.0010758048156276345, -0.0013393621193245053, 0.006657571066170931, -0.014196212403476238, -0.05827825143933296, -0.11914648115634918, -0.058860380202531815, -0.02795335464179516, 0.09364768117666245, -0.07079306989908218, 0.0020874538458883762, 0.0065265740267932415, 0.13892291486263275, 0.04460856691002846, -0.03687482699751854, 0.0404537096619606, -0.059706609696149826, -0.048140060156583786, -0.02660994417965412, 0.00009050016524270177, -0.06225487217307091, -0.1037517860531807, 0.0933908075094223, -0.05316571891307831, 0.02844890020787716, 0.10519175976514816, -0.01309486385434866, -0.01799347810447216, 0.08754914999008179, -0.021920057013630867, 0.022023620083928108, -0.06101483851671219, -0.04081053286790848, 0.18081608414649963, 0.021393265575170517, 0.05218213424086571, -0.09399500489234924, 0.00044325407361611724, 0.025108151137828827, -0.1358264982700348, 0.01833347976207733, 0.04634811729192734, 0.12022151798009872, -0.04767128452658653, 0.06326084583997726, 0.056923430413007736, 0.06816551089286804, 0.07554402947425842, 0.024142427369952202, -0.0816102847456932, -0.09289565682411194, 0.0008770943968556821, -0.03141385316848755, 0.04232894256711006, -0.07952902466058731, 0.035635076463222504, 0.07225646823644638, 0.02542412281036377, 0.024546030908823013, -0.15281540155410767, -0.008219577372074127, 0.0022762196604162455, -0.09389940649271011, -0.055149540305137634, 0.015475865453481674, 0.009702892042696476, 0.0899326428771019, -0.002250075340270996, -0.013304402120411396, 0.0021617505699396133, 0.001336779328994453, -0.1414986252784729, 0.1362134963274002, 0.030511872842907906, -0.21754226088523865, -0.10870958864688873, 0.013757430016994476, -0.03740064054727554, 0.013068128377199173, 0.0033172317780554295, -0.03983200713992119, -0.042005740106105804, -0.0718504786491394, 0.13789252936840057, 0.037165720015764236, -0.037249814718961716, 0.03355610743165016, -0.006202572491019964, 0.030240772292017937, -0.08761342614889145, -0.0489705353975296, -0.007197876460850239, -0.07029443234205246, 0.0564955435693264, -0.00842798687517643, 0.07745623588562012, 0.11246147006750107, -0.008085624314844608, 0.03190769627690315, 0.026147684082388878, 0.30490103363990784, -0.08268165588378906, 0.04660018905997276, 0.1993703842163086, -0.009507937356829643, 0.07738479971885681, 0.13443401455879211, 0.025534389540553093, -0.08657228946685791, 0.02912507764995098, -0.010448395274579525, -0.06448944658041, -0.1226237341761589, -0.09366969764232635, -0.009881437756121159, -0.00012166189844720066, 0.06792749464511871, 0.0014282664051279426, -0.01386028341948986, 0.14506728947162628, -0.026963362470269203, 0.08094369620084763, 0.0045386324636638165, 0.07337362319231033, -0.026795245707035065, -0.0029286195058375597, 0.09328540414571762, -0.06081560626626015, -0.11634477972984314, 0.13327491283416748, 0.007289544679224491, 0.16972966492176056, -0.012415125966072083, 0.008982338942587376, 0.07566425204277039, 0.06067970395088196, 0.05248205363750458, 0.04091796278953552, 0.012120039202272892, -0.010053904727101326, -0.09050733596086502, -0.033120788633823395, -0.03338808938860893, 0.0872945711016655, 0.11875655502080917, -0.04486523196101189, -0.061181653290987015, 0.05300244316458702, 0.042626455426216125, 0.1339038610458374, 0.16046017408370972, -0.3194594383239746, -0.013742712326347828, 0.11212997883558273, -0.08353463560342789, -0.05783197283744812, 0.005419669207185507, 0.0689154788851738, -0.24713605642318726, 0.03439728543162346, -0.03848753124475479, 0.11184407025575638, -0.15768350660800934, 0.013810373842716217, -0.09560918062925339, 0.013853535056114197, 0.02563292719423771, 0.1157975122332573, -0.18890388309955597, 0.1334155946969986, -0.02449268288910389, 0.011439519934356213, -0.09209262579679489, 0.029977280646562576, 0.014736379496753216, 0.11477166414260864, 0.14639446139335632, 0.02690471149981022, 0.053038619458675385, -0.07908929884433746, -0.1418479084968567, 0.05696612596511841, -0.010737400501966476, -0.16099250316619873, 0.09450839459896088, -0.037119124084711075, 0.014017554000020027, -0.031350553035736084, -0.03750825300812721, -0.10255606472492218, -0.04583370313048363, 0.0787462592124939, -0.06213618442416191, 0.08828075975179672, -0.05468365550041199, -0.041616152971982956, 0.02438046969473362, 0.1060132160782814, 0.012909057550132275, -0.14634276926517487, -0.0978010818362236, 0.03786798566579819, 0.12122129648923874, -0.10405204445123672, 0.01365638803690672, -0.01992904581129551, 0.10112310200929642, 0.024669384583830833, -0.0755225121974945, 0.058007605373859406, -0.06573820114135742, -0.17185474932193756, -0.014342037960886955, 0.047949742525815964, 0.003518658922985196, 0.053031519055366516, 0.050263676792383194, 0.013405178673565388, -0.06510359048843384, -0.16926977038383484, 0.014260978437960148, 0.14384080469608307, -0.041106607764959335, 0.08173106610774994, 0.12284711003303528, -0.038498301059007645, -0.05651794373989105, 0.06546304374933243, 0.15340107679367065, 0.19263620674610138, -0.09075527638196945, 0.06388784945011139, 0.13369028270244598, -0.020751984789967537, -0.18525849282741547, -0.07439690828323364, 0.01883130520582199, 0.027209600433707237, 0.014628635719418526, -0.12198233604431152, 0.06133611500263214, -0.028011268004775047, -0.05201256647706032, -0.033058080822229385, -0.2729940414428711, -0.07713273167610168, 0.0802299752831459, 0.07983355969190598, 0.09263506531715393, -0.09170450270175934, -0.07006285339593887, 0.017135487869381905, -0.3031659722328186, 0.12885043025016785, 0.002603407483547926, 0.045192379504442215, 0.018382688984274864, 0.16345907747745514, 0.0471218079328537, -0.04829929396510124, 0.19190646708011627, -0.03879617154598236, 0.030197132378816605, -0.09523697942495346, -0.09671644866466522, 0.05761948227882385, -0.022000130265951157, 0.2236868441104889, 0.009557950310409069, 0.09154641628265381, -0.07894187420606613, -0.07224688678979874, -0.10442668944597244, 0.02392713539302349, -0.0018440423300489783, -0.10072875022888184, -0.07204823940992355, 0.009463534690439701, -0.027319807559251785, 0.00934693869203329, -0.021165665239095688, 0.013308389112353325, -0.04381457343697548, 0.020287225022912025, 0.14924588799476624, 0.02311938814818859, -0.05877452716231346, -0.06331157684326172, -0.03443176671862602, 0.1046612337231636, -0.2052379697561264, 0.06347938627004623, 0.08014940470457077, -0.0016620068345218897, 0.14609861373901367, 0.010050298646092415, -0.11201199144124985, 0.09697966277599335, 0.07495269179344177, -0.16705915331840515, -0.29896080493927, -0.05802211910486221, -0.1356716752052307, -0.06798849999904633, 0.031098853796720505, 0.1626286804676056, -0.04864351823925972, -0.060416605323553085, -0.004628513008356094, 0.05605672299861908, -0.0179127249866724, 0.02786906249821186, 0.08039478957653046, 0.021055137738585472, -0.13368411362171173, 0.03078790009021759, 0.03677055612206459, 0.03690505772829056, 0.04365909844636917, -0.028031695634126663, -0.08537934720516205, -0.060158055275678635, -0.15404647588729858, 0.09409362077713013, -0.06372752785682678, -0.0011193560203537345, -0.09017512947320938, -0.05450054630637169, -0.002635083394125104, -0.17836585640907288, 0.05220239982008934, 0.03723449632525444, -0.014662643894553185, 0.00809774361550808, -0.050384946167469025, 0.06259063631296158, 0.10439734905958176, -0.0017362125217914581, -0.17930060625076294, 0.07239366322755814, 0.038015224039554596, 0.1278359293937683, -0.06758757680654526, -0.024666665121912956, -0.052392639219760895, -0.02885602042078972, -0.17872165143489838, -0.038297366350889206, -0.1589241325855255, 0.0003662213566713035, 0.025028318166732788, -0.16040828824043274, -0.049208033829927444, 0.020496463403105736, -0.09804405272006989, -0.018446316942572594, -0.01166686974465847, 0.05437859147787094, -0.11990384012460709, 0.00879913941025734, 0.14289937913417816, -0.0404968298971653, 0.13773790001869202, 0.02686513401567936, -0.04257563129067421, 0.024506453424692154, -0.04112052172422409, 0.005784198641777039, 0.024911893531680107, 0.012933110818266869, 0.05685149505734444, -0.14894336462020874, 0.03674168512225151, 0.008824564516544342, 0.07404223084449768, 0.014208586886525154, 0.04986416921019554, -0.10870996117591858, 0.0019095547031611204, 0.040584053844213486, -0.10447850823402405, -0.05549618974328041, 0.06285557150840759, 0.00693141994997859, 0.0715479627251625, 0.13129577040672302, 0.023782484233379364, 0.11270550638437271, -0.1495514065027237, 0.00814176257699728, -0.03169923275709152, 0.007707301061600447, -0.03906400874257088, -0.05876579508185387, 0.006225991062819958, -0.052621517330408096, 0.14842955768108368, 0.06960728019475937, 0.026841411367058754, -0.0389094315469265, 0.16461648046970367, 0.1915903091430664, 0.008037291467189789, 0.07142450660467148, 0.005947183817625046, -0.004793253261595964, -0.08493151515722275, 0.0128254359588027, 0.005293190013617277, -0.09038446843624115, 0.06145254150032997, 0.023134680464863777, 0.004441781900823116, 0.04967505484819412, 0.06923598796129227, 0.055053964257240295, -0.20548522472381592, -0.039217881858348846, -0.03017263114452362, 0.04581001028418541, -0.001192739699035883, 0.07087312638759613, 0.1378880888223648, -0.0995418056845665, 0.01272972859442234, 0.08736472576856613, -0.05417141318321228, -0.1427505761384964, -0.2535955607891083, -0.09557005017995834, -0.16036468744277954, 0.03749938681721687, -0.11208359152078629, 0.03856267035007477, 0.0660870149731636, 0.015710962936282158, -0.06560419499874115, 0.08989354968070984, 0.052775267511606216, -0.0677073672413826, 0.022723151370882988, 0.015162507072091103, 0.010329533368349075, 0.0724954903125763, 0.01453087106347084, 0.0035289344377815723, 0.10199524462223053, 0.06399992108345032, 0.035985954105854034, -0.07448695600032806, 0.03011057898402214, -0.08445215970277786, -0.10527662932872772, -0.0857958272099495, -0.031262025237083435, -0.036663658916950226, 0.05134924128651619, 0.04904516786336899, 0.024579742923378944, -0.006992330774664879, 0.21811887621879578, -0.06582726538181305, -0.06464537978172302, -0.2541777193546295, 0.2485533207654953, -0.1025632843375206, 0.060874711722135544, 0.02848765440285206, -0.07134115695953369, -0.034051503986120224, 0.23038050532341003, 0.2068714052438736, -0.07117125391960144, -0.015281603671610355, -0.0248374342918396, -0.013325763866305351, 0.027858346700668335, 0.09484612941741943, -0.07331899553537369, 0.1933462917804718, -0.06043670326471329, 0.09916109591722488, -0.0704357922077179, -0.074148029088974, -0.010079603642225266, 0.05892220139503479, 0.01816689968109131, -0.002977331867441535, -0.07276790589094162, 0.14849942922592163, -0.04581853002309799, -0.002452983520925045, -0.11220981925725937, -0.03420259803533554, -0.09189704805612564, 0.053206585347652435, -0.018172822892665863, 0.044625524431467056, 0.10013136267662048, -0.04368717223405838, 0.009547242894768715, 0.20618556439876556, 0.019032180309295654, -0.15676316618919373, 0.016257112845778465, 0.119962178170681, -0.04997606202960014, 0.19632084667682648, 0.006269044242799282, 0.07365228235721588, 0.09864705055952072, -0.05813128128647804, -0.15512500703334808, 0.039108697324991226, 0.0219768937677145, 0.03511442244052887, 0.011004650965332985, 0.06659763306379318, 0.020686542615294456, -0.01569310389459133, 0.10420318692922592, -0.032877203077077866, -0.010272026993334293, -0.03384249657392502, 0.020149430260062218, -0.12153879553079605, 0.02869434468448162, -0.07707031816244125, 0.06858378648757935, 0.13533014059066772, -0.0750749260187149, 0.019183125346899033, -0.02745845727622509, 0.09623052924871445, -0.019655868411064148, 0.042925331741571426, 0.007519928272813559, -0.13542596995830536, -0.04292378947138786, 0.0595335029065609, 0.054518427699804306, -0.15497824549674988, -0.051456306129693985, -0.03054196573793888, -0.029743904247879982, 0.012065745890140533, 0.11467324942350388, 0.16439318656921387, 0.0025028761010617018, 0.009368888102471828, -0.09473808109760284, -0.08772408962249756, 0.10185086727142334, -0.09538216143846512, -0.07424216717481613 ]
null
null
transformers
# Generating Rap song Lyrics like Eminem Using GPT2 ### I have built a custom model for it using data from Kaggle Creating a new finetuned model using data lyrics from leading hip-hop stars ### My model can be accessed at: gagan3012/rap-writer ``` from transformers import AutoTokenizer, AutoModelWithLMHead tokenizer = AutoTokenizer.from_pretrained("gagan3012/rap-writer") model = AutoModelWithLMHead.from_pretrained("gagan3012/rap-writer") ```
{}
text-generation
gagan3012/rap-writer
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "autotrain_compatible", "endpoints_compatible", "has_space", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us
# Generating Rap song Lyrics like Eminem Using GPT2 ### I have built a custom model for it using data from Kaggle Creating a new finetuned model using data lyrics from leading hip-hop stars ### My model can be accessed at: gagan3012/rap-writer
[ "# Generating Rap song Lyrics like Eminem Using GPT2", "### I have built a custom model for it using data from Kaggle \n\nCreating a new finetuned model using data lyrics from leading hip-hop stars", "### My model can be accessed at: gagan3012/rap-writer" ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n", "# Generating Rap song Lyrics like Eminem Using GPT2", "### I have built a custom model for it using data from Kaggle \n\nCreating a new finetuned model using data lyrics from leading hip-hop stars", "### My model can be accessed at: gagan3012/rap-writer" ]
[ 54, 14, 33, 17 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n# Generating Rap song Lyrics like Eminem Using GPT2### I have built a custom model for it using data from Kaggle \n\nCreating a new finetuned model using data lyrics from leading hip-hop stars### My model can be accessed at: gagan3012/rap-writer" ]
[ -0.00805413257330656, 0.06665698438882828, -0.0017286957008764148, 0.03970538452267647, 0.08871903270483017, -0.030060764402151108, 0.09420587122440338, 0.03151325881481171, -0.055426158010959625, 0.004508565180003643, 0.11270253360271454, 0.2093261033296585, 0.031123634427785873, 0.19164736568927765, 0.0867040604352951, -0.22476692497730255, -0.0009697380010038614, 0.05654694139957428, 0.050565384328365326, 0.12212793529033661, 0.06260503828525543, 0.0037397844716906548, 0.07468543201684952, 0.09048733115196228, -0.2907525300979614, -0.023774469271302223, -0.045479584485292435, -0.0633367970585823, -0.03213099390268326, 0.022547917440533638, -0.050942111760377884, 0.16839566826820374, 0.01658041402697563, -0.091623455286026, 0.035698309540748596, 0.009371941909193993, -0.06604704260826111, 0.010581293143332005, 0.03130590543150902, -0.14192496240139008, 0.1297595053911209, 0.014185444451868534, -0.016855228692293167, 0.021250758320093155, -0.1819695085287094, -0.08757387846708298, -0.03576882556080818, -0.029549481347203255, 0.18881818652153015, 0.045957379043102264, -0.03927021473646164, -0.014359746128320694, -0.010834853164851665, 0.0020946902222931385, 0.09940610826015472, -0.307640016078949, -0.0445241741836071, 0.20249657332897186, 0.13956207036972046, -0.05695357173681259, -0.040158968418836594, 0.07024402916431427, -0.0009023974416777492, 0.0460704043507576, 0.1405419558286667, -0.07985008507966995, -0.08855243772268295, -0.028797348961234093, -0.08513979613780975, 0.007250786293298006, 0.17760665714740753, -0.031435053795576096, -0.037385232746601105, -0.11974319070577621, 0.03395162522792816, -0.06360401213169098, -0.05181266739964485, -0.040297962725162506, -0.0473148450255394, 0.03648259863257408, -0.12213828414678574, -0.09815884381532669, -0.07789352536201477, -0.09214427322149277, -0.05354126915335655, 0.0994061604142189, -0.03508447855710983, -0.044435083866119385, -0.09827081114053726, 0.0883425772190094, -0.018374565988779068, -0.08788230270147324, 0.049028001725673676, -0.1212637722492218, 0.0012666302500292659, -0.011799992062151432, -0.03654143959283829, -0.10723117738962173, 0.002689126180484891, 0.10195336490869522, 0.22261740267276764, -0.04011799767613411, 0.1339598447084427, 0.05118986591696739, 0.1427663415670395, 0.0027706108521670103, -0.07153178751468658, -0.06016097590327263, 0.06584857404232025, -0.06470538675785065, 0.07259675860404968, -0.10506517440080643, -0.22706714272499084, 0.021750029176473618, -0.05013787001371384, -0.012725439853966236, 0.009262762032449245, 0.07825521379709244, -0.10154470801353455, -0.054323710501194, 0.06650245934724808, 0.006973872892558575, 0.0026297515723854303, -0.09206568449735641, 0.0690089762210846, 0.0175260528922081, -0.017708929255604744, 0.11137273907661438, -0.0028039736207574606, -0.0032520031090825796, -0.08950033783912659, -0.028769640251994133, -0.013062962330877781, 0.03963356092572212, 0.04880724847316742, 0.042671240866184235, 0.11289776116609573, -0.130026713013649, -0.14796309173107147, -0.06809727102518082, 0.056854408234357834, -0.05829695239663124, -0.19638684391975403, -0.08118753880262375, -0.00460326811298728, 0.03658214956521988, -0.08031634986400604, -0.014560705050826073, -0.002560292836278677, 0.05926350876688957, -0.06861020624637604, 0.1520972102880478, -0.021105466410517693, 0.12280414998531342, -0.12414833158254623, -0.03228786587715149, -0.20177620649337769, 0.0745372623205185, 0.006461530923843384, 0.05103034898638725, -0.06299503147602081, -0.022971751168370247, -0.07667098194360733, 0.02921944484114647, 0.031671371310949326, 0.13541501760482788, -0.1525629162788391, -0.08703860640525818, 0.15057769417762756, -0.02040809765458107, -0.12227700650691986, 0.11326191574335098, -0.002830419223755598, 0.15769705176353455, 0.13181011378765106, 0.2846190631389618, 0.012100251391530037, -0.011337288655340672, 0.03314198553562164, 0.07277388870716095, -0.05627371743321419, -0.01851990446448326, 0.06648373603820801, 0.0599573515355587, 0.0002541402936913073, 0.012887531891465187, 0.13865362107753754, 0.11497339606285095, -0.03651417791843414, -0.029004791751503944, 0.011872794479131699, -0.019793400540947914, -0.039138805121183395, -0.01397590059787035, 0.001350786304101348, 0.05736924707889557, -0.11418261379003525, -0.019525503739714622, 0.06178290396928787, -0.0001964770199265331, -0.02887560799717903, -0.06255626678466797, 0.14360550045967102, -0.06197679042816162, -0.03452926129102707, -0.10036832839250565, 0.038166821002960205, -0.08268986642360687, -0.008501396514475346, 0.11149907112121582, -0.06933024525642395, -0.032505426555871964, -0.0709344670176506, 0.029452087357640266, 0.028955448418855667, 0.04806964471936226, 0.03797874599695206, -0.08985745161771774, -0.0713842585682869, 0.006064391229301691, -0.021750446408987045, 0.05370171368122101, 0.0204731747508049, -0.024943741038441658, -0.032033566385507584, 0.09060537815093994, 0.03549901023507118, -0.0016760098515078425, 0.1431463211774826, 0.07155557721853256, -0.07613774389028549, -0.016155295073986053, 0.03875880315899849, 0.0448884479701519, -0.03330598026514053, 0.17270907759666443, -0.03381839022040367, 0.03116466850042343, 0.1827920377254486, -0.12940220534801483, -0.03078446350991726, 0.05581121891736984, 0.029605774208903313, 0.005461086053401232, -0.0077238441444933414, 0.03458116576075554, 0.1019512265920639, -0.06442201137542725, 0.10047011077404022, -0.03937382251024246, -0.025380821898579597, 0.0064838239923119545, -0.09974098950624466, -0.019145162776112556, 0.06954944133758545, 0.03958898410201073, -0.05517485365271568, 0.18576973676681519, 0.3048633337020874, 0.053187962621450424, 0.20108161866664886, 0.048800691962242126, -0.018934359773993492, -0.06808914244174957, -0.11464201658964157, -0.08645056933164597, -0.0008525591110810637, -0.08422374725341797, -0.04896169528365135, -0.012354748323559761, -0.01722697727382183, 0.06240742653608322, -0.050019729882478714, -0.14171551167964935, 0.0036742372903972864, 0.03467947244644165, -0.06774045526981354, 0.14620524644851685, -0.022101223468780518, 0.03182853013277054, -0.012052635662257671, -0.029285354539752007, 0.09718555957078934, 0.013749165460467339, -0.04868060350418091, 0.0853436291217804, -0.14188046753406525, -0.20270687341690063, 0.021132785826921463, -0.0999145582318306, 0.008460838347673416, -0.009153846651315689, 0.07187989354133606, -0.08365335315465927, 0.054928991943597794, 0.008531304076313972, 0.11340229958295822, -0.163746640086174, -0.06864078342914581, -0.03365499898791313, -0.027721090242266655, -0.1685841828584671, 0.024329133331775665, -0.03986780717968941, 0.0290447399020195, -0.05458942428231239, 0.12529206275939941, -0.18040721118450165, 0.054075803607702255, 0.21213406324386597, 0.10880433022975922, -0.015070742927491665, -0.013258803635835648, 0.2735725939273834, -0.15467415750026703, 0.07714514434337616, 0.1694301962852478, 0.12127173691987991, 0.039140891283750534, 0.06154254451394081, -0.0047433930449187756, -0.07213743031024933, 0.03556244447827339, -0.055245425552129745, -0.07290304452180862, -0.07892238348722458, -0.11038174480199814, -0.09207253158092499, -0.0021310928277671337, -0.06563933938741684, -0.024953436106443405, 0.0460454560816288, 0.11996078491210938, 0.03762179613113403, 0.13485771417617798, -0.03594546392560005, 0.021240917965769768, 0.13026875257492065, -0.0401466004550457, 0.08081541210412979, -0.012583224102854729, -0.1380275934934616, 0.06117650493979454, 0.04697154462337494, 0.05737010017037392, 0.11166106164455414, 0.07875944674015045, 0.04820520803332329, 0.1374562680721283, 0.07004000246524811, 0.044396545737981796, 0.057347092777490616, -0.008985837921500206, -0.06544900685548782, -0.037940289825201035, -0.03690693527460098, 0.06931436061859131, -0.03445622697472572, -0.21835313737392426, 0.005522665102034807, -0.005112190265208483, -0.028216077014803886, -0.015016649849712849, 0.1331651359796524, -0.14811395108699799, 0.014171790331602097, 0.09282585233449936, 0.08396048098802567, -0.06981289386749268, 0.11378534138202667, 0.041733287274837494, -0.05593317374587059, 0.043588731437921524, 0.025383291766047478, 0.06185577064752579, 0.018428971990942955, 0.003582214703783393, -0.02045775018632412, -0.032176654785871506, 0.008571385405957699, 0.09764716774225235, -0.22866812348365784, 0.10061712563037872, 0.018563788384199142, 0.0385720394551754, -0.07380691170692444, -0.021157091483473778, 0.02867508865892887, 0.014340233989059925, 0.24381008744239807, 0.034294359385967255, -0.21706052124500275, -0.060008108615875244, -0.1159144714474678, 0.039287663996219635, 0.010961631312966347, -0.09934451431035995, -0.017151210457086563, 0.030424395576119423, 0.041372936218976974, -0.07778937369585037, 0.027137229219079018, -0.1118098646402359, -0.1812155842781067, 0.05158354714512825, 0.07981729507446289, 0.06262098997831345, 0.040827471762895584, -0.02188548631966114, 0.012243879027664661, 0.036974091082811356, 0.1535831093788147, -0.10783729702234268, -0.08956166356801987, -0.0785638615489006, 0.0511448048055172, -0.04190564528107643, 0.001271011889912188, 0.028414523229002953, -0.0033308109268546104, -0.02639051526784897, -0.1564140021800995, 0.10318015515804291, -0.06269120424985886, -0.020872347056865692, -0.07170116156339645, 0.12374047189950943, 0.0699063315987587, 0.03219025209546089, 0.1403231918811798, 0.02356075495481491, -0.10031699389219284, -0.10647591948509216, 0.064108707010746, -0.05641859024763107, 0.022806130349636078, -0.03379463404417038, -0.04116413742303848, 0.05234726518392563, 0.019905935972929, -0.0803285762667656, 0.19859643280506134, 0.11186059564352036, -0.051211707293987274, 0.10964181274175644, 0.1417967528104782, -0.08583147823810577, -0.28357580304145813, -0.17311698198318481, -0.050763897597789764, 0.09075833857059479, 0.14932069182395935, -0.30291810631752014, 0.05752218887209892, 0.0299115851521492, -0.006390051916241646, 0.057322852313518524, -0.36947113275527954, -0.04304134473204613, 0.1602199375629425, -0.06838271766901016, 0.23195451498031616, -0.0850193127989769, -0.045508258044719696, -0.058680154383182526, -0.09838713705539703, 0.2109292596578598, -0.10860046744346619, 0.14995214343070984, 0.035534199327230453, 0.1933312863111496, 0.05826879292726517, 0.032994650304317474, 0.026138179004192352, 0.2001863569021225, 0.030717456713318825, -0.05675380676984787, -0.10458554327487946, 0.03451451286673546, 0.0019765556789934635, 0.013282024301588535, -0.08316376805305481, 0.016799917444586754, -0.20289663970470428, -0.0934498980641365, -0.09430515766143799, -0.018213288858532906, -0.0405634380877018, -0.06968969851732254, -0.08097928017377853, 0.1240328848361969, -0.01816055178642273, -0.0381661131978035, 0.0473969504237175, -0.08112846314907074, 0.05004718154668808, -0.08487387746572495, 0.037242379039525986, -0.027624642476439476, -0.04182023927569389, 0.03334595263004303, -0.08666280657052994, 0.06766095757484436, -0.19867561757564545, 0.01926562376320362, 0.06945707648992538, 0.055795010179281235, -0.01350108627229929, 0.05221671983599663, -0.11626515537500381, 0.02467115968465805, 0.1653663069009781, -0.1569318175315857, -0.004748887848109007, -0.11246564984321594, 0.05847112089395523, 0.17546410858631134, -0.03973150998353958, 0.08214317262172699, 0.017614787444472313, -0.1313716471195221, 0.029829613864421844, 0.039834242314100266, -0.046089883893728256, 0.004072097595781088, 0.03099946863949299, -0.03430705517530441, -0.09068986028432846, 0.0020939193200320005, -0.007445253431797028, 0.008059820160269737, -0.005049992818385363, 0.010306054726243019, -0.1099032312631607, -0.08413327485322952, -0.016130151227116585, -0.01261814869940281, -0.04850275442004204, 0.07170376926660538, 0.11888451129198074, -0.09782660752534866, -0.07482252269983292, 0.06168828159570694, 0.07838458567857742, 0.07095757126808167, 0.026185113936662674, -0.07254026085138321, -0.05282950773835182, 0.034572578966617584, 0.041695933789014816, -0.013805593363940716, -0.10905708372592926, 0.055318061262369156, -0.00045633563422597945, 0.16802653670310974, -0.11349963396787643, -0.027882631868124008, -0.10721779614686966, -0.0004932220908813179, 0.008455289527773857, -0.09712251275777817, -0.17714731395244598, -0.05969017744064331, 0.0034399915020912886, 0.06920330971479416, -0.08607493340969086, -0.03874945268034935, -0.11141936480998993, -0.030305543914437294, -0.056665919721126556, 0.02255689539015293, -0.01222404558211565, 0.0023805934470146894, 0.033254239708185196, -0.022950369864702225, 0.06850875169038773, 0.07670687884092331, 0.03654158487915993, -0.0031966818496584892, -0.15194515883922577, -0.023407019674777985, -0.01446541678160429, 0.0670863538980484, 0.06606891006231308, -0.06734522432088852, -0.02287386916577816, 0.0069444552063941956, 0.059028156101703644, 0.015741299837827682, -0.02944275364279747, -0.07587143778800964, -0.05627180635929108, 0.004066201392561197, -0.06742078810930252, -0.03825848177075386, 0.027105562388896942, 0.10975765436887741, 0.022619210183620453, 0.09258292615413666, -0.02197602018713951, 0.06697934120893478, -0.06358520686626434, 0.05436529591679573, -0.06376567482948303, -0.07946758717298508, -0.018000150099396706, -0.0722920149564743, 0.011115930043160915, -0.002443079138174653, 0.19305071234703064, 0.08486415445804596, -0.12169951945543289, -0.03466005623340607, 0.05839117243885994, 0.10926923900842667, -0.03106572851538658, 0.038779985159635544, 0.04770343378186226, -0.04222758114337921, -0.05892977863550186, 0.09221015870571136, -0.04399019479751587, 0.09542325139045715, -0.028887351974844933, -0.0935973972082138, 0.09593192487955093, 0.08199973404407501, 0.0074447025544941425, -0.01710151508450508, -0.09036751091480255, -0.04395315796136856, 0.000649825728032738, 0.06333403289318085, -0.004521216731518507, 0.09598685055971146, 0.031018473207950592, 0.026564590632915497, 0.0714581310749054, -0.03291894495487213, -0.07081419974565506, -0.15311628580093384, -0.2916485071182251, -0.04245568811893463, -0.11483888328075409, -0.030321842059493065, -0.09362631291151047, 0.008166450075805187, 0.14719577133655548, -0.01637839525938034, -0.017485830932855606, 0.04677969217300415, -0.023783322423696518, -0.13184989988803864, 0.10720223933458328, -0.06770380586385727, 0.0017660012235864997, -0.0461556501686573, 0.06525444984436035, 0.018209991976618767, 0.02946087159216404, 0.04921369627118111, 0.021475588902831078, 0.00033336132764816284, 0.09776710718870163, -0.1710214763879776, -0.05395933613181114, -0.11202818900346756, 0.04753042757511139, 0.0662398487329483, -0.007952507585287094, 0.06053532660007477, -0.020561669021844864, 0.01672571338713169, 0.15353375673294067, 0.02742934226989746, 0.08916531503200531, -0.08731471747159958, 0.15215137600898743, -0.08004730194807053, 0.029828161001205444, -0.015857866033911705, 0.061717115342617035, 0.030440203845500946, 0.27778393030166626, 0.1749309003353119, -0.12385223060846329, -0.05927656590938568, -0.02690744213759899, -0.013377930968999863, -0.01374972052872181, 0.0042964378371834755, 0.03448919579386711, 0.22202332317829132, -0.07485826313495636, -0.01686796359717846, -0.1470489352941513, -0.07605598121881485, -0.02722550556063652, 0.008164134807884693, 0.08011732995510101, -0.07406966388225555, 0.035195473581552505, 0.1917814463376999, -0.2807765305042267, 0.1288827508687973, -0.09982363879680634, -0.06866038590669632, -0.11830036342144012, 0.012730836868286133, -0.06672485172748566, 0.09346511214971542, 0.14677636325359344, -0.05562186613678932, -0.009732590988278389, 0.05830236151814461, 0.040688756853342056, -0.1255311220884323, 0.07655194401741028, 0.03098270110785961, 0.022324470803141594, 0.08402157574892044, -0.004190674517303705, 0.1166965439915657, 0.05947024002671242, 0.06936319172382355, -0.04727460443973541, 0.043860744684934616, -0.03098616749048233, 0.06152599677443504, 0.06267384439706802, 0.07985928654670715, 0.0019050639821216464, -0.08628340810537338, 0.03895357996225357, 0.05389104411005974, 0.06490454077720642, -0.026681525632739067, 0.011457736603915691, -0.0020894508343189955, -0.009379146620631218, -0.09290852397680283, 0.09659010916948318, 0.0407472662627697, -0.05785062164068222, -0.08620359748601913, 0.06279981881380081, -0.0018226153915748, -0.006378989201039076, 0.05120845139026642, -0.011586344800889492, -0.0013158415677025914, -0.05478912591934204, 0.013565197587013245, -0.023614974692463875, -0.11941657215356827, 0.08607260137796402, -0.09453480690717697, -0.009826654568314552, -0.07651437073945999, 0.08537538349628448, 0.055164989084005356, 0.03445355221629143, 0.002721157157793641, 0.12385866791009903, -0.006145811174064875, -0.008021186105906963, -0.10184914618730545, -0.11376316100358963 ]
null
null
transformers
--- Summarisation model summarsiation
{}
text2text-generation
gagan3012/summarsiation
[ "transformers", "pytorch", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "has_space", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us
--- Summarisation model summarsiation
[]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n" ]
[ 52 ]
[ "passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n" ]
[ 0.0019048319663852453, -0.0037844888865947723, -0.004846867639571428, 0.016280045732855797, 0.15925085544586182, 0.01139024831354618, 0.09671302884817123, 0.14897321164608002, -0.04899326339364052, -0.0070660836063325405, 0.1372009962797165, 0.1665002405643463, -0.01529724895954132, 0.10409914702177048, -0.08531556278467178, -0.2723868191242218, 0.041050851345062256, 0.05884578078985214, 0.006763700861483812, 0.12957002222537994, 0.08201905339956284, -0.07262634485960007, 0.09340167045593262, -0.03144669905304909, -0.1715334951877594, 0.06047246605157852, 0.04992982745170593, -0.1397479921579361, 0.11693664640188217, 0.04595603793859482, 0.11842931061983109, 0.037635594606399536, -0.05701420456171036, -0.10396002233028412, 0.0293845534324646, 0.04007481783628464, -0.07444078475236893, 0.06619872897863388, 0.10958369076251984, -0.08855950087308884, 0.11609432846307755, 0.014332751743495464, -0.01690986007452011, 0.0453108474612236, -0.14978627860546112, -0.0115242013707757, -0.017286943271756172, 0.009983301162719727, 0.03019011951982975, 0.10428427159786224, -0.025953028351068497, 0.13894023001194, -0.11443798243999481, 0.1249191015958786, 0.1755915880203247, -0.3218105733394623, 0.0008721072808839381, 0.07177659124135971, 0.0954747423529625, 0.08060836046934128, -0.017075497657060623, 0.049279846251010895, 0.03689901903271675, 0.03278602287173271, 0.05441928654909134, -0.05992512032389641, -0.1813686043024063, 0.0644410029053688, -0.10202427208423615, -0.05823950096964836, 0.24249997735023499, -0.06648484617471695, 0.08229529857635498, -0.0315452516078949, -0.13439735770225525, -0.09164708107709885, 0.0146126514300704, 0.011662522330880165, -0.05338037386536598, 0.059678707271814346, 0.014187299646437168, -0.06957265734672546, -0.1573067158460617, 0.007388781290501356, -0.20604579150676727, 0.12037379294633865, -0.0022145791444927454, 0.04438640549778938, -0.21971583366394043, 0.10112471133470535, 0.01954837702214718, -0.10843217372894287, 0.07814245671033859, -0.08621292561292648, 0.01956222951412201, -0.009294086135923862, -0.09783979505300522, -0.1523585468530655, 0.05506528168916702, 0.06974567472934723, -0.016469081863760948, -0.0013621869729831815, -0.05712024122476578, 0.08282897621393204, 0.026742184534668922, 0.09264246374368668, -0.039046403020620346, -0.03138374537229538, 0.026040101423859596, -0.10890994966030121, 0.004530361853539944, -0.08059261739253998, -0.16648244857788086, -0.08678217977285385, 0.08933352679014206, 0.06936588138341904, 0.0427776537835598, 0.10737014561891556, -0.01597854308784008, -0.017108382657170296, 0.02555423602461815, -0.08640097081661224, 0.002975356997922063, -0.0034396217670291662, 0.010684369131922722, 0.11860962957143784, 0.037888817489147186, 0.004787639249116182, -0.1508452594280243, 0.05129704251885414, -0.08456986397504807, -0.004201768897473812, -0.0386526882648468, -0.11637408286333084, 0.02888437733054161, -0.11346439272165298, 0.0019459343748167157, -0.1819058209657669, -0.10794739425182343, 0.013547050766646862, -0.003563873004168272, -0.02874688245356083, -0.05685725063085556, -0.015953734517097473, -0.060097914189100266, 0.07545679062604904, -0.07323034107685089, 0.033851344138383865, -0.04320577159523964, 0.10142723470926285, -0.05439102277159691, 0.08872391283512115, -0.15116755664348602, 0.08348871767520905, -0.11385006457567215, -0.026207635179162025, -0.06486699730157852, 0.05204048007726669, 0.037419021129608154, 0.10208529978990555, -0.023686395958065987, -0.04558229446411133, -0.09794006496667862, 0.05766747146844864, -0.010321461595594883, 0.17991861701011658, -0.1119103655219078, -0.08101113885641098, 0.2097437083721161, -0.0516032800078392, -0.14293880760669708, 0.08471720665693283, 0.014890666119754314, 0.04416865482926369, 0.05551055818796158, 0.21729257702827454, 0.03978389874100685, -0.025182973593473434, 0.06714358180761337, 0.11787192523479462, -0.10201745480298996, -0.07721184939146042, 0.013209090568125248, -0.014080026187002659, -0.07248388230800629, 0.03355661779642105, 0.10408741980791092, 0.07307033240795135, -0.04470960423350334, -0.03571781888604164, -0.06436226516962051, -0.008570763282477856, 0.11809982359409332, -0.001294085755944252, 0.13646534085273743, -0.07103265821933746, -0.04472484812140465, 0.027163831517100334, -0.02947130613029003, -0.02112656459212303, 0.059094879776239395, -0.005023777950555086, 0.13337215781211853, -0.03037242405116558, 0.03793719783425331, -0.20518949627876282, -0.08568139374256134, -0.027907684445381165, 0.1735931932926178, 0.008581404574215412, 0.14691676199436188, 0.05126412957906723, -0.038134533911943436, -0.01040117908269167, -0.006090906914323568, 0.12786434590816498, 0.009149910882115364, -0.08664630353450775, -0.04733043536543846, 0.0470752976834774, -0.06807634979486465, -0.042599860578775406, -0.06205863505601883, 0.030012013390660286, 0.03637655824422836, 0.12896181643009186, 0.013993069529533386, 0.06803536415100098, -0.008032101206481457, 0.02921360358595848, -0.10367264598608017, 0.013655728660523891, 0.08631832152605057, -0.013092545792460442, -0.05325587838888168, 0.23344485461711884, -0.22350485622882843, 0.23208631575107574, 0.21852712333202362, -0.2824490964412689, -0.0005818685167469084, -0.036813754588365555, -0.03419727459549904, 0.020097149536013603, 0.03740953654050827, -0.051420778036117554, 0.03739655017852783, -0.024677753448486328, 0.1943303644657135, -0.06214848905801773, -0.04623418301343918, -0.003677424043416977, -0.051297854632139206, -0.028216710314154625, 0.05521993711590767, 0.05055946111679077, -0.14452490210533142, 0.17969056963920593, 0.26677438616752625, 0.015687527135014534, 0.19574841856956482, 0.005315606482326984, -0.04260627552866936, 0.07293140143156052, -0.017192890867590904, -0.05518830940127373, -0.08852692693471909, -0.1755826324224472, -0.030800864100456238, 0.08786051720380783, 0.043464839458465576, 0.10033265501260757, -0.11051545292139053, -0.0307454951107502, 0.011240042746067047, 0.006229858845472336, -0.0198612529784441, 0.1188352108001709, 0.0916038304567337, 0.14202114939689636, -0.0043418025597929955, -0.005152907222509384, 0.10298898071050644, 0.026397770270705223, -0.10556304454803467, 0.17032410204410553, -0.14493925869464874, -0.3398903012275696, -0.14305217564105988, -0.1283554881811142, -0.029550181701779366, 0.0481751449406147, 0.126591756939888, -0.100038543343544, -0.015225457958877087, -0.041716817766427994, 0.07587543874979019, -0.08503478020429611, 0.048016101121902466, -0.1092776209115982, 0.05474958196282387, -0.05765443667769432, -0.0870586484670639, -0.04337740316987038, -0.0012005938915535808, -0.042839743196964264, 0.1490975320339203, -0.09574799239635468, 0.06705930083990097, 0.20642054080963135, -0.021416770294308662, 0.04550161957740784, -0.0442466177046299, 0.19283124804496765, -0.07007778435945511, 0.028251156210899353, 0.2250925600528717, -0.032153669744729996, 0.07288595288991928, 0.15226730704307556, -0.022663576528429985, -0.03787560015916824, 0.04395657777786255, -0.026733849197626114, -0.09023230522871017, -0.2322804033756256, -0.11889786273241043, -0.14082705974578857, 0.07175826281309128, 0.05754096433520317, 0.05656171590089798, 0.14769874513149261, 0.056594591587781906, -0.0090337498113513, 0.03087019920349121, -0.009781209751963615, 0.07582470029592514, 0.2039855271577835, -0.029816200956702232, 0.1504884958267212, -0.0629919245839119, -0.11681561917066574, 0.09275535494089127, 0.06313134729862213, 0.10160941630601883, 0.02251110039651394, 0.026960816234350204, 0.010898066684603691, 0.09511017054319382, 0.13752858340740204, 0.1379387527704239, 0.03932773321866989, -0.006873026490211487, -0.02523285523056984, -0.03255249559879303, 0.00002126315848727245, 0.054425422102212906, 0.045209161937236786, -0.1523430347442627, -0.09001755714416504, -0.10640548169612885, 0.08357469737529755, 0.08635000139474869, 0.10505597293376923, -0.22463750839233398, 0.023784343153238297, 0.07206699997186661, -0.03994642570614815, -0.11989019811153412, 0.07098391652107239, 0.05299854651093483, -0.1024707555770874, 0.046546246856451035, -0.031229954212903976, 0.11318814754486084, 0.022051602602005005, 0.10550439357757568, -0.053203485906124115, -0.09618035703897476, 0.009638491086661816, 0.10481701791286469, -0.30516061186790466, 0.20729410648345947, -0.007429636083543301, -0.09295950829982758, -0.11187880486249924, -0.018728306517004967, 0.0025576867628842592, 0.12164662778377533, 0.07349345833063126, 0.001384903327561915, -0.06238449364900589, -0.07514331489801407, 0.018887696787714958, 0.005249361507594585, 0.13861951231956482, -0.00704901572316885, 0.013407570309937, -0.061070434749126434, -0.0075938948430120945, 0.011643931269645691, 0.0650859847664833, 0.0014231993118301034, -0.17192824184894562, 0.07567143440246582, 0.05458882823586464, 0.04693356528878212, 0.01592378132045269, -0.033653389662504196, -0.10869722068309784, 0.19578181207180023, -0.017877740785479546, -0.08899964392185211, -0.1214602142572403, -0.048978541046381, 0.06916724890470505, -0.0666830986738205, 0.05096784606575966, -0.0720483809709549, 0.02991739846765995, -0.06392006576061249, -0.2284102737903595, 0.12943212687969208, -0.07299447804689407, -0.04307695850729942, -0.0322282649576664, 0.15847639739513397, -0.12298990786075592, 0.01999419741332531, 0.01866958849132061, 0.02402704581618309, -0.12511111795902252, -0.06121746823191643, -0.011848983354866505, -0.028766747564077377, 0.07283125072717667, 0.03457988426089287, -0.08438634872436523, -0.06576003134250641, -0.0019198559457436204, -0.0033527102787047625, 0.3267287611961365, 0.10421230643987656, -0.08351656049489975, 0.1654823273420334, 0.06374907493591309, -0.07078107446432114, -0.32191118597984314, -0.07873771339654922, -0.10027248412370682, -0.005710241850465536, 0.01002343650907278, -0.1268543303012848, 0.04427546262741089, -0.029947733506560326, -0.0038497543428093195, 0.08380930125713348, -0.24521243572235107, -0.09818269312381744, 0.13724389672279358, -0.020412998273968697, 0.3158648908138275, -0.12522555887699127, -0.07609208673238754, -0.03624487668275833, -0.109134241938591, 0.17277181148529053, -0.1096419245004654, 0.09867527335882187, -0.026042291894555092, 0.12853428721427917, 0.06134048104286194, -0.04277019575238228, 0.06854512542486191, -0.01833241991698742, 0.008027022704482079, -0.13450272381305695, -0.06482996791601181, 0.07512287050485611, -0.030992954969406128, 0.047261811792850494, -0.07413084805011749, 0.042952511459589005, -0.1277935802936554, -0.01882348023355007, -0.11093898862600327, 0.061318669468164444, 0.022534433752298355, -0.0656307116150856, -0.017191549763083458, -0.06914012879133224, 0.025176333263516426, -0.01658732071518898, 0.19708415865898132, -0.06335504353046417, 0.1823943704366684, 0.1869804412126541, 0.09760983288288116, -0.10386087000370026, 0.04452233761548996, -0.032749176025390625, -0.06730343401432037, 0.07338140159845352, -0.14253178238868713, 0.052002448588609695, 0.10644328594207764, -0.050541721284389496, 0.05814075469970703, 0.1106620654463768, 0.014361315406858921, -0.013961630873382092, 0.15362074971199036, -0.2562159597873688, 0.005441658664494753, -0.0962069109082222, -0.04770717769861221, 0.03046395815908909, 0.03032153844833374, 0.18230471014976501, 0.01120354700833559, -0.026954354718327522, -0.005538483150303364, -0.0018120072782039642, -0.053923096507787704, 0.05512459948658943, 0.036821864545345306, 0.03323718532919884, -0.11189843714237213, 0.06060293689370155, 0.05815809965133667, -0.1448555886745453, 0.029695017263293266, 0.19576966762542725, -0.11733770370483398, -0.13482601940631866, -0.0004756708804052323, 0.0677732452750206, -0.14123328030109406, -0.009141461923718452, -0.05016467347741127, -0.11345820873975754, 0.08574269711971283, 0.19972220063209534, 0.05710255727171898, 0.08909596502780914, -0.033956222236156464, -0.058453518897295, -0.029195280745625496, 0.010194899514317513, 0.011630593799054623, 0.028619157150387764, -0.10208519548177719, 0.13118597865104675, -0.04724050313234329, 0.16715489327907562, -0.09716197103261948, -0.04702884331345558, -0.152531698346138, 0.002426156075671315, -0.1466960608959198, -0.06943263113498688, -0.05659385770559311, -0.06354662030935287, -0.009821423329412937, -0.02434639073908329, -0.05047421529889107, -0.05024503916501999, -0.12260215729475021, 0.011571573093533516, -0.05621323361992836, 0.04525769129395485, -0.07972045242786407, -0.013484488241374493, 0.055132120847702026, -0.030423389747738838, 0.11408451944589615, 0.08748376369476318, -0.11662473529577255, 0.10743166506290436, -0.09839223325252533, -0.12656080722808838, 0.09486260265111923, 0.018033714964985847, 0.058434270322322845, 0.09707959741353989, 0.00891159288585186, 0.05598178133368492, 0.04158338904380798, 0.04411861300468445, 0.009987978264689445, -0.11124832183122635, 0.0380651019513607, -0.052996713668107986, -0.13563407957553864, -0.06263524293899536, -0.026792006567120552, 0.030286112800240517, 0.01089276373386383, 0.10767155885696411, -0.051219791173934937, 0.10564509779214859, -0.07189621031284332, 0.019965166226029396, -0.01300421915948391, -0.1647084802389145, -0.045782022178173065, -0.07341710478067398, 0.03664537891745567, -0.009205256588757038, 0.20625713467597961, 0.077974833548069, 0.02349081262946129, 0.037151604890823364, 0.08264945447444916, 0.0019355815602466464, 0.016970688477158546, 0.21777592599391937, 0.06962968409061432, -0.07049708813428879, -0.11852742731571198, 0.05963381752371788, 0.01887577399611473, 0.06965018808841705, 0.15576517581939697, 0.056794699281454086, -0.011962602846324444, 0.11193615943193436, -0.024064647033810616, -0.0068133813329041, -0.12810677289962769, -0.16301329433918, -0.004318809602409601, 0.09310556203126907, -0.052356306463479996, 0.04415218159556389, 0.17003647983074188, -0.018104281276464462, 0.0336272232234478, -0.03094765916466713, -0.05335594341158867, -0.16834230720996857, -0.14789271354675293, -0.07995127886533737, -0.11087332665920258, -0.02718210592865944, -0.10139597207307816, 0.08172602951526642, 0.06748189777135849, 0.05592411756515503, -0.0499156229197979, 0.11055595427751541, 0.05528182536363602, -0.11725020408630371, 0.07114271819591522, -0.027335863560438156, 0.09812980890274048, -0.009088918566703796, -0.006258330307900906, -0.08307218551635742, 0.010819291695952415, -0.03894173353910446, 0.049928583204746246, -0.04747631028294563, 0.008504346944391727, -0.16299764811992645, -0.10957249253988266, -0.04291011765599251, 0.0614963173866272, -0.019686607643961906, 0.1551779955625534, 0.01387507189065218, -0.03461948782205582, 0.013804392889142036, 0.2583024799823761, -0.09972049295902252, -0.0851818099617958, -0.041016120463609695, 0.2084883451461792, 0.061222467571496964, 0.07373743504285812, -0.02859536185860634, -0.031048838049173355, -0.11416899412870407, 0.34497207403182983, 0.31482934951782227, -0.07161965221166611, 0.0398726724088192, 0.016563279554247856, 0.030142735689878464, 0.11637284606695175, 0.14022967219352722, 0.0916035920381546, 0.23573338985443115, -0.0693872794508934, -0.010092688724398613, -0.022344037890434265, 0.0073167746886610985, -0.10196781903505325, 0.13043533265590668, 0.043621718883514404, -0.09041506797075272, -0.026318738237023354, 0.07670903205871582, -0.21243104338645935, 0.1389734297990799, -0.08875156939029694, -0.19405682384967804, -0.05883593484759331, 0.01516592688858509, 0.1496192216873169, -0.016031023114919662, 0.09138526022434235, -0.020647617056965828, -0.0722113624215126, 0.029204411432147026, 0.01666286215186119, -0.19871759414672852, 0.033451639115810394, 0.06371412426233292, -0.11542545258998871, -0.029661864042282104, -0.017291566357016563, 0.02199423685669899, 0.08295346796512604, 0.07082275301218033, -0.04837881773710251, 0.0535416342318058, 0.005592434201389551, -0.019829990342259407, 0.03503140062093735, 0.05262354016304016, 0.009726877324283123, -0.10966355353593826, 0.06299593299627304, -0.17635416984558105, 0.03804607316851616, -0.036695461720228195, -0.016934243962168694, -0.00767596485093236, -0.0276106558740139, -0.03190142661333084, 0.07071686536073685, 0.10571801662445068, -0.015258674509823322, -0.003450299147516489, -0.0712389275431633, -0.058415062725543976, -0.014903778210282326, -0.11116986721754074, -0.09979519993066788, -0.09476316720247269, -0.09682852029800415, 0.10881722718477249, -0.0060325018130242825, -0.2213381677865982, 0.01311418879777193, -0.08191841095685959, 0.050175897777080536, -0.1892065852880478, 0.10583088546991348, 0.10422021895647049, -0.007532334886491299, 0.0014266044599935412, -0.007911108434200287, 0.04282023757696152, 0.09190382063388824, -0.1298302263021469, -0.07897894084453583 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-large-xls-r-300m-hindi This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common_voice dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2.dev0 - Tokenizers 0.11.0
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["common_voice"], "model-index": [{"name": "wav2vec2-large-xls-r-300m-hindi", "results": []}]}
automatic-speech-recognition
gagan3012/wav2vec2-large-xls-r-300m-hindi
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "dataset:common_voice", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us
# wav2vec2-large-xls-r-300m-hindi This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the common_voice dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2.dev0 - Tokenizers 0.11.0
[ "# wav2vec2-large-xls-r-300m-hindi\n\nThis model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the common_voice dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0003\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 2\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 500\n- num_epochs: 30\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.17.0.dev0\n- Pytorch 1.10.2+cu102\n- Datasets 1.18.2.dev0\n- Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us \n", "# wav2vec2-large-xls-r-300m-hindi\n\nThis model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the common_voice dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0003\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 2\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 500\n- num_epochs: 30\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.17.0.dev0\n- Pytorch 1.10.2+cu102\n- Datasets 1.18.2.dev0\n- Tokenizers 0.11.0" ]
[ 61, 49, 6, 12, 8, 3, 140, 4, 39 ]
[ "passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us \n# wav2vec2-large-xls-r-300m-hindi\n\nThis model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the common_voice dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0003\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 2\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 500\n- num_epochs: 30\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.17.0.dev0\n- Pytorch 1.10.2+cu102\n- Datasets 1.18.2.dev0\n- Tokenizers 0.11.0" ]
[ -0.12029562890529633, 0.14340563118457794, -0.0013717523543164134, 0.05236584320664406, 0.12963251769542694, 0.0014695152640342712, 0.07955561578273773, 0.11646801233291626, -0.08166693896055222, 0.0671023577451706, 0.06461470574140549, -0.005223055835813284, 0.08425532281398773, 0.11689576506614685, -0.02499585598707199, -0.21804368495941162, 0.0004132949688937515, 0.01105158869177103, -0.05098409205675125, 0.09551060199737549, 0.11991573125123978, -0.0760030448436737, 0.04220406338572502, 0.043772101402282715, -0.15551458299160004, 0.03254007175564766, -0.04015062376856804, -0.05773875489830971, 0.09554438292980194, 0.05399437993764877, 0.07215193659067154, 0.012390648946166039, 0.11896750330924988, -0.22441205382347107, 0.005439940840005875, 0.0652477964758873, 0.02511867880821228, 0.07519792020320892, 0.06725624203681946, 0.014491723850369453, 0.08959265053272247, -0.1368243247270584, 0.09707584977149963, 0.043342042714357376, -0.05648418143391609, -0.20828503370285034, -0.09139695018529892, 0.08816128224134445, 0.1218658909201622, 0.12557479739189148, -0.014224921353161335, 0.12106840312480927, -0.03155740350484848, 0.07793785631656647, 0.16294214129447937, -0.24595749378204346, -0.060669731348752975, 0.010788273066282272, 0.06258440762758255, 0.045768558979034424, -0.10220039635896683, -0.003957860637456179, 0.05672381818294525, 0.03425956889986992, 0.08173301815986633, -0.0023710790555924177, -0.03722812980413437, -0.02621864713728428, -0.11792309582233429, -0.022084685042500496, 0.14853662252426147, 0.08927443623542786, -0.058184295892715454, -0.1240408793091774, -0.01185617409646511, -0.10995320230722427, -0.017339423298835754, -0.039420198649168015, 0.008220921270549297, -0.024373142048716545, -0.0630829930305481, -0.054454267024993896, -0.08697158098220825, -0.055029407143592834, 0.04162152484059334, 0.096168152987957, 0.011009521782398224, -0.017463872209191322, -0.018675686791539192, 0.07317086309194565, 0.005503397900611162, -0.1246638298034668, -0.003815228817984462, 0.020936625078320503, -0.11189785599708557, -0.05806540697813034, -0.04393969103693962, -0.08691567927598953, -0.019306398928165436, 0.11031603068113327, -0.015003124251961708, 0.08322495967149734, -0.024458572268486023, -0.006503286771476269, -0.006033965386450291, 0.15236899256706238, -0.03962118923664093, -0.07852853834629059, 0.001148562179878354, 0.10847736895084381, 0.00582697382196784, -0.019598981365561485, -0.08306773006916046, -0.03404892981052399, 0.0914231389760971, 0.048122212290763855, -0.032781731337308884, -0.012165925465524197, -0.05840649828314781, -0.03199870139360428, 0.0688665583729744, -0.11686860769987106, 0.03953268751502037, -0.012897180393338203, -0.04483112320303917, -0.03932616487145424, 0.016770394518971443, 0.04042348265647888, -0.012306654825806618, 0.04911867156624794, -0.054440926760435104, -0.013102283701300621, -0.05094422772526741, -0.05649203434586525, 0.018041634932160378, -0.006360988132655621, 0.002487432211637497, -0.08414831757545471, -0.15513838827610016, -0.060350582003593445, 0.03533690422773361, -0.0654502734541893, -0.07109235972166061, -0.01972026750445366, -0.015959948301315308, 0.027877509593963623, -0.01955050230026245, 0.15292058885097504, -0.03812038525938988, 0.06376882642507553, -0.00019983341917395592, 0.001719775958918035, 0.02778376080095768, 0.05854293331503868, -0.047680437564849854, 0.030699348077178, -0.07977166771888733, 0.09483002871274948, -0.0832797959446907, 0.018728232011198997, -0.13901670277118683, -0.09698319435119629, -0.0064824409782886505, -0.029703667387366295, 0.0847688764333725, 0.08574382215738297, -0.18515713512897491, -0.04227156937122345, 0.11149166524410248, -0.07452656328678131, -0.09311046451330185, 0.13933533430099487, -0.026894429698586464, 0.02285565622150898, 0.06336135417222977, 0.1332082897424698, 0.13943935930728912, -0.12083282321691513, -0.014698010869324207, -0.012931331992149353, 0.09760241955518723, 0.025986118242144585, 0.08127615600824356, -0.02928488329052925, 0.016653699800372124, -0.015421503223478794, -0.02265186235308647, 0.03132956847548485, -0.0632718876004219, -0.0719204694032669, -0.029227910563349724, -0.10375077277421951, 0.027786564081907272, 0.022379428148269653, 0.011838559992611408, -0.07050299644470215, -0.12018012255430222, 0.07267943769693375, 0.14128443598747253, -0.059287622570991516, 0.014478242956101894, -0.07964379340410233, 0.008346409536898136, -0.05213107541203499, -0.037418439984321594, -0.1913917511701584, -0.03641033545136452, 0.04357314482331276, -0.08250237256288528, 0.03182912990450859, 0.008058012463152409, 0.055260464549064636, 0.04953588545322418, -0.055972710251808167, -0.02735350839793682, -0.11022386699914932, 0.02278929017484188, -0.09274240583181381, -0.15529248118400574, -0.08055213838815689, -0.04304414987564087, 0.2204894721508026, -0.23071734607219696, -0.004674679134041071, 0.028120828792452812, 0.1529318392276764, 0.011092579923570156, -0.06443049013614655, 0.009111136198043823, 0.049194253981113434, 0.0126855643466115, -0.09582701325416565, 0.014071096666157246, 0.031246652826666832, -0.1308191567659378, -0.04784900322556496, -0.10934759676456451, 0.05386357754468918, 0.06141326576471329, 0.08996596932411194, -0.08904393762350082, -0.05530611798167229, -0.046221453696489334, -0.0491010919213295, -0.07377500087022781, -0.02327495627105236, 0.23775020241737366, 0.030864190310239792, 0.11746469885110855, -0.048441097140312195, -0.05533549189567566, 0.01600644551217556, 0.030518123880028725, -0.03846341744065285, 0.06030499190092087, 0.00959834922105074, -0.19020473957061768, 0.06413920223712921, 0.07346335053443909, -0.027602510526776314, 0.13967779278755188, -0.042952075600624084, -0.08710484206676483, -0.036954861134290695, -0.0006090368260629475, -0.010678417049348354, 0.09367919713258743, -0.10552023351192474, 0.005419078748673201, 0.03146449103951454, -0.0005446167779155076, 0.024480201303958893, -0.1414068192243576, 0.011891099624335766, 0.05999743565917015, -0.026130646467208862, -0.02071184292435646, -0.02504797652363777, -0.0008758765761740506, 0.05427124723792076, 0.029803676530718803, -0.014131061732769012, 0.020284589380025864, -0.0173393152654171, -0.0961311012506485, 0.14857560396194458, -0.10928797721862793, -0.19614310562610626, -0.13015373051166534, 0.03969684615731239, -0.03495481610298157, -0.02989240549504757, 0.024526821449398994, -0.10522343963384628, -0.06174110993742943, -0.07563501596450806, -0.020418696105480194, -0.05722222477197647, -0.0019934168085455894, 0.09618724137544632, -0.006513394881039858, 0.08676795661449432, -0.11892569810152054, 0.017389968037605286, 0.014026680961251259, -0.022167742252349854, -0.03827651962637901, 0.04211582988500595, 0.08217264711856842, 0.10729296505451202, 0.0013586417771875858, 0.017927955836057663, -0.036899372935295105, 0.22993382811546326, -0.10165401548147202, 0.005166220478713512, 0.14158476889133453, -0.004244903568178415, 0.03389592468738556, 0.09608393162488937, 0.021717295050621033, -0.088821180164814, 0.04014386609196663, 0.03515291586518288, -0.012990587390959263, -0.23697732388973236, -0.046118345111608505, -0.04341597482562065, -0.10725700110197067, 0.11230883747339249, 0.03834244981408119, 0.011559151113033295, 0.0345299132168293, -0.02038196474313736, 0.04924146085977554, 0.006010173819959164, 0.07086668908596039, 0.08183225989341736, 0.04572172090411186, 0.09568127244710922, -0.02425406500697136, 0.004847608972340822, 0.059325192123651505, -0.005083018448203802, 0.20444315671920776, 0.007559481542557478, 0.14551587402820587, 0.014386474154889584, 0.13750514388084412, -0.00255402410402894, 0.024493252858519554, 0.013094133697450161, -0.009111711755394936, 0.018909461796283722, -0.0507630854845047, -0.049271270632743835, 0.04019974172115326, 0.07153251022100449, 0.022214408963918686, -0.09062784165143967, 0.0007296240073628724, -0.022450651973485947, 0.29514026641845703, 0.04914794862270355, -0.2742305397987366, -0.10156404972076416, 0.0022429979871958494, -0.041002415120601654, -0.0798906534910202, 0.0005049998289905488, 0.10264964401721954, -0.13263645768165588, 0.09131336957216263, -0.04940832033753395, 0.09063780307769775, -0.047805242240428925, -0.002904811641201377, 0.05241479352116585, 0.0754866898059845, -0.0037812129594385624, 0.07478196173906326, -0.17724202573299408, 0.20696671307086945, 0.02602815441787243, 0.10824580490589142, -0.05676891282200813, 0.0464160330593586, 0.002042506355792284, 0.03342151641845703, 0.07131695747375488, -0.010857120156288147, -0.035238541662693024, -0.16585983335971832, -0.07173275947570801, 0.02666800282895565, 0.08780217170715332, -0.039104655385017395, 0.0644681453704834, -0.038820769637823105, -0.004084727726876736, 0.05165275186300278, -0.02472379431128502, -0.1898316740989685, -0.1607481986284256, 0.026996340602636337, 0.047763511538505554, 0.053930964320898056, -0.09758772701025009, -0.1109406128525734, -0.029683014377951622, 0.1927749365568161, 0.04059541970491409, -0.03633810579776764, -0.1452271044254303, 0.11174321174621582, 0.13165108859539032, -0.05254857614636421, 0.040767472237348557, 0.04071133956313133, 0.17414826154708862, 0.00393838481977582, -0.04292251169681549, 0.0546572208404541, -0.07152152061462402, -0.150070920586586, -0.04016019403934479, 0.17048637568950653, 0.06643325835466385, 0.05618304759263992, 0.027520649135112762, 0.023587753996253014, 0.016894506290555, -0.0673104003071785, 0.039964959025382996, 0.04075378552079201, 0.04843737185001373, 0.05274846777319908, -0.006392944138497114, -0.027542181313037872, -0.0520494319498539, -0.056309137493371964, 0.14984098076820374, 0.2333814948797226, -0.06846687197685242, 0.06993257254362106, 0.10282368212938309, -0.05572419986128807, -0.10087630897760391, 0.05137728899717331, 0.13562141358852386, 0.042711291462183, 0.07225289195775986, -0.18118329346179962, 0.05659356713294983, 0.09714584797620773, -0.02087125927209854, -0.029368208721280098, -0.27898091077804565, -0.12779752910137177, 0.10039924830198288, 0.07663605362176895, -0.04622577130794525, -0.09847793728113174, -0.051841724663972855, -0.07590948790311813, -0.1661485880613327, 0.07681426405906677, -0.07516408711671829, 0.08952426165342331, 0.021869221702218056, 0.06545212119817734, 0.028191160410642624, -0.02546798065304756, 0.15557880699634552, -0.00014605288743041456, 0.05231572687625885, -0.028330184519290924, 0.0830201655626297, 0.046248361468315125, -0.04927768185734749, 0.052782781422138214, -0.05170853063464165, 0.052580464631319046, -0.14607004821300507, -0.04238496720790863, -0.06020355224609375, 0.04685322940349579, -0.042653389275074005, -0.0542508102953434, -0.028937257826328278, 0.055473048239946365, 0.057500146329402924, -0.042280133813619614, 0.03968809172511101, 0.006446478888392448, 0.11615899950265884, 0.1224408894777298, 0.11989060044288635, -0.008682054467499256, -0.11257990449666977, -0.019860753789544106, -0.038898151367902756, 0.05955633521080017, -0.06565218418836594, 0.02851812355220318, 0.10992639511823654, 0.04440954327583313, 0.169782817363739, -0.0023811173159629107, -0.0765928104519844, 0.011273624375462532, 0.018780719488859177, -0.038288217037916183, -0.16340720653533936, -0.03900976851582527, 0.04182295873761177, -0.14620855450630188, -0.038148265331983566, 0.12521634995937347, -0.06386785209178925, -0.02941996604204178, -0.0068135857582092285, 0.01946411281824112, -0.04436914622783661, 0.18634545803070068, 0.015435053035616875, 0.07845859229564667, -0.07205817848443985, 0.09111179411411285, 0.09914196282625198, -0.12141374498605728, 0.07556019723415375, 0.029147770255804062, -0.07143042236566544, -0.02398700825870037, 0.034550465643405914, 0.11688973009586334, -0.0036297719925642014, -0.043072499334812164, -0.059575870633125305, -0.11259999126195908, 0.042570605874061584, 0.04343990609049797, 0.016033576801419258, -0.027767598628997803, -0.042132824659347534, 0.015406436286866665, -0.12689410150051117, 0.08900053054094315, 0.06395717710256577, 0.054884448647499084, -0.141330748796463, 0.08427441120147705, 0.01437430176883936, 0.018597643822431564, -0.0022720838896930218, 0.00477502541616559, -0.04768179729580879, -0.0027851201593875885, -0.17338977754116058, -0.02302955836057663, -0.022988738492131233, 0.01514060515910387, -0.031110605224967003, -0.05715898051857948, -0.037178557366132736, 0.0487484484910965, -0.06318394839763641, -0.07507239282131195, 0.01299070194363594, 0.06266852468252182, -0.1149887666106224, -0.01021052896976471, 0.040701862424612045, -0.09336478263139725, 0.06936386227607727, 0.04637470468878746, 0.029644446447491646, 0.014087354764342308, -0.05727047100663185, -0.011231007985770702, 0.03345458582043648, 0.028712237253785133, 0.057614732533693314, -0.15532778203487396, -0.020138926804065704, -0.0028117336332798004, 0.012842245399951935, 0.011810116469860077, 0.08910386264324188, -0.09732476621866226, -0.08051407337188721, -0.061957892030477524, -0.02072432078421116, -0.054878782480955124, 0.05003603175282478, 0.11194606870412827, 0.06767506152391434, 0.15635022521018982, -0.08018367737531662, 0.05565667897462845, -0.20474568009376526, -0.03182976320385933, -0.03382433205842972, -0.006835281383246183, -0.03470156341791153, -0.04253353923559189, 0.08248895406723022, -0.043932896107435226, 0.07767936587333679, -0.037992365658283234, 0.10849910229444504, 0.05289551988244057, -0.07064487040042877, -0.012915484607219696, 0.0022740033455193043, 0.17603817582130432, 0.06809510290622711, -0.0028601454105228186, 0.10691101849079132, -0.036898255348205566, 0.06545108556747437, 0.11573532223701477, 0.10574682801961899, 0.15137261152267456, 0.014599485322833061, 0.04476460814476013, 0.06510929018259048, -0.12405593693256378, -0.14754362404346466, 0.07264331728219986, -0.03961290419101715, 0.12507298588752747, -0.037147097289562225, 0.16059744358062744, 0.091526560485363, -0.1759120672941208, 0.05030041188001633, -0.06209305301308632, -0.10603214800357819, -0.06423554569482803, -0.05487232655286789, -0.07495748996734619, -0.1133410856127739, 0.011073101311922073, -0.100442074239254, 0.030207736417651176, 0.036922235041856766, 0.011700053699314594, 0.015653619542717934, 0.14980748295783997, -0.02668278105556965, -0.0198642797768116, 0.08942396193742752, -0.0019023786298930645, -0.015328731387853622, -0.07580263912677765, -0.05265801027417183, 0.07778788357973099, 0.029845628887414932, 0.08590326458215714, -0.043107327073812485, -0.0349498987197876, 0.04288947954773903, 0.032278310507535934, -0.0965259000658989, 0.032335326075553894, 0.0042704027146101, 0.057794973254203796, 0.07863198965787888, 0.05655355751514435, 0.019649112597107887, -0.038367681205272675, 0.24343162775039673, -0.06470122933387756, -0.05792904645204544, -0.13953743875026703, 0.12394370138645172, 0.014148220419883728, -0.017736351117491722, 0.06300252676010132, -0.10509124398231506, 0.0055309319868683815, 0.0996483862400055, 0.10482911765575409, -0.035751067101955414, -0.008695519529283047, -0.015718087553977966, -0.02246176451444626, -0.06418190896511078, 0.10089701414108276, 0.11814378947019577, -0.013477045111358166, -0.04984022304415703, 0.03677065297961235, -0.03427749127149582, -0.06985055655241013, -0.03767012804746628, 0.082198865711689, -0.0029977597296237946, 0.01370973140001297, -0.006845769006758928, 0.12130625545978546, 0.04785344749689102, -0.23246558010578156, -0.01937783882021904, -0.15758377313613892, -0.19561657309532166, -0.017786797136068344, 0.05236836150288582, 0.00910322368144989, 0.04572441428899765, -0.0016288532642647624, 0.01490327250212431, 0.14313311874866486, -0.01527281291782856, -0.005318735260516405, -0.12366364151239395, 0.11454688012599945, -0.08638756722211838, 0.18448348343372345, -0.009334860369563103, 0.07203710079193115, 0.09841737896203995, 0.05364321172237396, -0.12867067754268646, 0.024615518748760223, 0.07738050073385239, -0.03585074841976166, 0.05422576889395714, 0.18723788857460022, -0.046217601746320724, 0.134948268532753, 0.042593732476234436, -0.1418369710445404, -0.012220213189721107, -0.09584346413612366, 0.014363386668264866, -0.07500370591878891, 0.013873041607439518, -0.05813218280673027, 0.16881020367145538, 0.16083061695098877, -0.068543441593647, -0.018289662897586823, -0.05673127993941307, 0.02241186611354351, 0.060081616044044495, 0.13942912220954895, -0.03489873185753822, -0.22773627936840057, 0.02122701331973076, 0.009654329158365726, 0.03346836566925049, -0.2350054681301117, -0.10498160868883133, 0.041591424494981766, -0.06954964250326157, -0.01890704035758972, 0.1133013367652893, 0.03597294166684151, 0.01919795386493206, -0.038416050374507904, -0.12225348502397537, -0.043091822415590286, 0.12612158060073853, -0.16038250923156738, -0.035589348524808884 ]
null
null
transformers
# Wav2Vec2-Large-XLSR-53-Chuvash Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on Chuvash using the [Common Voice](https://huggingface.co/datasets/common_voice) When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ```python import torch import torchaudio from datasets import load_dataset from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor test_dataset = load_dataset("common_voice", "cv", split="test") processor = Wav2Vec2Processor.from_pretrained("gagan3012/wav2vec2-xlsr-chuvash") model = Wav2Vec2ForCTC.from_pretrained("gagan3012/wav2vec2-xlsr-chuvash") resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): \\tspeech_array, sampling_rate = torchaudio.load(batch["path"]) \\tbatch["speech"] = resampler(speech_array).squeeze().numpy() \\treturn batch test_dataset = test_dataset.map(speech_file_to_array_fn) inputs = processor(test_dataset["speech"][:2], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): \\tlogits = model(inputs.input_values, attention_mask=inputs.attention_mask).logits predicted_ids = torch.argmax(logits, dim=-1) print("Prediction:", processor.batch_decode(predicted_ids)) print("Reference:", test_dataset["sentence"][:2]) ``` #### Results: Prediction: ['проектпа килӗшӳллӗн тӗлӗ мероприяти иртермелле', 'твăра çак планета минтӗ пуяни калленнана'] Reference: ['Проектпа килӗшӳллӗн, тӗрлӗ мероприяти ирттермелле.', 'Çак планета питĕ пуян иккен.'] ## Evaluation The model can be evaluated as follows on the Chuvash test data of Common Voice. ```python import torch import torchaudio from datasets import load_dataset, load_metric from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor import re !mkdir cer !wget -O cer/cer.py https://huggingface.co/ctl/wav2vec2-large-xlsr-cantonese/raw/main/cer.py test_dataset = load_dataset("common_voice", "cv", split="test") #TODO: replace {lang_id} in your language code here. Make sure the code is one of the *ISO codes* of [this](https://huggingface.co/languages) site. wer = load_metric("wer") cer = load_metric("cer") processor = Wav2Vec2Processor.from_pretrained("gagan3012/wav2vec2-xlsr-chuvash") model = Wav2Vec2ForCTC.from_pretrained("gagan3012/wav2vec2-xlsr-chuvash") model.to("cuda") chars_to_ignore_regex = '[\\\\,\\\\?\\\\.\\\\!\\\\-\\\\;\\\\:\\\\"\\\\“]' # TODO: adapt this list to include all special characters you removed from the data resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): \\tbatch["sentence"] = re.sub(chars_to_ignore_regex, '', batch["sentence"]).lower() \\tspeech_array, sampling_rate = torchaudio.load(batch["path"]) \\tbatch["speech"] = resampler(speech_array).squeeze().numpy() \\treturn batch test_dataset = test_dataset.map(speech_file_to_array_fn) # Preprocessing the datasets. # We need to read the aduio files as arrays def evaluate(batch): \\tinputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True) \\twith torch.no_grad(): \\t\\tlogits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits \\tpred_ids = torch.argmax(logits, dim=-1) \\tbatch["pred_strings"] = processor.batch_decode(pred_ids) \\treturn batch result = test_dataset.map(evaluate, batched=True, batch_size=8) print("WER: {:2f}".format(100 * wer.compute(predictions=result["pred_strings"], references=result["sentence"]))) print("CER: {:2f}".format(100 * cer.compute(predictions=result["pred_strings"], references=result["sentence"]))) ``` **Test Result**: 48.40 % ## Training The script used for training can be found [here](https://colab.research.google.com/drive/1A7Y20c1QkSHfdOmLXPMiOEpwlTjDZ7m5?usp=sharing)
{"language": "cv", "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech", "xlsr-fine-tuning-week"], "datasets": ["common_voice"], "metrics": ["wer"], "model-index": [{"name": "wav2vec2-xlsr-chuvash by Gagan Bhatia", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Speech Recognition"}, "dataset": {"name": "Common Voice cv", "type": "common_voice", "args": "cv"}, "metrics": [{"type": "wer", "value": 48.4, "name": "Test WER"}]}]}]}
automatic-speech-recognition
gagan3012/wav2vec2-xlsr-chuvash
[ "transformers", "pytorch", "jax", "wav2vec2", "automatic-speech-recognition", "audio", "speech", "xlsr-fine-tuning-week", "cv", "dataset:common_voice", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "cv" ]
TAGS #transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #cv #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us
# Wav2Vec2-Large-XLSR-53-Chuvash Fine-tuned facebook/wav2vec2-large-xlsr-53 on Chuvash using the Common Voice When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: #### Results: Prediction: ['проектпа килӗшӳллӗн тӗлӗ мероприяти иртермелле', 'твăра çак планета минтӗ пуяни калленнана'] Reference: ['Проектпа килӗшӳллӗн, тӗрлӗ мероприяти ирттермелле.', 'Çак планета питĕ пуян иккен.'] ## Evaluation The model can be evaluated as follows on the Chuvash test data of Common Voice. Test Result: 48.40 % ## Training The script used for training can be found here
[ "# Wav2Vec2-Large-XLSR-53-Chuvash \n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Chuvash using the Common Voice\n\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:", "#### Results: \n\nPrediction: ['проектпа килӗшӳллӗн тӗлӗ мероприяти иртермелле', 'твăра çак планета минтӗ пуяни калленнана']\n\nReference: ['Проектпа килӗшӳллӗн, тӗрлӗ мероприяти ирттермелле.', 'Çак планета питĕ пуян иккен.']", "## Evaluation\n\nThe model can be evaluated as follows on the Chuvash test data of Common Voice.\n\n\n\nTest Result: 48.40 %", "## Training\n\nThe script used for training can be found here" ]
[ "TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #cv #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us \n", "# Wav2Vec2-Large-XLSR-53-Chuvash \n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Chuvash using the Common Voice\n\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:", "#### Results: \n\nPrediction: ['проектпа килӗшӳллӗн тӗлӗ мероприяти иртермелле', 'твăра çак планета минтӗ пуяни калленнана']\n\nReference: ['Проектпа килӗшӳллӗн, тӗрлӗ мероприяти ирттермелле.', 'Çак планета питĕ пуян иккен.']", "## Evaluation\n\nThe model can be evaluated as follows on the Chuvash test data of Common Voice.\n\n\n\nTest Result: 48.40 %", "## Training\n\nThe script used for training can be found here" ]
[ 81, 64, 20, 98, 30, 11 ]
[ "passage: TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #cv #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us \n# Wav2Vec2-Large-XLSR-53-Chuvash \n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Chuvash using the Common Voice\n\nWhen using this model, make sure that your speech input is sampled at 16kHz.## Usage\n\nThe model can be used directly (without a language model) as follows:#### Results: \n\nPrediction: ['проектпа килӗшӳллӗн тӗлӗ мероприяти иртермелле', 'твăра çак планета минтӗ пуяни калленнана']\n\nReference: ['Проектпа килӗшӳллӗн, тӗрлӗ мероприяти ирттермелле.', 'Çак планета питĕ пуян иккен.']## Evaluation\n\nThe model can be evaluated as follows on the Chuvash test data of Common Voice.\n\n\n\nTest Result: 48.40 %## Training\n\nThe script used for training can be found here" ]
[ -0.13130874931812286, 0.0766790360212326, -0.004165677819401026, -0.04018952324986458, 0.05316861718893051, -0.04254359006881714, 0.18578925728797913, 0.1134696900844574, 0.05130983516573906, 0.05822701007127762, 0.039184607565402985, 0.029134953394532204, 0.07743483781814575, 0.09404581040143967, -0.016242356970906258, -0.20538544654846191, 0.01713976263999939, -0.030163215473294258, 0.0847267434000969, 0.10562913864850998, 0.1193595603108406, -0.0623042918741703, 0.03844381868839264, 0.03934476152062416, -0.04164518415927887, 0.012279894202947617, 0.010688631795346737, -0.0949104055762291, 0.10460203886032104, 0.07776179164648056, 0.04426753893494606, 0.03663690388202667, 0.02283954992890358, -0.23611783981323242, 0.01516634225845337, 0.017695674672722816, 0.0033064635936170816, -0.0035260405857115984, 0.059816475957632065, -0.022422829642891884, 0.28746843338012695, 0.025099903345108032, -0.04251802712678909, 0.0792485773563385, -0.0862240195274353, -0.08086400479078293, -0.01994929276406765, 0.0562753826379776, 0.12349678575992584, 0.11779560893774033, -0.08851656317710876, 0.1335134506225586, -0.08439628779888153, 0.10723628103733063, 0.06070952117443085, -0.16770240664482117, -0.033636391162872314, 0.10975044965744019, 0.027307601645588875, 0.04130270704627037, -0.07537712901830673, 0.018682118505239487, 0.050164543092250824, 0.009136423468589783, -0.13843998312950134, -0.05896070972084999, -0.06643129140138626, 0.0027478192932903767, -0.14742106199264526, -0.02325683832168579, 0.186292365193367, 0.008042840287089348, -0.06148794665932655, -0.10676749050617218, 0.011331746354699135, -0.026835858821868896, -0.03885969892144203, -0.04761900007724762, -0.00995766930282116, 0.019008418545126915, 0.011957571841776371, 0.002422334160655737, -0.11298766732215881, -0.09567741304636002, 0.019652636721730232, 0.07308945059776306, 0.01590072177350521, -0.016871841624379158, -0.084962859749794, 0.03244725987315178, -0.06340522319078445, -0.09079955518245697, -0.04407421499490738, -0.0035343461204320192, -0.04495452344417572, -0.017116477712988853, -0.04293324053287506, -0.10656812787055969, 0.12236995995044708, 0.019435426220297813, 0.02447054162621498, 0.06578230857849121, -0.018701665103435516, 0.02074238657951355, 0.055198244750499725, 0.11626604944467545, -0.0785725936293602, -0.06636996567249298, -0.018497778102755547, 0.06788277626037598, -0.03897624462842941, 0.005133488681167364, -0.026828033849596977, -0.06440650671720505, 0.03735974803566933, 0.07032550871372223, 0.017504878342151642, 0.001396268722601235, -0.053661756217479706, -0.01701716147363186, 0.06518781930208206, -0.0869954377412796, -0.050368208438158035, 0.071685291826725, -0.022221604362130165, 0.10751382261514664, 0.011306392960250378, 0.042061708867549896, -0.04858417063951492, -0.030935492366552353, 0.0068983458913862705, 0.06821483373641968, -0.01583152823150158, -0.0531289279460907, 0.018917594105005264, 0.04330601170659065, -0.011899671517312527, -0.11212067306041718, -0.04133448749780655, -0.11609137803316116, -0.00850757211446762, -0.017601918429136276, 0.017793437466025352, -0.10279079526662827, -0.04585787281394005, -0.02946152351796627, -0.03132883459329605, 0.010717054829001427, -0.05617445334792137, 0.04675408452749252, 0.04655538871884346, 0.05249744653701782, 0.09089614450931549, 0.05713054910302162, -0.0931541919708252, -0.005990488920360804, -0.032242245972156525, 0.11702375113964081, -0.10626755654811859, -0.02570197731256485, -0.14362776279449463, -0.051225680857896805, -0.03876573592424393, 0.07468081265687943, 0.0803518146276474, 0.14990293979644775, -0.25855332612991333, -0.060886450111866, 0.2004828304052353, -0.1015433743596077, -0.08432324230670929, 0.2159992754459381, 0.005952687934041023, 0.08131545782089233, 0.10749198496341705, 0.18702977895736694, 0.1224813312292099, -0.19111409783363342, -0.032958678901195526, 0.03431457281112671, 0.012961686588823795, 0.05812210217118263, 0.057100873440504074, -0.0781116634607315, -0.008218537084758282, 0.02660061977803707, -0.030840925872325897, 0.0060433452017605305, -0.05332079529762268, -0.05849932134151459, -0.005612561013549566, -0.08572370558977127, 0.05087408795952797, 0.040255967527627945, -0.026575619354844093, -0.046722836792469025, -0.051404207944869995, -0.05017489567399025, 0.1319890171289444, -0.07535699009895325, 0.007753316778689623, -0.11931394040584564, 0.0954023003578186, -0.011893930844962597, -0.006713601760566235, -0.12120446562767029, 0.15405131876468658, 0.004274562932550907, 0.0535200797021389, 0.06266418844461441, 0.12353269010782242, 0.035429149866104126, 0.0005485214060172439, -0.0033848718740046024, -0.02361520193517208, 0.045679740607738495, -0.002690610010176897, -0.06801164895296097, -0.11594254523515701, 0.022244658321142197, -0.0463348813354969, 0.12248758971691132, -0.1918438971042633, -0.02323819510638714, 0.0604025162756443, 0.028417782858014107, -0.017865726724267006, -0.0067266542464494705, 0.07491404563188553, 0.06536339223384857, -0.011313182301819324, 0.02093791775405407, 0.005385753232985735, -0.010438643395900726, -0.04979970306158066, 0.09674011915922165, -0.11412986367940903, -0.07697514444589615, 0.08603770285844803, -0.03445345163345337, -0.06361748278141022, -0.004305823240429163, -0.0238945409655571, -0.010980414226651192, -0.01541050337255001, 0.01749076321721077, 0.2549152076244354, 0.012672996148467064, 0.06824874132871628, -0.06817860901355743, 0.024293916299939156, 0.01499936543405056, -0.10970072448253632, 0.007899175398051739, 0.14286567270755768, -0.0008946973248384893, -0.03553565964102745, 0.0342797115445137, -0.030462104827165604, -0.0705207884311676, 0.24841240048408508, -0.04781213402748108, -0.11393552273511887, -0.034178465604782104, 0.029097329825162888, -0.017379188910126686, 0.035229120403528214, -0.18178467452526093, -0.026202857494354248, 0.03409157320857048, 0.037666261196136475, 0.0592983141541481, -0.1140860766172409, 0.013610927388072014, 0.009698877111077309, -0.1118256077170372, -0.10677136480808258, 0.0937563106417656, -0.03371910750865936, 0.04882196709513664, -0.08063926547765732, -0.03322244808077812, -0.08470775187015533, -0.05676768720149994, -0.17519162595272064, 0.14885424077510834, -0.06914383172988892, -0.15971212089061737, -0.11372938752174377, 0.015049612149596214, -0.02140912227332592, 0.019686702638864517, 0.07743816822767258, -0.09889883548021317, -0.015050916001200676, -0.045330118387937546, 0.09714839607477188, 0.006864278111606836, -0.05204775929450989, -0.0651177391409874, 0.059380851686000824, 0.022631749510765076, -0.11667194962501526, -0.017168570309877396, -0.03019022010266781, -0.04535289853811264, 0.02395123988389969, -0.04609587416052818, 0.012003716081380844, 0.14308017492294312, 0.06083940342068672, -0.010258723050355911, -0.008063477464020252, 0.24882261455059052, -0.11585317552089691, -0.0026241610758006573, 0.1662520319223404, -0.0007311312947422266, -0.0069559914991259575, 0.11259493231773376, 0.0018548117950558662, -0.08495401591062546, 0.02374245971441269, -0.0360858179628849, -0.04317273199558258, -0.23889042437076569, -0.11819452792406082, -0.058533087372779846, 0.022633204236626625, -0.03492848575115204, -0.014473969116806984, 0.02770155295729637, 0.03902848809957504, -0.03191937506198883, -0.03646411374211311, 0.025180352851748466, 0.004042805638164282, 0.099045030772686, -0.0005484735011123121, 0.0830177366733551, -0.07451952248811722, -0.023053687065839767, 0.03298817574977875, 0.010870339348912239, 0.11962152272462845, 0.07776355743408203, 0.0973701998591423, 0.10589992254972458, 0.08786525577306747, 0.05230940505862236, -0.007750760298222303, -0.024721501395106316, -0.03850021958351135, 0.029644900932908058, -0.0816287100315094, -0.0847334936261177, 0.08199107646942139, 0.18322743475437164, -0.07290467619895935, -0.024479372426867485, 0.043346744030714035, 0.04872991889715195, 0.18211689591407776, 0.1042461097240448, -0.1587088406085968, -0.10396319627761841, -0.013254649937152863, -0.0356188639998436, 0.0016418197192251682, 0.041521355509757996, 0.09688129276037216, -0.13555429875850677, 0.040809448808431625, 0.04611369967460632, 0.07169338315725327, -0.04850520193576813, 0.03272135928273201, -0.12955595552921295, -0.023521143943071365, -0.02169390209019184, 0.09617620706558228, -0.30019861459732056, 0.2672881484031677, 0.012024863623082638, 0.0913025364279747, -0.049226995557546616, 0.0038181645795702934, 0.040835995227098465, 0.02593502588570118, 0.12822796404361725, 0.009157556109130383, -0.06950896978378296, -0.14649246633052826, -0.07058273255825043, 0.03336585313081741, 0.04459531232714653, 0.0069268085062503815, 0.07654597610235214, -0.007891505025327206, -0.018160229548811913, -0.02607390284538269, -0.04976139962673187, -0.1711382269859314, -0.03402910381555557, 0.03283519297838211, 0.07329618185758591, 0.04353834688663483, -0.012833083048462868, -0.04130184277892113, -0.05140082910656929, 0.01723114401102066, -0.14127258956432343, -0.03663626313209534, -0.03993706405162811, -0.05831170082092285, 0.12792067229747772, -0.1043977290391922, -0.025516420602798462, 0.038458216935396194, 0.07406862080097198, -0.024170920252799988, -0.011210042051970959, 0.053441308438777924, -0.13386137783527374, -0.14699430763721466, -0.009792927652597427, 0.17902779579162598, 0.07923094183206558, 0.09575697034597397, 0.06293805688619614, 0.041010599583387375, -0.030558714643120766, -0.0575832724571228, 0.00445350119844079, 0.07538917660713196, -0.11076990514993668, 0.03611787408590317, 0.006272442638874054, -0.17723330855369568, -0.14606592059135437, -0.08336107432842255, 0.10631369799375534, 0.2169535905122757, -0.011072095483541489, 0.10669038444757462, 0.18668249249458313, -0.08589673787355423, -0.1860184222459793, -0.04170963168144226, 0.09105020761489868, 0.09162962436676025, -0.012781263329088688, -0.1669689267873764, 0.03888889402151108, 0.014298886992037296, -0.0313413143157959, 0.012667727656662464, -0.2636623680591583, -0.13240352272987366, 0.17343199253082275, -0.029814990237355232, -0.007553836330771446, -0.08777215331792831, -0.06988892704248428, -0.03537757694721222, -0.036156754940748215, -0.056652866303920746, -0.041811540722846985, 0.0830015167593956, 0.02788419835269451, 0.06002407148480415, 0.06262007355690002, -0.017520355060696602, 0.1421150118112564, 0.09114264696836472, -0.006769110448658466, -0.02636098489165306, -0.01032519992440939, -0.04617796465754509, 0.005839335732161999, 0.1606825441122055, -0.04115966334939003, 0.027322186157107353, -0.150015726685524, -0.0637681633234024, -0.0711146667599678, 0.061995282769203186, 0.0005257092998363078, -0.027936678379774094, -0.0040192375890910625, -0.004794151987880468, -0.0002035505312960595, 0.010038714855909348, -0.05689701810479164, -0.14612522721290588, 0.05177402123808861, 0.19715861976146698, 0.1705828160047531, 0.037822429090738297, -0.02554292604327202, 0.00485969427973032, -0.0029186177998781204, 0.0844932347536087, -0.09302926808595657, -0.004121736157685518, 0.08276307582855225, 0.04862720146775246, 0.1071062758564949, -0.027606332674622536, -0.11511413007974625, 0.05023316666483879, 0.0496489480137825, -0.07161875814199448, -0.1018151044845581, -0.036425068974494934, -0.0077011967077851295, -0.04257364198565483, 0.007090606261044741, 0.1205405592918396, -0.07895468920469284, 0.001679872046224773, -0.04103118181228638, 0.03042846918106079, -0.09021097421646118, 0.18852044641971588, 0.05329444259405136, 0.07760018110275269, -0.07952820509672165, 0.002520896727219224, 0.013796706683933735, -0.04235455393791199, 0.06586898118257523, -0.011287652887403965, -0.04331027343869209, -0.09862051904201508, -0.04071078822016716, 0.10944113880395889, 0.040467534214258194, -0.08515941351652145, -0.051366545259952545, -0.05951924994587898, 0.0009035872644744813, 0.09905990213155746, 0.07426189631223679, 0.02946748398244381, -0.047640085220336914, -0.026653915643692017, -0.08209507912397385, 0.06462618708610535, 0.08585955947637558, -0.016925593838095665, -0.0446401908993721, 0.15798267722129822, 0.02223624661564827, 0.027605179697275162, -0.026993870735168457, -0.06437322497367859, -0.07331833243370056, 0.07782240957021713, -0.10957150161266327, 0.0164593905210495, -0.11179155856370926, -0.006530773360282183, 0.036086756736040115, -0.0629926398396492, -0.0302569679915905, 0.037554629147052765, -0.09881992638111115, 0.041416820138692856, -0.031019315123558044, 0.13336102664470673, -0.10995601117610931, 0.035147760063409805, 0.06083454191684723, -0.0734691470861435, 0.05965159088373184, 0.10189137607812881, -0.06671860814094543, 0.10994704812765121, -0.20011186599731445, 0.01599407196044922, 0.02701457031071186, 0.05827096104621887, 0.008335993625223637, -0.11665689945220947, 0.04209330677986145, 0.036909960210323334, 0.07588155567646027, 0.005198417231440544, 0.06630709022283554, -0.07523573935031891, -0.009627113118767738, -0.04815598949790001, -0.04885081201791763, -0.028500866144895554, 0.03955969214439392, 0.04761480167508125, 0.06765413284301758, 0.12843680381774902, -0.11078166961669922, 0.0895853340625763, -0.10709387063980103, -0.014643372036516666, -0.046698447316884995, -0.022050345316529274, -0.08041856437921524, -0.09121308475732803, 0.10257060080766678, -0.04908396676182747, 0.10794606059789658, 0.012839745730161667, 0.10395188629627228, -0.006652303971350193, -0.10724755376577377, 0.031311940401792526, 0.03957977145910263, 0.1864362210035324, 0.07177182286977768, 0.009248530492186546, -0.097364641726017, -0.004858276341110468, -0.003220105078071356, 0.023609044030308723, -0.029177939519286156, 0.13243862986564636, -0.053952816873788834, 0.08555684238672256, 0.08598890155553818, -0.0508195199072361, 0.0009624160593375564, -0.06565631181001663, -0.1768166571855545, 0.034688808023929596, -0.038503535091876984, 0.1756296455860138, 0.11978132277727127, -0.1450423151254654, 0.07518208026885986, -0.05395834147930145, -0.0760403573513031, -0.11947472393512726, -0.1167718917131424, -0.07999929040670395, -0.12497083097696304, 0.04654031991958618, -0.09876497834920883, 0.031532078981399536, 0.06700834631919861, 0.06492411345243454, -0.04264993593096733, 0.18606771528720856, 0.02784944325685501, -0.07491081207990646, 0.13822275400161743, -0.06573577225208282, 0.0010009981924667954, -0.062200795859098434, 0.00888073816895485, 0.11714418977499008, -0.05058395490050316, 0.04857062175869942, 0.015850167721509933, -0.10524553805589676, -0.0016954002203419805, -0.06266342103481293, -0.09354618936777115, 0.00932818278670311, -0.019993577152490616, 0.05685725435614586, 0.11034940183162689, 0.10201532393693924, -0.03782396391034126, 0.024265676736831665, 0.13146550953388214, -0.035790521651506424, -0.14791575074195862, -0.14652228355407715, 0.20917090773582458, 0.02424195408821106, 0.05481373518705368, -0.038139209151268005, -0.04426675662398338, 0.007216861937195063, 0.19619032740592957, 0.16692538559436798, 0.00858552660793066, 0.02341960370540619, -0.01859721913933754, 0.019295334815979004, -0.0013042987557128072, 0.049075040966272354, 0.056992821395397186, 0.2015356570482254, -0.02191496267914772, 0.0122973108664155, -0.07509102672338486, -0.0810425654053688, -0.01305802445858717, 0.056573156267404556, -0.0033380344975739717, -0.08687656372785568, -0.00043933032429777086, 0.18313848972320557, -0.12068627774715424, -0.11558373272418976, -0.10651107877492905, -0.12969617545604706, -0.1004772037267685, -0.019257061183452606, -0.009530622512102127, 0.14941143989562988, 0.05347819626331329, -0.05140155181288719, -0.0280451700091362, 0.1151038408279419, -0.000646221567876637, -0.04446765035390854, -0.02951309271156788, 0.031727708876132965, -0.06972837448120117, 0.026806019246578217, 0.012761702761054039, 0.17291949689388275, 0.07392395287752151, 0.07025076448917389, 0.0002550484787207097, 0.18097782135009766, 0.062006328254938126, -0.08066203445196152, 0.0614728219807148, 0.11975162476301193, -0.015752214938402176, 0.09403332322835922, 0.08128045499324799, -0.11158067733049393, 0.07494963705539703, -0.05119676887989044, 0.044888272881507874, -0.12059508264064789, 0.099802665412426, -0.07037966698408127, 0.10974639654159546, 0.11762232333421707, -0.08011310547590256, -0.04234221950173378, -0.022791435942053795, 0.08449720591306686, -0.011317608878016472, 0.006760754156857729, -0.07662247866392136, -0.21847161650657654, 0.013722417876124382, -0.041672226041555405, 0.03528355434536934, -0.2227526605129242, -0.011046190746128559, 0.011943608522415161, -0.031343013048172, 0.015415189787745476, 0.08434135466814041, 0.06350398808717728, 0.012940643355250359, -0.007221465930342674, -0.03166961669921875, 0.037518855184316635, 0.13675859570503235, -0.12685953080654144, -0.12671609222888947 ]
null
null
transformers
# Wav2Vec2-Large-XLSR-53-khmer Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on Khmer using the [Common Voice](https://huggingface.co/datasets/common_voice), and [OpenSLR Kh](http://www.openslr.org/42/). When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ```python import torch import torchaudio from datasets import load_dataset from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor !wget https://www.openslr.org/resources/42/km_kh_male.zip !unzip km_kh_male.zip !ls km_kh_male colnames=['path','sentence'] df = pd.read_csv('/content/km_kh_male/line_index.tsv',sep='\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\t',header=None,names = colnames) df['path'] = '/content/km_kh_male/wavs/'+df['path'] +'.wav' train, test = train_test_split(df, test_size=0.1) test.to_csv('/content/km_kh_male/line_index_test.csv') test_dataset = load_dataset('csv', data_files='/content/km_kh_male/line_index_test.csv',split = 'train') processor = Wav2Vec2Processor.from_pretrained("gagan3012/wav2vec2-xlsr-nepali") model = Wav2Vec2ForCTC.from_pretrained("gagan3012/wav2vec2-xlsr-nepali") resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): \\\\\\\\\\\\\\\\tspeech_array, sampling_rate = torchaudio.load(batch["path"]) \\\\\\\\\\\\\\\\tbatch["speech"] = resampler(speech_array).squeeze().numpy() \\\\\\\\\\\\\\\\treturn batch test_dataset = test_dataset.map(speech_file_to_array_fn) inputs = processor(test_dataset["speech"][:2], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): \\\\\\\\\\\\\\\\tlogits = model(inputs.input_values, attention_mask=inputs.attention_mask).logits predicted_ids = torch.argmax(logits, dim=-1) print("Prediction:", processor.batch_decode(predicted_ids)) print("Reference:", test_dataset["sentence"][:2]) ``` #### Result Prediction: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ'] Reference: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ'] ## Evaluation The model can be evaluated as follows on the {language} test data of Common Voice. # TODO: replace #TODO: replace language with your {language}, *e.g.* French ```python import torch import torchaudio from datasets import load_dataset, load_metric from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor import re from sklearn.model_selection import train_test_split import pandas as pd from datasets import load_dataset !wget https://www.openslr.org/resources/42/km_kh_male.zip !unzip km_kh_male.zip !ls km_kh_male colnames=['path','sentence'] df = pd.read_csv('/content/km_kh_male/line_index.tsv',sep='\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\t',header=None,names = colnames) df['path'] = '/content/km_kh_male/wavs/'+df['path'] +'.wav' train, test = train_test_split(df, test_size=0.1) test.to_csv('/content/km_kh_male/line_index_test.csv') test_dataset = load_dataset('csv', data_files='/content/km_kh_male/line_index_test.csv',split = 'train') wer = load_metric("wer") cer = load_metric("cer") processor = Wav2Vec2Processor.from_pretrained("gagan3012/wav2vec2-xlsr-khmer") model = Wav2Vec2ForCTC.from_pretrained("gagan3012/wav2vec2-xlsr-khmer") model.to("cuda") chars_to_ignore_regex = '[\\\\,\\\\?\\\\.\\\\!\\\\-\\\\;\\\\:\\\\"\\\\“]' resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): \\tbatch["text"] = re.sub(chars_to_ignore_regex, '', batch["text"]).lower() \\tspeech_array, sampling_rate = torchaudio.load(batch["path"]) \\tbatch["speech"] = resampler(speech_array).squeeze().numpy() \\treturn batch test_dataset = test_dataset.map(speech_file_to_array_fn) # Preprocessing the datasets. # We need to read the aduio files as arrays def evaluate(batch): \\tinputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True) \\twith torch.no_grad(): \\t\\tlogits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits \\tpred_ids = torch.argmax(logits, dim=-1) \\tbatch["pred_strings"] = processor.batch_decode(pred_ids) \\treturn batch cer = load_metric("cer") result = test_dataset.map(evaluate, batched=True, batch_size=8) print("WER: {:2f}".format(100 * wer.compute(predictions=result["pred_strings"], references=result["text"]))) print("CER: {:2f}".format(100 * cer.compute(predictions=result["pred_strings"], references=result["text"]))) ``` **Test Result**: 24.96 % WER: 24.962519 CER: 6.950925 ## Training The script used for training can be found [here](https://colab.research.google.com/drive/1yo_OTMH8FHQrAKCkKdQGMqpkj-kFhS_2?usp=sharing)
{"language": "km", "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech", "xlsr-fine-tuning-week"], "datasets": ["OpenSLR", "common_voice"], "metrics": ["wer"], "model-index": [{"name": "wav2vec2-xlsr-Khmer by Gagan Bhatia", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Speech Recognition"}, "dataset": {"name": "OpenSLR km", "type": "OpenSLR", "args": "km"}, "metrics": [{"type": "wer", "value": 24.96, "name": "Test WER"}]}]}]}
automatic-speech-recognition
gagan3012/wav2vec2-xlsr-khmer
[ "transformers", "pytorch", "jax", "wav2vec2", "automatic-speech-recognition", "audio", "speech", "xlsr-fine-tuning-week", "km", "dataset:OpenSLR", "dataset:common_voice", "license:apache-2.0", "model-index", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "km" ]
TAGS #transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #km #dataset-OpenSLR #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #has_space #region-us
# Wav2Vec2-Large-XLSR-53-khmer Fine-tuned facebook/wav2vec2-large-xlsr-53 on Khmer using the Common Voice, and OpenSLR Kh. When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: #### Result Prediction: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ'] Reference: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ'] ## Evaluation The model can be evaluated as follows on the {language} test data of Common Voice. # TODO: replace #TODO: replace language with your {language}, *e.g.* French Test Result: 24.96 % WER: 24.962519 CER: 6.950925 ## Training The script used for training can be found here
[ "# Wav2Vec2-Large-XLSR-53-khmer \n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Khmer using the Common Voice, and OpenSLR Kh. \n\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:", "#### Result \n\nPrediction: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ']\n\nReference: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ']", "## Evaluation\n\nThe model can be evaluated as follows on the {language} test data of Common Voice. # TODO: replace #TODO: replace language with your {language}, *e.g.* French\n\n\n\n\nTest Result: 24.96 % \n\nWER: 24.962519\nCER: 6.950925", "## Training\n\nThe script used for training can be found here" ]
[ "TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #km #dataset-OpenSLR #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #has_space #region-us \n", "# Wav2Vec2-Large-XLSR-53-khmer \n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Khmer using the Common Voice, and OpenSLR Kh. \n\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:", "#### Result \n\nPrediction: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ']\n\nReference: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ']", "## Evaluation\n\nThe model can be evaluated as follows on the {language} test data of Common Voice. # TODO: replace #TODO: replace language with your {language}, *e.g.* French\n\n\n\n\nTest Result: 24.96 % \n\nWER: 24.962519\nCER: 6.950925", "## Training\n\nThe script used for training can be found here" ]
[ 91, 68, 20, 65, 66, 11 ]
[ "passage: TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #km #dataset-OpenSLR #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #has_space #region-us \n# Wav2Vec2-Large-XLSR-53-khmer \n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Khmer using the Common Voice, and OpenSLR Kh. \n\nWhen using this model, make sure that your speech input is sampled at 16kHz.## Usage\n\nThe model can be used directly (without a language model) as follows:#### Result \n\nPrediction: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ']\n\nReference: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ']## Evaluation\n\nThe model can be evaluated as follows on the {language} test data of Common Voice. # TODO: replace #TODO: replace language with your {language}, *e.g.* French\n\n\n\n\nTest Result: 24.96 % \n\nWER: 24.962519\nCER: 6.950925## Training\n\nThe script used for training can be found here" ]
[ -0.09031026065349579, 0.0717727318406105, -0.0068450854159891605, -0.035807736217975616, 0.05218857154250145, -0.06387702375650406, 0.09556276351213455, 0.09012695401906967, 0.01115777250379324, 0.022562764585018158, 0.022436413913965225, 0.0441143698990345, 0.08825872838497162, 0.056533750146627426, -0.005276513751596212, -0.22472630441188812, 0.06472983956336975, 0.034275054931640625, 0.027172120288014412, 0.06963658332824707, 0.13342902064323425, -0.03571128845214844, 0.036613959819078445, 0.0432443730533123, 0.012594586238265038, 0.04913643002510071, 0.009909894317388535, -0.0707576721906662, 0.12601712346076965, 0.06052879989147186, 0.05740708112716675, 0.08868679404258728, -0.025422589853405952, -0.1798829734325409, 0.015648141503334045, 0.008996610529720783, -0.03627024590969086, -0.03130534663796425, 0.014446442946791649, -0.012094429694116116, 0.17928731441497803, -0.007958732545375824, -0.04160159453749657, 0.06910715252161026, -0.06594399362802505, -0.1545686572790146, -0.04831152781844139, -0.007628717925399542, 0.09481565654277802, 0.12503552436828613, -0.07165523618459702, 0.11936558783054352, -0.07670584321022034, 0.0890340730547905, 0.1119571328163147, -0.1672837734222412, 0.002599429339170456, 0.09966295212507248, 0.06507834792137146, 0.026054423302412033, -0.11321428418159485, 0.02149507775902748, 0.006927070673555136, -0.002215775428339839, -0.06689896434545517, -0.110443614423275, -0.09526380896568298, -0.009287431836128235, -0.1021905317902565, 0.008131474256515503, 0.24551209807395935, -0.003792740171775222, -0.03911741450428963, -0.05847678706049919, -0.009263665415346622, 0.016127722337841988, -0.047518935054540634, -0.10418875515460968, 0.0313713513314724, 0.03140300512313843, 0.03284519538283348, -0.13395321369171143, -0.07379616796970367, -0.06367341428995132, -0.02221180684864521, -0.00023386464454233646, 0.0649089366197586, -0.04500745236873627, -0.044218823313713074, 0.03667990490794182, -0.17250466346740723, -0.04750092327594757, -0.07068534940481186, 0.00021066746558062732, -0.04725924879312515, 0.005834732670336962, -0.058368053287267685, -0.12901009619235992, 0.11259075254201889, 0.03208816051483154, -0.0092603899538517, 0.07559164613485336, -0.09188584238290787, 0.059057869017124176, -0.005741084460169077, 0.14135544002056122, -0.088735431432724, -0.0702386125922203, 0.008320578373968601, 0.062312688678503036, -0.04421067237854004, 0.019205858930945396, -0.009813047014176846, -0.07219375669956207, 0.06681806594133377, 0.06925547868013382, 0.0195151474326849, 0.03807660937309265, -0.03649573028087616, -0.009358190931379795, 0.07622548937797546, -0.10098162293434143, -0.03220197930932045, 0.008002613671123981, 0.004526922479271889, 0.09395436197519302, 0.084395632147789, 0.05122825503349304, -0.08659690618515015, -0.028898222371935844, 0.015225561335682869, -0.007127090357244015, -0.05738217756152153, -0.02458750642836094, -0.015408233739435673, 0.01366026233881712, -0.00021223687508609146, -0.02764020673930645, -0.07543933391571045, -0.09896676242351532, 0.012701796367764473, 0.01402722205966711, -0.048590078949928284, -0.15271109342575073, -0.08464019000530243, -0.03680587187409401, -0.013861197978258133, 0.0016847621882334352, -0.02572665922343731, 0.06258897483348846, -0.0005947002791799605, 0.024113986641168594, -0.06087857484817505, 0.07973932474851608, -0.025897415354847908, -0.002331528114154935, -0.09431666880846024, 0.12002813816070557, -0.132862389087677, -0.16313929855823517, -0.12574923038482666, -0.05748118460178375, -0.07424410432577133, 0.07103399187326431, 0.045881953090429306, 0.11044321954250336, -0.24190181493759155, -0.07656484842300415, 0.24061542749404907, -0.10308150202035904, -0.025178521871566772, 0.19415320456027985, -0.003072653664276004, 0.10079794377088547, 0.10294616222381592, 0.15747904777526855, 0.08730259537696838, -0.16256271302700043, -0.03025883063673973, 0.026965033262968063, -0.02945689670741558, 0.050740789622068405, 0.07798003405332565, -0.07866815477609634, -0.02102196216583252, 0.03315654769539833, -0.08252966403961182, -0.03081413544714451, -0.013179664500057697, -0.03669394552707672, 0.010791346430778503, -0.015468910336494446, 0.06048412621021271, 0.01408936083316803, -0.00612728763371706, 0.0072459205985069275, -0.07420038431882858, 0.053028106689453125, 0.11085391789674759, -0.08013895153999329, 0.052851997315883636, -0.07013174146413803, 0.06844307482242584, 0.048756953328847885, 0.0005304207443259656, -0.07752761244773865, 0.08757971227169037, 0.013377204537391663, 0.047919292002916336, 0.09723605960607529, 0.1605919450521469, 0.009416031651198864, -0.014561161398887634, -0.09350791573524475, -0.022247232496738434, 0.005064614582806826, -0.016452286392450333, -0.009821760468184948, -0.08715901523828506, 0.07489179819822311, -0.014781716279685497, 0.04256637394428253, -0.10412325710058212, -0.010340631008148193, 0.041547682136297226, 0.03651810437440872, -0.018429666757583618, 0.026695335283875465, -0.0052520944736897945, 0.04732108861207962, 0.04398031905293465, 0.020609449595212936, -0.0057684145867824554, 0.028047047555446625, -0.0376397930085659, 0.1857193112373352, -0.15332245826721191, -0.08671997487545013, 0.10429519414901733, -0.08859646320343018, -0.04656878858804703, 0.07085848599672318, -0.01946338266134262, -0.022111043334007263, 0.024978669360280037, 0.0043787406757473946, 0.29355934262275696, 0.0014777763281017542, 0.10657768696546555, -0.09105103462934494, -0.013387959450483322, -0.026188332587480545, -0.05769157409667969, 0.0035139606334269047, 0.07145635038614273, 0.022178884595632553, -0.05521517992019653, 0.07017505168914795, -0.06739311665296555, -0.10401848703622818, 0.2460927665233612, -0.030228465795516968, -0.06785102188587189, -0.04648314043879509, 0.05773621052503586, -0.02868160977959633, 0.05019218847155571, -0.20550629496574402, 0.017986446619033813, 0.026304427534341812, 0.0264581348747015, 0.09740371257066727, -0.07201070338487625, 0.04290008917450905, -0.014111638069152832, -0.10041852295398712, -0.048597417771816254, 0.13853177428245544, 0.010956864804029465, 0.0472944900393486, -0.08704936504364014, -0.02186398208141327, -0.03551425784826279, -0.03967117518186569, -0.10395298898220062, 0.1378590613603592, -0.11451974511146545, -0.18514244258403778, -0.13593006134033203, 0.03289470821619034, -0.05460288003087044, 0.010496405884623528, 0.10233031213283539, -0.06023970618844032, -0.03087848797440529, -0.013351126573979855, 0.08322115987539291, -0.05143750458955765, -0.06467120349407196, -0.09949282556772232, 0.016809051856398582, 0.008523331955075264, -0.14660656452178955, -0.036570265889167786, -0.013541151769459248, -0.1401870995759964, -0.010409221053123474, -0.054937779903411865, -0.024343544617295265, 0.0563807487487793, 0.016604850068688393, -0.025919368490576744, -0.009318734519183636, 0.18184302747249603, -0.12382858246564865, 0.03900231048464775, 0.07858777046203613, 0.02315615490078926, 0.04574262350797653, 0.1323293149471283, -0.01150008849799633, -0.03819011524319649, -0.026081889867782593, 0.09888099879026413, 0.01703670807182789, -0.23880702257156372, -0.09460356086492538, -0.07506461441516876, -0.0711684301495552, -0.103042833507061, 0.023843765258789062, 0.07517426460981369, 0.006157597992569208, -0.061979446560144424, -0.01582879200577736, 0.06573975086212158, 0.011009343899786472, 0.14449182152748108, -0.04697003960609436, 0.06204180046916008, -0.03999748453497887, -0.04333357512950897, 0.055488087236881256, 0.035470426082611084, 0.14465291798114777, 0.04680095240473747, 0.05164496600627899, 0.13630741834640503, 0.06237392872571945, -0.005830529611557722, -0.018595540896058083, -0.057266753166913986, 0.017594285309314728, -0.026499254629015923, -0.030226491391658783, 0.008311525918543339, 0.038700640201568604, 0.173552468419075, -0.09230384230613708, -0.017624327912926674, 0.03612041845917702, 0.05275212228298187, 0.1811387687921524, -0.0034358338452875614, -0.12525804340839386, -0.013885363936424255, -0.011277702637016773, -0.05083036422729492, -0.06254498660564423, 0.031726568937301636, 0.043963655829429626, -0.13868191838264465, 0.059433307498693466, 0.028222307562828064, 0.02056232839822769, -0.025664260610938072, 0.020453277975320816, -0.11557632684707642, -0.02186855673789978, 0.007888957858085632, 0.08785205334424973, -0.2698283791542053, 0.29571759700775146, 0.007857944816350937, 0.05070716515183449, -0.04238824173808098, 0.018137335777282715, 0.041585881263017654, 0.038790762424468994, 0.16873624920845032, 0.02330421842634678, -0.056699056178331375, -0.04257466271519661, -0.03700963035225868, 0.04386640712618828, 0.053377822041511536, 0.01922815665602684, 0.0540781207382679, -0.023040251806378365, -0.03614271432161331, -0.009197465144097805, -0.019848067313432693, -0.1469317376613617, -0.08925272524356842, 0.06904450058937073, 0.0845395028591156, 0.09022891521453857, -0.02449377253651619, -0.04611371085047722, -0.06395844370126724, 0.12785467505455017, -0.0962323322892189, -0.0488932728767395, -0.036685798317193985, -0.08578420430421829, 0.15846160054206848, -0.11257067322731018, 0.04631879925727844, 0.034268394112586975, 0.03605341166257858, -0.06348435580730438, 0.020512746647000313, 0.07898525893688202, -0.07357440143823624, -0.050081148743629456, 0.029091928154230118, 0.19136789441108704, 0.08600632101297379, 0.0628434419631958, 0.031376421451568604, 0.02138940803706646, 0.01654474064707756, -0.09195606410503387, -0.03426486626267433, 0.03528793156147003, -0.025140248239040375, 0.04305775463581085, -0.07496043294668198, -0.1917942315340042, -0.11197834461927414, -0.024072160944342613, 0.14092668890953064, 0.24087217450141907, 0.01725546270608902, 0.1605607569217682, 0.16905423998832703, -0.1105290874838829, -0.19740325212478638, -0.10200786590576172, 0.06555015593767166, 0.04187603294849396, -0.027592618018388748, -0.16250315308570862, -0.04905172809958458, -0.020415494218468666, -0.02848406508564949, -0.08790883421897888, -0.1802128702402115, -0.13093990087509155, 0.16908197104930878, -0.048109378665685654, 0.15630801022052765, -0.062306392937898636, -0.03954003378748894, -0.0013083284720778465, 0.033949676901102066, -0.030699322000145912, -0.006537752691656351, 0.10963089764118195, 0.039245348423719406, 0.016197366639971733, 0.05143488943576813, -0.00979649182409048, 0.11900696903467178, 0.017538385465741158, 0.004591608420014381, -0.010524806566536427, -0.05448736622929573, 0.021674590185284615, 0.0061612254939973354, 0.14593985676765442, -0.010981406085193157, 0.050589434802532196, -0.07330195605754852, -0.06551862508058548, -0.10536956787109375, 0.08632302284240723, -0.0206641536206007, -0.0028946807142347097, -0.004643748980015516, 0.048234689980745316, 0.05637552589178085, 0.014780526980757713, -0.0886954665184021, -0.12054047733545303, 0.01888354681432247, 0.13880713284015656, 0.13109742105007172, 0.03912694752216339, -0.05492592602968216, 0.04667995125055313, 0.010426539927721024, 0.07857140898704529, -0.014901013113558292, 0.048488304018974304, 0.04107026755809784, 0.0061953687109053135, 0.1156332790851593, -0.0019810334779322147, -0.12888751924037933, 0.02827584370970726, 0.023954305797815323, -0.05963708832859993, -0.16647857427597046, -0.07014303654432297, -0.012710658833384514, 0.00518486462533474, -0.031837135553359985, 0.0876828283071518, -0.039205439388751984, -0.028881197795271873, -0.03090747632086277, 0.036438848823308945, -0.09615965932607651, 0.06577101349830627, -0.0026603497099131346, 0.05691799148917198, -0.06161379814147949, 0.04867655038833618, 0.02687901072204113, -0.08386867493391037, 0.09733770042657852, -0.0014256907161325216, -0.03483058884739876, -0.06566999852657318, 0.03765133395791054, 0.12265872955322266, 0.06259848922491074, -0.07140211015939713, -0.07712357491254807, -0.024280648678541183, -0.004190861713141203, 0.1317855715751648, 0.035547953099012375, 0.02326146885752678, 0.006873519625514746, -0.026541465893387794, -0.05169661343097687, 0.05966288968920708, 0.10603935271501541, -0.026118770241737366, 0.004554902669042349, 0.03677360713481903, 0.03964994475245476, 0.055291466414928436, -0.025535136461257935, -0.03617394343018532, -0.06988556683063507, 0.01905364729464054, -0.03897788003087044, 0.014792012050747871, -0.07158731669187546, -0.015345326624810696, 0.003713815240189433, -0.07324837148189545, -0.021618090569972992, 0.06129489466547966, -0.11068819463253021, 0.020210113376379013, -0.0036943417508155107, 0.13255642354488373, -0.07805155962705612, 0.030179910361766815, 0.12113143503665924, -0.052158109843730927, 0.04314166679978371, 0.07480596005916595, -0.061389777809381485, 0.12898357212543488, -0.12598182260990143, -0.046368151903152466, 0.008677465841174126, 0.051015034317970276, -0.015826571732759476, -0.1461227536201477, 0.030583670362830162, 0.05299866572022438, 0.10737267136573792, -0.05292796343564987, 0.07059577107429504, -0.08438267558813095, 0.00028541439678519964, -0.08137473464012146, -0.06993262469768524, 0.00016614212654531002, 0.050259556621313095, 0.05113993585109711, 0.029504360631108284, 0.10457015782594681, -0.12699706852436066, 0.04008237272500992, -0.09284776449203491, 0.026397403329610825, -0.03876969590783119, -0.02072797529399395, -0.001210670336149633, -0.11762241274118423, 0.058247193694114685, -0.060433730483055115, 0.14472384750843048, -0.044928889721632004, 0.08997765928506851, -0.0002871597243938595, -0.1260180026292801, -0.0439167357981205, 0.019904736429452896, 0.154211163520813, 0.032092802226543427, 0.012719377875328064, -0.03990739956498146, 0.006496367044746876, -0.017541060224175453, 0.03151283785700798, 0.021314702928066254, 0.20020601153373718, 0.020974863320589066, 0.08847158402204514, 0.0708487331867218, -0.0757264494895935, -0.013089786283671856, -0.0011977077228948474, -0.08297327160835266, -0.016443492844700813, -0.042763907462358475, 0.10536927729845047, 0.12310614436864853, -0.14596803486347198, 0.17260968685150146, 0.02993553690612316, -0.08859830349683762, -0.13881385326385498, -0.14142701029777527, -0.1024717390537262, -0.1031772792339325, 0.0324387364089489, -0.0884280577301979, 0.06570441275835037, 0.011006775312125683, 0.06330287456512451, -0.0500619113445282, 0.18135139346122742, -0.0781383365392685, -0.11437679082155228, 0.0958922803401947, -0.07529572397470474, 0.02022688463330269, -0.045280590653419495, 0.013865690678358078, 0.07064549624919891, 0.04463883116841316, 0.01865852251648903, 0.031244074925780296, -0.03830075263977051, 0.01089894026517868, -0.07492411881685257, -0.027283133938908577, -0.001401794026605785, -0.01827542670071125, 0.079151950776577, 0.12235129624605179, 0.06625504791736603, -0.08033788204193115, 0.04429306834936142, 0.11555267125368118, -0.013483047485351562, -0.19716446101665497, -0.1417911946773529, 0.1518913060426712, 0.07988478243350983, 0.00754649518057704, -0.017459744587540627, -0.06232917681336403, -0.0016331239603459835, 0.2171836495399475, 0.20449519157409668, 0.08420567214488983, 0.04347340017557144, 0.01670886017382145, 0.018052924424409866, 0.03271446004509926, -0.019486667588353157, 0.06951718777418137, 0.1399938464164734, -0.03140726312994957, -0.02927592396736145, -0.0595070905983448, -0.012740839272737503, -0.009529056027531624, 0.14776483178138733, 0.016895964741706848, -0.11905714124441147, 0.035623498260974884, 0.11610522866249084, -0.11216118186712265, -0.17556115984916687, -0.15239456295967102, -0.11439750343561172, -0.1070404052734375, -0.02350134775042534, -0.0397891141474247, 0.07957620918750763, 0.0253765732049942, -0.0028379601426422596, -0.04360226169228554, 0.16272257268428802, 0.030411574989557266, -0.06637834757566452, -0.03835621103644371, 0.03520621731877327, -0.043549880385398865, 0.10264942049980164, 0.03136748448014259, 0.10804636776447296, 0.05004901438951492, 0.02708638273179531, 0.054103583097457886, 0.17014972865581512, 0.04712589085102081, -0.06517840921878815, 0.04706868901848793, 0.13399335741996765, -0.019495418295264244, 0.09667413681745529, 0.02497819811105728, -0.05483398213982582, 0.1063331738114357, -0.11487641930580139, -0.07886353135108948, -0.09359253942966461, 0.14202894270420074, -0.14262282848358154, 0.088338203728199, 0.10708250850439072, -0.014799343422055244, 0.0028079096227884293, -0.06536862999200821, 0.0656260997056961, 0.0016950045246630907, 0.011890720576047897, -0.0670551210641861, -0.1918824315071106, 0.05392531678080559, -0.020779721438884735, 0.05954376235604286, -0.18713900446891785, -0.002396916737779975, 0.03407222405076027, -0.02073672041296959, -0.004820875357836485, 0.07074405997991562, 0.012144510634243488, -0.020372049883008003, -0.0094069829210639, -0.09420368075370789, 0.02153138630092144, 0.09053683280944824, -0.14697407186031342, -0.06654892861843109 ]
null
null
transformers
# Wav2Vec2-Large-XLSR-53-Nepali Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on Nepali using the [Common Voice](https://huggingface.co/datasets/common_voice), and [OpenSLR ne](http://www.openslr.org/43/). When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ```python import torch import torchaudio from datasets import load_dataset from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor !wget https://www.openslr.org/resources/43/ne_np_female.zip !unzip ne_np_female.zip !ls ne_np_female colnames=['path','sentence'] df = pd.read_csv('/content/ne_np_female/line_index.tsv',sep='\\t',header=None,names = colnames) df['path'] = '/content/ne_np_female/wavs/'+df['path'] +'.wav' train, test = train_test_split(df, test_size=0.1) test.to_csv('/content/ne_np_female/line_index_test.csv') test_dataset = load_dataset('csv', data_files='/content/ne_np_female/line_index_test.csv',split = 'train') processor = Wav2Vec2Processor.from_pretrained("gagan3012/wav2vec2-xlsr-nepali") model = Wav2Vec2ForCTC.from_pretrained("gagan3012/wav2vec2-xlsr-nepali") resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): \tspeech_array, sampling_rate = torchaudio.load(batch["path"]) \tbatch["speech"] = resampler(speech_array).squeeze().numpy() \treturn batch test_dataset = test_dataset.map(speech_file_to_array_fn) inputs = processor(test_dataset["speech"][:2], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): \tlogits = model(inputs.input_values, attention_mask=inputs.attention_mask).logits predicted_ids = torch.argmax(logits, dim=-1) print("Prediction:", processor.batch_decode(predicted_ids)) print("Reference:", test_dataset["sentence"][:2]) ``` #### Result Prediction: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ'] Reference: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ'] ## Evaluation The model can be evaluated as follows on the {language} test data of Common Voice. # TODO: replace #TODO: replace language with your {language}, *e.g.* French ```python import torch import torchaudio from datasets import load_dataset, load_metric from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor import re !wget https://www.openslr.org/resources/43/ne_np_female.zip !unzip ne_np_female.zip !ls ne_np_female colnames=['path','sentence'] df = pd.read_csv('/content/ne_np_female/line_index.tsv',sep='\\t',header=None,names = colnames) df['path'] = '/content/ne_np_female/wavs/'+df['path'] +'.wav' train, test = train_test_split(df, test_size=0.1) test.to_csv('/content/ne_np_female/line_index_test.csv') test_dataset = load_dataset('csv', data_files='/content/ne_np_female/line_index_test.csv',split = 'train') wer = load_metric("wer") processor = Wav2Vec2Processor.from_pretrained("gagan3012/wav2vec2-xlsr-nepali") model = Wav2Vec2ForCTC.from_pretrained("gagan3012/wav2vec2-xlsr-nepali") model.to("cuda") chars_to_ignore_regex = '[\\,\\?\\.\\!\\-\\;\\:\\"\\“]' resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): \tbatch["sentence"] = re.sub(chars_to_ignore_regex, '', batch["sentence"]).lower() \tspeech_array, sampling_rate = torchaudio.load(batch["path"]) \tbatch["speech"] = resampler(speech_array).squeeze().numpy() \treturn batch test_dataset = test_dataset.map(speech_file_to_array_fn) # Preprocessing the datasets. # We need to read the aduio files as arrays def evaluate(batch): \tinputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True) \twith torch.no_grad(): \t\tlogits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits \tpred_ids = torch.argmax(logits, dim=-1) \tbatch["pred_strings"] = processor.batch_decode(pred_ids) \treturn batch result = test_dataset.map(evaluate, batched=True, batch_size=8) print("WER: {:2f}".format(100 * wer.compute(predictions=result["pred_strings"], references=result["sentence"]))) ``` **Test Result**: 05.97 % ## Training The script used for training can be found [here](https://colab.research.google.com/drive/1AHnYWXb5cwfMEa2o4O3TSdasAR3iVBFP?usp=sharing)
{"language": "ne", "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech", "xlsr-fine-tuning-week"], "datasets": ["OpenSLR", "common_voice"], "metrics": ["wer"], "model-index": [{"name": "wav2vec2-xlsr-nepali", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Speech Recognition"}, "dataset": {"name": "OpenSLR ne", "type": "OpenSLR", "args": "ne"}, "metrics": [{"type": "wer", "value": 5.97, "name": "Test WER"}]}]}]}
automatic-speech-recognition
gagan3012/wav2vec2-xlsr-nepali
[ "transformers", "pytorch", "jax", "wav2vec2", "automatic-speech-recognition", "audio", "speech", "xlsr-fine-tuning-week", "ne", "dataset:OpenSLR", "dataset:common_voice", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "ne" ]
TAGS #transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #ne #dataset-OpenSLR #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us
# Wav2Vec2-Large-XLSR-53-Nepali Fine-tuned facebook/wav2vec2-large-xlsr-53 on Nepali using the Common Voice, and OpenSLR ne. When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: #### Result Prediction: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ'] Reference: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ'] ## Evaluation The model can be evaluated as follows on the {language} test data of Common Voice. # TODO: replace #TODO: replace language with your {language}, *e.g.* French Test Result: 05.97 % ## Training The script used for training can be found here
[ "# Wav2Vec2-Large-XLSR-53-Nepali \n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Nepali using the Common Voice, and OpenSLR ne. \n\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:", "#### Result \n\nPrediction: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ']\n\nReference: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ']", "## Evaluation\n\nThe model can be evaluated as follows on the {language} test data of Common Voice. # TODO: replace #TODO: replace language with your {language}, *e.g.* French\n\n\n\n\nTest Result: 05.97 %", "## Training\n\nThe script used for training can be found here" ]
[ "TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #ne #dataset-OpenSLR #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us \n", "# Wav2Vec2-Large-XLSR-53-Nepali \n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Nepali using the Common Voice, and OpenSLR ne. \n\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:", "#### Result \n\nPrediction: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ']\n\nReference: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ']", "## Evaluation\n\nThe model can be evaluated as follows on the {language} test data of Common Voice. # TODO: replace #TODO: replace language with your {language}, *e.g.* French\n\n\n\n\nTest Result: 05.97 %", "## Training\n\nThe script used for training can be found here" ]
[ 87, 68, 20, 65, 53, 11 ]
[ "passage: TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #ne #dataset-OpenSLR #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us \n# Wav2Vec2-Large-XLSR-53-Nepali \n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Nepali using the Common Voice, and OpenSLR ne. \n\nWhen using this model, make sure that your speech input is sampled at 16kHz.## Usage\n\nThe model can be used directly (without a language model) as follows:#### Result \n\nPrediction: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ']\n\nReference: ['पारानाको ब्राजिली राज्यमा रहेको राजधानी', 'देवराज जोशी त्रिभुवन विश्वविद्यालयबाट शिक्षाशास्त्रमा स्नातक हुनुहुन्छ']## Evaluation\n\nThe model can be evaluated as follows on the {language} test data of Common Voice. # TODO: replace #TODO: replace language with your {language}, *e.g.* French\n\n\n\n\nTest Result: 05.97 %## Training\n\nThe script used for training can be found here" ]
[ -0.12455910444259644, 0.11881386488676071, -0.0063506183214485645, -0.013370373286306858, 0.06634875386953354, -0.0642174631357193, 0.10430872440338135, 0.08425108343362808, 0.029825523495674133, 0.023526854813098907, 0.034375596791505814, 0.0845237672328949, 0.08786876499652863, 0.12016873061656952, -0.013083911500871181, -0.19167691469192505, 0.0765218734741211, 0.0287136472761631, 0.02974671497941017, 0.09323534369468689, 0.14290106296539307, -0.03640109673142433, 0.02430778741836548, 0.0436631515622139, -0.016149695962667465, 0.06484686583280563, 0.017104007303714752, -0.07989310473203659, 0.11105785518884659, 0.06076861917972565, 0.06965996325016022, 0.09672929346561432, -0.0046432288363575935, -0.21723809838294983, 0.02016323246061802, 0.008220315910875797, -0.02960589900612831, -0.03489513695240021, 0.05852728709578514, -0.011967809870839119, 0.13920313119888306, 0.04253268986940384, -0.045054398477077484, 0.08805352449417114, -0.09163475036621094, -0.09592270851135254, -0.014860406517982483, 0.023960981518030167, 0.08092962205410004, 0.15071018040180206, -0.0770961120724678, 0.1597219556570053, -0.07461846619844437, 0.07560282200574875, 0.07029148936271667, -0.11575700342655182, 0.007985454984009266, 0.10197879374027252, 0.05600915849208832, 0.05023695155978203, -0.10852878540754318, 0.011874525807797909, 0.02865014038980007, -0.022733375430107117, -0.09031253308057785, -0.08877858519554138, -0.09125674515962601, -0.005365752149373293, -0.10434870421886444, -0.002756671281531453, 0.2025112360715866, -0.005381112918257713, -0.0557607039809227, -0.06009845808148384, -0.021077420562505722, 0.010921050794422626, -0.05497516691684723, -0.10849513858556747, 0.021866891533136368, 0.07327745109796524, 0.05454972758889198, -0.12393858283758163, -0.07288720458745956, -0.05925637111067772, 0.012807423248887062, 0.02330867573618889, 0.05834069103002548, -0.025805044919252396, -0.08954007923603058, 0.015065050683915615, -0.12391351163387299, -0.04178445041179657, -0.07061704993247986, -0.035299401730298996, -0.05722050741314888, -0.018706688657402992, -0.06462562084197998, -0.11298879235982895, 0.10119584202766418, 0.04172331094741821, 0.0032212596852332354, 0.08560178428888321, -0.0988674983382225, 0.05025474354624748, -0.0018808384193107486, 0.12922333180904388, -0.09565093368291855, -0.08716553449630737, -0.0028523001819849014, 0.06687899678945541, -0.05520792305469513, 0.01893877610564232, 0.005204228218644857, -0.07124359160661697, 0.0276839267462492, 0.09146863967180252, 0.020670920610427856, 0.03912542387843132, -0.05909721180796623, -0.025853650644421577, 0.07422604411840439, -0.10625294595956802, -0.036746796220541, 0.01866483874619007, -0.013837726786732674, 0.1339382529258728, 0.06169735640287399, 0.050113651901483536, -0.09944435209035873, -0.09807497262954712, 0.004523451440036297, 0.006317696068435907, -0.07555124908685684, -0.015825843438506126, -0.020005933940410614, 0.004805581644177437, -0.01074330322444439, -0.03852345049381256, -0.14476945996284485, -0.0940624475479126, 0.01720752753317356, 0.00970930140465498, -0.030756842344999313, -0.1723259687423706, -0.027738040313124657, -0.05216497927904129, -0.031106848269701004, -0.008160661906003952, -0.02225419133901596, 0.054455969482660294, 0.02802744321525097, 0.03471178561449051, -0.03866269439458847, 0.07382494956254959, -0.047276854515075684, 0.000015372152120107785, -0.0329366959631443, 0.13390888273715973, -0.13491471111774445, -0.11437730491161346, -0.16276749968528748, -0.04718484729528427, -0.05575435981154442, 0.08045927435159683, 0.03640633821487427, 0.16103596985340118, -0.2717951536178589, -0.08112344145774841, 0.249333918094635, -0.12186485528945923, -0.03395708650350571, 0.1896199733018875, -0.004168759565800428, 0.14351871609687805, 0.11395709961652756, 0.11753873527050018, 0.07988450676202774, -0.16389065980911255, 0.012534819543361664, -0.014600020833313465, -0.05217335745692253, 0.01751457154750824, 0.07388852536678314, -0.07571513950824738, -0.057898253202438354, 0.03497331216931343, -0.08150447905063629, -0.016208456829190254, -0.02167711965739727, -0.05683985352516174, 0.009275812655687332, -0.03694844990968704, 0.0375206433236599, 0.031082291156053543, -0.004586850758641958, 0.01351968850940466, -0.06719924509525299, 0.024836242198944092, 0.11726352572441101, -0.09197350591421127, 0.06469614803791046, -0.11696667969226837, 0.08605408668518066, 0.02876845933496952, -0.006615151185542345, -0.09465721994638443, 0.12781977653503418, 0.01918732561171055, 0.03537129983305931, 0.08809935301542282, 0.15375059843063354, 0.021256694570183754, -0.042588092386722565, -0.07566159963607788, -0.023654265329241753, 0.02572428248822689, -0.043601144105196, -0.0007637420785613358, -0.07350342720746994, 0.06754592806100845, -0.029920488595962524, 0.07356486469507217, -0.1561150699853897, -0.015417871996760368, -0.04077434539794922, 0.029171815142035484, -0.00790038425475359, 0.0059881960041821, 0.005920237395912409, 0.037015803158283234, 0.027717651799321175, 0.008162682875990868, 0.005594999063760042, 0.018173282966017723, -0.05867745354771614, 0.1553269475698471, -0.11922802031040192, -0.11358778923749924, 0.09778906404972076, -0.13544662296772003, -0.03098667785525322, 0.04794565588235855, -0.021315742284059525, -0.007995082065463066, 0.02798442542552948, 0.047544632107019424, 0.24055804312229156, 0.00423077866435051, 0.10204054415225983, -0.08152706921100616, 0.022365575656294823, -0.043520502746105194, -0.0686834454536438, -0.027797849848866463, 0.05233341455459595, 0.05165738984942436, -0.06525690853595734, 0.07135649770498276, -0.07374483346939087, -0.10331880301237106, 0.23898756504058838, -0.011930407956242561, -0.09196247160434723, -0.010795467533171177, 0.04053588584065437, -0.02688491903245449, -0.020772425457835197, -0.1856606900691986, 0.02708353102207184, 0.03906416520476341, 0.009638041257858276, 0.11774027347564697, -0.036407895386219025, 0.02730984427034855, -0.006034239195287228, -0.10012519359588623, -0.07326217740774155, 0.12257508933544159, -0.014044356532394886, 0.032626207917928696, -0.10912622511386871, -0.014568480663001537, -0.025224296376109123, -0.023739619180560112, -0.12205653637647629, 0.12185348570346832, -0.10872437804937363, -0.22390641272068024, -0.17235547304153442, -0.007924801670014858, -0.03856811672449112, 0.038660526275634766, 0.07949168235063553, -0.0922057181596756, -0.06568200141191483, -0.018581151962280273, 0.09761087596416473, -0.03830743581056595, -0.08133422583341599, -0.050144195556640625, 0.03212207183241844, 0.03522976115345955, -0.14784860610961914, -0.00938507542014122, 0.014119086787104607, -0.10451120138168335, -0.005065284203737974, -0.09248092025518417, -0.026978464797139168, 0.06456850469112396, 0.03600127249956131, -0.013984824530780315, 0.01589088886976242, 0.21411311626434326, -0.11773031204938889, 0.03640745207667351, 0.13270415365695953, 0.01628050021827221, 0.013032875023782253, 0.1433386653661728, -0.029762884601950645, -0.07786943763494492, -0.019244663417339325, 0.07524436712265015, 0.030037814751267433, -0.2714739441871643, -0.09797988831996918, -0.06760501116514206, -0.08553025871515274, -0.06184477359056473, 0.024918990209698677, 0.08956439048051834, 0.028398960828781128, -0.06709665060043335, -0.0713348537683487, 0.04383876919746399, 0.018413355574011803, 0.10327423363924026, -0.040584858506917953, 0.07429700344800949, -0.03658904880285263, -0.033687107264995575, 0.03346351161599159, 0.044618815183639526, 0.12542109191417694, 0.09342414140701294, 0.08615846186876297, 0.11379076540470123, 0.0490739569067955, 0.01644209958612919, -0.006482305470854044, -0.02520747110247612, -0.008634379133582115, -0.009427153505384922, -0.010492924600839615, -0.0429447740316391, 0.04223257303237915, 0.19614362716674805, -0.09513118118047714, 0.0062971883453428745, 0.03289436548948288, 0.039729923009872437, 0.13422338664531708, 0.008594371378421783, -0.1803530603647232, -0.0012004985474050045, -0.004627089947462082, -0.0794549435377121, -0.07408439368009567, 0.027190668508410454, -0.008443200960755348, -0.0991479828953743, 0.055818840861320496, 0.03314988687634468, 0.04834989085793495, 0.008116855286061764, 0.01126168668270111, -0.1308612823486328, -0.028653576970100403, 0.007289953995496035, 0.08443965017795563, -0.2895289957523346, 0.29877975583076477, 0.011622280813753605, 0.07498149573802948, -0.031079549342393875, 0.019336102530360222, 0.004187942948192358, 0.057298727333545685, 0.16615544259548187, 0.004948670044541359, -0.029289042577147484, -0.03142526000738144, -0.038846202194690704, 0.028843073174357414, 0.04806479811668396, 0.04738825932145119, 0.07588443160057068, -0.01973310112953186, -0.014428095892071724, -0.012506709434092045, -0.008754083886742592, -0.19589343667030334, -0.0799688920378685, 0.0727333277463913, 0.1022339016199112, 0.12015870958566666, -0.019794434309005737, -0.053830694407224655, -0.10777058452367783, 0.12991835176944733, -0.07047706097364426, -0.036297813057899475, -0.05977710336446762, -0.0647469162940979, 0.16074922680854797, -0.10772574692964554, 0.05164181813597679, 0.052963417023420334, 0.06549125164747238, -0.05433545634150505, 0.013721797615289688, 0.06487584114074707, -0.08739108592271805, -0.031519196927547455, 0.013473114930093288, 0.16557204723358154, 0.1199878454208374, 0.04240056127309799, 0.04848363250494003, 0.001354703912511468, 0.02016827091574669, -0.06200478971004486, -0.040928591042757034, 0.04555132985115051, -0.050798770040273666, 0.029285810887813568, -0.04615071043372154, -0.20177224278450012, -0.11797410249710083, -0.037694547325372696, 0.143447607755661, 0.2312084287405014, 0.044892165809869766, 0.15686629712581635, 0.19062426686286926, -0.11205200850963593, -0.1837313175201416, -0.06853360682725906, 0.0771256536245346, 0.04030967131257057, -0.04409768059849739, -0.16480374336242676, -0.05642732232809067, -0.021072354167699814, -0.027718031778931618, -0.10687652975320816, -0.161274254322052, -0.11843690276145935, 0.18036329746246338, -0.024126842617988586, 0.15223433077335358, -0.06840097159147263, -0.028684211894869804, -0.042043574154376984, -0.019949954003095627, -0.019251007586717606, 0.04304107278585434, 0.08928974717855453, 0.01793309673666954, 0.045137953013181686, 0.055162202566862106, -0.02976827882230282, 0.11622938513755798, 0.024495016783475876, -0.009759082458913326, -0.007452012039721012, 0.015554802492260933, 0.030230915173888206, 0.02647136151790619, 0.1347648948431015, 0.02002357505261898, 0.027623655274510384, -0.06078703701496124, -0.10503605753183365, -0.0847954973578453, 0.07754345238208771, -0.018437262624502182, -0.005025454331189394, 0.036680836230516434, 0.0610358752310276, 0.05839259549975395, 0.02383541874587536, -0.06890048831701279, -0.16093365848064423, 0.018301820382475853, 0.16817574203014374, 0.1641702800989151, 0.04896429553627968, -0.024104632437229156, 0.04016398638486862, 0.003714604303240776, 0.08180925995111465, -0.023113567382097244, 0.06238066405057907, 0.027991902083158493, -0.005155471153557301, 0.1294572800397873, -0.01488612312823534, -0.08917680382728577, 0.020465992391109467, 0.040447626262903214, -0.0058806100860238075, -0.16274100542068481, -0.038711678236722946, -0.009104597382247448, -0.03103957138955593, -0.045525528490543365, 0.07046117633581161, -0.03651911020278931, -0.029205530881881714, -0.031611260026693344, 0.03159518539905548, -0.10995586216449738, 0.10174708068370819, 0.0150208855047822, 0.035271722823381424, -0.07129313796758652, 0.07145442813634872, 0.016913803294301033, -0.09986139833927155, 0.06665994971990585, -0.0281673651188612, -0.03460795432329178, -0.074893519282341, 0.04797634482383728, 0.10601986944675446, 0.006615060847252607, -0.08321423828601837, -0.07148809731006622, -0.0016626077704131603, 0.020954595878720284, 0.08236370235681534, 0.021236788481473923, 0.0006121272454038262, -0.01840893179178238, -0.029384322464466095, -0.06540472060441971, 0.01638091541826725, 0.1416342556476593, -0.028258906677365303, -0.010055119171738625, 0.054715171456336975, 0.012564099393785, 0.017336707562208176, -0.030586589127779007, -0.029276149347424507, -0.06517571955919266, 0.05181184411048889, -0.0821700245141983, 0.019278863444924355, -0.07644549012184143, -0.00801184494048357, 0.004453623201698065, -0.07571970671415329, -0.02606196328997612, 0.04218917340040207, -0.12287061661481857, 0.02741081453859806, 0.015386374667286873, 0.12002155184745789, -0.06698708981275558, 0.045205507427453995, 0.12181185930967331, -0.039535701274871826, 0.03603104501962662, 0.0877213180065155, -0.08574923872947693, 0.14087729156017303, -0.17317929863929749, -0.029335225000977516, 0.0033378126099705696, 0.05131232365965843, -0.013929126784205437, -0.12701931595802307, 0.03693310171365738, 0.07815789431333542, 0.07714412361383438, -0.056643299758434296, 0.05391557514667511, -0.06468129903078079, 0.020335912704467773, -0.06390166282653809, -0.06424710899591446, 0.008004795759916306, 0.05905640870332718, 0.04915577545762062, 0.025111904367804527, 0.12615683674812317, -0.10668592154979706, 0.029960209503769875, -0.08482697606086731, 0.04172947630286217, -0.030914196744561195, 0.010572751984000206, 0.001625214354135096, -0.1122027337551117, 0.06350896507501602, -0.06721372902393341, 0.14610496163368225, -0.06930305808782578, 0.08724178373813629, -0.0035669193603098392, -0.14997142553329468, -0.02643262967467308, 0.016409123316407204, 0.23390720784664154, 0.07500098645687103, 0.0390663743019104, -0.012887375429272652, -0.016804292798042297, 0.0101004121825099, 0.06797070056200027, 0.008030825294554234, 0.2410033494234085, -0.048911213874816895, 0.09505824744701385, 0.021788649260997772, -0.07703094184398651, 0.015085120685398579, -0.006788854952901602, -0.11956426501274109, -0.0272990595549345, -0.028731223195791245, 0.11718028783798218, 0.15638688206672668, -0.12159785628318787, 0.1479436606168747, 0.030536020174622536, -0.0936662033200264, -0.14448106288909912, -0.1431046575307846, -0.11472704261541367, -0.11380603164434433, 0.07180899381637573, -0.08828423917293549, 0.07636859267950058, -0.027765942737460136, 0.06251773983240128, -0.0789228081703186, 0.15173889696598053, -0.028252240270376205, -0.12747976183891296, 0.06807281076908112, -0.08782114088535309, 0.019880261272192, -0.03328467905521393, 0.0001982858666451648, 0.0617704913020134, 0.05478331074118614, 0.034169476479291916, 0.028219209983944893, -0.03898879885673523, 0.035029008984565735, -0.08805818855762482, -0.03080022521317005, -0.003541818354278803, -0.012711147777736187, 0.060896698385477066, 0.14488919079303741, 0.0458202064037323, -0.05989069864153862, 0.039828382432460785, 0.10056646913290024, -0.004658139776438475, -0.20392760634422302, -0.1537940502166748, 0.15718181431293488, 0.0816613957285881, 0.03413790464401245, -0.021136203780770302, -0.04906483739614487, -0.008801109157502651, 0.2440282702445984, 0.18310460448265076, 0.05910841003060341, 0.027266113087534904, 0.022869093343615532, 0.01961703784763813, 0.003684146562591195, 0.006643852684646845, 0.05576290190219879, 0.059884704649448395, -0.027508750557899475, -0.031746167689561844, -0.052824392914772034, -0.010947402566671371, -0.021876726299524307, 0.16977672278881073, -0.0073158289305865765, -0.12847094237804413, 0.016574257984757423, 0.12788473069667816, -0.1563890129327774, -0.11718537658452988, -0.15418840944766998, -0.05168654024600983, -0.1145024225115776, -0.05030302703380585, -0.04708944261074066, 0.07909107208251953, 0.02053125388920307, 0.0037794734816998243, -0.04899876192212105, 0.1522340476512909, 0.026943834498524666, -0.0510127879679203, -0.04747450351715088, 0.04600636288523674, -0.0491509884595871, 0.09322486072778702, 0.040001340210437775, 0.08707458525896072, 0.039901528507471085, 0.04253505542874336, 0.03053680807352066, 0.19338077306747437, 0.04958655312657356, -0.04670333117246628, 0.07503854483366013, 0.1339213252067566, -0.02227424457669258, 0.09239233285188675, 0.030297700315713882, -0.05796479061245918, 0.09676475077867508, -0.10443470627069473, -0.06032055616378784, -0.10892169177532196, 0.14717315137386322, -0.1355237364768982, 0.09662231802940369, 0.07564979791641235, -0.017877252772450447, -0.0015358391683548689, -0.0842004045844078, 0.09693550318479538, 0.004665891639888287, 0.03173131123185158, -0.08202502131462097, -0.22655165195465088, 0.026094473898410797, -0.033866386860609055, 0.059283554553985596, -0.19029980897903442, 0.022994522005319595, 0.006805791053920984, -0.03611867502331734, 0.007117721252143383, 0.09299538284540176, 0.030240150168538094, -0.015027444809675217, 0.012992977164685726, -0.09056230634450912, 0.05765097960829735, 0.10216966271400452, -0.13769790530204773, -0.07257520407438278 ]
null
null
transformers
# Wav2Vec2-Large-XLSR-53-Punjabi Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on Punjabi using the [Common Voice](https://huggingface.co/datasets/common_voice) When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ```python import torch import torchaudio from datasets import load_dataset from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor test_dataset = load_dataset("common_voice", "pa-IN", split="test") processor = Wav2Vec2Processor.from_pretrained("gagan3012/wav2vec2-xlsr-punjabi") model = Wav2Vec2ForCTC.from_pretrained("gagan3012/wav2vec2-xlsr-punjabi") resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): \\\\tspeech_array, sampling_rate = torchaudio.load(batch["path"]) \\\\tbatch["speech"] = resampler(speech_array).squeeze().numpy() \\\\treturn batch test_dataset = test_dataset.map(speech_file_to_array_fn) inputs = processor(test_dataset["speech"][:2], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): \\\\tlogits = model(inputs.input_values, attention_mask=inputs.attention_mask).logits predicted_ids = torch.argmax(logits, dim=-1) print("Prediction:", processor.batch_decode(predicted_ids)) print("Reference:", test_dataset["sentence"][:2]) ``` #### Results: Prediction: ['ਹਵਾ ਲਾਤ ਵਿੱਚ ਪੰਦ ਛੇ ਇਖਲਾਟਕੀ ਮੁਜਰਮ ਸਨ', 'ਮੈ ਇ ਹਾ ਪੈਸੇ ਲੇਹੜ ਨਹੀਂ ਸੀ ਚੌਨਾ'] Reference: ['ਹਵਾਲਾਤ ਵਿੱਚ ਪੰਜ ਛੇ ਇਖ਼ਲਾਕੀ ਮੁਜਰਮ ਸਨ', 'ਮੈਂ ਇਹ ਪੈਸੇ ਲੈਣੇ ਨਹੀਂ ਸੀ ਚਾਹੁੰਦਾ'] ## Evaluation The model can be evaluated as follows on the {language} test data of Common Voice. # TODO: replace #TODO: replace language with your {language}, *e.g.* French ```python import torch import torchaudio from datasets import load_dataset, load_metric from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor import re test_dataset = load_dataset("common_voice", "pa-IN", split="test") #TODO: replace {lang_id} in your language code here. Make sure the code is one of the *ISO codes* of [this](https://huggingface.co/languages) site. wer = load_metric("wer") processor = Wav2Vec2Processor.from_pretrained("gagan3012/wav2vec2-xlsr-punjabi") model = Wav2Vec2ForCTC.from_pretrained("gagan3012/wav2vec2-xlsr-punjabi") model.to("cuda") chars_to_ignore_regex = '[\\\\\\\\\\\\\\\\,\\\\\\\\\\\\\\\\?\\\\\\\\\\\\\\\\.\\\\\\\\\\\\\\\\!\\\\\\\\\\\\\\\\-\\\\\\\\\\\\\\\\;\\\\\\\\\\\\\\\\:\\\\\\\\\\\\\\\\"\\\\\\\\\\\\\\\\“]' # TODO: adapt this list to include all special characters you removed from the data resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): \\\\\\\\tbatch["sentence"] = re.sub(chars_to_ignore_regex, '', batch["sentence"]).lower() \\\\\\\\tspeech_array, sampling_rate = torchaudio.load(batch["path"]) \\\\\\\\tbatch["speech"] = resampler(speech_array).squeeze().numpy() \\\\\\\\treturn batch test_dataset = test_dataset.map(speech_file_to_array_fn) # Preprocessing the datasets. # We need to read the aduio files as arrays def evaluate(batch): \\\\\\\\tinputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True) \\\\\\\\twith torch.no_grad(): \\\\\\\\t\\\\\\\\tlogits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits \\\\\\\\tpred_ids = torch.argmax(logits, dim=-1) \\\\\\\\tbatch["pred_strings"] = processor.batch_decode(pred_ids) \\\\\\\\treturn batch result = test_dataset.map(evaluate, batched=True, batch_size=8) print("WER: {:2f}".format(100 * wer.compute(predictions=result["pred_strings"], references=result["sentence"]))) ``` **Test Result**: 58.05 % ## Training The script used for training can be found [here](https://colab.research.google.com/drive/1A7Y20c1QkSHfdOmLXPMiOEpwlTjDZ7m5?usp=sharing)
{"language": "pa-IN", "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech", "xlsr-fine-tuning-week"], "datasets": ["common_voice"], "metrics": ["wer"], "model-index": [{"name": "wav2vec2-xlsr-punjabi", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Speech Recognition"}, "dataset": {"name": "Common Voice pa", "type": "common_voice", "args": "pa-IN"}, "metrics": [{"type": "wer", "value": 58.06, "name": "Test WER"}]}]}]}
automatic-speech-recognition
gagan3012/wav2vec2-xlsr-punjabi
[ "transformers", "pytorch", "jax", "wav2vec2", "automatic-speech-recognition", "audio", "speech", "xlsr-fine-tuning-week", "dataset:common_voice", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "pa-IN" ]
TAGS #transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us
# Wav2Vec2-Large-XLSR-53-Punjabi Fine-tuned facebook/wav2vec2-large-xlsr-53 on Punjabi using the Common Voice When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: #### Results: Prediction: ['ਹਵਾ ਲਾਤ ਵਿੱਚ ਪੰਦ ਛੇ ਇਖਲਾਟਕੀ ਮੁਜਰਮ ਸਨ', 'ਮੈ ਇ ਹਾ ਪੈਸੇ ਲੇਹੜ ਨਹੀਂ ਸੀ ਚੌਨਾ'] Reference: ['ਹਵਾਲਾਤ ਵਿੱਚ ਪੰਜ ਛੇ ਇਖ਼ਲਾਕੀ ਮੁਜਰਮ ਸਨ', 'ਮੈਂ ਇਹ ਪੈਸੇ ਲੈਣੇ ਨਹੀਂ ਸੀ ਚਾਹੁੰਦਾ'] ## Evaluation The model can be evaluated as follows on the {language} test data of Common Voice. # TODO: replace #TODO: replace language with your {language}, *e.g.* French Test Result: 58.05 % ## Training The script used for training can be found here
[ "# Wav2Vec2-Large-XLSR-53-Punjabi \n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Punjabi using the Common Voice\n\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:", "#### Results: \n\nPrediction: ['ਹਵਾ ਲਾਤ ਵਿੱਚ ਪੰਦ ਛੇ ਇਖਲਾਟਕੀ ਮੁਜਰਮ ਸਨ', 'ਮੈ ਇ ਹਾ ਪੈਸੇ ਲੇਹੜ ਨਹੀਂ ਸੀ ਚੌਨਾ']\n\nReference: ['ਹਵਾਲਾਤ ਵਿੱਚ ਪੰਜ ਛੇ ਇਖ਼ਲਾਕੀ ਮੁਜਰਮ ਸਨ', 'ਮੈਂ ਇਹ ਪੈਸੇ ਲੈਣੇ ਨਹੀਂ ਸੀ ਚਾਹੁੰਦਾ']", "## Evaluation\n\nThe model can be evaluated as follows on the {language} test data of Common Voice. # TODO: replace #TODO: replace language with your {language}, *e.g.* French\n\n\n\n\nTest Result: 58.05 %", "## Training\n\nThe script used for training can be found here" ]
[ "TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us \n", "# Wav2Vec2-Large-XLSR-53-Punjabi \n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Punjabi using the Common Voice\n\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:", "#### Results: \n\nPrediction: ['ਹਵਾ ਲਾਤ ਵਿੱਚ ਪੰਦ ਛੇ ਇਖਲਾਟਕੀ ਮੁਜਰਮ ਸਨ', 'ਮੈ ਇ ਹਾ ਪੈਸੇ ਲੇਹੜ ਨਹੀਂ ਸੀ ਚੌਨਾ']\n\nReference: ['ਹਵਾਲਾਤ ਵਿੱਚ ਪੰਜ ਛੇ ਇਖ਼ਲਾਕੀ ਮੁਜਰਮ ਸਨ', 'ਮੈਂ ਇਹ ਪੈਸੇ ਲੈਣੇ ਨਹੀਂ ਸੀ ਚਾਹੁੰਦਾ']", "## Evaluation\n\nThe model can be evaluated as follows on the {language} test data of Common Voice. # TODO: replace #TODO: replace language with your {language}, *e.g.* French\n\n\n\n\nTest Result: 58.05 %", "## Training\n\nThe script used for training can be found here" ]
[ 78, 61, 20, 78, 54, 11 ]
[ "passage: TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us \n# Wav2Vec2-Large-XLSR-53-Punjabi \n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Punjabi using the Common Voice\n\nWhen using this model, make sure that your speech input is sampled at 16kHz.## Usage\n\nThe model can be used directly (without a language model) as follows:#### Results: \n\nPrediction: ['ਹਵਾ ਲਾਤ ਵਿੱਚ ਪੰਦ ਛੇ ਇਖਲਾਟਕੀ ਮੁਜਰਮ ਸਨ', 'ਮੈ ਇ ਹਾ ਪੈਸੇ ਲੇਹੜ ਨਹੀਂ ਸੀ ਚੌਨਾ']\n\nReference: ['ਹਵਾਲਾਤ ਵਿੱਚ ਪੰਜ ਛੇ ਇਖ਼ਲਾਕੀ ਮੁਜਰਮ ਸਨ', 'ਮੈਂ ਇਹ ਪੈਸੇ ਲੈਣੇ ਨਹੀਂ ਸੀ ਚਾਹੁੰਦਾ']## Evaluation\n\nThe model can be evaluated as follows on the {language} test data of Common Voice. # TODO: replace #TODO: replace language with your {language}, *e.g.* French\n\n\n\n\nTest Result: 58.05 %## Training\n\nThe script used for training can be found here" ]
[ -0.10072465240955353, 0.0811477079987526, -0.007928336039185524, -0.006018593907356262, 0.06667797267436981, -0.03607632964849472, 0.1362556666135788, 0.11470868438482285, 0.06403021514415741, 0.05895455926656723, 0.03567960858345032, 0.1054590567946434, 0.10088484734296799, 0.14775945246219635, -0.009895138442516327, -0.18387189507484436, 0.06648242473602295, 0.010864593088626862, 0.020019084215164185, 0.09690520167350769, 0.1503177434206009, -0.04221450164914131, 0.023754796013236046, 0.014707807451486588, -0.05913906544446945, 0.07501045614480972, -0.0015069945948198438, -0.06706828624010086, 0.1144135370850563, 0.08433296531438828, 0.07140009105205536, 0.08709391206502914, 0.02192901074886322, -0.2596508860588074, 0.019876867532730103, 0.010153763927519321, -0.051027849316596985, -0.010918700136244297, 0.08949937671422958, -0.01904023066163063, 0.052840519696474075, -0.005643466487526894, -0.01846349984407425, 0.0752609446644783, -0.09301797300577164, -0.11958683282136917, -0.035891540348529816, 0.030804572626948357, 0.09553808718919754, 0.13295923173427582, -0.08132906258106232, 0.1249159425497055, -0.07697460800409317, 0.09087511897087097, 0.15212571620941162, -0.17476427555084229, 0.01998131349682808, 0.09661086648702621, 0.07205694168806076, 0.018136097118258476, -0.09753021597862244, -0.012785904109477997, 0.003263008315116167, -0.022702926769852638, -0.07023399323225021, -0.06428781896829605, -0.03236616775393486, -0.01165065634995699, -0.10344643145799637, -0.030866194516420364, 0.1839689016342163, 0.02945731207728386, -0.027579382061958313, -0.09019860625267029, -0.05865451321005821, -0.01050275843590498, -0.05754737928509712, -0.1125086098909378, 0.022287121042609215, 0.05230337753891945, 0.06582764536142349, -0.0747632160782814, -0.08253350108861923, -0.04948686435818672, 0.01641000434756279, 0.056376952677965164, 0.07393403351306915, -0.02602837048470974, -0.10506834089756012, 0.026312127709388733, -0.11272866278886795, -0.06700630486011505, -0.055735714733600616, -0.015303713269531727, -0.0975913256406784, -0.00744723342359066, -0.08319161832332611, -0.09704083204269409, 0.10464312136173248, 0.11666718125343323, 0.025756705552339554, 0.11059444397687912, -0.08316755294799805, 0.04555964842438698, 0.003865330247208476, 0.10677774995565414, -0.07949094474315643, -0.09940911084413528, -0.006272200029343367, 0.10782944411039352, -0.044817667454481125, 0.001702816691249609, -0.033146560192108154, -0.023908058181405067, 0.061944060027599335, 0.08364680409431458, 0.009165897034108639, 0.03386214002966881, -0.06920658051967621, -0.019000252708792686, 0.08365952968597412, -0.1591508686542511, -0.013053137809038162, 0.05774335190653801, -0.014343745075166225, 0.09536521881818771, 0.044918350875377655, 0.026696229353547096, -0.10397190600633621, -0.06963784247636795, -0.0020739322062581778, 0.011572496965527534, -0.06708429008722305, -0.04893180727958679, -0.012474337592720985, 0.056461550295352936, -0.019443342462182045, -0.04578709974884987, -0.14776082336902618, -0.07432428002357483, 0.010310661047697067, -0.03528221324086189, -0.031218988820910454, -0.1472209244966507, -0.042682796716690063, -0.028256656602025032, -0.024051113054156303, 0.04396138712763786, -0.02881498821079731, 0.06424946337938309, 0.041807662695646286, 0.0412072092294693, 0.03097594901919365, 0.07118108123540878, -0.04249683395028114, -0.020985661074519157, -0.05024290829896927, 0.13953019678592682, -0.11526188254356384, -0.1202392578125, -0.16219140589237213, -0.052837420254945755, -0.03666536137461662, 0.06614508479833603, 0.012815063819289207, 0.15680557489395142, -0.25511234998703003, -0.08047043532133102, 0.25491419434547424, -0.11947087198495865, -0.006504856050014496, 0.15186722576618195, -0.010512237437069416, 0.08287755399942398, 0.07958108931779861, 0.13234376907348633, 0.05287546291947365, -0.13411030173301697, 0.015367085114121437, -0.00021619067410938442, -0.019104335457086563, 0.06978752464056015, 0.07983462512493134, -0.07141049951314926, -0.018971022218465805, 0.008886370807886124, -0.07366412878036499, -0.03404530882835388, -0.02905389480292797, -0.051301080733537674, 0.019551094621419907, -0.013268545269966125, 0.06782425940036774, 0.00548573536798358, -0.003697357838973403, 0.018443116918206215, -0.10411771386861801, 0.009880252182483673, 0.11418884992599487, -0.05943869799375534, 0.051899004727602005, -0.09782776236534119, 0.10409722477197647, 0.01583591103553772, 0.005599562078714371, -0.13169977068901062, 0.04631786420941353, 0.028845105320215225, -0.01567474938929081, 0.0576898455619812, 0.15890370309352875, 0.025875546038150787, -0.006580604240298271, -0.05165708810091019, -0.03822809085249901, -0.016620371490716934, -0.029757387936115265, -0.02600567415356636, -0.10878683626651764, 0.005620191805064678, -0.05930408462882042, 0.07032721489667892, -0.13732179999351501, -0.011981461197137833, 0.008918548934161663, 0.07654132694005966, -0.022281112149357796, -0.002502838848158717, -0.0036951221991330385, 0.01502455398440361, 0.02098655328154564, 0.006933483760803938, 0.006136753596365452, -0.007002864498645067, -0.054111648350954056, 0.11443599313497543, -0.16616308689117432, -0.08796152472496033, 0.11107950657606125, -0.10478640347719193, -0.04315682500600815, 0.02574061043560505, -0.01873837411403656, -0.011945294216275215, 0.019859066233038902, -0.030219096690416336, 0.19183579087257385, 0.033569034188985825, 0.12384070456027985, -0.06405945867300034, -0.01838529482483864, -0.030480004847049713, -0.049831073731184006, -0.016756737604737282, 0.06040370836853981, 0.029393181204795837, -0.060952991247177124, 0.09828819334506989, -0.035248398780822754, -0.029853196814656258, 0.19271808862686157, 0.010427328757941723, -0.0903400182723999, -0.024556241929531097, 0.05563736706972122, -0.04095377027988434, -0.00752888061106205, -0.1498362272977829, 0.009562039747834206, 0.0382440909743309, 0.028622297570109367, 0.07522419840097427, -0.06049666181206703, 0.027441630139946938, -0.006259584333747625, -0.10651097446680069, -0.03774719312787056, 0.09532126039266586, 0.02041276916861534, 0.06190579757094383, -0.07576912641525269, -0.03312048688530922, 0.009301002137362957, -0.048667944967746735, -0.12873411178588867, 0.15419219434261322, -0.10992369800806046, -0.22526411712169647, -0.15135705471038818, -0.04567015916109085, -0.02218322642147541, 0.03110123611986637, 0.10153063386678696, -0.11726842075586319, -0.0696277767419815, -0.009503191336989403, 0.0961674302816391, -0.031077681109309196, -0.08151507377624512, -0.04062133654952049, -0.018854010850191116, 0.06067748740315437, -0.1488579511642456, -0.02201199159026146, 0.013123765587806702, -0.09887616336345673, 0.020662855356931686, -0.05024576559662819, -0.03243856504559517, 0.06784563511610031, 0.039145201444625854, 0.0010155567433685064, -0.010935148224234581, 0.21892043948173523, -0.09353531897068024, 0.039592843502759933, 0.13554227352142334, -0.005109976977109909, 0.042093757539987564, 0.15188992023468018, -0.007019143085926771, -0.09814861416816711, -0.031157327815890312, 0.08786541223526001, 0.03590376302599907, -0.27454498410224915, -0.08379639685153961, -0.051849015057086945, -0.12242184579372406, -0.018647972494363785, 0.05234035477042198, 0.062173888087272644, 0.0021819546818733215, -0.06604986637830734, -0.06168830767273903, 0.06662049889564514, 0.04107103496789932, 0.12061312049627304, -0.033630941063165665, 0.09605445712804794, -0.00697771180421114, -0.020322147756814957, 0.02989603579044342, 0.051163140684366226, 0.15337078273296356, 0.10236761718988419, 0.10487687587738037, 0.10115636140108109, 0.05638618767261505, -0.0014903182163834572, 0.008516108617186546, 0.0054636611603200436, 0.017018819227814674, -0.021374473348259926, -0.016105586662888527, -0.016869017854332924, 0.04603830352425575, 0.1332928091287613, -0.10564818978309631, -0.021255183964967728, 0.012843488715589046, 0.06136263534426689, 0.14890000224113464, 0.02158844657242298, -0.19318340718746185, 0.024716246873140335, -0.007496440317481756, -0.11080151051282883, -0.05679831653833389, 0.017433324828743935, -0.02459881082177162, -0.10923445969820023, 0.06293708086013794, -0.005857742391526699, 0.048291414976119995, 0.007968773134052753, 0.01671021804213524, -0.06841936707496643, -0.02501453086733818, 0.027583520859479904, 0.09811151772737503, -0.3068525195121765, 0.2819809913635254, 0.00964196678251028, 0.05978722125291824, -0.03606097027659416, 0.04922652617096901, 0.0008174047688953578, 0.016244348138570786, 0.17315272986888885, 0.00008436525968136266, -0.02362477220594883, -0.06600779294967651, -0.0463312566280365, 0.028964977711439133, 0.07569605112075806, -0.002444078214466572, 0.08152678608894348, -0.024245604872703552, -0.02539052814245224, -0.008279792033135891, 0.03222236409783363, -0.14724679291248322, -0.09845516085624695, 0.0973581001162529, 0.019099973142147064, 0.10905461758375168, -0.037818778306245804, -0.06481260806322098, -0.1118067055940628, 0.11700725555419922, -0.10241823643445969, -0.05316000431776047, -0.07509739696979523, 0.015387223102152348, 0.1292363852262497, -0.10618927329778671, 0.05555776506662369, 0.04244436323642731, 0.07648614794015884, -0.05164649337530136, -0.02832242101430893, 0.07646293938159943, -0.08322667330503464, -0.09835126250982285, -0.0019317454425618052, 0.18240976333618164, 0.09428969770669937, 0.03090807795524597, 0.040760837495326996, 0.01327759400010109, 0.02456408180296421, -0.05673472210764885, -0.013156038708984852, 0.03040521778166294, -0.03721880540251732, 0.02234143763780594, -0.04772118479013443, -0.20765621960163116, -0.11860372871160507, -0.01772802695631981, 0.13431507349014282, 0.24474510550498962, 0.018304532393813133, 0.14430665969848633, 0.1981651335954666, -0.12498103827238083, -0.16593784093856812, -0.11118489503860474, 0.08302865922451019, 0.02242852933704853, -0.014483348466455936, -0.17730861902236938, -0.020334521308541298, -0.004022683948278427, -0.022301435470581055, -0.11779424548149109, -0.2355482280254364, -0.1286781132221222, 0.14816299080848694, -0.01816864125430584, 0.17394742369651794, -0.11687743663787842, -0.0580616295337677, -0.015619506128132343, -0.042071808129549026, -0.00381585955619812, 0.05316830053925514, 0.09601698070764542, 0.013970743864774704, 0.07434225082397461, 0.044900454580783844, -0.025312794372439384, 0.12301107496023178, 0.04914555326104164, -0.003231292823329568, -0.020296407863497734, -0.009529989212751389, 0.043957628309726715, 0.015798743814229965, 0.10858509689569473, -0.027420463040471077, 0.014385904185473919, -0.10301721096038818, -0.07736744731664658, -0.10003648698329926, 0.06624598056077957, -0.0518096387386322, 0.005884097423404455, 0.02819935604929924, 0.026302559301257133, 0.08372388780117035, 0.010452352464199066, -0.04691440612077713, -0.11028434336185455, 0.04310237616300583, 0.1960226446390152, 0.11389417201280594, 0.07098901271820068, -0.11440149694681168, 0.01749289035797119, -0.004494460765272379, 0.0806499794125557, -0.05214459076523781, 0.055868711322546005, 0.04832456633448601, -0.008254892192780972, 0.15629839897155762, 0.006694911979138851, -0.09880604594945908, 0.007858345285058022, 0.03603416308760643, -0.02639816887676716, -0.19230708479881287, -0.006957334000617266, 0.017609233036637306, -0.04041675478219986, -0.08112278580665588, 0.09513454884290695, -0.02014106512069702, -0.06091991439461708, -0.026842275634407997, 0.05317287519574165, -0.09045775234699249, 0.1319788098335266, 0.0025781928561627865, 0.08135606348514557, -0.08609534800052643, 0.07008865475654602, 0.030035922303795815, -0.09414592385292053, 0.0780194103717804, 0.022101106122136116, -0.04785096272826195, -0.09405425935983658, 0.027275418862700462, 0.14789217710494995, -0.000853695732075721, -0.055741533637046814, -0.07990293204784393, -0.032297052443027496, 0.033170029520988464, 0.09644077718257904, 0.01090185809880495, -0.0030744418036192656, 0.004674786236137152, -0.02604776807129383, -0.03438575565814972, 0.06052043288946152, 0.12995783984661102, -0.016181839630007744, -0.014507441781461239, 0.07101926952600479, 0.01151933241635561, 0.02624056115746498, -0.023310646414756775, -0.0042455620132386684, -0.0847257450222969, 0.019978230819106102, -0.11192499846220016, 0.023304235190153122, -0.0598258338868618, -0.013308783993124962, 0.00012813096691388637, -0.049355048686265945, 0.014375395141541958, 0.04750339314341545, -0.11819043010473251, 0.001020635711029172, 0.005607777275145054, 0.10919924080371857, -0.11599962413311005, 0.021197594702243805, 0.08145564794540405, -0.06010500714182854, 0.060194116085767746, 0.08807548880577087, -0.07836556434631348, 0.1323234587907791, -0.11508640646934509, -0.04191550984978676, -0.0005274987779557705, 0.05774277448654175, -0.007064280100166798, -0.143591970205307, 0.03745916485786438, 0.04151071235537529, 0.04533909633755684, -0.04384666308760643, 0.10221539437770844, -0.10194236785173416, 0.007434216793626547, -0.0812973603606224, -0.0680226981639862, -0.012508846819400787, 0.06337038427591324, 0.047481145709753036, 0.029526133090257645, 0.1316843330860138, -0.11032305657863617, 0.053977515548467636, -0.12485180795192719, 0.027170509099960327, -0.007572015281766653, -0.00598463648930192, -0.003341405186802149, -0.10356613993644714, 0.07023356854915619, -0.08543411642313004, 0.14013716578483582, -0.06345383077859879, 0.09128846228122711, 0.024885695427656174, -0.12183612585067749, -0.045028336346149445, 0.01931064762175083, 0.15333174169063568, 0.04225870966911316, 0.041977886110544205, -0.037434112280607224, -0.014515426009893417, 0.001383059425279498, 0.024143187329173088, 0.047680534422397614, 0.24406011402606964, -0.007429202552884817, 0.08060817420482635, 0.04628578945994377, -0.1278674155473709, -0.04903586581349373, 0.02152443490922451, -0.10504516959190369, 0.01405570562928915, -0.04505153000354767, 0.11023350805044174, 0.16413727402687073, -0.1081569641828537, 0.15945099294185638, -0.007322953548282385, -0.0721159502863884, -0.12305581569671631, -0.11973289400339127, -0.1083475649356842, -0.1066787987947464, 0.06000179424881935, -0.09182150661945343, 0.09073150902986526, 0.02772696688771248, 0.05119609460234642, -0.035959068685770035, 0.14374376833438873, -0.010777614079415798, -0.1156216710805893, 0.03891918808221817, -0.07235640287399292, 0.023368773981928825, 0.003073371946811676, -0.00219031423330307, 0.07962244004011154, 0.05063032731413841, 0.033776264637708664, 0.048910390585660934, -0.008851633407175541, 0.035968758165836334, -0.11906229704618454, -0.05393444746732712, 0.008179683238267899, -0.0015311399474740028, 0.031270790845155716, 0.1641078144311905, 0.04399801418185234, -0.055667322129011154, 0.015356862917542458, 0.09386304765939713, -0.011045651510357857, -0.17615030705928802, -0.13762390613555908, 0.1780657023191452, 0.07927985489368439, 0.0530579499900341, -0.012838788330554962, -0.053946685045957565, -0.010916746221482754, 0.20441938936710358, 0.1678614467382431, 0.05611640587449074, 0.027878720313310623, 0.05788511782884598, 0.01809368096292019, 0.027672655880451202, -0.010946298018097878, 0.0722632184624672, 0.07781533896923065, -0.04502924904227257, -0.0010805808706209064, -0.045460015535354614, -0.02986113913357258, -0.03167854994535446, 0.16211998462677002, -0.007278012577444315, -0.09232216328382492, 0.0029792117420583963, 0.09344444423913956, -0.06686791032552719, -0.1539369374513626, -0.07689078897237778, -0.07172758132219315, -0.10761687904596329, -0.042410723865032196, -0.012189815752208233, 0.07555363327264786, 0.048229388892650604, 0.01835019886493683, -0.05302845686674118, 0.17918415367603302, 0.013671435415744781, -0.05799856409430504, -0.0610833577811718, 0.050820738077163696, -0.10934986174106598, 0.08483725041151047, 0.04420952871441841, 0.12109117209911346, 0.05416790395975113, 0.04516309127211571, 0.006899907719343901, 0.16760243475437164, 0.061000362038612366, -0.02192353829741478, 0.07518719881772995, 0.14790569245815277, -0.01310072559863329, 0.08927654474973679, 0.01890186406672001, -0.08075112104415894, 0.06527423113584518, -0.06761342287063599, -0.08122394979000092, -0.08761346340179443, 0.14026325941085815, -0.11107795685529709, 0.07762867212295532, 0.12027152627706528, -0.01149834506213665, 0.0012389780022203922, -0.08792039752006531, 0.07451915740966797, 0.0112583814188838, 0.03332196921110153, -0.0853290855884552, -0.23306752741336823, 0.051665276288986206, -0.0362197570502758, 0.0712009146809578, -0.1615045964717865, -0.023566175252199173, 0.006824177224189043, -0.018101921305060387, -0.03286812826991081, 0.11576689034700394, 0.04489102587103844, -0.008857602253556252, -0.0005223507178016007, -0.10688862204551697, 0.02964855171740055, 0.1335882544517517, -0.16446217894554138, -0.057716820389032364 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xls-r-300m-hi This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the MOZILLA-FOUNDATION/COMMON_VOICE_8_0 - HI dataset. It achieves the following results on the evaluation set: - Loss: 0.7522 - Wer: 1.0091 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 7.5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 50.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | 5.0417 | 2.59 | 500 | 5.1484 | 1.0 | | 3.3722 | 5.18 | 1000 | 3.3380 | 1.0001 | | 1.9752 | 7.77 | 1500 | 1.3910 | 1.0074 | | 1.5868 | 10.36 | 2000 | 1.0298 | 1.0084 | | 1.4413 | 12.95 | 2500 | 0.9313 | 1.0175 | | 1.3296 | 15.54 | 3000 | 0.8966 | 1.0194 | | 1.2746 | 18.13 | 3500 | 0.8875 | 1.0097 | | 1.2147 | 20.73 | 4000 | 0.8746 | 1.0089 | | 1.1774 | 23.32 | 4500 | 0.8383 | 1.0198 | | 1.129 | 25.91 | 5000 | 0.7848 | 1.0167 | | 1.0995 | 28.5 | 5500 | 0.7992 | 1.0210 | | 1.0665 | 31.09 | 6000 | 0.7878 | 1.0107 | | 1.0321 | 33.68 | 6500 | 0.7653 | 1.0082 | | 1.0068 | 36.27 | 7000 | 0.7635 | 1.0065 | | 0.9916 | 38.86 | 7500 | 0.7728 | 1.0090 | | 0.9735 | 41.45 | 8000 | 0.7688 | 1.0070 | | 0.9745 | 44.04 | 8500 | 0.7455 | 1.0097 | | 0.9677 | 46.63 | 9000 | 0.7605 | 1.0099 | | 0.9313 | 49.22 | 9500 | 0.7527 | 1.0097 | ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2.dev0 - Tokenizers 0.11.0
{"language": ["hi"], "license": "apache-2.0", "tags": ["automatic-speech-recognition", "mozilla-foundation/common_voice_8_0", "generated_from_trainer"], "datasets": ["common_voice"], "model-index": [{"name": "xls-r-300m-hi", "results": []}]}
automatic-speech-recognition
gagan3012/xls-r-300m-hi
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "mozilla-foundation/common_voice_8_0", "generated_from_trainer", "hi", "dataset:common_voice", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "hi" ]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #mozilla-foundation/common_voice_8_0 #generated_from_trainer #hi #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us
xls-r-300m-hi ============= This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the MOZILLA-FOUNDATION/COMMON\_VOICE\_8\_0 - HI dataset. It achieves the following results on the evaluation set: * Loss: 0.7522 * Wer: 1.0091 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 7.5e-05 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * gradient\_accumulation\_steps: 4 * total\_train\_batch\_size: 32 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 2000 * num\_epochs: 50.0 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.17.0.dev0 * Pytorch 1.10.2+cu102 * Datasets 1.18.2.dev0 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 7.5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 2000\n* num\\_epochs: 50.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.2.dev0\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #mozilla-foundation/common_voice_8_0 #generated_from_trainer #hi #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 7.5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 2000\n* num\\_epochs: 50.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.2.dev0\n* Tokenizers 0.11.0" ]
[ 79, 160, 4, 39 ]
[ "passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #mozilla-foundation/common_voice_8_0 #generated_from_trainer #hi #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 7.5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 2000\n* num\\_epochs: 50.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.2.dev0\n* Tokenizers 0.11.0" ]
[ -0.12411151081323624, 0.1543121188879013, -0.003862259676679969, 0.024612732231616974, 0.10550402104854584, 0.008292148821055889, 0.09603679180145264, 0.14901535212993622, -0.06877940148115158, 0.12423531711101532, 0.09662747383117676, 0.0918274000287056, 0.10007340461015701, 0.1463412493467331, -0.018551157787442207, -0.26886600255966187, 0.036534469574689865, -0.03381148725748062, -0.08796457201242447, 0.09958140552043915, 0.08878125995397568, -0.10730811208486557, 0.0356898307800293, 0.01115917693823576, -0.0938040241599083, -0.017349576577544212, -0.039995480328798294, -0.061661962419748306, 0.08431363105773926, 0.05061886087059975, 0.03608431667089462, 0.029618944972753525, 0.07193496823310852, -0.2787189185619354, 0.009219200350344181, 0.06166490912437439, 0.03674368932843208, 0.054542962461709976, 0.10252958536148071, -0.007064370904117823, 0.10731688886880875, -0.08091923594474792, 0.03497331961989403, 0.05339924618601799, -0.08033770322799683, -0.27000707387924194, -0.10628578066825867, 0.01376500353217125, 0.1293608397245407, 0.09257948398590088, -0.040931276977062225, 0.0493798591196537, -0.04688390716910362, 0.07801788300275803, 0.2330797016620636, -0.23173175752162933, -0.06252162903547287, -0.018151545897126198, 0.038637422025203705, 0.03356701135635376, -0.10344024747610092, 0.008657233789563179, 0.02945495955646038, 0.0095041673630476, 0.07339884340763092, 0.018509075045585632, 0.07025105506181717, 0.004602391272783279, -0.1299954354763031, -0.0517367385327816, 0.14937616884708405, 0.09467504918575287, -0.017767438665032387, -0.11474084109067917, -0.025576479732990265, -0.17273752391338348, -0.049145691096782684, 0.010850531980395317, 0.023154571652412415, -0.025157511234283447, -0.0725991502404213, 0.016284946352243423, -0.03775440901517868, -0.05708695203065872, 0.02541809342801571, 0.12863710522651672, 0.03892246261239052, -0.0426139160990715, 0.026618195697665215, 0.0646548643708229, 0.06534937024116516, -0.15915727615356445, -0.022056326270103455, 0.0441533699631691, -0.09185189008712769, 0.0006507643265649676, -0.02288075163960457, 0.020454462617635727, 0.06537199765443802, 0.1255417764186859, 0.02018558420240879, 0.0949082151055336, 0.004809416830539703, 0.007112777791917324, -0.06568729877471924, 0.15002411603927612, -0.0759405642747879, -0.12959307432174683, -0.0320279486477375, 0.10259812325239182, 0.028567280620336533, -0.028215250000357628, -0.08664143830537796, 0.03551812097430229, 0.10475191473960876, 0.049767058342695236, 0.019304657354950905, 0.006800802890211344, -0.07232306897640228, -0.028252726420760155, -0.0018602844793349504, -0.11613243073225021, 0.05216389149427414, 0.040091145783662796, -0.059530992060899734, 0.019789431244134903, -0.04101897031068802, 0.03136139735579491, -0.03709588199853897, 0.08447384089231491, -0.050626251846551895, -0.011908846907317638, -0.06839492172002792, -0.09394611418247223, 0.041213225573301315, -0.048423074185848236, -0.0018580700270831585, -0.06798587739467621, -0.07879049330949783, -0.07197512686252594, 0.04724820703268051, -0.06533107161521912, -0.06738872826099396, -0.10049029439687729, -0.0780387669801712, 0.06982302665710449, -0.02694655768573284, 0.1472252458333969, -0.06358947604894638, 0.09082811325788498, 0.009028728120028973, 0.05875677615404129, 0.06746304035186768, 0.07027913630008698, -0.013249676674604416, 0.041504230350255966, -0.13171537220478058, 0.10083271563053131, -0.10292085260152817, 0.05156542360782623, -0.14676527678966522, -0.08505582809448242, -0.003513042815029621, -0.0044577219523489475, 0.10957145690917969, 0.1259939819574356, -0.1769210249185562, -0.09716719388961792, 0.14341747760772705, -0.05239272862672806, -0.09674205631017685, 0.1388108879327774, -0.005216808523982763, -0.054757826030254364, 0.03630892559885979, 0.18836447596549988, 0.10085056722164154, -0.09640274196863174, -0.02434091456234455, -0.06288570910692215, 0.13570110499858856, 0.02827863208949566, 0.10750740021467209, -0.06416811048984528, 0.01802448183298111, -0.0016716265818104148, -0.011715646833181381, 0.07483148574829102, -0.07950899749994278, -0.07072685658931732, -0.009146175347268581, -0.07310967147350311, -0.0116347661241889, 0.04712292179465294, 0.009373380802571774, -0.09326392412185669, -0.11978176236152649, -0.028054146096110344, 0.11232854425907135, -0.10464457422494888, 0.028662359341979027, -0.08761037141084671, 0.09365160018205643, -0.029175765812397003, 0.0032569144386798143, -0.13908378779888153, 0.016910552978515625, 0.04287807643413544, -0.04038199409842491, -0.00476235244423151, -0.0629844143986702, 0.05423694849014282, 0.033166732639074326, -0.03322441503405571, -0.061858806759119034, -0.03210555016994476, 0.000610672403126955, -0.05106527730822563, -0.24368055164813995, -0.05622861906886101, -0.023833777755498886, 0.17313161492347717, -0.15816649794578552, -0.012676887214183807, 0.04289711266756058, 0.14737337827682495, 0.02226865477859974, -0.0595240592956543, 0.011137736961245537, 0.06718040257692337, -0.03464833274483681, -0.07803545892238617, 0.025354882702231407, 0.01268541719764471, -0.11110744625329971, 0.0163203664124012, -0.11862918734550476, 0.06793703883886337, 0.11265569925308228, 0.023474495857954025, -0.04075482115149498, -0.059855274856090546, -0.03944362699985504, -0.043605219572782516, -0.029903370887041092, -0.0020616762340068817, 0.13952504098415375, 0.010129780508577824, 0.09452524036169052, -0.07859798520803452, -0.02859850414097309, 0.04595126211643219, 0.027561094611883163, -0.036177270114421844, 0.1340089589357376, 0.0718059092760086, -0.04785139858722687, 0.1076928898692131, 0.07290484011173248, -0.026944570243358612, 0.14265815913677216, -0.06581251323223114, -0.08688733726739883, -0.04108024761080742, 0.014471099711954594, 0.021560026332736015, 0.10324937850236893, -0.18647685647010803, -0.02585047297179699, 0.026967793703079224, 0.032268259674310684, 0.01863606832921505, -0.16269178688526154, 0.01046487782150507, 0.037735361605882645, -0.07552672922611237, -0.0028118344489485025, -0.003093593055382371, -0.029190855100750923, 0.08001117408275604, 0.025814997032284737, -0.0735611766576767, -0.03234482184052467, -0.040713001042604446, -0.0948939323425293, 0.1508353352546692, -0.12757058441638947, -0.14230145514011383, -0.10105184465646744, -0.05590660497546196, -0.03548014163970947, -0.006818023510277271, 0.054100580513477325, -0.10303287953138351, -0.0425691194832325, -0.06919436156749725, 0.01184080820530653, -0.06479522585868835, 0.04310918226838112, 0.053908079862594604, -0.002713002497330308, 0.030177311971783638, -0.08444844186306, 0.008923456072807312, -0.007442751433700323, 0.01791178435087204, -0.011569857597351074, 0.0067466204054653645, 0.11344931274652481, 0.15905611217021942, 0.07005401700735092, 0.04722807556390762, -0.03684253618121147, 0.17949111759662628, -0.12956762313842773, -0.00309249316342175, 0.10308831930160522, 0.03670080751180649, 0.03676026687026024, 0.15435051918029785, 0.029390724375844002, -0.09275730699300766, 0.017307370901107788, 0.020293889567255974, -0.013618906028568745, -0.2169748991727829, -0.046836934983730316, -0.08350982517004013, -0.03644653409719467, 0.0941179171204567, 0.024457179009914398, -0.006107225548475981, 0.016602156683802605, -0.028073281049728394, 0.00613754615187645, 0.02734081819653511, 0.05380319803953171, 0.07764209061861038, 0.04606131091713905, 0.11136964708566666, -0.017857953906059265, -0.00823199562728405, 0.03829127550125122, -0.014569018967449665, 0.24185451865196228, 0.02049882896244526, 0.19489431381225586, 0.04068131744861603, 0.15818220376968384, -0.0010538653004914522, 0.030251018702983856, 0.02438320405781269, 0.003985360264778137, 0.01175265945494175, -0.059131328016519547, -0.04989980533719063, 0.042968910187482834, 0.1362471729516983, 0.017396626994013786, -0.11231544613838196, 0.02071642503142357, 0.02653535082936287, 0.3475891947746277, 0.09687482565641403, -0.27815794944763184, -0.066129170358181, 0.01936604641377926, -0.055553141981363297, -0.036847710609436035, 0.038420725613832474, 0.1063586175441742, -0.056903012096881866, 0.09416534751653671, -0.038275592029094696, 0.08680088818073273, -0.08185304701328278, -0.005868215579539537, 0.054573677480220795, 0.09041275084018707, 0.003776581259444356, 0.0554441437125206, -0.2667340040206909, 0.2632381319999695, -0.003595722373574972, 0.08558212965726852, -0.0541711188852787, 0.042036741971969604, 0.037317726761102676, -0.025875596329569817, 0.07686637341976166, 0.0003384268784429878, -0.14422687888145447, -0.12956008315086365, -0.1087469682097435, 0.011179255321621895, 0.12451782822608948, -0.05664511024951935, 0.11139044165611267, -0.04021487757563591, -0.0388440266251564, 0.03068157285451889, -0.04503532499074936, -0.11742842942476273, -0.1213330551981926, 0.033578503876924515, 0.05079187825322151, 0.0901150107383728, -0.0782289206981659, -0.09251751005649567, -0.0694500058889389, 0.15721222758293152, -0.10316141694784164, -0.009166390635073185, -0.13092701137065887, 0.07536135613918304, 0.13743437826633453, -0.06423642486333847, 0.04533663019537926, 0.010650836862623692, 0.12946724891662598, 0.02372400276362896, -0.016430215910077095, 0.1111147478222847, -0.08121974766254425, -0.18110156059265137, -0.06761983782052994, 0.16069768369197845, 0.02361919730901718, 0.06214814633131027, -0.01445283554494381, 0.03064015880227089, -0.010374183766543865, -0.07008583098649979, 0.092009536921978, 0.09228076040744781, 0.029108598828315735, 0.059864018112421036, -0.00017823762027546763, -0.012018037959933281, -0.08764557540416718, -0.08658736199140549, 0.12955650687217712, 0.26790851354599, -0.07738975435495377, 0.049109239131212234, 0.05069315433502197, -0.058281295001506805, -0.15284080803394318, -0.015075240284204483, 0.11060932278633118, 0.046999912708997726, -0.03813010826706886, -0.20946794748306274, -0.0023068978916853666, 0.06485264748334885, -0.023463312536478043, 0.06343285739421844, -0.3309203088283539, -0.1270918846130371, 0.08071491122245789, 0.06919064372777939, 0.005913892760872841, -0.16097582876682281, -0.07789528369903564, -0.04245220869779587, -0.0875297486782074, 0.03547930344939232, -0.004354468546807766, 0.12054996937513351, 0.013725615106523037, 0.004718783777207136, 0.012376529164612293, -0.0452708974480629, 0.15944623947143555, 0.0028896143194288015, 0.02028820291161537, -0.015335156582295895, 0.04486802592873573, -0.023809116333723068, -0.06637230515480042, -0.010364048182964325, -0.06761156767606735, 0.03012389875948429, -0.13269703090190887, -0.03181591257452965, -0.06938107311725616, 0.014726291410624981, -0.028232412412762642, -0.020722394809126854, -0.044823311269283295, 0.04039767384529114, 0.1029965952038765, 0.012021586298942566, 0.13294285535812378, -0.05060790479183197, 0.115770123898983, 0.09373214840888977, 0.09320742636919022, -0.03402718901634216, -0.06868886947631836, -0.012692281045019627, -0.019634807482361794, 0.037600077688694, -0.11191883683204651, 0.029063932597637177, 0.13528577983379364, 0.039657846093177795, 0.14288024604320526, 0.04974519833922386, -0.0914398804306984, 0.022727258503437042, 0.06294472515583038, -0.07190138846635818, -0.16166464984416962, -0.026534447446465492, 0.033810459077358246, -0.10887865722179413, -0.009291703812777996, 0.106547512114048, -0.024434318765997887, -0.0009153406717814505, 0.005561428144574165, 0.04657965153455734, -0.021383872255682945, 0.21219351887702942, 0.007996531203389168, 0.07300939410924911, -0.10569928586483002, 0.08832450956106186, 0.049594637006521225, -0.14135242998600006, 0.0591924823820591, 0.0721006914973259, -0.060681190341711044, -0.014279605820775032, 0.01390281692147255, 0.09388317167758942, 0.04224115237593651, -0.06364651769399643, -0.10773833096027374, -0.14726576209068298, 0.10671333968639374, 0.04986744374036789, 0.02441631816327572, 0.017746377736330032, -0.016558486968278885, 0.017729412764310837, -0.0915529727935791, 0.096822589635849, 0.08144264668226242, 0.05279918387532234, -0.1281723976135254, 0.07317859679460526, 0.009033665992319584, -0.002517710207030177, 0.00030559953302145004, -0.012837800197303295, -0.09591127187013626, 0.023691562935709953, -0.11922629177570343, -0.011593649163842201, -0.0867854580283165, -0.009870863519608974, 0.015215292572975159, -0.06506721675395966, -0.06892850995063782, 0.016506463289260864, -0.10695614665746689, -0.06169978156685829, -0.033691491931676865, 0.06684672087430954, -0.09173169732093811, -0.014050749130547047, 0.023790504783391953, -0.12831826508045197, 0.09517930448055267, 0.04894276708364487, 0.029964376240968704, 0.0045854318886995316, -0.07231191545724869, -0.01752944104373455, 0.032313015311956406, 0.008015314117074013, 0.03479496017098427, -0.20110450685024261, -0.010629733093082905, -0.003845921717584133, 0.005833582021296024, -0.013025534339249134, 0.04272269457578659, -0.10994397848844528, -0.03069905750453472, -0.05439263582229614, -0.034640468657016754, -0.04587898775935173, 0.06895466893911362, 0.10125548392534256, 0.019229736179113388, 0.15092957019805908, -0.07527454197406769, 0.0475212000310421, -0.2011300027370453, 0.013580824248492718, -0.034787844866514206, -0.06489428132772446, -0.05021971836686134, -0.026412133127450943, 0.09986776858568192, -0.04899897053837776, 0.07725328952074051, -0.04875252768397331, 0.05438263341784477, 0.027174493297934532, -0.10994257032871246, 0.022511808201670647, 0.04988696053624153, 0.19558729231357574, 0.06174834445118904, -0.02000177651643753, 0.08371346443891525, 0.0015463362215086818, 0.08193279057741165, 0.17591674625873566, 0.12421157956123352, 0.14587076008319855, 0.10527832806110382, 0.11034520715475082, 0.061973605304956436, -0.14312079548835754, -0.14394564926624298, 0.15965864062309265, -0.06944569945335388, 0.15539269149303436, -0.003978334832936525, 0.18347282707691193, 0.1035131886601448, -0.19880248606204987, 0.047456249594688416, -0.03532172366976738, -0.0733628123998642, -0.10372861474752426, -0.07843092083930969, -0.09260286390781403, -0.1954781413078308, 0.010948163457214832, -0.10784906893968582, 0.06543098390102386, 0.03265968710184097, 0.045991845428943634, 0.04070897027850151, 0.06246437132358551, 0.044050246477127075, -0.014249596744775772, 0.12226632982492447, 0.005926258862018585, -0.040500905364751816, -0.04764556512236595, -0.12404555827379227, 0.050477035343647, -0.04369578883051872, 0.07600466161966324, -0.016459273174405098, -0.10774163901805878, 0.08043079823255539, 0.018689539283514023, -0.10757670551538467, 0.026554882526397705, -0.026128778234124184, 0.056269560009241104, 0.11704088747501373, 0.04659898951649666, -0.015275906771421432, 0.004241234622895718, 0.19996406137943268, -0.0943513959646225, -0.03778072074055672, -0.13626213371753693, 0.17189954221248627, 0.004831789061427116, 0.011786578223109245, 0.02229013666510582, -0.0805426687002182, -0.025287311524152756, 0.16337768733501434, 0.1279820054769516, -0.016341209411621094, -0.03465522453188896, 0.026146387681365013, -0.011368056759238243, -0.04305853694677353, 0.07146934419870377, 0.12481958419084549, 0.04511090740561485, -0.03509969264268875, -0.027508128434419632, -0.05145856365561485, -0.06755542010068893, -0.027347324416041374, 0.07426853477954865, 0.007853780873119831, -0.027285398915410042, -0.0052182236686348915, 0.11815650016069412, -0.05615914613008499, -0.1406397670507431, 0.0531947985291481, -0.18368832767009735, -0.185907244682312, -0.023388704285025597, 0.06113864853978157, 0.05213770270347595, 0.04987455531954765, 0.001966026844456792, -0.0169865470379591, 0.1304602175951004, 0.007921915501356125, -0.05599088966846466, -0.10493365675210953, 0.062019214034080505, -0.10923858731985092, 0.16489766538143158, -0.04038712754845619, 0.015854474157094955, 0.13256214559078217, 0.08457498252391815, -0.09584265947341919, 0.03721858933568001, 0.08579987287521362, -0.10375852137804031, 0.06811854988336563, 0.17331741750240326, -0.04644802585244179, 0.15217135846614838, 0.06186912581324577, -0.05878134071826935, 0.024562032893300056, -0.08816298842430115, -0.03177056834101677, -0.05596037209033966, 0.006586499512195587, -0.05558647960424423, 0.14495854079723358, 0.1607825756072998, -0.07318143546581268, -0.017427891492843628, -0.025381749495863914, 0.014967692084610462, 0.013724480755627155, 0.14169727265834808, -0.03437090665102005, -0.27623701095581055, 0.012248705141246319, -0.003062873613089323, 0.03309418633580208, -0.2026996910572052, -0.07173600792884827, 0.013758531771600246, -0.055428922176361084, -0.06669189035892487, 0.12377560138702393, 0.05243320018053055, 0.025442559272050858, -0.07700192928314209, -0.11665551364421844, -0.017623892053961754, 0.18074075877666473, -0.1657392978668213, -0.06294626742601395 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xls-r-300m-pa This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the MOZILLA-FOUNDATION/COMMON_VOICE_8_0 - PA-IN dataset. It achieves the following results on the evaluation set: - Loss: 1.0443 - Wer: 0.5715 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 7.5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 500.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:------:|:-----:|:---------------:|:------:| | 4.6694 | 19.22 | 500 | 4.0455 | 1.0 | | 3.3907 | 38.45 | 1000 | 3.2836 | 1.0 | | 2.0866 | 57.67 | 1500 | 1.2788 | 0.7715 | | 1.4106 | 76.9 | 2000 | 0.7866 | 0.6891 | | 1.1711 | 96.15 | 2500 | 0.6556 | 0.6272 | | 1.038 | 115.37 | 3000 | 0.6195 | 0.5680 | | 0.8989 | 134.6 | 3500 | 0.6563 | 0.5602 | | 0.8021 | 153.82 | 4000 | 0.6644 | 0.5327 | | 0.7161 | 173.07 | 4500 | 0.6844 | 0.5253 | | 0.6449 | 192.3 | 5000 | 0.7018 | 0.5331 | | 0.5659 | 211.52 | 5500 | 0.7451 | 0.5465 | | 0.5118 | 230.75 | 6000 | 0.7857 | 0.5386 | | 0.4385 | 249.97 | 6500 | 0.8062 | 0.5382 | | 0.3984 | 269.22 | 7000 | 0.8316 | 0.5621 | | 0.3666 | 288.45 | 7500 | 0.8736 | 0.5504 | | 0.3256 | 307.67 | 8000 | 0.9133 | 0.5688 | | 0.289 | 326.9 | 8500 | 0.9556 | 0.5684 | | 0.2663 | 346.15 | 9000 | 0.9344 | 0.5708 | | 0.2445 | 365.37 | 9500 | 0.9472 | 0.5590 | | 0.2289 | 384.6 | 10000 | 0.9713 | 0.5672 | | 0.2048 | 403.82 | 10500 | 0.9978 | 0.5762 | | 0.1857 | 423.07 | 11000 | 1.0230 | 0.5798 | | 0.1751 | 442.3 | 11500 | 1.0409 | 0.5755 | | 0.1688 | 461.52 | 12000 | 1.0445 | 0.5727 | | 0.1633 | 480.75 | 12500 | 1.0484 | 0.5739 | | 0.1488 | 499.97 | 13000 | 1.0443 | 0.5715 | ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2.dev0 - Tokenizers 0.11.0
{"language": ["pa-IN"], "license": "apache-2.0", "tags": ["automatic-speech-recognition", "mozilla-foundation/common_voice_8_0", "generated_from_trainer"], "datasets": ["common_voice"], "model-index": [{"name": "xls-r-300m-pa", "results": []}]}
automatic-speech-recognition
gagan3012/xls-r-300m-pa
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "mozilla-foundation/common_voice_8_0", "generated_from_trainer", "dataset:common_voice", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "pa-IN" ]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #mozilla-foundation/common_voice_8_0 #generated_from_trainer #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us
xls-r-300m-pa ============= This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the MOZILLA-FOUNDATION/COMMON\_VOICE\_8\_0 - PA-IN dataset. It achieves the following results on the evaluation set: * Loss: 1.0443 * Wer: 0.5715 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 7.5e-05 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * gradient\_accumulation\_steps: 4 * total\_train\_batch\_size: 32 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 2000 * num\_epochs: 500.0 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.17.0.dev0 * Pytorch 1.10.2+cu102 * Datasets 1.18.2.dev0 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 7.5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 2000\n* num\\_epochs: 500.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.2.dev0\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #mozilla-foundation/common_voice_8_0 #generated_from_trainer #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 7.5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 2000\n* num\\_epochs: 500.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.2.dev0\n* Tokenizers 0.11.0" ]
[ 77, 160, 4, 39 ]
[ "passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #mozilla-foundation/common_voice_8_0 #generated_from_trainer #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 7.5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 2000\n* num\\_epochs: 500.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.2.dev0\n* Tokenizers 0.11.0" ]
[ -0.12321756035089493, 0.14380420744419098, -0.004304243251681328, 0.028821488842368126, 0.10690420866012573, 0.005487707909196615, 0.09184662252664566, 0.15203234553337097, -0.07896053791046143, 0.12377707660198212, 0.09512977302074432, 0.09689455479383469, 0.08983458578586578, 0.15711013972759247, -0.026030993089079857, -0.2697265148162842, 0.035903334617614746, -0.02579757757484913, -0.0784120187163353, 0.09754812717437744, 0.0887315645813942, -0.10750622302293777, 0.02845027856528759, 0.011159791611135006, -0.09161160886287689, -0.011145015247166157, -0.042109500616788864, -0.06659077107906342, 0.09215331822633743, 0.04581378027796745, 0.03394317254424095, 0.029263198375701904, 0.06915847957134247, -0.27116158604621887, 0.0072790831327438354, 0.06056175380945206, 0.03798166662454605, 0.0538501963019371, 0.10671965777873993, -0.0007734967512078583, 0.0866212248802185, -0.07903465628623962, 0.04250582680106163, 0.052758172154426575, -0.07902998477220535, -0.2854451835155487, -0.10648641735315323, 0.017931709066033363, 0.13447445631027222, 0.08627734333276749, -0.03956707566976547, 0.05424036830663681, -0.044521454721689224, 0.08543967455625534, 0.24164614081382751, -0.24224615097045898, -0.05817326158285141, -0.02977459877729416, 0.034701187163591385, 0.02482803910970688, -0.0964784249663353, 0.00534772640094161, 0.026319829747080803, 0.02076786383986473, 0.08530636876821518, 0.009437687695026398, 0.04886355623602867, -0.00993428286164999, -0.1343509554862976, -0.04725387319922447, 0.1451081782579422, 0.08771073073148727, -0.02484690211713314, -0.11736151576042175, -0.029509834945201874, -0.179926797747612, -0.04429251700639725, 0.0004875822633039206, 0.03257785737514496, -0.02959159016609192, -0.07804044336080551, 0.014295169152319431, -0.040139976888895035, -0.06184595450758934, 0.018634336069226265, 0.14354091882705688, 0.04248572513461113, -0.04122449830174446, 0.023241644725203514, 0.06870207190513611, 0.028145503252744675, -0.1525697410106659, -0.02013702504336834, 0.05312143266201019, -0.09244433790445328, -0.005464266054332256, -0.024746106937527657, -0.0048699709586799145, 0.05966624245047569, 0.14856569468975067, -0.0031547476537525654, 0.09822052717208862, 0.006005325820297003, 0.012943288311362267, -0.06427974998950958, 0.14946751296520233, -0.06694773584604263, -0.11467622220516205, -0.03356846049427986, 0.10078027844429016, 0.024256180971860886, -0.023780662566423416, -0.08174658566713333, 0.02541046403348446, 0.11129047721624374, 0.047141481190919876, 0.0035610347986221313, 0.00876366626471281, -0.07047517597675323, -0.022310232743620872, -0.009082538075745106, -0.11723783612251282, 0.05400454252958298, 0.030776675790548325, -0.05693439766764641, 0.02030012756586075, -0.02759351208806038, 0.030493631958961487, -0.033761344850063324, 0.10896727442741394, -0.04145710915327072, -0.009702461771667004, -0.06798029690980911, -0.08985086530447006, 0.037935521453619, -0.060719653964042664, -0.006138937547802925, -0.07214498519897461, -0.08154185861349106, -0.07254185527563095, 0.0428452230989933, -0.0626857578754425, -0.06771210581064224, -0.09410952031612396, -0.07556448131799698, 0.0680735632777214, -0.03195472061634064, 0.1538366973400116, -0.06316423416137695, 0.08843132853507996, 0.022157147526741028, 0.05606715381145477, 0.05528440698981285, 0.07167728990316391, -0.013913956470787525, 0.043601419776678085, -0.15227852761745453, 0.09798053652048111, -0.09505404531955719, 0.0404977984726429, -0.14795400202274323, -0.08008372038602829, -0.017203904688358307, -0.0024015067610889673, 0.11248462647199631, 0.12077822536230087, -0.18054069578647614, -0.09976895898580551, 0.14726804196834564, -0.05600389838218689, -0.08698150515556335, 0.1386619210243225, -0.006273836828768253, -0.050705067813396454, 0.03728906437754631, 0.18760864436626434, 0.0971788838505745, -0.08894053846597672, -0.025334253907203674, -0.07347763329744339, 0.12013919651508331, 0.030997591093182564, 0.09630193561315536, -0.07485277950763702, 0.0376218780875206, -0.0033947378396987915, -0.007116599939763546, 0.06408140063285828, -0.07516557723283768, -0.06003762036561966, -0.01665683090686798, -0.0712365061044693, -0.0032308101654052734, 0.03270750865340233, 0.015394330956041813, -0.09926475584506989, -0.11761899292469025, -0.01820996031165123, 0.10668092966079712, -0.0979381576180458, 0.03562210127711296, -0.08515586704015732, 0.08445841073989868, -0.02339644730091095, 0.002385617233812809, -0.15430809557437897, 0.031273964792490005, 0.04870045557618141, -0.046294599771499634, 0.010188912972807884, -0.06568146497011185, 0.05745910853147507, 0.02971014752984047, -0.03318633884191513, -0.06264521181583405, -0.030703594908118248, -0.0036615144927054644, -0.06084847450256348, -0.2309018075466156, -0.054542385041713715, -0.02497648261487484, 0.14930617809295654, -0.14981123805046082, -0.007099548354744911, 0.04710741713643074, 0.15552100539207458, 0.02640528604388237, -0.05717027187347412, 0.011018048040568829, 0.07104908674955368, -0.02487703040242195, -0.074574775993824, 0.024423109367489815, 0.008060833439230919, -0.11493244767189026, 0.00982011016458273, -0.12574264407157898, 0.06373218446969986, 0.10562658309936523, 0.0194531437009573, -0.057215020060539246, -0.06690705567598343, -0.03783326596021652, -0.04868444800376892, -0.03743818774819374, -0.01136629469692707, 0.1745719313621521, 0.017334802076220512, 0.10134121030569077, -0.07523595541715622, -0.034304894506931305, 0.04111536964774132, 0.02536361664533615, -0.029452577233314514, 0.12057551741600037, 0.08121547847986221, -0.055263616144657135, 0.10154836624860764, 0.07277096807956696, -0.0368012972176075, 0.13819393515586853, -0.06362668424844742, -0.08635883778333664, -0.038079630583524704, 0.019911643117666245, 0.018210383132100105, 0.0892772302031517, -0.18008820712566376, -0.025565393269062042, 0.029344791546463966, 0.03538808226585388, 0.01725327968597412, -0.1680535525083542, 0.0236784890294075, 0.042122840881347656, -0.07785578072071075, 0.007730378303676844, -0.0032229258213192225, -0.01792527176439762, 0.08007887005805969, 0.02258172817528248, -0.07675428688526154, -0.023531774058938026, -0.03342254459857941, -0.08788806945085526, 0.14474636316299438, -0.12230901420116425, -0.14586572349071503, -0.10961276292800903, -0.04556750878691673, -0.027774494141340256, -0.007010387256741524, 0.06267459690570831, -0.09332334250211716, -0.0455288328230381, -0.06380195170640945, 0.017022570595145226, -0.06014878675341606, 0.045328568667173386, 0.044866468757390976, -0.005496737081557512, 0.0414767749607563, -0.08966928720474243, 0.006322792731225491, -0.004494480323046446, 0.012082950212061405, -0.006644873879849911, 0.020606525242328644, 0.09991791099309921, 0.15172630548477173, 0.06154559552669525, 0.043607186526060104, -0.0434110090136528, 0.1679387092590332, -0.13053542375564575, -0.005552033428102732, 0.10446107387542725, 0.026228126138448715, 0.04232117161154747, 0.155375137925148, 0.02953854762017727, -0.09364908933639526, 0.017173243686556816, 0.027941010892391205, -0.014000597409904003, -0.2150702178478241, -0.03527811914682388, -0.07633045315742493, -0.02957555092871189, 0.09808260947465897, 0.03148771822452545, -0.0062765637412667274, 0.019654814153909683, -0.025257417932152748, 0.007499093655496836, 0.024808689951896667, 0.05874072387814522, 0.08726584166288376, 0.04583831503987312, 0.11848637461662292, -0.011444206349551678, -0.012954591773450375, 0.041768740862607956, -0.019072428345680237, 0.2435096651315689, 0.026287462562322617, 0.18781496584415436, 0.03886764496564865, 0.1523798108100891, -0.0017620401922613382, 0.03132013604044914, 0.023539148271083832, 0.0071528032422065735, 0.009632401168346405, -0.05389998480677605, -0.04408784955739975, 0.03874058648943901, 0.13582025468349457, 0.01930168829858303, -0.1126561313867569, 0.015097190625965595, 0.02622847445309162, 0.3732735514640808, 0.08282554149627686, -0.2794751822948456, -0.05945131927728653, 0.013372021727263927, -0.07297825813293457, -0.0404842309653759, 0.03476506844162941, 0.1104699969291687, -0.06768602877855301, 0.09825578331947327, -0.041667453944683075, 0.08882235735654831, -0.08069875836372375, -0.0038031351286917925, 0.05715078487992287, 0.09294994175434113, 0.003691100049763918, 0.04189377650618553, -0.26470622420310974, 0.2568330764770508, 0.00007094912871252745, 0.08434141427278519, -0.05010303854942322, 0.05096025764942169, 0.03482885658740997, -0.028168680146336555, 0.06950407475233078, -0.010542485862970352, -0.13942766189575195, -0.1300952434539795, -0.10751690715551376, 0.005003081168979406, 0.13440479338169098, -0.058906905353069305, 0.11353511363267899, -0.03616274893283844, -0.03613228723406792, 0.031317368149757385, -0.047358471900224686, -0.1170516312122345, -0.10751478374004364, 0.032334908843040466, 0.0775766372680664, 0.07911107689142227, -0.07774022966623306, -0.09226623177528381, -0.07363197207450867, 0.15545083582401276, -0.10144975036382675, -0.01412358321249485, -0.12716606259346008, 0.06417593359947205, 0.1441291719675064, -0.061113838106393814, 0.0558200441300869, 0.013861965388059616, 0.12540027499198914, 0.01981380395591259, -0.009577360935509205, 0.11609672755002975, -0.07593904435634613, -0.18943025171756744, -0.06750646978616714, 0.1705419272184372, 0.016837038099765778, 0.05941038578748703, -0.02005600556731224, 0.037114717066287994, -0.0008472666959278286, -0.06244170293211937, 0.09203621000051498, 0.075980544090271, 0.027280882000923157, 0.06547944247722626, -0.007242342922836542, -0.024588672444224358, -0.08444782346487045, -0.08755171298980713, 0.13345927000045776, 0.27531513571739197, -0.07761113345623016, 0.04983168840408325, 0.04761330783367157, -0.05484899505972862, -0.13890346884727478, -0.014370974153280258, 0.12269977480173111, 0.04623027890920639, -0.022292667999863625, -0.2084539532661438, -0.0027513005770742893, 0.07913867384195328, -0.02570679597556591, 0.08006880432367325, -0.33186036348342896, -0.13618186116218567, 0.0847887396812439, 0.06110875681042671, -0.0027800556272268295, -0.16108550131320953, -0.08000606298446655, -0.037634339183568954, -0.07855013757944107, 0.025467371568083763, -0.019806884229183197, 0.1302243024110794, 0.016868585720658302, 0.026829034090042114, 0.01245603896677494, -0.03913978859782219, 0.14909863471984863, 0.007000655867159367, 0.02607496827840805, -0.017660774290561676, 0.038488563150167465, -0.04679983854293823, -0.06585729122161865, -0.0019001781474798918, -0.07169982045888901, 0.022670872509479523, -0.12721186876296997, -0.0323956124484539, -0.06264542043209076, 0.019198814406991005, -0.022233787924051285, -0.018534870818257332, -0.04093391075730324, 0.029831642284989357, 0.08166273683309555, 0.008799048140645027, 0.12915150821208954, -0.057146910578012466, 0.11333850771188736, 0.11695284396409988, 0.09852619469165802, -0.03569689393043518, -0.0933966338634491, -0.007230717223137617, -0.023277070373296738, 0.03623003140091896, -0.09557344764471054, 0.033199090510606766, 0.13277067244052887, 0.03710759058594704, 0.14781971275806427, 0.04744727537035942, -0.08941645175218582, 0.027017856016755104, 0.057816918939352036, -0.07160875201225281, -0.15605366230010986, -0.012175997719168663, 0.04153936356306076, -0.11085603386163712, 0.005591155029833317, 0.12228059768676758, -0.02154800482094288, -0.003520744852721691, 0.011073630303144455, 0.03746945783495903, -0.02558054029941559, 0.22274982929229736, 0.0010746413609012961, 0.06972800940275192, -0.1000400111079216, 0.09624690562486649, 0.05257958173751831, -0.12725162506103516, 0.06157306209206581, 0.07355082035064697, -0.056429117918014526, -0.009287701919674873, 0.013943095691502094, 0.09261032193899155, 0.046103883534669876, -0.06454609334468842, -0.10812301188707352, -0.14233210682868958, 0.10637009888887405, 0.061396125704050064, 0.022921675816178322, 0.012827431783080101, -0.022019652649760246, 0.02332882769405842, -0.0862680971622467, 0.08643841743469238, 0.09102583676576614, 0.05895509943366051, -0.1267324537038803, 0.09656374156475067, 0.010537122376263142, -0.009218433871865273, -0.00009527136717224494, -0.009579673409461975, -0.08876752853393555, 0.025036904960870743, -0.10649207979440689, -0.010460276156663895, -0.07136685401201248, -0.01626519113779068, 0.02154703624546528, -0.059976253658533096, -0.0654141902923584, 0.02033262886106968, -0.10525523871183395, -0.060809846967458725, -0.03103613294661045, 0.0651000440120697, -0.09268791228532791, -0.018342841416597366, 0.022680852562189102, -0.12305097281932831, 0.09541067481040955, 0.05285037308931351, 0.022906381636857986, 0.00044983456609770656, -0.0803300142288208, -0.02127608098089695, 0.04352208971977234, 0.00012334156781435013, 0.030471649020910263, -0.1992674171924591, -0.014185384847223759, -0.01056567020714283, 0.01610761694610119, -0.011761032044887543, 0.040117908269166946, -0.11602625250816345, -0.030626384541392326, -0.05197618156671524, -0.04089280217885971, -0.046821944415569305, 0.05485110729932785, 0.09791518747806549, 0.02777734026312828, 0.16067413985729218, -0.08035679161548615, 0.04831130802631378, -0.19662494957447052, 0.019613567739725113, -0.03440459817647934, -0.06707555055618286, -0.05002206936478615, -0.021616658195853233, 0.0970083475112915, -0.05288330093026161, 0.07703527063131332, -0.0541590191423893, 0.053893279284238815, 0.03153333067893982, -0.12048913538455963, 0.012775937095284462, 0.046371940523386, 0.17564044892787933, 0.06379136443138123, -0.027409492060542107, 0.07370122522115707, 0.004065928049385548, 0.0829031839966774, 0.15324090421199799, 0.12880836427211761, 0.1452338546514511, 0.08833517134189606, 0.11300040781497955, 0.06494323909282684, -0.13845069706439972, -0.13890859484672546, 0.14762039482593536, -0.05772460252046585, 0.15258720517158508, -0.007877718657255173, 0.20867158472537994, 0.09971512109041214, -0.19272944331169128, 0.05568891763687134, -0.02868036925792694, -0.07171276956796646, -0.10452268272638321, -0.06725028902292252, -0.09485211968421936, -0.18887829780578613, 0.009015290066599846, -0.09889153391122818, 0.06894183158874512, 0.030111204832792282, 0.04254072532057762, 0.04106505960226059, 0.07138380408287048, 0.023271987214684486, -0.01685960963368416, 0.12674133479595184, 0.005713709630072117, -0.033845625817775726, -0.05406486615538597, -0.12952889502048492, 0.0517970509827137, -0.03256966546177864, 0.08001183718442917, -0.019615642726421356, -0.1071624681353569, 0.06668373197317123, 0.019462427124381065, -0.10450654476881027, 0.023431334644556046, -0.023169182240962982, 0.05354873463511467, 0.10654565691947937, 0.0452192947268486, -0.014144779182970524, -0.0022018004674464464, 0.19826745986938477, -0.09951003640890121, -0.037301335483789444, -0.13103871047496796, 0.15697899460792542, 0.011231478303670883, 0.009757980704307556, 0.01803874410688877, -0.08685478568077087, -0.026044731959700584, 0.14895515143871307, 0.12593989074230194, -0.010498062707483768, -0.02902442216873169, 0.019155316054821014, -0.010156306438148022, -0.03283505514264107, 0.06216251477599144, 0.12401130050420761, 0.03823792189359665, -0.030066953971982002, -0.014442468993365765, -0.04301324486732483, -0.05914076045155525, -0.031195491552352905, 0.06765100359916687, 0.003249567002058029, -0.018437866121530533, -0.008576036430895329, 0.11358999460935593, -0.041763998568058014, -0.148550882935524, 0.038439784198999405, -0.1765700876712799, -0.18467660248279572, -0.01789545826613903, 0.057805757969617844, 0.04779209941625595, 0.05381929129362106, 0.0017646694323047996, -0.020950937643647194, 0.12676256895065308, 0.003479807171970606, -0.05253615975379944, -0.11350289732217789, 0.06061200052499771, -0.11514224112033844, 0.16806894540786743, -0.03772568702697754, 0.02650108002126217, 0.1253902018070221, 0.08242452889680862, -0.0863146111369133, 0.04035790637135506, 0.0761777013540268, -0.10730373859405518, 0.06596158444881439, 0.18015095591545105, -0.050066910684108734, 0.15556596219539642, 0.05378793552517891, -0.07485675066709518, 0.02295561321079731, -0.07728170603513718, -0.03698469325900078, -0.06655512005090714, 0.005670666228979826, -0.05874231457710266, 0.1402333378791809, 0.1649203896522522, -0.06816602498292923, -0.019268793985247612, -0.029137806966900826, 0.016246715560555458, 0.02507832460105419, 0.12709346413612366, -0.041856274008750916, -0.27679798007011414, 0.02122277207672596, 0.007466454058885574, 0.027456579729914665, -0.2103341668844223, -0.08156061172485352, 0.012411888688802719, -0.05494140461087227, -0.0662076398730278, 0.11208673566579819, 0.0405886285007, 0.030464131385087967, -0.07320781797170639, -0.13092133402824402, -0.016430344432592392, 0.18298707902431488, -0.17379042506217957, -0.06052718684077263 ]
null
null
transformers
# Model Trained Using AutoNLP - Problem type: Binary Classification - Model ID: 19984005 - CO2 Emissions (in grams): 20.790169878009916 ## Validation Metrics - Loss: 0.06693269312381744 - Accuracy: 0.9789 - Precision: 0.9843244336569579 - Recall: 0.9733 - AUC: 0.99695552 - F1: 0.9787811745776348 ## Usage You can use cURL to access this model: ``` $ curl -X POST -H "Authorization: Bearer YOUR_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoNLP"}' https://api-inference.huggingface.co/models/gagandeepkundi/autonlp-text-classification-19984005 ``` Or Python API: ``` from transformers import AutoModelForSequenceClassification, AutoTokenizer model = AutoModelForSequenceClassification.from_pretrained("gagandeepkundi/autonlp-text-classification-19984005", use_auth_token=True) tokenizer = AutoTokenizer.from_pretrained("gagandeepkundi/autonlp-text-classification-19984005", use_auth_token=True) inputs = tokenizer("I love AutoNLP", return_tensors="pt") outputs = model(**inputs) ```
{"language": "es", "tags": "autonlp", "datasets": ["gagandeepkundi/autonlp-data-text-classification"], "widget": [{"text": "I love AutoNLP \ud83e\udd17"}], "co2_eq_emissions": 20.790169878009916}
text-classification
gagandeepkundi/latam-question-quality
[ "transformers", "pytorch", "roberta", "text-classification", "autonlp", "es", "dataset:gagandeepkundi/autonlp-data-text-classification", "co2_eq_emissions", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "es" ]
TAGS #transformers #pytorch #roberta #text-classification #autonlp #es #dataset-gagandeepkundi/autonlp-data-text-classification #co2_eq_emissions #autotrain_compatible #endpoints_compatible #region-us
# Model Trained Using AutoNLP - Problem type: Binary Classification - Model ID: 19984005 - CO2 Emissions (in grams): 20.790169878009916 ## Validation Metrics - Loss: 0.06693269312381744 - Accuracy: 0.9789 - Precision: 0.9843244336569579 - Recall: 0.9733 - AUC: 0.99695552 - F1: 0.9787811745776348 ## Usage You can use cURL to access this model: Or Python API:
[ "# Model Trained Using AutoNLP\n\n- Problem type: Binary Classification\n- Model ID: 19984005\n- CO2 Emissions (in grams): 20.790169878009916", "## Validation Metrics\n\n- Loss: 0.06693269312381744\n- Accuracy: 0.9789\n- Precision: 0.9843244336569579\n- Recall: 0.9733\n- AUC: 0.99695552\n- F1: 0.9787811745776348", "## Usage\n\nYou can use cURL to access this model:\n\n\n\nOr Python API:" ]
[ "TAGS\n#transformers #pytorch #roberta #text-classification #autonlp #es #dataset-gagandeepkundi/autonlp-data-text-classification #co2_eq_emissions #autotrain_compatible #endpoints_compatible #region-us \n", "# Model Trained Using AutoNLP\n\n- Problem type: Binary Classification\n- Model ID: 19984005\n- CO2 Emissions (in grams): 20.790169878009916", "## Validation Metrics\n\n- Loss: 0.06693269312381744\n- Accuracy: 0.9789\n- Precision: 0.9843244336569579\n- Recall: 0.9733\n- AUC: 0.99695552\n- F1: 0.9787811745776348", "## Usage\n\nYou can use cURL to access this model:\n\n\n\nOr Python API:" ]
[ 71, 41, 67, 17 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #text-classification #autonlp #es #dataset-gagandeepkundi/autonlp-data-text-classification #co2_eq_emissions #autotrain_compatible #endpoints_compatible #region-us \n# Model Trained Using AutoNLP\n\n- Problem type: Binary Classification\n- Model ID: 19984005\n- CO2 Emissions (in grams): 20.790169878009916## Validation Metrics\n\n- Loss: 0.06693269312381744\n- Accuracy: 0.9789\n- Precision: 0.9843244336569579\n- Recall: 0.9733\n- AUC: 0.99695552\n- F1: 0.9787811745776348## Usage\n\nYou can use cURL to access this model:\n\n\n\nOr Python API:" ]
[ -0.13447600603103638, 0.15254123508930206, -0.0016582254320383072, 0.04068990424275398, 0.07281842827796936, 0.028601793572306633, 0.0818631574511528, 0.08279933780431747, 0.011779683642089367, 0.051071468740701675, 0.15558399260044098, 0.1506713628768921, 0.0156637504696846, 0.22519925236701965, -0.12578295171260834, -0.13803847134113312, 0.094149649143219, 0.0179402194917202, 0.0685075968503952, 0.11808565258979797, 0.10328847914934158, -0.10483618825674057, 0.1228623241186142, 0.04817447438836098, -0.15106959640979767, -0.031310658901929855, 0.053434986621141434, -0.13107813894748688, 0.10601650178432465, 0.13656564056873322, 0.18305864930152893, 0.054832734167575836, 0.09536272287368774, -0.12724484503269196, -0.010414237156510353, -0.035918671637773514, -0.05643734708428383, 0.10582316666841507, 0.008419710211455822, -0.06065674126148224, -0.03663138672709465, -0.0037166851107031107, 0.07818521559238434, 0.016343872994184494, -0.08645614981651306, -0.004098021890968084, -0.031185168772935867, -0.037481945008039474, 0.13356630504131317, 0.10361188650131226, -0.01970105990767479, 0.2806895077228546, -0.14620408415794373, 0.05257200822234154, 0.030516555532813072, -0.1593577116727829, -0.011156625114381313, 0.05742541700601578, -0.05738574638962746, -0.06971614062786102, -0.033380795270204544, 0.04669013246893883, 0.10431306809186935, -0.00821925699710846, 0.0449477843940258, -0.04552774503827095, -0.06160302087664604, -0.0067728557623922825, -0.11572308093309402, -0.06880856305360794, 0.24087822437286377, 0.07043659687042236, -0.09950854629278183, -0.0009060283191502094, -0.09479448199272156, -0.14949244260787964, -0.04217018932104111, -0.06733471155166626, -0.007404601667076349, -0.06394269317388535, -0.047108523547649384, 0.09550058841705322, -0.1600627303123474, -0.0421348437666893, -0.17108748853206635, 0.1488579511642456, -0.01619901694357395, 0.025793075561523438, -0.013613137416541576, 0.09827471524477005, -0.14156213402748108, -0.07753695547580719, -0.05139590799808502, -0.010539999231696129, -0.08933109790086746, -0.041163619607686996, -0.03620695322751999, 0.08679662644863129, 0.03368169814348221, 0.2138427346944809, 0.059871092438697815, -0.002466253936290741, 0.09060985594987869, 0.005926032084971666, 0.012216760776937008, 0.22458888590335846, -0.05971982702612877, -0.014278420247137547, 0.07741286605596542, -0.08387523144483566, 0.062083370983600616, -0.06030460447072983, -0.12190768867731094, -0.11625868082046509, 0.14091411232948303, 0.03855528309941292, 0.008421613834798336, 0.022149836644530296, -0.12288839370012283, -0.01571831852197647, 0.12851041555404663, -0.041121501475572586, 0.03151547908782959, -0.017279237508773804, -0.047530584037303925, 0.03516228497028351, 0.10191090404987335, 0.048443179577589035, -0.03782786428928375, 0.06707210093736649, -0.11579618602991104, -0.016529586166143417, -0.02318665012717247, -0.10314241051673889, 0.054103244096040726, -0.06472698599100113, 0.05302289128303528, -0.23114216327667236, -0.07796528935432434, -0.00622726883739233, -0.04114042595028877, -0.020629925653338432, -0.06132553145289421, -0.05887884274125099, -0.03980230912566185, 0.04731769859790802, -0.019434785470366478, -0.0269989762455225, -0.061899472028017044, 0.002510189078748226, 0.0375392809510231, 0.045043375343084335, -0.16497555375099182, 0.023628706112504005, -0.0906074270606041, -0.011472806334495544, -0.15103866159915924, 0.03074527531862259, -0.006392898038029671, 0.0043692560866475105, -0.13664546608924866, -0.05168439447879791, 0.11929940432310104, -0.03489605337381363, 0.05808641016483307, 0.16650508344173431, -0.04368450492620468, -0.035187702625989914, -0.0046874843537807465, -0.06585010886192322, -0.08684049546718597, 0.10120736807584763, -0.037708621472120285, 0.0040193405002355576, 0.052341341972351074, -0.04686246067285538, 0.13342714309692383, -0.11502872407436371, -0.046388763934373856, 0.03328089043498039, -0.006056595593690872, -0.08925516903400421, 0.10620208084583282, 0.013611269183456898, -0.1382841318845749, 0.00018073485989589244, 0.03843579813838005, 0.03256182000041008, -0.07313674688339233, -0.1154869943857193, -0.031968310475349426, -0.01581697352230549, 0.040050335228443146, -0.05500555410981178, 0.056931525468826294, -0.034960631281137466, -0.09854013472795486, -0.0417928546667099, 0.13435949385166168, -0.010341890156269073, 0.0015235586324706674, -0.148587167263031, 0.09504818916320801, -0.16463051736354828, -0.06168526038527489, -0.14531579613685608, -0.07108397781848907, -0.0015823333524167538, 0.016759738326072693, -0.008082397282123566, -0.023273175582289696, 0.0111360102891922, 0.050389472395181656, 0.023052651435136795, -0.019994111731648445, 0.03716235235333443, -0.0017030293820425868, -0.12855401635169983, -0.08245803415775299, 0.015738051384687424, -0.01174601074308157, 0.31067436933517456, -0.09176350384950638, -0.01544173713773489, 0.03292283043265343, 0.06969179213047028, -0.03126794844865799, 0.05564762279391289, -0.012204703874886036, 0.026721101254224777, -0.07730726897716522, 0.0033350472804158926, 0.001752821495756507, -0.03420461341738701, -0.18150493502616882, 0.03500102460384369, -0.1814856082201004, 0.19086985290050507, 0.1793137937784195, -0.0323915109038353, -0.07653825730085373, 0.009960376657545567, 0.030910754576325417, 0.004698268603533506, -0.061669789254665375, 0.0027223757933825254, 0.12034238874912262, -0.0011985874734818935, 0.08197849243879318, -0.08435026556253433, -0.024305589497089386, 0.07069005817174911, -0.07676959782838821, -0.018202094361186028, 0.1535760462284088, 0.09069471061229706, -0.18531712889671326, 0.08762065321207047, 0.08616642653942108, -0.13321825861930847, 0.0028823381289839745, 0.0381956584751606, -0.058754049241542816, -0.02813303843140602, -0.06191183254122734, 0.03127354755997658, 0.06642688065767288, -0.019504327327013016, 0.06688462197780609, 0.09152504801750183, -0.026631053537130356, -0.0052734981290996075, -0.1602056324481964, -0.007784754503518343, 0.03258448466658592, 0.006257924716919661, -0.06585144996643066, -0.010542012751102448, 0.008463747799396515, 0.1364251673221588, 0.01895788311958313, -0.1533314287662506, 0.044715844094753265, 0.02588275633752346, -0.15116769075393677, 0.2790200710296631, -0.1095304936170578, -0.2717389762401581, -0.17332707345485687, -0.035868093371391296, -0.016476429998874664, 0.04994652420282364, 0.0552365817129612, -0.06159370392560959, -0.12975139915943146, -0.02900320664048195, -0.026257850229740143, -0.0037483610212802887, 0.06916821002960205, -0.0404052771627903, -0.06558417528867722, 0.006462525576353073, -0.08930076658725739, -0.03345920890569687, -0.02553979866206646, -0.01784939505159855, 0.17037348449230194, -0.07568042725324631, 0.13803493976593018, 0.2027701884508133, -0.08071716874837875, -0.03847036138176918, 0.024510588496923447, 0.23030681908130646, -0.0815969780087471, 0.014868916012346745, 0.13068342208862305, -0.0011967255268245935, 0.01511505339294672, 0.11342249065637589, -0.007847083732485771, -0.086739681661129, 0.02411210909485817, -0.030315671116113663, -0.05120199918746948, -0.2239675670862198, -0.16011562943458557, 0.009390046820044518, -0.044853463768959045, 0.0959387794137001, 0.019199850037693977, 0.16126243770122528, 0.16338412463665009, -0.016312556341290474, 0.06590583175420761, -0.05207838863134384, 0.09372264891862869, 0.18701498210430145, 0.037000056356191635, 0.14372973144054413, -0.059432704001665115, -0.10170459747314453, 0.08976226300001144, -0.03531564399600029, 0.06574315577745438, 0.09925679117441177, 0.032173041254282, -0.024608932435512543, 0.08073143661022186, 0.09845539182424545, 0.12053263932466507, 0.11722207069396973, -0.035145558416843414, -0.025858696550130844, -0.04472584277391434, -0.09304890781641006, 0.0793856829404831, 0.09340788424015045, 0.008513708598911762, -0.10789341479539871, 0.006919384002685547, 0.021551160141825676, 0.016048675402998924, 0.16443224251270294, -0.46391236782073975, -0.09841679781675339, 0.04461175948381424, -0.026140639558434486, -0.09698983281850815, -0.02634943649172783, -0.059577394276857376, -0.1556549221277237, 0.04191180318593979, 0.022079650312662125, 0.09171273559331894, -0.05046907812356949, -0.016941377893090248, -0.16032911837100983, 0.016899334266781807, -0.033698782324790955, 0.09354059398174286, -0.25268858671188354, 0.22710467875003815, 0.057456810027360916, 0.010990921407938004, -0.07765931636095047, 0.010183880105614662, -0.006665410008281469, 0.13615013659000397, 0.13551391661167145, 0.013972931541502476, 0.03871200606226921, -0.22900646924972534, -0.20246362686157227, 0.08177939802408218, -0.01700708456337452, -0.032290004193782806, 0.07269063591957092, 0.03455822169780731, -0.03229555860161781, 0.026595814153552055, -0.013690008781850338, -0.07755067944526672, -0.04118786379694939, 0.045328497886657715, 0.1324339210987091, 0.004203731659799814, 0.01354573667049408, -0.09947028756141663, -0.020712021738290787, 0.15367251634597778, -0.0787106305360794, -0.062091972678899765, -0.13643787801265717, 0.006445432547479868, 0.12552562355995178, -0.12010829895734787, 0.07537604868412018, -0.0424923412501812, 0.09805513173341751, -0.012166888453066349, -0.06817968934774399, 0.1316176801919937, -0.0667886808514595, -0.10751481354236603, 0.0052216555923223495, 0.0959848165512085, 0.048160769045352936, 0.08053936064243317, 0.07186415046453476, 0.0310895424336195, -0.09220913797616959, -0.1449040174484253, 0.014026669785380363, 0.06962840259075165, 0.1120351180434227, 0.08648525923490524, 0.05575823411345482, -0.14757172763347626, -0.06541517376899719, 0.04968224838376045, 0.17036937177181244, 0.2387121021747589, -0.06810152530670166, -0.00656583346426487, 0.15664799511432648, -0.00661261472851038, -0.20103971660137177, -0.017709046602249146, -0.00836032722145319, 0.09386202692985535, -0.18759849667549133, -0.05450625345110893, 0.10446621477603912, 0.12653324007987976, -0.04434274882078171, 0.0038724688347429037, -0.15762829780578613, -0.16033688187599182, 0.29457026720046997, 0.03756365925073624, 0.18626651167869568, -0.04250812157988548, -0.004127769730985165, -0.1421348601579666, -0.21718370914459229, 0.1952698975801468, -0.007240418344736099, 0.08954259008169174, -0.0055177961476147175, 0.1375560760498047, 0.03201489895582199, -0.015131666325032711, 0.21683289110660553, 0.03232165053486824, 0.01511146780103445, 0.0014762090286239982, -0.0966988354921341, -0.030079778283834457, -0.050866372883319855, 0.11832018196582794, 0.06663764268159866, 0.04270285367965698, -0.1556783765554428, -0.05347897857427597, -0.007412292063236237, 0.12590157985687256, -0.018162686377763748, -0.07687321305274963, -0.054587118327617645, -0.031439222395420074, -0.021954113617539406, -0.054645948112010956, 0.09521596133708954, -0.009052255190908909, 0.03075173869729042, 0.09287427365779877, 0.16666099429130554, -0.03955848515033722, 0.04795282334089279, 0.0447411946952343, -0.10122206807136536, 0.10856204479932785, -0.15541201829910278, 0.08175928145647049, 0.12455657869577408, -0.02087196707725525, 0.09101640433073044, 0.0408208929002285, -0.08278921246528625, -0.006828421261161566, 0.07771751284599304, -0.14860080182552338, 0.08798223733901978, 0.0028736128006130457, 0.005380753427743912, -0.021893059834837914, 0.07769890129566193, 0.14617882668972015, -0.05648209899663925, -0.052244700491428375, 0.009890173561871052, -0.015148996375501156, -0.021929558366537094, 0.22829721868038177, 0.062496889382600784, 0.07377849519252777, -0.13529492914676666, 0.046831078827381134, 0.045096077024936676, -0.03558945655822754, 0.01746317744255066, -0.046065546572208405, -0.12982460856437683, -0.08414240926504135, -0.004548175260424614, 0.09516186267137527, -0.31534048914909363, -0.08676221966743469, -0.03162135183811188, -0.08195380121469498, 0.0692586824297905, 0.2443096786737442, 0.12702544033527374, 0.051664408296346664, 0.0038440998177975416, -0.10087959468364716, -0.10054574906826019, -0.030939649790525436, 0.0534910149872303, 0.05114327743649483, -0.0871899202466011, 0.11361782997846603, -0.01420174352824688, 0.0835270807147026, -0.05282573401927948, -0.025419650599360466, -0.1309695541858673, 0.002737616654485464, -0.07438557595014572, 0.07858788222074509, -0.04636429250240326, 0.02375989779829979, 0.011920254677534103, -0.07080765068531036, -0.08831125497817993, 0.01790793612599373, -0.08403276652097702, -0.001495140721090138, -0.0072782025672495365, 0.03003641590476036, -0.06617090106010437, -0.05782127380371094, 0.05867442116141319, -0.029246702790260315, 0.07923034578561783, 0.13890761137008667, 0.06734052300453186, 0.09030940383672714, -0.14400754868984222, 0.005555631592869759, 0.15770511329174042, 0.0303849745541811, 0.12052861601114273, -0.1840667575597763, 0.05101122707128525, 0.056371964514255524, 0.018194086849689484, 0.03999694809317589, 0.09084766358137131, -0.11368520557880402, -0.0036813300102949142, -0.10098027437925339, -0.11656972765922546, -0.12211062759160995, 0.000285684916889295, 0.10269557684659958, 0.01665055938065052, 0.07861901819705963, -0.020292289555072784, 0.04668765887618065, -0.08762221783399582, 0.04415947198867798, -0.08295751363039017, -0.06653770804405212, -0.09672988206148148, -0.04894617944955826, 0.059282731264829636, -0.029922490939497948, 0.11559800803661346, -0.09745955467224121, 0.1606612205505371, -0.0008538000984117389, 0.07870013266801834, 0.02562076784670353, 0.009302262216806412, 0.14172258973121643, 0.1458745002746582, -0.008880316279828548, 0.0504821352660656, 0.0891568511724472, 0.11809016764163971, -0.05585111305117607, 0.06963326036930084, -0.05745097994804382, 0.005631318315863609, 0.193465918302536, -0.008794233202934265, -0.09139040857553482, -0.02588271163403988, -0.04462194815278053, -0.11617127805948257, -0.0030069611966609955, 0.06258714944124222, 0.021961750462651253, 0.12055636942386627, -0.08756040036678314, -0.05395127832889557, -0.031709738075733185, -0.06678589433431625, -0.2157716453075409, -0.04024120792746544, -0.17297258973121643, -0.049818217754364014, -0.019246967509388924, -0.11983627080917358, -0.08267858624458313, 0.08511125296354294, 0.055234428495168686, -0.03295949101448059, 0.051340021193027496, -0.0740966722369194, -0.031654082238674164, -0.004226196091622114, 0.01850728876888752, 0.039598122239112854, -0.03982819989323616, -0.0537402518093586, 0.031514834612607956, 0.045503221452236176, 0.051417071372270584, -0.012961965054273605, 0.06463057547807693, 0.1279970109462738, -0.005303368903696537, -0.10447318106889725, -0.05125473812222481, 0.03934040665626526, 0.06015203520655632, 0.012910728342831135, 0.0175277441740036, 0.03180958330631256, 0.023676857352256775, 0.17667247354984283, -0.07016992568969727, -0.010170217603445053, -0.13675126433372498, 0.29085442423820496, -0.024815451353788376, 0.08836212009191513, 0.05310645326972008, -0.04750236123800278, -0.020884443074464798, 0.13404902815818787, 0.0965738371014595, -0.019242923706769943, 0.021188199520111084, -0.00598514499142766, -0.002172438893467188, -0.0012576160952448845, 0.0014601723523810506, 0.05453918129205704, 0.14468760788440704, -0.11232065409421921, -0.006672990042716265, -0.010991916060447693, 0.01006621215492487, 0.01246911846101284, 0.026092572137713432, -0.016010623425245285, -0.023759251460433006, -0.07982534170150757, 0.06268375366926193, -0.07391146570444107, 0.0240751001983881, 0.0686650350689888, -0.117221400141716, -0.14028048515319824, 0.03132259473204613, -0.10381729155778885, -0.013184194453060627, 0.11337950080633163, -0.10155270248651505, -0.05966304615139961, 0.024281449615955353, 0.04420354962348938, -0.19934168457984924, -0.05898485705256462, 0.05533250421285629, 0.1791120022535324, 0.15831321477890015, 0.03596927225589752, 0.20677439868450165, 0.1289789229631424, 0.048773471266031265, -0.08775731921195984, 0.14491356909275055, 0.037214357405900955, -0.07352680712938309, 0.14898565411567688, -0.0008464193088002503, -0.017002882435917854, 0.08002186566591263, 0.07907777279615402, -0.14338770508766174, 0.028117220848798752, -0.08999509364366531, 0.03491155803203583, -0.06666839867830276, -0.016860635951161385, -0.11999229341745377, 0.09808820486068726, 0.07742418348789215, -0.06622271239757538, -0.0638047456741333, -0.011779449880123138, 0.10807527601718903, 0.04366699978709221, -0.14039020240306854, -0.015383901074528694, -0.09790745377540588, 0.08495927602052689, -0.03141404315829277, 0.04049116000533104, -0.16700120270252228, 0.0005714084836654365, -0.06714705377817154, -0.047905899584293365, -0.045591771602630615, 0.07218804210424423, -0.033406276255846024, 0.0224628746509552, -0.0561470091342926, -0.11307048797607422, -0.006078596226871014, 0.08013374358415604, -0.0764738917350769, -0.18012380599975586 ]
null
null
transformers
# Sentiment Classification for hinglish text: `gk-hinglish-sentiment` ## Model description Trained small amount of reviews dataset ## Intended uses & limitations I wanted something to work well with hinglish data as it is being used in India mostly. The training data was not much as expected #### How to use ```python #sample code from transformers import BertTokenizer, BertForSequenceClassification tokenizerg = BertTokenizer.from_pretrained("/content/model") modelg = BertForSequenceClassification.from_pretrained("/content/model") text = "kuch bhi type karo hinglish mai" encoded_input = tokenizerg(text, return_tensors='pt') output = modelg(**encoded_input) print(output) #output contains 3 lables LABEL_0 = Negative ,LABEL_1 = Nuetral ,LABEL_2 = Positive ``` #### Limitations and bias The data contains only hinglish codemixed text it and was very much limited may be I will Update this model if I can get good amount of data ## Training data Training data contains labeled data for 3 labels link to the pre-trained model card with description of the pre-training data. I have Tuned below model https://huggingface.co/rohanrajpal/bert-base-multilingual-codemixed-cased-sentiment ### BibTeX entry and citation info ```@inproceedings{khanuja-etal-2020-gluecos, title = "{GLUEC}o{S}: An Evaluation Benchmark for Code-Switched {NLP}", author = "Khanuja, Simran and Dandapat, Sandipan and Srinivasan, Anirudh and Sitaram, Sunayana and Choudhury, Monojit", booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics", month = jul, year = "2020", address = "Online", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/2020.acl-main.329", pages = "3575--3585" } ```
{"license": "apache-2.0", "tags": ["sentiment", "multilingual", "hindi codemix", "hinglish"], "datasets": ["sail"], "language_bcp47": ["hi-en"]}
text-classification
ganeshkharad/gk-hinglish-sentiment
[ "transformers", "pytorch", "jax", "safetensors", "bert", "text-classification", "sentiment", "multilingual", "hindi codemix", "hinglish", "dataset:sail", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #safetensors #bert #text-classification #sentiment #multilingual #hindi codemix #hinglish #dataset-sail #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
# Sentiment Classification for hinglish text: 'gk-hinglish-sentiment' ## Model description Trained small amount of reviews dataset ## Intended uses & limitations I wanted something to work well with hinglish data as it is being used in India mostly. The training data was not much as expected #### How to use #### Limitations and bias The data contains only hinglish codemixed text it and was very much limited may be I will Update this model if I can get good amount of data ## Training data Training data contains labeled data for 3 labels link to the pre-trained model card with description of the pre-training data. I have Tuned below model URL ### BibTeX entry and citation info
[ "# Sentiment Classification for hinglish text: 'gk-hinglish-sentiment'", "## Model description\n\nTrained small amount of reviews dataset", "## Intended uses & limitations\n\nI wanted something to work well with hinglish data as it is being used in India mostly.\nThe training data was not much as expected", "#### How to use", "#### Limitations and bias\n\nThe data contains only hinglish codemixed text it and was very much limited may be I will Update this model if I can get good amount of data", "## Training data\n\nTraining data contains labeled data for 3 labels\n\nlink to the pre-trained model card with description of the pre-training data.\nI have Tuned below model\n\nURL", "### BibTeX entry and citation info" ]
[ "TAGS\n#transformers #pytorch #jax #safetensors #bert #text-classification #sentiment #multilingual #hindi codemix #hinglish #dataset-sail #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Sentiment Classification for hinglish text: 'gk-hinglish-sentiment'", "## Model description\n\nTrained small amount of reviews dataset", "## Intended uses & limitations\n\nI wanted something to work well with hinglish data as it is being used in India mostly.\nThe training data was not much as expected", "#### How to use", "#### Limitations and bias\n\nThe data contains only hinglish codemixed text it and was very much limited may be I will Update this model if I can get good amount of data", "## Training data\n\nTraining data contains labeled data for 3 labels\n\nlink to the pre-trained model card with description of the pre-training data.\nI have Tuned below model\n\nURL", "### BibTeX entry and citation info" ]
[ 76, 20, 11, 36, 5, 39, 39, 11 ]
[ "passage: TAGS\n#transformers #pytorch #jax #safetensors #bert #text-classification #sentiment #multilingual #hindi codemix #hinglish #dataset-sail #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Sentiment Classification for hinglish text: 'gk-hinglish-sentiment'## Model description\n\nTrained small amount of reviews dataset## Intended uses & limitations\n\nI wanted something to work well with hinglish data as it is being used in India mostly.\nThe training data was not much as expected#### How to use#### Limitations and bias\n\nThe data contains only hinglish codemixed text it and was very much limited may be I will Update this model if I can get good amount of data## Training data\n\nTraining data contains labeled data for 3 labels\n\nlink to the pre-trained model card with description of the pre-training data.\nI have Tuned below model\n\nURL### BibTeX entry and citation info" ]
[ -0.11011651903390884, 0.14895471930503845, 0.0017037816578522325, 0.08473934233188629, 0.08066248148679733, 0.024449340999126434, 0.1029137521982193, 0.1392749547958374, -0.04162008315324783, 0.024849167093634605, 0.025277048349380493, -0.01351164747029543, 0.10128940641880035, 0.16128334403038025, -0.006241241004317999, -0.26852092146873474, 0.0409235805273056, 0.0004244340234436095, 0.013529281131923199, 0.10920552909374237, 0.12802420556545258, -0.11280661076307297, 0.0769961029291153, 0.0004897924955002964, -0.10481490939855576, 0.010378049686551094, -0.04960929974913597, -0.03466230258345604, 0.06048839911818504, 0.007489390671253204, 0.10093207657337189, 0.032869596034288406, 0.10327066481113434, -0.2286578118801117, 0.0446445494890213, 0.017262687906622887, 0.0007563924300484359, 0.06285475939512253, 0.039818793535232544, -0.038824547082185745, 0.18310464918613434, -0.04526492953300476, 0.06823598593473434, 0.054748378694057465, -0.07645022869110107, -0.11016366630792618, -0.09001288563013077, 0.1338745653629303, 0.0783805251121521, 0.1059640496969223, -0.03311603143811226, 0.08466704189777374, -0.09492061287164688, 0.034979213029146194, 0.09942220151424408, -0.18592870235443115, 0.007995308376848698, 0.1428210586309433, 0.03444189578294754, 0.10466252267360687, -0.1669418066740036, -0.015316309407353401, 0.09144335985183716, -0.004005326423794031, 0.0016136865597218275, -0.02270563133060932, -0.05705713853240013, 0.02201901376247406, -0.06139875203371048, 0.0036158361472189426, 0.1321856826543808, -0.008863190189003944, -0.052346061915159225, -0.11415858566761017, 0.007596632931381464, -0.1269889920949936, -0.01185210794210434, 0.04507140442728996, 0.03363199159502983, 0.013934498652815819, 0.16214686632156372, -0.04081864655017853, -0.12638424336910248, -0.014498088508844376, 0.05435972288250923, 0.01447997335344553, 0.029756315052509308, 0.008043362759053707, -0.09603457152843475, 0.05923287197947502, -0.013032219372689724, -0.07834898680448532, -0.06825119256973267, -0.026775987818837166, -0.03580799698829651, -0.01144812535494566, -0.04134447127580643, -0.043936002999544144, -0.015733752399683, 0.05365478992462158, -0.1210334300994873, 0.07246311008930206, -0.04000142961740494, 0.01367331575602293, 0.09267356991767883, 0.07897153496742249, -0.09863945841789246, 0.018508857116103172, 0.06777659803628922, 0.07348603755235672, 0.09531258046627045, -0.032240431755781174, -0.07577890902757645, 0.0032908853609114885, 0.03602796792984009, 0.07961423695087433, -0.10339846462011337, 0.029409585520625114, -0.07058608531951904, -0.03621382266283035, 0.1017632707953453, -0.14468657970428467, -0.019584622234106064, -0.008216929621994495, -0.04194033890962601, -0.06032616272568703, 0.04097175598144531, 0.009668301790952682, 0.0009059603326022625, 0.02304452657699585, -0.09500539302825928, 0.04455481842160225, -0.10314861685037613, -0.0789259746670723, -0.029322080314159393, -0.11409644037485123, -0.022311000153422356, -0.05001264810562134, -0.26969149708747864, -0.040562085807323456, 0.029325127601623535, -0.09744726121425629, 0.010014968924224377, -0.08209715783596039, -0.0083425622433424, 0.032408397644758224, -0.006371765397489071, 0.07795925438404083, -0.05696309357881546, 0.05900399014353752, -0.04882630705833435, 0.0848114863038063, -0.01689661666750908, 0.05397610366344452, -0.13761204481124878, 0.025934386998414993, -0.02507500909268856, 0.039829984307289124, -0.10187385976314545, -0.00563831627368927, -0.08710142970085144, -0.058553703129291534, 0.029735196381807327, 0.026606803759932518, -0.005918145179748535, 0.14942567050457, -0.21201777458190918, -0.011043909005820751, 0.13824249804019928, -0.18724071979522705, -0.09263891726732254, 0.05924459919333458, -0.06583596765995026, 0.14481952786445618, 0.04624338075518608, 0.14678482711315155, 0.06459823995828629, -0.039940670132637024, -0.044734686613082886, 0.028238670900464058, 0.010569659993052483, 0.042940255254507065, 0.03466178849339485, -0.030497359111905098, -0.03288804739713669, 0.0016153340693563223, -0.004628919996321201, -0.042913686484098434, -0.0428704172372818, -0.09597958624362946, -0.002168654929846525, -0.07789355516433716, 0.04557494446635246, 0.07774603366851807, 0.03761100396513939, -0.0432799868285656, -0.08004902303218842, 0.036686576902866364, 0.1654079705476761, -0.040241457521915436, 0.03651369735598564, -0.09636647254228592, 0.08979182690382004, -0.09586505591869354, -0.057081278413534164, -0.15732987225055695, 0.047258410602808, 0.022023657336831093, -0.014116799458861351, 0.03679487481713295, 0.07044865936040878, 0.05535157769918442, 0.033186472952365875, -0.07571329921483994, -0.04269232228398323, -0.005021608900278807, 0.03512425720691681, -0.10649579018354416, -0.24240857362747192, -0.019072895869612694, -0.04812668263912201, 0.24995781481266022, -0.23961615562438965, -0.003547310596331954, 0.07639423757791519, 0.1299855262041092, 0.0342647023499012, -0.04261422157287598, 0.1195875033736229, 0.017390374094247818, -0.0565420500934124, -0.08376199752092361, 0.032104428857564926, 0.006108064204454422, -0.061524294316768646, 0.050185732543468475, -0.09380689263343811, -0.07606122642755508, 0.0977397933602333, 0.010831661522388458, -0.06853270530700684, 0.0038970333989709616, -0.009674823842942715, -0.01853562332689762, -0.07218901813030243, 0.05629713460803032, 0.07605573534965515, -0.0006868555792607367, 0.13949339091777802, -0.07274647057056427, -0.08392277359962463, -0.021381665021181107, -0.05626380443572998, -0.052941519767045975, 0.07703560590744019, 0.03963283821940422, -0.13899534940719604, 0.14773277938365936, 0.039076607674360275, -0.014952403493225574, 0.21469227969646454, -0.008089855313301086, -0.07927417010068893, 0.02106281742453575, -0.03382044658064842, -0.047438349574804306, 0.16375236213207245, -0.1005602478981018, -0.022607313469052315, 0.031113017350435257, -0.01680683344602585, 0.038298871368169785, -0.08823269605636597, -0.015512843616306782, 0.009007113054394722, -0.039337094873189926, -0.045741308480501175, 0.018529118970036507, -0.04263071343302727, 0.10584335774183273, 0.008171798661351204, 0.057878609746694565, 0.032167330384254456, -0.015562665648758411, -0.13005203008651733, 0.16045549511909485, -0.1348964422941208, -0.22844430804252625, -0.09982170164585114, 0.0596885122358799, 0.009975693188607693, -0.005258181598037481, 0.0004212323692627251, -0.2019144594669342, -0.09081823378801346, -0.11169256269931793, -0.020570311695337296, -0.06051740422844887, -0.028615031391382217, -0.1312512904405594, -0.011274921707808971, 0.00988677516579628, -0.13567708432674408, 0.028897693380713463, 0.021420905366539955, -0.037516988813877106, 0.046226926147937775, -0.01393873617053032, 0.0265456885099411, 0.15223874151706696, -0.03902767226099968, 0.024675102904438972, -0.018239108845591545, 0.12361615151166916, -0.0672340840101242, 0.09541736543178558, 0.0784909650683403, 0.026287073269486427, 0.00971631333231926, 0.16621871292591095, -0.007325123064219952, -0.10099892318248749, 0.01989494264125824, 0.06897319108247757, -0.0028088358230888844, -0.24031996726989746, -0.08206072449684143, -0.04955059662461281, -0.06531845778226852, 0.034367531538009644, 0.0530252531170845, 0.09034586697816849, 0.07147897034883499, -0.06896597892045975, 0.02179824560880661, -0.06614183634519577, 0.07383548468351364, -0.024055520072579384, 0.02664880082011223, 0.07235585898160934, -0.08311749994754791, 0.07260673493146896, 0.11076744645833969, 0.010847180150449276, 0.2118268460035324, -0.04084618762135506, 0.14571870863437653, 0.10061341524124146, 0.1322021186351776, 0.05113285034894943, 0.005610309541225433, -0.10040590167045593, -0.023872079327702522, -0.0034543427173048258, -0.048118844628334045, -0.07678686827421188, 0.10773013532161713, -0.016191724687814713, 0.0614316463470459, -0.04782015457749367, -0.05807383358478546, 0.002669024746865034, 0.11866000294685364, 0.07266757637262344, -0.17550891637802124, -0.10404466092586517, 0.07426513731479645, -0.017796004191040993, -0.07677967846393585, 0.027276117354631424, 0.08383530378341675, -0.07754573225975037, 0.07501150667667389, -0.04839281737804413, 0.07405126094818115, -0.12114713340997696, -0.04466729238629341, -0.04612940177321434, -0.012092065997421741, -0.04988464340567589, 0.1321752369403839, -0.24043865501880646, 0.21249602735042572, 0.03888741135597229, 0.07135169953107834, -0.051810555160045624, -0.03868313133716583, 0.028902294114232063, 0.12693944573402405, 0.07784352451562881, 0.022578509524464607, 0.03755385801196098, -0.15189532935619354, -0.005973072722554207, 0.007516910787671804, 0.07084109634160995, 0.004959377925843, 0.08640047907829285, -0.01843167282640934, 0.017018435522913933, -0.010629643686115742, -0.09637478739023209, -0.19986675679683685, -0.14258389174938202, 0.062352076172828674, -0.02592073567211628, 0.02001095935702324, -0.07659276574850082, -0.10648319125175476, 0.02396589331328869, 0.1581098586320877, -0.08043550699949265, -0.11637861281633377, -0.11128321290016174, 0.10572060197591782, 0.048781607300043106, -0.05176859721541405, 0.04795307666063309, 0.005315869115293026, 0.1204460859298706, -0.013383524492383003, -0.029501743614673615, 0.04757252708077431, -0.054329678416252136, -0.145344078540802, -0.08004546910524368, 0.11194836348295212, 0.126335009932518, 0.03385884314775467, 0.013669249601662159, 0.025773363187909126, 0.027427859604358673, -0.08935105800628662, 0.021082255989313126, 0.13228854537010193, 0.07585938274860382, 0.09321379661560059, -0.04007936269044876, -0.04365327209234238, -0.0005914807552471757, -0.11821837723255157, 0.08532961457967758, 0.13758660852909088, -0.05732313543558121, 0.199515238404274, 0.1509612649679184, -0.13154570758342743, -0.16698889434337616, 0.07852131873369217, -0.041963472962379456, 0.07555999606847763, -0.027582598850131035, -0.1775694191455841, 0.054605238139629364, 0.012174099683761597, -0.03639068454504013, -0.000754020526073873, -0.19345787167549133, -0.10497473180294037, 0.08096688240766525, 0.13895654678344727, 0.0352853462100029, -0.1287258118391037, -0.08439427614212036, -0.044112879782915115, -0.12401866167783737, 0.10919132828712463, -0.123200424015522, 0.05460028350353241, -0.05807236209511757, 0.05746900662779808, 0.022142773494124413, -0.0423554889857769, 0.14005251228809357, 0.013700314797461033, 0.06257063150405884, -0.10650843381881714, -0.018859606236219406, 0.09881211817264557, -0.007466333452612162, 0.14953172206878662, -0.04414396360516548, 0.04597979784011841, -0.18682122230529785, -0.025323709473013878, -0.06491289287805557, 0.025006191805005074, -0.07697490602731705, -0.10152176022529602, -0.09363140165805817, 0.1588529348373413, 0.029454436153173447, -0.004878702573478222, 0.02707020565867424, -0.07386358827352524, 0.05035033077001572, 0.05116860568523407, 0.21743544936180115, 0.027586430311203003, 0.041914649307727814, -0.032036278396844864, -0.03215549886226654, 0.06711931526660919, -0.2463936060667038, -0.007859122939407825, 0.043510597199201584, 0.03937249630689621, 0.1871194988489151, -0.009405788965523243, -0.05756577476859093, 0.052012454718351364, 0.10507369041442871, -0.039942964911460876, -0.14357849955558777, 0.011144906282424927, -0.007840157486498356, -0.10038477182388306, -0.018466832116246223, 0.056922826915979385, -0.10441011190414429, -0.04192099720239639, -0.0250852070748806, 0.06027838960289955, -0.03435419127345085, 0.08998747169971466, 0.15966418385505676, 0.042578041553497314, -0.08649203181266785, 0.09479927271604538, 0.0844031274318695, -0.0703277587890625, 0.08142094314098358, 0.1016070544719696, -0.1412060260772705, -0.09337478131055832, -0.0202882532030344, 0.2176515907049179, -0.03471535071730614, -0.04717050865292549, -0.06568007171154022, -0.05969351530075073, 0.006490401457995176, 0.08084110170602798, 0.019082413986325264, 0.04438788443803787, -0.06671857088804245, -0.024177107959985733, -0.13543601334095, 0.08975057303905487, 0.0326988510787487, 0.05303558334708214, -0.10451344400644302, 0.06370202451944351, 0.048319533467292786, 0.03458067774772644, -0.028750263154506683, -0.020621666684746742, -0.04904673993587494, -0.012069902382791042, -0.1700633466243744, 0.060946159064769745, -0.06277210265398026, 0.00844039861112833, -0.042327798902988434, -0.06782194972038269, -0.012294464744627476, 0.01923675276339054, -0.051502589136362076, 0.008652706630527973, 0.0016287051839753985, 0.08092372864484787, -0.12786749005317688, -0.013535958714783192, 0.041060544550418854, -0.022155797109007835, 0.08245047181844711, -0.008929150179028511, -0.021037284284830093, 0.05950482189655304, -0.14352062344551086, 0.06240759789943695, -0.044233933091163635, 0.01091086957603693, 0.049939073622226715, -0.13574731349945068, -0.009382291696965694, 0.011950559914112091, 0.002395294839516282, 0.04115282744169235, 0.13409999012947083, -0.08694258332252502, 0.04032820090651512, -0.09423203766345978, 0.035456132143735886, -0.05821262672543526, 0.15786880254745483, 0.10202871263027191, -0.010429298505187035, 0.12329665571451187, -0.08186686038970947, 0.021862203255295753, -0.1374802440404892, 0.0019082998624071479, -0.028755517676472664, -0.01163881178945303, -0.17590223252773285, -0.03587917983531952, 0.04732298105955124, -0.0678880587220192, 0.11033192276954651, 0.023081181570887566, 0.01744004897773266, 0.05330529063940048, -0.013849136419594288, 0.030014120042324066, 0.005646503064781427, 0.2327987402677536, 0.037951789796352386, -0.00823636632412672, 0.03693629056215286, 0.006455308757722378, -0.009087838232517242, 0.046748410910367966, 0.09452935308218002, 0.15495073795318604, 0.053716808557510376, 0.06388985365629196, -0.054050546139478683, -0.12460768967866898, -0.059498436748981476, 0.11716003715991974, -0.06348079442977905, 0.05513453483581543, -0.02542722038924694, 0.0661725401878357, 0.24062183499336243, -0.14619116485118866, 0.07046124339103699, -0.05487637221813202, -0.11422606557607651, -0.04244283586740494, -0.13484929502010345, -0.07579965144395828, -0.07717911899089813, 0.020573550835251808, -0.12029773741960526, 0.029454268515110016, 0.12759941816329956, 0.054407984018325806, -0.015596522018313408, 0.17342865467071533, -0.017446953803300858, -0.0032578720711171627, 0.033289942890405655, 0.017681235447525978, -0.000830429547931999, -0.0927174910902977, 0.0050493162125349045, 0.0052505237981677055, -0.04304954409599304, 0.020034508779644966, 0.02439875528216362, 0.024441629648208618, 0.0077980817295610905, -0.014407348819077015, -0.08894862979650497, 0.050459958612918854, 0.029975857585668564, 0.07329777628183365, 0.1646231710910797, 0.03958634287118912, 0.006921559106558561, -0.007370812352746725, 0.21198873221874237, -0.03301255777478218, -0.08035994321107864, -0.09284455329179764, 0.15477068722248077, 0.04123422130942345, -0.010878893546760082, 0.059721674770116806, -0.1393931657075882, 0.03832901641726494, 0.15900832414627075, 0.1638994812965393, -0.08242863416671753, -0.00018812173220794648, -0.0244046151638031, 0.005733228288590908, -0.04486912116408348, 0.09110430628061295, 0.023622483015060425, 0.061751220375299454, -0.10680557042360306, 0.033761825412511826, -0.062023282051086426, -0.04176441207528114, -0.0428578145802021, 0.11306673288345337, -0.007548397406935692, -0.038365431129932404, -0.06908781081438065, 0.06477413326501846, 0.0032216378021985292, -0.17689640820026398, 0.05164894834160805, -0.10701833665370941, -0.1581723690032959, -0.038958389312028885, -0.03658657148480415, 0.03655251860618591, 0.000049134952860185876, 0.03017076849937439, 0.02110322192311287, 0.14862370491027832, 0.04861077666282654, -0.09380633383989334, -0.13420934975147247, 0.13374866545200348, -0.08163338154554367, 0.16357803344726562, -0.01896577887237072, 0.060871969908475876, 0.09311705827713013, 0.026718685403466225, -0.10381889343261719, 0.06262878328561783, 0.04615583270788193, 0.11252973228693008, 0.0007910921121947467, 0.18908683955669403, 0.003219463163986802, 0.039155907928943634, 0.05016901716589928, -0.06206659600138664, 0.05899425223469734, -0.06736349314451218, -0.07145468145608902, -0.09363831579685211, 0.10324480384588242, -0.03911568596959114, 0.16512735188007355, 0.1006571501493454, -0.0638880804181099, -0.04752890765666962, -0.05784275755286217, 0.03903897851705551, 0.0017931177280843258, 0.06109023094177246, 0.004698011092841625, -0.12421345710754395, -0.0029921161476522684, -0.03129110857844353, 0.07668249309062958, -0.31257590651512146, -0.011352410539984703, -0.0033766431733965874, -0.002395466435700655, -0.08166536688804626, 0.12391093373298645, 0.04630453512072563, 0.056418877094984055, -0.035572607070207596, -0.019414367154240608, -0.016676289960741997, 0.10236041247844696, -0.10221491754055023, -0.054835256189107895 ]
null
null
transformers
## Generating Chinese poetry by topic. ```python from transformers import * tokenizer = BertTokenizer.from_pretrained("gaochangkuan/model_dir") model = AutoModelWithLMHead.from_pretrained("gaochangkuan/model_dir") prompt= '''<s>田园躬耕''' length= 84 stop_token='</s>' temperature = 1.2 repetition_penalty=1.3 k= 30 p= 0.95 device ='cuda' seed=2020 no_cuda=False prompt_text = prompt if prompt else input("Model prompt >>> ") encoded_prompt = tokenizer.encode( '<s>'+prompt_text+'<sep>', add_special_tokens=False, return_tensors="pt" ) encoded_prompt = encoded_prompt.to(device) output_sequences = model.generate( input_ids=encoded_prompt, max_length=length, min_length=10, do_sample=True, early_stopping=True, num_beams=10, temperature=temperature, top_k=k, top_p=p, repetition_penalty=repetition_penalty, bad_words_ids=None, bos_token_id=tokenizer.bos_token_id, pad_token_id=tokenizer.pad_token_id, eos_token_id=tokenizer.eos_token_id, length_penalty=1.2, no_repeat_ngram_size=2, num_return_sequences=1, attention_mask=None, decoder_start_token_id=tokenizer.bos_token_id,) generated_sequence = output_sequences[0].tolist() text = tokenizer.decode(generated_sequence) text = text[: text.find(stop_token) if stop_token else None] print(''.join(text).replace(' ','').replace('<pad>','').replace('<s>','')) ```
{}
text-generation
gaochangkuan/model_dir
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
## Generating Chinese poetry by topic.
[ "## Generating Chinese poetry by topic." ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## Generating Chinese poetry by topic." ]
[ 50, 9 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## Generating Chinese poetry by topic." ]
[ 0.07378317415714264, 0.059403397142887115, -0.009274330921471119, 0.07456928491592407, 0.12417822331190109, -0.0015131225809454918, 0.17722734808921814, 0.08922062069177628, -0.06128685921430588, -0.04370338097214699, 0.1534433811903, 0.1726551502943039, 0.01961582712829113, -0.054326966404914856, 0.035195596516132355, -0.35446110367774963, -0.004737752955406904, 0.146819069981575, 0.06867927312850952, 0.13952437043190002, 0.09472241252660751, -0.05334277078509331, 0.09765265882015228, 0.07734508812427521, -0.06898780912160873, -0.053679417818784714, -0.0479736328125, -0.09816231578588486, 0.07669010013341904, -0.001170364674180746, 0.05517858639359474, 0.005957972723990679, -0.028127193450927734, -0.08633101731538773, 0.003530217567458749, -0.09346308559179306, -0.09867937117815018, -0.023795047774910927, 0.16836866736412048, -0.13493840396404266, 0.11911685764789581, 0.051007263362407684, -0.01918845623731613, 0.05241144821047783, -0.10364106297492981, -0.029521014541387558, 0.016922201961278915, 0.16625553369522095, 0.17775410413742065, 0.15130408108234406, -0.04852597787976265, 0.04489917680621147, -0.060550522059202194, 0.06548584997653961, 0.08401919156312943, -0.3885590732097626, -0.03885768726468086, 0.12454943358898163, 0.10029632598161697, 0.022413000464439392, -0.18010364472866058, 0.028814472258090973, -0.008858395740389824, 0.01873512752354145, -0.022574689239263535, -0.15563301742076874, -0.05812445655465126, 0.09177650511264801, -0.07271192967891693, 0.07433826476335526, 0.21944931149482727, -0.06942654401063919, 0.02379779703915119, -0.0491618886590004, -0.060916099697351456, -0.016995297744870186, -0.07227783650159836, -0.06676505506038666, -0.024185508489608765, 0.08013539016246796, 0.08740435540676117, -0.12638413906097412, -0.1182960495352745, -0.011648174375295639, -0.07169272750616074, -0.1325794905424118, -0.006667347624897957, -0.03393002226948738, -0.15294846892356873, 0.03330625593662262, 0.081951804459095, -0.055297691375017166, 0.016394348815083504, -0.08067959547042847, 0.04968179017305374, 0.042868200689554214, 0.05491194874048233, -0.1094440221786499, 0.008451657369732857, 0.10902269184589386, 0.15810440480709076, 0.07719594240188599, 0.04769591987133026, 0.0697113499045372, 0.09320417791604996, 0.03725022077560425, 0.016194462776184082, -0.0018329716986045241, -0.015430518426001072, -0.10769830644130707, -0.0016662090783938766, -0.012016336433589458, -0.16837459802627563, -0.00039815998752601445, 0.008123500272631645, -0.00876533892005682, -0.03267960250377655, 0.0805865153670311, -0.05209975689649582, -0.01712203025817871, 0.0685722604393959, 0.010518236085772514, -0.02985134907066822, -0.007796783931553364, 0.04806821420788765, 0.21058297157287598, -0.015096167102456093, 0.0559280626475811, -0.08450549095869064, 0.11849798262119293, -0.06829223036766052, 0.027120785787701607, 0.05176573991775513, 0.050467319786548615, 0.026532895863056183, -0.04084761068224907, 0.058157507330179214, -0.10582731664180756, -0.11942407488822937, -0.014091101475059986, 0.03910883888602257, -0.05803276598453522, -0.08760951459407806, -0.07647765427827835, -0.09402953833341599, -0.022410215809941292, -0.015418044291436672, -0.049299076199531555, -0.03340991958975792, 0.0954420417547226, -0.03564286231994629, 0.049893252551555634, -0.05798766016960144, 0.049772799015045166, -0.16008523106575012, -0.012549901381134987, -0.043922897428274155, 0.08827852457761765, 0.06528869271278381, 0.12395554035902023, 0.06424503773450851, 0.04140956327319145, -0.019950635731220245, -0.005435351748019457, -0.10978271812200546, 0.2227783203125, -0.15604403614997864, -0.10393333435058594, 0.16261173784732819, -0.05602044239640236, -0.10952801257371902, 0.11667442321777344, 0.001498973579145968, 0.12167612463235855, 0.06302674859762192, 0.18199551105499268, -0.07259368896484375, 0.03261179476976395, 0.035847436636686325, 0.09316319227218628, -0.04682350531220436, 0.04356013983488083, 0.10162217915058136, 0.07978245615959167, 0.03591103479266167, 0.05014045909047127, 0.07295544445514679, -0.027980607002973557, -0.010982089675962925, -0.10750158876180649, -0.034779489040374756, 0.024526923894882202, 0.06300435960292816, -0.03351979702711105, 0.1353229135274887, -0.026226112619042397, -0.024215005338191986, -0.04202618822455406, -0.01325639896094799, 0.05018090084195137, 0.0698511153459549, -0.05065396428108215, 0.05787665769457817, 0.15882433950901031, 0.0727485790848732, -0.10000181943178177, 0.10629260540008545, -0.024088451638817787, 0.10727198421955109, 0.04537355154752731, 0.03912816196680069, -0.02091112919151783, -0.03840883448719978, -0.04145682975649834, 0.09352655708789825, 0.07255510985851288, -0.045706335455179214, -0.07112587988376617, -0.08746643364429474, 0.10747382044792175, -0.025339817628264427, 0.07568403333425522, -0.19774878025054932, -0.013418333604931831, 0.00941446889191866, 0.05605190992355347, -0.030307842418551445, 0.05883379653096199, 0.0514347068965435, 0.026104599237442017, -0.12716640532016754, 0.0658961832523346, 0.0840202197432518, 0.025511484593153, -0.15644559264183044, 0.2843323349952698, -0.06912735849618912, 0.037262964993715286, 0.2047608643770218, -0.3241133689880371, -0.047862812876701355, -0.08844811469316483, -0.04919296130537987, -0.008921320550143719, 0.06912723928689957, 0.11340242624282837, 0.13396404683589935, -0.06089254096150398, 0.14944279193878174, -0.06470236927270889, 0.0029411581344902515, -0.03909458592534065, -0.08002657443284988, -0.013380002230405807, 0.16260185837745667, 0.0800720602273941, -0.06001846864819527, 0.2195592075586319, 0.1777707189321518, 0.018375102430582047, 0.1855705827474594, 0.08156708627939224, 0.07139342278242111, 0.014597193337976933, 0.0023465955164283514, -0.06805438548326492, 0.0046702162362635136, -0.21615396440029144, -0.014159157872200012, 0.016314750537276268, -0.0674767792224884, 0.0399567149579525, -0.15741078555583954, -0.14177343249320984, -0.07006918638944626, 0.0037002142053097486, 0.03427257388830185, 0.05094698816537857, -0.01777532510459423, 0.040646012872457504, -0.03118211403489113, 0.08412908017635345, 0.08012522011995316, 0.05862148851156235, -0.09926335513591766, 0.09582997113466263, -0.09388332068920135, -0.29976969957351685, -0.004402665421366692, -0.13806027173995972, 0.003893279703333974, 0.05823161453008652, 0.08197218179702759, -0.1768009513616562, -0.007240114733576775, -0.015388195402920246, 0.005557592958211899, -0.15148571133613586, -0.06384103745222092, -0.07895805686712265, 0.12011050432920456, -0.16500376164913177, 0.06512395292520523, -0.013606402091681957, -0.03467823192477226, -0.07937140762805939, 0.12756800651550293, -0.13331134617328644, 0.008548211306333542, 0.13901500403881073, 0.1330016553401947, -0.030982306227087975, -0.026474153622984886, 0.16787205636501312, -0.13079428672790527, 0.04870704561471939, 0.11119330674409866, -0.017704008147120476, 0.07423306256532669, 0.07011685520410538, 0.004591259639710188, -0.047032032161951065, 0.05684612691402435, 0.03122173249721527, -0.030850743874907494, -0.255045086145401, -0.024560371413826942, -0.14320063591003418, 0.19112449884414673, -0.04477856308221817, 0.1057572141289711, 0.10227607935667038, 0.13635002076625824, -0.0023028189316391945, 0.11973502486944199, 0.001473330776207149, 0.064569391310215, 0.2015989124774933, 0.001368291792459786, 0.07515885680913925, -0.06301524490118027, -0.06710293889045715, 0.040825989097356796, 0.05993882566690445, 0.05291176214814186, 0.1401621252298355, 0.07342954725027084, 0.056070566177368164, 0.2285267412662506, 0.15943990647792816, -0.041586775332689285, -0.002572324126958847, -0.0334487147629261, -0.07937317341566086, -0.0823047012090683, 0.02345254085958004, 0.13819418847560883, -0.0012919498840346932, -0.19003289937973022, 0.0741884633898735, 0.040791191160678864, 0.08115977793931961, 0.08310011774301529, 0.06345392018556595, -0.043867889791727066, 0.039149533957242966, 0.11663733422756195, -0.01927816867828369, -0.05231683701276779, 0.13647933304309845, -0.051424089819192886, -0.24993248283863068, 0.07623285055160522, -0.04949132353067398, 0.16475655138492584, -0.04045172408223152, 0.077780582010746, -0.077279232442379, -0.20656359195709229, 0.061386480927467346, 0.06021690368652344, -0.35175201296806335, 0.18525417149066925, -0.003692023688927293, -0.0648394450545311, -0.052848707884550095, -0.04108677804470062, 0.014488964341580868, 0.0461423322558403, 0.17777714133262634, -0.029070552438497543, 0.006800380069762468, 0.010090556927025318, -0.03858660161495209, 0.05346846953034401, 0.11350304633378983, -0.030232340097427368, -0.05725023150444031, 0.006778911221772432, -0.02745799720287323, -0.0847703218460083, 0.11398250609636307, -0.05271972343325615, -0.19287194311618805, 0.04732023552060127, -0.012442744337022305, 0.07427062094211578, 0.10104230046272278, -0.014217104762792587, 0.03874858841300011, 0.13221359252929688, -0.0807994157075882, -0.1116875410079956, -0.03687199950218201, -0.11840791255235672, -0.026244409382343292, -0.04135533794760704, 0.013838529586791992, -0.1165199726819992, -0.03933307155966759, -0.06098619103431702, -0.07001818716526031, 0.07597991824150085, -0.03419191390275955, -0.03765903785824776, -0.06761228293180466, 0.17671824991703033, -0.0022925653029233217, 0.08803817629814148, 0.016686050221323967, -0.01366477645933628, -0.07717088609933853, -0.1392911672592163, -0.007816074416041374, -0.1695513278245926, -0.027327440679073334, -0.04830673336982727, -0.1905023455619812, -0.05080355703830719, -0.15369750559329987, -0.19305334985256195, 0.13784123957157135, 0.19038167595863342, -0.0024015847593545914, 0.24734650552272797, 0.158353790640831, -0.09348507225513458, -0.261450856924057, -0.16853435337543488, -0.09094934910535812, -0.03530702367424965, -0.011829596012830734, -0.31345972418785095, 0.07758262008428574, -0.06052020192146301, -0.00040802123839966953, 0.09998849779367447, -0.20675288140773773, -0.1327417641878128, 0.0708891823887825, -0.08713535219430923, 0.18682031333446503, -0.19224396347999573, -0.13792532682418823, 0.04790807142853737, -0.14325520396232605, 0.14673596620559692, 0.104744553565979, 0.14081257581710815, -0.027727507054805756, 0.05811401456594467, -0.014038169756531715, 0.02672620490193367, 0.0787259116768837, 0.0761321410536766, -0.028433794155716896, -0.07973279803991318, -0.08672375231981277, 0.04832848906517029, 0.10114958882331848, -0.044074647128582, -0.08785517513751984, 0.01631041057407856, -0.19371703267097473, -0.0674651488661766, -0.07844910025596619, -0.07832790911197662, -0.021451888605952263, -0.10418176651000977, -0.09300804883241653, 0.059730157256126404, -0.071129210293293, -0.008047196082770824, 0.20250017940998077, 0.011960279196500778, 0.05673680454492569, -0.04481003060936928, 0.05136901140213013, -0.056229107081890106, 0.005326901562511921, -0.17419421672821045, -0.04508773237466812, 0.04892301559448242, -0.13194887340068817, -0.020845556631684303, 0.0908336341381073, -0.013548610731959343, 0.0867648646235466, 0.0758371353149414, -0.008147158659994602, 0.07169177383184433, 0.059862688183784485, -0.3093333840370178, -0.07083915174007416, -0.08852017670869827, -0.14351911842823029, 0.18818649649620056, -0.06604635715484619, 0.10713731497526169, -0.04909176379442215, -0.09902618825435638, 0.07976356148719788, -0.01286364160478115, -0.00314596900716424, -0.03219249099493027, -0.07935473322868347, -0.018651334568858147, -0.0625288188457489, 0.007246804423630238, 0.13519129157066345, -0.15265564620494843, -0.01502308901399374, 0.3012058138847351, -0.11722524464130402, -0.045982953161001205, -0.1173018366098404, 0.09737859666347504, -0.13527078926563263, 0.09315164387226105, -0.019622202962636948, -0.13377167284488678, 0.03462839499115944, 0.09457365423440933, 0.09544574469327927, 0.10096394270658493, -0.013361826539039612, -0.05153103917837143, 0.041014738380908966, 0.008637310937047005, 0.07207830995321274, -0.04943487420678139, -0.05524013191461563, 0.0038002522196620703, 0.080233633518219, 0.2116754651069641, -0.08896154165267944, -0.09263946861028671, -0.1470664143562317, 0.0428563617169857, -0.08836302161216736, 0.019377034157514572, -0.16001170873641968, -0.13272161781787872, -0.055246151983737946, -0.0400906465947628, -0.08367488533258438, -0.09136078506708145, -0.07906265556812286, -0.010839859023690224, -0.02341226302087307, 0.06021532788872719, -0.03855283185839653, -0.018813585862517357, 0.12364578992128372, -0.03787965700030327, 0.08910264819860458, 0.17500890791416168, -0.08450314402580261, 0.023832645267248154, -0.18586213886737823, 0.030689803883433342, 0.07085651159286499, 0.03779592365026474, -0.02245340496301651, 0.14533990621566772, -0.06085220351815224, 0.05881298705935478, 0.12556298077106476, 0.0695766881108284, 0.04984673112630844, -0.12417955696582794, -0.11136485636234283, -0.08503212779760361, -0.18034467101097107, -0.05002203956246376, -0.04423706233501434, 0.06108308583498001, -0.02650604583323002, 0.1309327781200409, -0.08999095112085342, 0.15718142688274384, -0.04321841523051262, 0.0528700053691864, -0.027040526270866394, -0.06470827013254166, -0.12278979271650314, -0.10979297012090683, -0.02499544993042946, -0.02737082727253437, 0.20933546125888824, -0.045306771993637085, -0.15189874172210693, 0.0949237197637558, 0.11563500016927719, 0.018147453665733337, 0.005956919863820076, 0.1323121339082718, 0.194880411028862, -0.11541805416345596, -0.0936809703707695, 0.09047011286020279, -0.05260741338133812, 0.0228708665817976, 0.048201583325862885, 0.058432020246982574, 0.1052132397890091, 0.12848950922489166, -0.04900895804166794, 0.08912058174610138, 0.006498181726783514, 0.016269082203507423, -0.06591968238353729, 0.02973637543618679, -0.05973924323916435, 0.1920899897813797, 0.19707508385181427, 0.029223497956991196, 0.08472736179828644, -0.017559928819537163, -0.11959799379110336, -0.12212255597114563, -0.1407116949558258, -0.050253089517354965, -0.04628077521920204, -0.01775268279016018, 0.006233192514628172, 0.02021676115691662, 0.03366154432296753, 0.07998956739902496, -0.10447468608617783, 0.0326942503452301, 0.22191746532917023, -0.10467225313186646, 0.12025705724954605, -0.02383769489824772, 0.07059957087039948, -0.0725654885172844, 0.015299106948077679, -0.09561005979776382, -0.03161159157752991, -0.05523008853197098, 0.05953917279839516, -0.013546749018132687, 0.01376114971935749, -0.11571979522705078, -0.0704435333609581, -0.05602224916219711, 0.028185369446873665, 0.003776038996875286, 0.12057080864906311, -0.022377997636795044, 0.008332417346537113, 0.03544076159596443, 0.19696034491062164, 0.0867689847946167, 0.06401260197162628, -0.05176214128732681, 0.10445841401815414, -0.06251522153615952, 0.01980535313487053, 0.02677372843027115, 0.08953390270471573, -0.04995796084403992, 0.3714083135128021, 0.2989039719104767, -0.1542336791753769, -0.028593305498361588, -0.00033268475090153515, 0.05587858706712723, 0.09816014021635056, -0.03944380581378937, 0.1306988000869751, 0.3620460033416748, -0.11891288310289383, -0.0671214759349823, -0.1862572580575943, -0.014448084868490696, -0.08954421430826187, 0.10134808719158173, 0.13938789069652557, 0.0012530835811048746, -0.03044394589960575, 0.10862648487091064, -0.3475922644138336, 0.15989893674850464, -0.09812553972005844, -0.13904105126857758, -0.045118704438209534, 0.0849338099360466, 0.012828986160457134, 0.09238886088132858, 0.06963640451431274, 0.0611313097178936, 0.008967170491814613, 0.03412449732422829, 0.05748085677623749, -0.1655062437057495, 0.07511071115732193, 0.13372445106506348, -0.04816906154155731, 0.08928737789392471, -0.020248346030712128, -0.06256580352783203, 0.030741283670067787, 0.0743768960237503, 0.03466007485985756, -0.005767039489001036, 0.016652395948767662, 0.001892654225230217, 0.00008131824142765254, 0.036804668605327606, 0.09148148447275162, -0.06412207335233688, 0.08167830109596252, 0.08462004363536835, 0.10244022309780121, 0.0854521170258522, -0.009488801471889019, -0.011658412404358387, 0.1509765237569809, -0.16429825127124786, -0.03175041452050209, 0.13272972404956818, -0.013735122047364712, -0.004333223681896925, -0.004934211727231741, -0.03645246848464012, -0.14544552564620972, -0.18989601731300354, -0.042144905775785446, -0.0771096795797348, 0.017953451722860336, 0.035979535430669785, -0.01799830049276352, -0.23908594250679016, -0.05031827092170715, -0.07304035872220993, 0.08048682659864426, -0.19506537914276123, 0.0024475937243551016, 0.015862133353948593, 0.01714175008237362, 0.005498567130416632, -0.110921211540699, 0.06879906356334686, -0.031155476346611977, -0.0496002621948719, -0.14037872850894928 ]
null
null
transformers
### What style is that? This model can help identify five architectural styles that were prominent in the early to mid 20th century. Check back for updates including more architectural styles and more accurate predictions as this model diversifies and improves its training. Upload a photograph of a building to the File Uploader on the right. The Image Classifier will predict its architectural style using a database of over 700 images. Scroll down to read more about each style. ### Classical Revival (1895 - 1950) The Classical Revival or Neoclassical style is one of the most commonly seen across the state and the country. This style was inspired by the World's Columbian Exposition in Chicago held in 1893 which promoted a renewed interest in the classical forms. This style encompasses many different styles, including Colonial Revival, Greek Revival, Neoclassical Revival and Mediterranean Revival. Colonial Revival is most commonly used in residential dwellings, while Greek and Neoclassical Revival styles are commonly used in commercial buildings like banks, post offices, and municipal buildings. ![classical revival architecture](images/ex_classical_revival_architecture.jpg) #### Queen Anne (1880-1910) The Queen Anne style was one of a number of popular architectural styles that emerged in the United States during the Victorian Period. It ranges from high style, like the image pictured here, to more vernacular styles that exhibit the Queen Anne form without its high style architectural details. ![queen anne architecture](images/ex_queen_anne_architecture.jpg) #### Craftsman Bungalow (1900-1930) The terms “craftsman” and “bungalow” are often used interchangably, however, “craftsman” refers to the Arts and Crafts movement and is considered an architectural style, whereas “bungalow” is the form of house. Bungalows often exhibit a craftsman style. ![craftsman bungalow architecture](images/ex_craftsman_bungalow_architecture.jpg) #### Tudor Cottage (1910-1950) Tudor homes are inspired by the Medieval period and can range is size and style. In general, the Tudor style features steeply pitched roofs, often with a cat-slide roof line, predominately brick construction, sometimes accented with half-timber framing, front-facing, prominently placed brick or stone chimneys, and tall windows with rectangular or diamond-shaped panes. Front doors are typically off-center with a round arch at the top of the door or doorway. ![tudor cottage architecture](images/ex_tudor_cottage_architecture.jpg) #### Mid-Century Modern Ranch (1930-1970) The Ranch style originated in southern California in the mid-1930s. In the 1940s, the Ranch was one of the small house types financed by the Federal Housing Administration (FHA), along with Minimal Traditional and other small house styles. The Ranch house began to pick up popularity as the financial controls that encouraged small house building lifted following WWII; by the 1950s it was the most predominant residential style in the country. ![mid-century modern ranch](images/ex_mid-century_modern_ranch.jpg) This model was created with HuggingPics🤗🖼️ Image Classifier! Make your own!: [the demo on Google Colab](https://colab.research.google.com/github/nateraw/huggingpics/blob/main/HuggingPics.ipynb).
{"tags": ["image-classification", "pytorch", "huggingpics"], "metrics": ["accuracy"]}
image-classification
gatecitypreservation/architectural_styles
[ "transformers", "pytorch", "tensorboard", "vit", "image-classification", "huggingpics", "model-index", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #vit #image-classification #huggingpics #model-index #autotrain_compatible #endpoints_compatible #has_space #region-us
### What style is that? This model can help identify five architectural styles that were prominent in the early to mid 20th century. Check back for updates including more architectural styles and more accurate predictions as this model diversifies and improves its training. Upload a photograph of a building to the File Uploader on the right. The Image Classifier will predict its architectural style using a database of over 700 images. Scroll down to read more about each style. ### Classical Revival (1895 - 1950) The Classical Revival or Neoclassical style is one of the most commonly seen across the state and the country. This style was inspired by the World's Columbian Exposition in Chicago held in 1893 which promoted a renewed interest in the classical forms. This style encompasses many different styles, including Colonial Revival, Greek Revival, Neoclassical Revival and Mediterranean Revival. Colonial Revival is most commonly used in residential dwellings, while Greek and Neoclassical Revival styles are commonly used in commercial buildings like banks, post offices, and municipal buildings. !classical revival architecture #### Queen Anne (1880-1910) The Queen Anne style was one of a number of popular architectural styles that emerged in the United States during the Victorian Period. It ranges from high style, like the image pictured here, to more vernacular styles that exhibit the Queen Anne form without its high style architectural details. !queen anne architecture #### Craftsman Bungalow (1900-1930) The terms “craftsman” and “bungalow” are often used interchangably, however, “craftsman” refers to the Arts and Crafts movement and is considered an architectural style, whereas “bungalow” is the form of house. Bungalows often exhibit a craftsman style. !craftsman bungalow architecture #### Tudor Cottage (1910-1950) Tudor homes are inspired by the Medieval period and can range is size and style. In general, the Tudor style features steeply pitched roofs, often with a cat-slide roof line, predominately brick construction, sometimes accented with half-timber framing, front-facing, prominently placed brick or stone chimneys, and tall windows with rectangular or diamond-shaped panes. Front doors are typically off-center with a round arch at the top of the door or doorway. !tudor cottage architecture #### Mid-Century Modern Ranch (1930-1970) The Ranch style originated in southern California in the mid-1930s. In the 1940s, the Ranch was one of the small house types financed by the Federal Housing Administration (FHA), along with Minimal Traditional and other small house styles. The Ranch house began to pick up popularity as the financial controls that encouraged small house building lifted following WWII; by the 1950s it was the most predominant residential style in the country. !mid-century modern ranch This model was created with HuggingPics️ Image Classifier! Make your own!: the demo on Google Colab.
[ "### What style is that?\n\nThis model can help identify five architectural styles that were prominent in the early to mid 20th century. Check back for updates including more architectural styles and more accurate predictions as this model diversifies and improves its training. \n\nUpload a photograph of a building to the File Uploader on the right. The Image Classifier will predict its architectural style using a database of over 700 images. Scroll down to read more about each style.", "### Classical Revival (1895 - 1950)\n\nThe Classical Revival or Neoclassical style is one of the most commonly seen across the state and the country. This style was inspired by the World's Columbian Exposition in Chicago held in 1893 which promoted a renewed interest in the classical forms. This style encompasses many different styles, including Colonial Revival, Greek Revival, Neoclassical Revival and Mediterranean Revival. Colonial Revival is most commonly used in residential dwellings, while Greek and Neoclassical Revival styles are commonly used in commercial buildings like banks, post offices, and municipal buildings. \n\n!classical revival architecture", "#### Queen Anne (1880-1910)\n\nThe Queen Anne style was one of a number of popular architectural styles that emerged in the United States during the Victorian Period. It ranges from high style, like the image pictured here, to more vernacular styles that exhibit the Queen Anne form without its high style architectural details.\n\n!queen anne architecture", "#### Craftsman Bungalow (1900-1930)\n\nThe terms “craftsman” and “bungalow” are often used interchangably, however, “craftsman” refers to the Arts and Crafts movement and is considered an architectural style, whereas “bungalow” is the form of house. Bungalows often exhibit a craftsman style.\n\n!craftsman bungalow architecture", "#### Tudor Cottage (1910-1950)\n\nTudor homes are inspired by the Medieval period and can range is size and style. In general, the Tudor style features steeply pitched roofs, often with a cat-slide roof line, predominately brick construction, sometimes accented with half-timber framing, front-facing, prominently placed brick or stone chimneys, and tall windows with rectangular or diamond-shaped panes. Front doors are typically off-center with a round arch at the top of the door or doorway. \n\n!tudor cottage architecture", "#### Mid-Century Modern Ranch (1930-1970)\n\nThe Ranch style originated in southern California in the mid-1930s. In the 1940s, the Ranch was one of the small house types financed by the Federal Housing Administration (FHA), along with Minimal Traditional and other small house styles. The Ranch house began to pick up popularity as the financial controls that encouraged small house building lifted following WWII; by the 1950s it was the most predominant residential style in the country.\n\n!mid-century modern ranch\n\nThis model was created with HuggingPics️ Image Classifier! \nMake your own!: the demo on Google Colab." ]
[ "TAGS\n#transformers #pytorch #tensorboard #vit #image-classification #huggingpics #model-index #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "### What style is that?\n\nThis model can help identify five architectural styles that were prominent in the early to mid 20th century. Check back for updates including more architectural styles and more accurate predictions as this model diversifies and improves its training. \n\nUpload a photograph of a building to the File Uploader on the right. The Image Classifier will predict its architectural style using a database of over 700 images. Scroll down to read more about each style.", "### Classical Revival (1895 - 1950)\n\nThe Classical Revival or Neoclassical style is one of the most commonly seen across the state and the country. This style was inspired by the World's Columbian Exposition in Chicago held in 1893 which promoted a renewed interest in the classical forms. This style encompasses many different styles, including Colonial Revival, Greek Revival, Neoclassical Revival and Mediterranean Revival. Colonial Revival is most commonly used in residential dwellings, while Greek and Neoclassical Revival styles are commonly used in commercial buildings like banks, post offices, and municipal buildings. \n\n!classical revival architecture", "#### Queen Anne (1880-1910)\n\nThe Queen Anne style was one of a number of popular architectural styles that emerged in the United States during the Victorian Period. It ranges from high style, like the image pictured here, to more vernacular styles that exhibit the Queen Anne form without its high style architectural details.\n\n!queen anne architecture", "#### Craftsman Bungalow (1900-1930)\n\nThe terms “craftsman” and “bungalow” are often used interchangably, however, “craftsman” refers to the Arts and Crafts movement and is considered an architectural style, whereas “bungalow” is the form of house. Bungalows often exhibit a craftsman style.\n\n!craftsman bungalow architecture", "#### Tudor Cottage (1910-1950)\n\nTudor homes are inspired by the Medieval period and can range is size and style. In general, the Tudor style features steeply pitched roofs, often with a cat-slide roof line, predominately brick construction, sometimes accented with half-timber framing, front-facing, prominently placed brick or stone chimneys, and tall windows with rectangular or diamond-shaped panes. Front doors are typically off-center with a round arch at the top of the door or doorway. \n\n!tudor cottage architecture", "#### Mid-Century Modern Ranch (1930-1970)\n\nThe Ranch style originated in southern California in the mid-1930s. In the 1940s, the Ranch was one of the small house types financed by the Federal Housing Administration (FHA), along with Minimal Traditional and other small house styles. The Ranch house began to pick up popularity as the financial controls that encouraged small house building lifted following WWII; by the 1950s it was the most predominant residential style in the country.\n\n!mid-century modern ranch\n\nThis model was created with HuggingPics️ Image Classifier! \nMake your own!: the demo on Google Colab." ]
[ 53, 98, 154, 78, 85, 133, 149 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #vit #image-classification #huggingpics #model-index #autotrain_compatible #endpoints_compatible #has_space #region-us \n### What style is that?\n\nThis model can help identify five architectural styles that were prominent in the early to mid 20th century. Check back for updates including more architectural styles and more accurate predictions as this model diversifies and improves its training. \n\nUpload a photograph of a building to the File Uploader on the right. The Image Classifier will predict its architectural style using a database of over 700 images. Scroll down to read more about each style.### Classical Revival (1895 - 1950)\n\nThe Classical Revival or Neoclassical style is one of the most commonly seen across the state and the country. This style was inspired by the World's Columbian Exposition in Chicago held in 1893 which promoted a renewed interest in the classical forms. This style encompasses many different styles, including Colonial Revival, Greek Revival, Neoclassical Revival and Mediterranean Revival. Colonial Revival is most commonly used in residential dwellings, while Greek and Neoclassical Revival styles are commonly used in commercial buildings like banks, post offices, and municipal buildings. \n\n!classical revival architecture#### Queen Anne (1880-1910)\n\nThe Queen Anne style was one of a number of popular architectural styles that emerged in the United States during the Victorian Period. It ranges from high style, like the image pictured here, to more vernacular styles that exhibit the Queen Anne form without its high style architectural details.\n\n!queen anne architecture#### Craftsman Bungalow (1900-1930)\n\nThe terms “craftsman” and “bungalow” are often used interchangably, however, “craftsman” refers to the Arts and Crafts movement and is considered an architectural style, whereas “bungalow” is the form of house. Bungalows often exhibit a craftsman style.\n\n!craftsman bungalow architecture" ]
[ -0.02939445525407791, 0.24283070862293243, -0.007097023073583841, 0.03478650748729706, 0.06001271307468414, -0.03651997819542885, -0.03424454107880592, 0.030116748064756393, -0.11338524520397186, 0.09789345413446426, -0.1278339922428131, -0.0712040513753891, 0.05590222775936127, -0.12392932176589966, 0.049670394510030746, -0.2795759439468384, 0.03799998760223389, -0.00947631150484085, -0.06781230121850967, 0.008678377605974674, 0.082927405834198, -0.07433317601680756, 0.06521454453468323, 0.0717325210571289, 0.030317343771457672, 0.019481541588902473, -0.11614267528057098, -0.05436941236257553, 0.1054503321647644, 0.06009884551167488, 0.10463017970323563, 0.012786468490958214, 0.059709735214710236, -0.0811649039387703, -0.010274319909512997, 0.1456226408481598, -0.03420756384730339, -0.023508818820118904, 0.161179780960083, -0.025405455380678177, 0.2500568628311157, 0.0056162429973483086, 0.006874264217913151, -0.031085193157196045, -0.07128124684095383, -0.22457866370677948, -0.10558114945888519, 0.12883250415325165, 0.019592322409152985, 0.024170061573386192, 0.054967500269412994, -0.03197849541902542, -0.07892896980047226, 0.0357246920466423, 0.16850407421588898, -0.12997719645500183, -0.02234438620507717, -0.05022734776139259, 0.13269029557704926, 0.19227845966815948, -0.0902203693985939, 0.009697667323052883, -0.10435333847999573, 0.059882041066884995, 0.006747263949364424, -0.034997425973415375, 0.09182102978229523, 0.04382571205496788, -0.1666913479566574, -0.026139630004763603, 0.17848487198352814, 0.00009693562606116757, -0.0033025529701262712, -0.011480661109089851, -0.025572150945663452, 0.25976070761680603, -0.009267053566873074, -0.11716167628765106, -0.004181000404059887, -0.0017385274404659867, 0.04206034913659096, -0.0798368975520134, -0.06237087398767471, 0.057551078498363495, -0.06966381520032883, 0.0711035430431366, 0.022531496360898018, 0.04739248752593994, -0.07305289059877396, 0.06291550397872925, 0.03906872868537903, -0.027634277939796448, 0.013691464439034462, -0.027633093297481537, 0.024402298033237457, -0.03459080681204796, 0.0009035590919665992, -0.06575506925582886, -0.055526312440633774, 0.03232068568468094, 0.023511996492743492, 0.016442732885479927, 0.0042473566718399525, -0.030742118135094643, 0.0679284855723381, 0.10220891237258911, -0.03761548176407814, -0.10259459167718887, -0.07234270125627518, -0.04600472375750542, -0.12880057096481323, 0.018787169829010963, 0.013601452112197876, 0.021756401285529137, -0.00239711650647223, -0.006845461204648018, 0.008398490026593208, -0.019944114610552788, 0.05990550294518471, -0.003969993442296982, 0.13574133813381195, -0.09030421823263168, 0.08527284860610962, -0.008537812158465385, 0.07126148045063019, 0.14245712757110596, -0.01353655569255352, -0.06817225366830826, -0.1282600611448288, 0.06202341243624687, -0.0533536933362484, -0.06511563807725906, -0.06652335077524185, -0.12341901659965515, -0.02658126875758171, -0.20182667672634125, -0.13469822704792023, -0.05296424403786659, -0.036446984857320786, -0.09101017564535141, 0.009840661659836769, 0.009312727488577366, 0.019454529508948326, 0.03354506567120552, -0.04335353896021843, -0.05800260975956917, 0.04402557387948036, -0.04923345521092415, -0.03141072764992714, -0.018641890957951546, -0.18500401079654694, -0.004218223039060831, 0.0032304502092301846, 0.053455110639333725, 0.013444953598082066, 0.03297360613942146, -0.1022099107503891, 0.02755461260676384, -0.04626195877790451, 0.02498517371714115, -0.02509124018251896, 0.0427936390042305, -0.028495030477643013, 0.012390286661684513, -0.09027483314275742, 0.010995425283908844, -0.10338756442070007, -0.05718499794602394, 0.1556592434644699, -0.17594578862190247, 0.039968859404325485, 0.08839800208806992, 0.005968327634036541, 0.009834119118750095, 0.1331648826599121, -0.1011963039636612, 0.1731625348329544, -0.061244502663612366, -0.0429992750287056, -0.013819686137139797, -0.08416713774204254, 0.04815094918012619, -0.02055824175477028, -0.08244702219963074, -0.024429086595773697, 0.006653529591858387, -0.08668553829193115, -0.07040272653102875, -0.017454592511057854, 0.042029377073049545, 0.04001937806606293, 0.04576528072357178, -0.04477643221616745, 0.038800857961177826, -0.03390634059906006, -0.012511544860899448, -0.14420118927955627, 0.01696321740746498, 0.02193359285593033, -0.04559721425175667, 0.057995159178972244, -0.025895925238728523, 0.044669367372989655, 0.17439493536949158, -0.01701161451637745, -0.1378398984670639, 0.035028599202632904, 0.06723392009735107, -0.11452949047088623, 0.0008345490787178278, 0.13550764322280884, 0.05592923238873482, 0.12299098819494247, -0.023880159482359886, 0.0726599395275116, -0.05768671631813049, 0.04770316556096077, -0.009011558257043362, -0.10741139203310013, -0.08793193101882935, -0.06731223315000534, -0.017508523538708687, -0.06332115083932877, -0.013341575860977173, 0.20433418452739716, 0.17206111550331116, 0.047274988144636154, -0.03550167381763458, -0.12387504428625107, 0.029481323435902596, -0.013424365781247616, -0.016536427661776543, -0.0007800287567079067, 0.012467949651181698, 0.033322904258966446, 0.046729303896427155, -0.17173436284065247, -0.21712608635425568, -0.011259907856583595, 0.01153702661395073, -0.017002785578370094, -0.08211738616228104, -0.036101315170526505, -0.004189744591712952, -0.036972854286432266, -0.13027000427246094, 0.1617361307144165, 0.03852498531341553, 0.019149702042341232, -0.036951255053281784, -0.05678379535675049, -0.02093621715903282, 0.04329846799373627, -0.13882365822792053, -0.006578199565410614, 0.12993820011615753, -0.09320911020040512, 0.009066294878721237, 0.16246233880519867, -0.014221803285181522, 0.07102810591459274, 0.02367166243493557, -0.0791812464594841, 0.005301772151142359, -0.03897838294506073, 0.03642472252249718, 0.05912403389811516, -0.0806470736861229, -0.015999021008610725, -0.014365904964506626, 0.0709603801369667, 0.059953708201646805, -0.08180102705955505, 0.06346690654754639, 0.02708624303340912, 0.04532231390476227, 0.11900552362203598, -0.07523403316736221, 0.09817051142454147, 0.0656636580824852, -0.015170715749263763, -0.04927187040448189, -0.04454243928194046, -0.04869532212615013, -0.001465909997932613, 0.10136234015226364, -0.12139862030744553, -0.21631944179534912, -0.10441038012504578, -0.0850188285112381, 0.06625165790319443, -0.016806932166218758, 0.03271789476275444, -0.10062656551599503, -0.11727288365364075, -0.10228181630373001, 0.16152015328407288, -0.02950963005423546, -0.04406821355223656, -0.0991937592625618, 0.09830445796251297, -0.11559982597827911, -0.13979211449623108, -0.006206365767866373, -0.016372429206967354, -0.09466201812028885, -0.003776601981371641, 0.00995127484202385, 0.07200455665588379, 0.08514530956745148, -0.02254529856145382, 0.006831768434494734, -0.009715922176837921, 0.02502923458814621, -0.06385210901498795, 0.08207683265209198, 0.06505578756332397, -0.021582409739494324, 0.12461047619581223, 0.10431943088769913, 0.02099229209125042, 0.00723612355068326, -0.050983864814043045, 0.08492708206176758, -0.016276542097330093, -0.05250019580125809, -0.06158158928155899, -0.07765574008226395, -0.12111856788396835, 0.02308136411011219, 0.08548061549663544, 0.010672301985323429, -0.024181488901376724, -0.08725688606500626, -0.08077453821897507, 0.06814036518335342, 0.017991291359066963, 0.24899697303771973, -0.025296133011579514, 0.0012974163983017206, -0.0394667349755764, -0.08786716312170029, 0.10633139312267303, 0.007986742071807384, 0.2759873569011688, 0.034216951578855515, 0.11961598694324493, 0.0537140890955925, -0.05767914280295372, 0.09033598005771637, -0.038556698709726334, 0.008656680583953857, 0.024169715121388435, -0.05161875858902931, -0.07246319204568863, 0.13766920566558838, 0.09082301706075668, 0.13873882591724396, -0.1427057385444641, -0.002252061851322651, -0.13071399927139282, 0.03607390448451042, 0.10337954014539719, -0.004683065228164196, -0.07436169683933258, 0.03198353573679924, 0.0553109273314476, -0.07499026507139206, -0.1361863762140274, 0.04150286316871643, 0.09280682355165482, -0.13189153373241425, 0.053409576416015625, -0.03213069587945938, 0.10500206053256989, -0.08311277627944946, -0.0430062934756279, 0.0323856957256794, 0.1408560872077942, -0.024446912109851837, 0.038173988461494446, 0.05083126574754715, 0.07784836739301682, -0.009858465753495693, -0.014209259301424026, -0.07759816199541092, 0.05983668565750122, 0.02647259831428528, 0.1298413872718811, 0.011478708125650883, 0.040117744356393814, -0.11762760579586029, 0.013360592536628246, 0.07483033835887909, 0.025743087753653526, 0.13700799643993378, -0.16164430975914001, 0.07874839007854462, 0.000030274602977442555, -0.02051835134625435, -0.03446561470627785, -0.08913034200668335, -0.10043209791183472, -0.20681683719158173, 0.034443728625774384, -0.10853876918554306, -0.016907108947634697, -0.04525836557149887, -0.04075018689036369, -0.1319606900215149, 0.027180947363376617, 0.03936593607068062, -0.017539475113153458, -0.10250585526227951, -0.12160567939281464, 0.1085190549492836, -0.035653356462717056, 0.07071233540773392, -0.039774518460035324, 0.04582750052213669, -0.01892416551709175, -0.01881830394268036, -0.03582312539219856, 0.0016893829451873899, -0.0974070131778717, -0.06704454869031906, 0.13503499329090118, 0.1867406815290451, -0.06766020506620407, -0.026049939915537834, 0.11758244782686234, -0.005356145557016134, -0.0776452049612999, -0.007534986361861229, 0.0771356150507927, 0.03786598891019821, 0.08643057942390442, -0.09525284171104431, 0.053766101598739624, -0.04659426584839821, -0.029532277956604958, 0.024140773341059685, 0.1411590576171875, -0.053635429590940475, 0.09387683123350143, 0.10454437136650085, -0.1315092295408249, -0.17064784467220306, -0.10189932584762573, -0.05086176469922066, 0.020024225115776062, 0.1296171098947525, -0.15981820225715637, 0.1430595964193344, 0.03841632977128029, 0.006198016926646233, -0.043461501598358154, -0.13766904175281525, -0.022584425285458565, 0.11777622997760773, 0.049359727650880814, -0.13993261754512787, -0.07581638544797897, -0.010298981331288815, 0.04639000818133354, -0.08119071274995804, 0.033736422657966614, 0.08307141065597534, 0.02991386316716671, -0.005120562389492989, -0.08608581125736237, 0.022699175402522087, -0.0005326336831785738, 0.08656331896781921, 0.0731232613325119, 0.031182873994112015, -0.1125754565000534, 0.008561475202441216, 0.2048894762992859, -0.007605630438774824, -0.02856498584151268, 0.02851666323840618, 0.019979234784841537, -0.007959850132465363, -0.022410819306969643, -0.06881444901227951, 0.08409285545349121, -0.03457101434469223, 0.004060823004692793, -0.09745040535926819, 0.11278373748064041, -0.023861635476350784, -0.03917641192674637, 0.12051697820425034, -0.052584242075681686, 0.13156487047672272, 0.10601683706045151, 0.08227793127298355, -0.06163151189684868, -0.060823507606983185, -0.03306940570473671, 0.0017410177970305085, 0.049035996198654175, 0.1344766765832901, 0.07381021231412888, 0.06782698631286621, 0.014608904719352722, 0.0868731439113617, 0.003959495574235916, -0.11186935752630234, -0.08567073196172714, 0.1073724702000618, -0.061782144010066986, -0.1673835963010788, -0.04661313444375992, 0.050998348742723465, -0.029878539964556694, 0.042544346302747726, 0.07512583583593369, 0.01638728938996792, 0.016127463430166245, -0.06530740857124329, 0.00611519068479538, 0.04966990277171135, -0.051899515092372894, 0.014274267479777336, -0.004503353498876095, 0.03912767022848129, 0.07501254230737686, 0.11198297142982483, -0.14199580252170563, -0.06359604746103287, 0.1967797726392746, -0.06462910026311874, -0.039300091564655304, 0.03766597434878349, 0.02585943415760994, 0.09467931091785431, -0.031259577721357346, -0.041126977652311325, -0.13847477734088898, 0.11181643605232239, 0.20353539288043976, -0.011378755792975426, 0.05694907531142235, -0.07833629101514816, 0.0503242053091526, -0.04141492769122124, 0.05110260099172592, -0.00821606907993555, -0.030105222016572952, -0.0793585255742073, 0.07582812756299973, 0.08119457960128784, -0.010439474135637283, -0.02954619564116001, -0.08133553713560104, -0.048757776618003845, -0.030798617750406265, -0.15269464254379272, -0.0775134414434433, -0.15960514545440674, 0.0050757634453475475, -0.04915208742022514, -0.016390101984143257, -0.0318521149456501, -0.02756948210299015, -0.018862802535295486, -0.01659834384918213, -0.0157796498388052, 0.1342022866010666, -0.07179861515760422, 0.05441133677959442, 0.1315016895532608, -0.06444298475980759, 0.07773345708847046, -0.11833792179822922, -0.042985498905181885, 0.04630930721759796, 0.056587785482406616, -0.059720516204833984, -0.030975572764873505, 0.07409913092851639, -0.03980546072125435, -0.007470575161278248, -0.020418329164385796, 0.000325859960867092, -0.027261702343821526, -0.012607752345502377, 0.0249063391238451, -0.05559961870312691, 0.06073116138577461, 0.005351461470127106, -0.1230873391032219, -0.06970009207725525, 0.04751601442694664, 0.12153370678424835, -0.016866784542798996, 0.0653417706489563, 0.037807755172252655, 0.0260559543967247, -0.14846192300319672, -0.03901391476392746, 0.0272802896797657, -0.013670037500560284, 0.07144106179475784, -0.07073071599006653, 0.01701725460588932, 0.044800419360399246, 0.152568519115448, -0.016738103702664375, -0.07349229604005814, 0.07789444923400879, 0.10541310906410217, -0.1264650970697403, -0.02526695653796196, -0.024739032611250877, 0.0028220899403095245, -0.02200894244015217, 0.017160674557089806, -0.11598676443099976, -0.11407268792390823, 0.023447226732969284, 0.20405200123786926, 0.06467685103416443, 0.09038899838924408, -0.005052638705819845, 0.10553043335676193, -0.05897923558950424, -0.027048995718359947, 0.11254464834928513, 0.12800531089305878, 0.05875866487622261, -0.06822659820318222, 0.015262464061379433, 0.17927350103855133, -0.0703200176358223, 0.1633264571428299, 0.056194134056568146, -0.012877791188657284, -0.05056789144873619, -0.19240136444568634, -0.009909430518746376, -0.009076382033526897, -0.012091124430298805, -0.056449126452207565, 0.10468639433383942, 0.07273267209529877, 0.007738855201750994, 0.016443556174635887, -0.04155449941754341, -0.12012959271669388, -0.1644660383462906, 0.01697652041912079, -0.020211197435855865, 0.1736556738615036, 0.038868047297000885, 0.06759099662303925, 0.00020182563457638025, 0.08245339244604111, -0.07998444885015488, 0.06525728106498718, -0.016274061053991318, 0.051610857248306274, -0.10507999360561371, -0.083089679479599, 0.022081108763813972, -0.07562481611967087, -0.04076208919286728, 0.04349702596664429, 0.050358790904283524, -0.06063069403171539, 0.01721126027405262, 0.15047968924045563, -0.012579088099300861, -0.0008083553984761238, -0.040499329566955566, 0.04087341949343681, 0.07144708186388016, 0.01879594475030899, 0.0797509178519249, -0.05104993283748627, 0.018564144149422646, -0.0034626880660653114, 0.08712152391672134, 0.02899264357984066, -0.0029096254147589207, 0.05279996991157532, -0.01961740106344223, 0.089702308177948, 0.07580875605344772, -0.053222186863422394, 0.2989063560962677, -0.038017790764570236, -0.008897453546524048, -0.06868132948875427, 0.028237344697117805, -0.12488913536071777, 0.11117605865001678, 0.011593797244131565, -0.02707597054541111, 0.009143410250544548, 0.14211855828762054, 0.032380443066358566, -0.05733694136142731, -0.034134745597839355, -0.06298030912876129, -0.11068303138017654, -0.0004904482630081475, 0.09231582283973694, -0.05704230070114136, 0.016422443091869354, 0.031108183786273003, -0.01771785132586956, 0.2615524232387543, 0.035910021513700485, -0.07108531892299652, -0.09536800533533096, 0.14065347611904144, -0.04386402294039726, 0.05062764137983322, 0.015189201571047306, -0.021164895966649055, 0.053065285086631775, 0.0034302647691220045, -0.01012511644512415, 0.027038732543587685, -0.016653921455144882, -0.0751756951212883, -0.021317217499017715, 0.002169936429709196, 0.008694695308804512, 0.12258781492710114, 0.06794612109661102, 0.008426048792898655, 0.1312684416770935, -0.04640047997236252, -0.0628560483455658, -0.004084748215973377, 0.11570122838020325, -0.06472083181142807, 0.06383325904607773, 0.10597871989011765, 0.05330284684896469, 0.0030027437023818493, -0.04900156706571579, -0.060075920075178146, 0.02722778730094433, 0.05553359165787697, -0.05325087532401085, -0.010095841251313686, -0.05266990140080452, -0.17576885223388672, -0.011627388186752796, 0.10273893922567368, -0.15971015393733978, 0.06771878153085709, -0.025662342086434364, -0.05707826465368271, 0.04217321425676346, 0.09637747704982758, 0.0023444690741598606, -0.05015820264816284, -0.021458111703395844, 0.003358380403369665, 0.07288414239883423, -0.15941552817821503, 0.06739400327205658 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-cola This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.7550 - Matthews Correlation: 0.5265 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.5296 | 1.0 | 535 | 0.5144 | 0.4215 | | 0.3504 | 2.0 | 1070 | 0.4903 | 0.5046 | | 0.2393 | 3.0 | 1605 | 0.6339 | 0.5058 | | 0.175 | 4.0 | 2140 | 0.7550 | 0.5265 | | 0.1259 | 5.0 | 2675 | 0.8688 | 0.5259 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.9.0+cu111 - Datasets 1.12.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["glue"], "metrics": ["matthews_correlation"], "model-index": [{"name": "distilbert-base-uncased-finetuned-cola", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "glue", "type": "glue", "args": "cola"}, "metrics": [{"type": "matthews_correlation", "value": 0.5264763891845121, "name": "Matthews Correlation"}]}]}]}
text-classification
gauravtripathy/distilbert-base-uncased-finetuned-cola
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "dataset:glue", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-cola ====================================== This model is a fine-tuned version of distilbert-base-uncased on the glue dataset. It achieves the following results on the evaluation set: * Loss: 0.7550 * Matthews Correlation: 0.5265 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 5 ### Training results ### Framework versions * Transformers 4.11.3 * Pytorch 1.9.0+cu111 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 67, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.10293228179216385, 0.10211790353059769, -0.0022867287043482065, 0.1229473426938057, 0.1666216105222702, 0.03385860100388527, 0.1261749565601349, 0.12640413641929626, -0.08447157591581345, 0.022485584020614624, 0.12115568667650223, 0.1584663689136505, 0.022172022610902786, 0.11729228496551514, -0.049910128116607666, -0.264028936624527, -0.01238543912768364, 0.0473504476249218, -0.05294135957956314, 0.1341211497783661, 0.09216577559709549, -0.12151941657066345, 0.09031782299280167, 0.011848635971546173, -0.1939038336277008, -0.0035890426952391863, -0.00151662842836231, -0.052654728293418884, 0.14732976257801056, 0.025929421186447144, 0.12400811165571213, -0.000053551211749436334, 0.08655817061662674, -0.1950768679380417, 0.010359424166381359, 0.04644487053155899, 0.004357806406915188, 0.09359845519065857, 0.045615773648023605, 0.005042934790253639, 0.11485105007886887, -0.08148791640996933, 0.054909348487854004, 0.021969391033053398, -0.11509831249713898, -0.20643767714500427, -0.08016394078731537, 0.03740183264017105, 0.07948257774114609, 0.10506565868854523, -0.005903282202780247, 0.118460513651371, -0.07930883765220642, 0.09264145791530609, 0.2199171632528305, -0.28675246238708496, -0.06649725139141083, 0.04456263780593872, 0.014260012656450272, 0.044055547565221786, -0.10034371167421341, -0.03775164112448692, 0.04636719822883606, 0.05252493545413017, 0.12677240371704102, -0.02886156365275383, -0.12205179780721664, 0.003943507093936205, -0.14108145236968994, -0.034052956849336624, 0.16836845874786377, 0.03960483521223068, -0.028071854263544083, -0.05562591552734375, -0.060361169278621674, -0.14716725051403046, -0.03625238314270973, -0.011698693968355656, 0.0464656800031662, -0.022566115483641624, -0.040680430829524994, -0.010814763605594635, -0.10871094465255737, -0.06288853287696838, -0.0769818127155304, 0.1077738031744957, 0.036126598715782166, 0.008027154952287674, -0.02833678387105465, 0.1123582124710083, -0.005455060862004757, -0.12297862768173218, 0.02397901937365532, 0.02143707685172558, 0.011964915320277214, -0.040639352053403854, -0.05304218456149101, -0.062216758728027344, 0.011138148605823517, 0.13088594377040863, -0.04630223661661148, 0.040777117013931274, 0.04977878928184509, 0.04933656007051468, -0.09065655618906021, 0.1921754628419876, -0.035307448357343674, -0.02882358245551586, 0.009318295866250992, 0.047068770974874496, 0.01864885725080967, -0.011798413470387459, -0.12454123049974442, 0.00527193071320653, 0.0894339308142662, 0.008874326013028622, -0.06095420941710472, 0.0741189494729042, -0.0564362034201622, -0.025115519762039185, 0.00436109583824873, -0.09189720451831818, 0.022407852113246918, -0.00012450087524484843, -0.07062164694070816, -0.020773189142346382, 0.035129353404045105, 0.016378128901124, -0.020720677450299263, 0.10922794044017792, -0.08774715662002563, 0.027437489479780197, -0.09356613457202911, -0.10951882600784302, 0.019443031400442123, -0.10545651614665985, 0.022130966186523438, -0.09493713080883026, -0.1870652139186859, -0.016704358160495758, 0.06112116947770119, -0.02395394630730152, -0.06191932410001755, -0.05484509840607643, -0.06839887797832489, 0.011917080730199814, -0.010059287771582603, 0.11795024573802948, -0.06372997909784317, 0.09127407521009445, 0.020180784165859222, 0.060755349695682526, -0.04344556853175163, 0.059769000858068466, -0.10313893109560013, 0.016086161136627197, -0.15210497379302979, 0.04057732969522476, -0.05068393796682358, 0.07048557698726654, -0.08246681094169617, -0.10405773669481277, 0.009960782714188099, -0.004616579506546259, 0.062258679419755936, 0.09257113933563232, -0.1871272623538971, -0.07496824860572815, 0.1566639393568039, -0.07179006934165955, -0.12156010419130325, 0.12136990576982498, -0.059911441057920456, 0.05673099681735039, 0.05800947919487953, 0.1781131625175476, 0.08270809054374695, -0.07719041407108307, 0.001962123205885291, 0.025267207995057106, 0.05142663046717644, -0.06747710704803467, 0.06939826160669327, 0.005594620015472174, 0.01946302130818367, 0.03593016415834427, -0.028980368748307228, 0.06400439888238907, -0.08611955493688583, -0.09907142072916031, -0.039731308817863464, -0.08256641775369644, 0.04218128323554993, 0.07525300979614258, 0.0685475766658783, -0.09082342684268951, -0.07639588415622711, 0.05042766034603119, 0.08240915834903717, -0.057524386793375015, 0.023655356839299202, -0.05030953884124756, 0.07552163302898407, -0.02680014632642269, -0.022723102942109108, -0.18102334439754486, -0.03833219036459923, 0.007940770126879215, 0.0008936069207265973, 0.016191819682717323, 0.0293242447078228, 0.060093678534030914, 0.06024206057190895, -0.047139789909124374, -0.016849525272846222, -0.030310533940792084, 0.0008023443515412509, -0.1282641589641571, -0.1906816065311432, -0.031042277812957764, -0.024314746260643005, 0.15747632086277008, -0.20678584277629852, 0.049052946269512177, -0.01716942898929119, 0.07036878913640976, 0.012303530238568783, -0.0057311453856527805, -0.03707874193787575, 0.07378889620304108, -0.044759832322597504, -0.05293324217200279, 0.08083145320415497, 0.0181035865098238, -0.08778280764818192, -0.049537383019924164, -0.09662133455276489, 0.1554623693227768, 0.1270475834608078, -0.10431350767612457, -0.07620560377836227, -0.021447433158755302, -0.06714988499879837, -0.03411936014890671, -0.04863967373967171, 0.024921713396906853, 0.1872512847185135, -0.00414161616936326, 0.15024179220199585, -0.06761927902698517, -0.043780919164419174, 0.017336107790470123, -0.036681704223155975, 0.017142344266176224, 0.12778300046920776, 0.13678133487701416, -0.06068835034966469, 0.1550617814064026, 0.145793154835701, -0.08979953825473785, 0.14537963271141052, -0.04109717905521393, -0.06405612826347351, -0.015771424397826195, -0.031084435060620308, -0.011116551235318184, 0.10088247805833817, -0.15166400372982025, 0.0012375368969514966, 0.033794231712818146, 0.01647959277033806, 0.025556163862347603, -0.2251734882593155, -0.04004596918821335, 0.034886136651039124, -0.04150136560201645, -0.0054308390244841576, -0.007015174720436335, 0.006398918107151985, 0.10080210864543915, 0.0011162234004586935, -0.08686558902263641, 0.03884623944759369, 0.00246254401281476, -0.08441213518381119, 0.21587513387203217, -0.08261082321405411, -0.17427998781204224, -0.13143739104270935, -0.07296444475650787, -0.04656561464071274, -0.00015476273256354034, 0.06702996045351028, -0.08790277689695358, -0.031083928421139717, -0.07283113896846771, 0.022400353103876114, 0.011085357517004013, 0.023634716868400574, 0.0048742955550551414, 0.004546612501144409, 0.06294062733650208, -0.11091221868991852, -0.014993567951023579, -0.056854572147130966, -0.04403455927968025, 0.04526297748088837, 0.031042957678437233, 0.11287714540958405, 0.15397539734840393, -0.013578025624155998, 0.010587431490421295, -0.029993969947099686, 0.24025335907936096, -0.05967717990279198, -0.01714526303112507, 0.14374007284641266, -0.012295611202716827, 0.052070144563913345, 0.12049724906682968, 0.0724264457821846, -0.07711335271596909, 0.0049663386307656765, 0.0349731408059597, -0.03659248352050781, -0.23019050061702728, -0.05864494666457176, -0.05804623290896416, 0.010861237533390522, 0.092123843729496, 0.024515606462955475, 0.028455622494220734, 0.07220529019832611, 0.041639309376478195, 0.07891497761011124, -0.039184849709272385, 0.05488339066505432, 0.13200291991233826, 0.03253066539764404, 0.12530697882175446, -0.04485247656702995, -0.06470288336277008, 0.043283671140670776, -0.007851089350879192, 0.22500012814998627, 0.00498597789555788, 0.12827841937541962, 0.061436161398887634, 0.16259542107582092, -0.005482909269630909, 0.07941409200429916, -0.010804400779306889, -0.034850314259529114, -0.01829976588487625, -0.03802775219082832, -0.03972357138991356, 0.025606950744986534, -0.06625261902809143, 0.06174756586551666, -0.11933648586273193, 0.015773549675941467, 0.059412408620119095, 0.2493380904197693, 0.035232312977313995, -0.3214189410209656, -0.09892550855875015, 0.0030548900831490755, -0.03374151512980461, -0.0217451099306345, 0.026999959722161293, 0.0956345722079277, -0.10153902322053909, 0.027490416541695595, -0.07547023147344589, 0.09686662256717682, -0.0532696507871151, 0.04829416051506996, 0.08392155915498734, 0.09046503156423569, 0.01247273851186037, 0.09351567178964615, -0.28535711765289307, 0.2724659740924835, -0.0007187346345745027, 0.05634443834424019, -0.07844468951225281, 0.010085172951221466, 0.04311038553714752, 0.0617624931037426, 0.08132167905569077, -0.012144510634243488, -0.024864478036761284, -0.18457560241222382, -0.07140568643808365, 0.028937863186001778, 0.06106057018041611, -0.03816510736942291, 0.08368834853172302, -0.03268950432538986, 0.007421617861837149, 0.07158444076776505, 0.0018232903676107526, -0.049662407487630844, -0.10900235176086426, -0.005491705145686865, 0.02214399352669716, -0.059970103204250336, -0.06015748903155327, -0.12070491909980774, -0.12658987939357758, 0.15638743340969086, -0.03237874060869217, -0.040809910744428635, -0.108126200735569, 0.08461818844079971, 0.061650779098272324, -0.08975368738174438, 0.04530291259288788, 0.0005929115577600896, 0.08017714321613312, 0.021392740309238434, -0.07426926493644714, 0.10096711665391922, -0.0765409916639328, -0.15710070729255676, -0.06625448912382126, 0.10508828610181808, 0.03252893313765526, 0.06573929637670517, -0.011538748629391193, 0.007097951602190733, -0.048221834003925323, -0.0899098813533783, 0.0157444067299366, 0.008682792074978352, 0.07937674224376678, 0.01812531054019928, -0.0769382119178772, 0.006771215703338385, -0.06016010046005249, -0.03258366137742996, 0.21024571359157562, 0.2158423513174057, -0.10153422504663467, 0.025718165561556816, 0.02286433055996895, -0.07310625165700912, -0.20149895548820496, 0.03209054470062256, 0.05646910145878792, 0.009728043340146542, 0.04095728322863579, -0.1810753047466278, 0.1385360211133957, 0.10775274783372879, -0.014298124238848686, 0.10412899404764175, -0.3202374279499054, -0.12213919311761856, 0.13664571940898895, 0.13395246863365173, 0.10184559971094131, -0.12939350306987762, -0.020932715386152267, -0.018083496019244194, -0.1386205106973648, 0.11839033663272858, -0.08920150995254517, 0.11923332512378693, -0.03470944985747337, 0.08193439245223999, 0.0019103640224784613, -0.0584154948592186, 0.11890982836484909, 0.029355058446526527, 0.09175474941730499, -0.059251703321933746, -0.03512052446603775, 0.03123394027352333, -0.04431236907839775, 0.035793647170066833, -0.09338604658842087, 0.030570238828659058, -0.10604918748140335, -0.026041310280561447, -0.06602125614881516, 0.045677412301301956, -0.04284219443798065, -0.06899946928024292, -0.03692661598324776, 0.02686753123998642, 0.049646906554698944, -0.008095373399555683, 0.12122344225645065, 0.028622204437851906, 0.1413784623146057, 0.09826763719320297, 0.07204247266054153, -0.06908541172742844, -0.07781636714935303, -0.02717282436788082, -0.011589732952415943, 0.05001619830727577, -0.13408835232257843, 0.02152666635811329, 0.15354293584823608, 0.01912223920226097, 0.1510246992111206, 0.0823773443698883, -0.01839156076312065, 0.0006340544205158949, 0.05673054978251457, -0.1661752462387085, -0.08904203027486801, -0.0139485327526927, -0.06544327735900879, -0.12135298550128937, 0.04187800735235214, 0.09335416555404663, -0.06716503202915192, -0.00854936707764864, -0.004332916811108589, 0.015022773295640945, -0.04731593281030655, 0.18519560992717743, 0.06192591413855553, 0.04738573357462883, -0.09876950085163116, 0.06990070641040802, 0.04750203713774681, -0.07180026918649673, 0.004038524813950062, 0.07468368113040924, -0.08844168484210968, -0.05540917441248894, 0.06618228554725647, 0.18912407755851746, -0.04818600043654442, -0.04585614427924156, -0.14010009169578552, -0.12294613569974899, 0.07862210273742676, 0.14014258980751038, 0.12017867714166641, 0.010635899379849434, -0.06837818771600723, 0.000588404422160238, -0.10836377739906311, 0.10513027757406235, 0.05070411041378975, 0.06309375166893005, -0.143156960606575, 0.1407182216644287, 0.017011260613799095, 0.05002648010849953, -0.020021822303533554, 0.02507937140762806, -0.09987478703260422, 0.005851608235388994, -0.09810540080070496, -0.013316075317561626, -0.03489955514669418, 0.012692613527178764, -0.006023561581969261, -0.04606341943144798, -0.05516805872321129, 0.010568820871412754, -0.10615460574626923, -0.02290933020412922, 0.025670474395155907, 0.06971292942762375, -0.10798588395118713, -0.037200186401605606, 0.02850211039185524, -0.061468616127967834, 0.0778270810842514, 0.04380037263035774, 0.016251100227236748, 0.04931725934147835, -0.1368095427751541, 0.016054028645157814, 0.0741824060678482, 0.031630221754312515, 0.06476709991693497, -0.09693238884210587, -0.007009484805166721, -0.005799370817840099, 0.03826693072915077, 0.019915807992219925, 0.07783228904008865, -0.14090614020824432, 0.0033579571172595024, -0.023016007617115974, -0.08148306608200073, -0.0679435208439827, 0.025627601891756058, 0.08968912810087204, 0.020695852115750313, 0.20147675275802612, -0.07614127546548843, 0.05207391083240509, -0.21476686000823975, 0.00602736184373498, -0.00898794736713171, -0.10838301479816437, -0.10405520349740982, -0.07142055779695511, 0.05415463447570801, -0.057870715856552124, 0.15364210307598114, 0.04755222052335739, 0.02116718515753746, 0.024372970685362816, -0.006664292421191931, 0.015061943791806698, 0.012095077894628048, 0.18828660249710083, 0.029968159273266792, -0.034546174108982086, 0.05746045336127281, 0.043523602187633514, 0.10385739058256149, 0.11247079074382782, 0.2020433396100998, 0.14116422832012177, -0.004939631558954716, 0.09197620302438736, 0.041565459221601486, -0.058717671781778336, -0.1596382111310959, 0.04675380513072014, -0.036912091076374054, 0.11001145839691162, -0.020111916586756706, 0.21756623685359955, 0.05966166406869888, -0.17037402093410492, 0.047409914433956146, -0.053513552993535995, -0.0871051624417305, -0.11304879188537598, -0.05292489752173424, -0.07837802916765213, -0.12711761891841888, -0.005404567811638117, -0.1170675978064537, -0.004286153707653284, 0.12756498157978058, 0.003694164101034403, -0.02805294096469879, 0.1554909348487854, 0.00789803545922041, 0.023206498473882675, 0.056500211358070374, 0.011705156415700912, -0.035210005939006805, -0.13210441172122955, -0.0585331991314888, -0.018046673387289047, -0.007394176442176104, 0.03254667669534683, -0.061976175755262375, -0.03837079554796219, 0.032222554087638855, -0.02231663465499878, -0.0932476818561554, 0.0044808583334088326, 0.013048709370195866, 0.05314519628882408, 0.04629126936197281, 0.010941000655293465, 0.019157880917191505, -0.0025506617967039347, 0.19922113418579102, -0.07066750526428223, -0.06752666085958481, -0.10790571570396423, 0.2306250035762787, 0.03267180919647217, -0.02204267680644989, 0.03523681312799454, -0.06521956622600555, 0.0032544927671551704, 0.24848921597003937, 0.2182394564151764, -0.08356104791164398, -0.007423927076160908, 0.016474084928631783, -0.008918780833482742, -0.022821171209216118, 0.10075689852237701, 0.1441466361284256, 0.05678163841366768, -0.09291009604930878, -0.04756271466612816, -0.05913683399558067, -0.01872420497238636, -0.03722444549202919, 0.0696612298488617, 0.04772054776549339, 0.007786301430314779, -0.03467464819550514, 0.05562012270092964, -0.07210247963666916, -0.08981996029615402, 0.05333121865987778, -0.2165629267692566, -0.16955634951591492, -0.01353386789560318, 0.09877856820821762, 0.0028938769828528166, 0.060236699879169464, -0.031175728887319565, -0.003417396219447255, 0.0939769595861435, -0.02075004205107689, -0.09734198451042175, -0.06696230918169022, 0.08778399229049683, -0.10811364650726318, 0.22142042219638824, -0.046577975153923035, 0.05537102371454239, 0.12515078485012054, 0.06956698000431061, -0.06979300081729889, 0.06339655071496964, 0.043599870055913925, -0.04075022414326668, 0.027245905250310898, 0.06929980963468552, -0.03538898378610611, 0.062272705137729645, 0.04901720955967903, -0.14046722650527954, 0.019744357094168663, -0.049150168895721436, -0.06616439670324326, -0.04572019726037979, -0.02336043119430542, -0.06187709793448448, 0.13226591050624847, 0.2158280462026596, -0.026524873450398445, -0.010707451961934566, -0.07004721462726593, 0.010463695041835308, 0.052571672946214676, 0.024181634187698364, -0.05712085962295532, -0.21054165065288544, 0.016438627615571022, 0.03813619911670685, -0.01870911940932274, -0.2456497699022293, -0.10105979442596436, -0.00005664302079821937, -0.07310589402914047, -0.09662215411663055, 0.07390806823968887, 0.08636944741010666, 0.05075787752866745, -0.05678851902484894, -0.040403302758932114, -0.07686646282672882, 0.145597442984581, -0.14400917291641235, -0.09283246099948883 ]
null
null
sentence-transformers
# gaussfer/test_simcse_new This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. <!--- Describe your model here --> ## Usage (Sentence-Transformers) Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed: ``` pip install -U sentence-transformers ``` Then you can use the model like this: ```python from sentence_transformers import SentenceTransformer sentences = ["This is an example sentence", "Each sentence is converted"] model = SentenceTransformer('gaussfer/test_simcse_new') embeddings = model.encode(sentences) print(embeddings) ``` ## Usage (HuggingFace Transformers) Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings. ```python from transformers import AutoTokenizer, AutoModel import torch #Mean Pooling - Take attention mask into account for correct averaging def mean_pooling(model_output, attention_mask): token_embeddings = model_output[0] #First element of model_output contains all token embeddings input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float() return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9) # Sentences we want sentence embeddings for sentences = ['This is an example sentence', 'Each sentence is converted'] # Load model from HuggingFace Hub tokenizer = AutoTokenizer.from_pretrained('gaussfer/test_simcse_new') model = AutoModel.from_pretrained('gaussfer/test_simcse_new') # Tokenize sentences encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt') # Compute token embeddings with torch.no_grad(): model_output = model(**encoded_input) # Perform pooling. In this case, mean pooling. sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask']) print("Sentence embeddings:") print(sentence_embeddings) ``` ## Evaluation Results <!--- Describe how your model was evaluated --> For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name=gaussfer/test_simcse_new) ## Training The model was trained with the parameters: **DataLoader**: `torch.utils.data.dataloader.DataLoader` of length 875 with parameters: ``` {'batch_size': 16, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'} ``` **Loss**: `sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss` with parameters: ``` {'scale': 20.0, 'similarity_fct': 'cos_sim'} ``` Parameters of the fit()-Method: ``` { "epochs": 40, "evaluation_steps": 0, "evaluator": "NoneType", "max_grad_norm": 1, "optimizer_class": "<class 'transformers.optimization.AdamW'>", "optimizer_params": { "lr": 2e-05 }, "scheduler": "WarmupLinear", "steps_per_epoch": null, "warmup_steps": 10000, "weight_decay": 0.01 } ``` ## Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False}) ) ``` ## Citing & Authors <!--- Describe where people can find more information -->
{"tags": ["sentence-transformers", "feature-extraction", "sentence-similarity", "transformers"], "pipeline_tag": "sentence-similarity"}
sentence-similarity
gaussfer/test_simcse_new
[ "sentence-transformers", "pytorch", "bert", "feature-extraction", "sentence-similarity", "transformers", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #sentence-transformers #pytorch #bert #feature-extraction #sentence-similarity #transformers #endpoints_compatible #region-us
# gaussfer/test_simcse_new This is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. ## Usage (Sentence-Transformers) Using this model becomes easy when you have sentence-transformers installed: Then you can use the model like this: ## Usage (HuggingFace Transformers) Without sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings. ## Evaluation Results For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL ## Training The model was trained with the parameters: DataLoader: 'URL.dataloader.DataLoader' of length 875 with parameters: Loss: 'sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss' with parameters: Parameters of the fit()-Method: ## Full Model Architecture ## Citing & Authors
[ "# gaussfer/test_simcse_new\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.", "## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:", "## Usage (HuggingFace Transformers)\nWithout sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.", "## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL", "## Training\nThe model was trained with the parameters:\n\nDataLoader:\n\n'URL.dataloader.DataLoader' of length 875 with parameters:\n\n\nLoss:\n\n'sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss' with parameters:\n \n\nParameters of the fit()-Method:", "## Full Model Architecture", "## Citing & Authors" ]
[ "TAGS\n#sentence-transformers #pytorch #bert #feature-extraction #sentence-similarity #transformers #endpoints_compatible #region-us \n", "# gaussfer/test_simcse_new\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.", "## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:", "## Usage (HuggingFace Transformers)\nWithout sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.", "## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL", "## Training\nThe model was trained with the parameters:\n\nDataLoader:\n\n'URL.dataloader.DataLoader' of length 875 with parameters:\n\n\nLoss:\n\n'sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss' with parameters:\n \n\nParameters of the fit()-Method:", "## Full Model Architecture", "## Citing & Authors" ]
[ 42, 54, 38, 64, 29, 86, 5, 6 ]
[ "passage: TAGS\n#sentence-transformers #pytorch #bert #feature-extraction #sentence-similarity #transformers #endpoints_compatible #region-us \n# gaussfer/test_simcse_new\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:## Usage (HuggingFace Transformers)\nWithout sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL## Training\nThe model was trained with the parameters:\n\nDataLoader:\n\n'URL.dataloader.DataLoader' of length 875 with parameters:\n\n\nLoss:\n\n'sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss' with parameters:\n \n\nParameters of the fit()-Method:## Full Model Architecture## Citing & Authors" ]
[ -0.014248175546526909, 0.10685956478118896, -0.008519032970070839, 0.046612419188022614, 0.11791115999221802, 0.033638764172792435, 0.1582903116941452, 0.10150571167469025, -0.004114984069019556, 0.0839126706123352, 0.014466779306530952, 0.10808076709508896, -0.005580904893577099, 0.01042559277266264, 0.03966708853840828, -0.2650858461856842, 0.040063224732875824, -0.06133168563246727, 0.008104490116238594, 0.06980270892381668, 0.11220075190067291, -0.07190477102994919, 0.04334639012813568, -0.003020794130861759, -0.07226690649986267, 0.025749465450644493, -0.028529038652777672, -0.037364013493061066, 0.07319687306880951, 0.054593827575445175, 0.047873903065919876, 0.0032655121758580208, 0.0006732414476573467, -0.20666375756263733, 0.016095317900180817, 0.09011313319206238, -0.020402561873197556, 0.052498530596494675, 0.01599261164665222, -0.05197662487626076, 0.13818401098251343, -0.1027395948767662, 0.0753326565027237, 0.05369314178824425, -0.12679053843021393, -0.07233051210641861, -0.03486417606472969, 0.01136433333158493, 0.08798273652791977, 0.09705425053834915, -0.055493954569101334, 0.1268204152584076, -0.05323118716478348, 0.08828724175691605, 0.14437037706375122, -0.253923624753952, -0.040863800793886185, 0.029400696977972984, 0.04419904202222824, 0.03804892301559448, -0.11931687593460083, 0.01437280885875225, -0.0388181135058403, 0.042044103145599365, 0.05209362134337425, -0.018421098589897156, 0.05191781744360924, 0.003467363538220525, -0.10749363154172897, 0.009359084069728851, 0.13760949671268463, 0.024183647707104683, -0.016535267233848572, -0.18938714265823364, -0.07867047935724258, 0.0930437445640564, -0.036969222128391266, -0.02125481143593788, 0.03521627560257912, 0.053286973387002945, -0.040873974561691284, -0.11169779300689697, -0.0773078054189682, -0.013177968561649323, -0.05339909344911575, 0.021150927990674973, 0.003895407309755683, -0.047766778618097305, -0.009722388349473476, 0.06973567605018616, -0.025944987311959267, -0.10819008946418762, -0.028676575049757957, -0.03893807902932167, -0.11680558323860168, -0.007507968228310347, -0.05346643179655075, -0.09295598417520523, 0.04838975518941879, 0.1596001535654068, 0.0697999969124794, 0.0389130599796772, -0.04515935853123665, 0.049188099801540375, 0.01824118383228779, 0.1375124454498291, -0.028924304991960526, -0.07023298740386963, -0.024517923593521118, 0.015330540016293526, -0.0015324176056310534, -0.033997125923633575, -0.037033021450042725, 0.004914883524179459, 0.023472674190998077, 0.06241679564118385, 0.06684797257184982, 0.06744097173213959, -0.02844531275331974, -0.056347426027059555, 0.03589298203587532, -0.12651191651821136, 0.0319998636841774, 0.038866836577653885, -0.019335635006427765, 0.05551881343126297, 0.08100488781929016, -0.02823202684521675, -0.08685106039047241, 0.0064568184316158295, -0.08652609586715698, -0.013817626982927322, -0.05507220700383186, -0.12336458265781403, -0.004641203675419092, 0.0037342559080570936, -0.06537781655788422, -0.07990396022796631, -0.11862118542194366, -0.07238870859146118, 0.048851560801267624, -0.05193595960736275, -0.003725756425410509, -0.10899338126182556, -0.007931123487651348, 0.008691397495567799, 0.01748407445847988, -0.04811714217066765, 0.007036968600004911, 0.01787310279905796, -0.04626834765076637, 0.0527334026992321, 0.0659368559718132, 0.041543878614902496, -0.09866675734519958, 0.011714018881320953, -0.17129062116146088, 0.16952957212924957, -0.05982271209359169, 0.0549502931535244, -0.11112648248672485, 0.02720760926604271, 0.014592093415558338, 0.052310794591903687, -0.0041204579174518585, 0.12409263104200363, -0.1741727739572525, -0.08046068996191025, 0.16461659967899323, -0.06838331371545792, -0.08485894650220871, 0.09924355149269104, -0.05377618595957756, 0.10940819978713989, 0.1356954723596573, 0.12946315109729767, 0.09835493564605713, -0.06080666184425354, 0.004380438011139631, 0.01436325628310442, -0.03577311709523201, 0.13582591712474823, 0.04624254256486893, -0.07270938903093338, 0.1020532175898552, 0.002879558829590678, -0.05042111128568649, 0.01848459430038929, 0.008059277199208736, -0.042338717728853226, 0.00509650306776166, -0.027333512902259827, 0.044177114963531494, -0.03811937943100929, 0.015925804153084755, 0.0038301092572510242, -0.12101750075817108, 0.11151523888111115, 0.06520133465528488, -0.08921509236097336, 0.033218927681446075, -0.07066906243562698, 0.0002803766692522913, -0.0062004579231143, 0.007918065413832664, -0.19566954672336578, -0.1320604383945465, 0.018053825944662094, -0.02600347436964512, 0.11377398669719696, 0.020569253712892532, 0.056166186928749084, 0.04412268102169037, -0.03789648041129112, 0.009131241589784622, 0.02295304276049137, 0.008473418653011322, -0.06800811737775803, -0.13876047730445862, -0.013838149607181549, -0.04603279009461403, 0.05565597489476204, -0.102837935090065, 0.03317287936806679, 0.018153591081500053, 0.10108785331249237, 0.0434676893055439, -0.036795683205127716, -0.007492818403989077, -0.04068978130817413, -0.0026082382537424564, -0.049907341599464417, 0.056918755173683167, 0.024834826588630676, -0.12048593163490295, 0.06468082219362259, -0.17028796672821045, -0.13727645576000214, 0.07193955779075623, -0.027065368369221687, -0.05754060670733452, -0.017159195616841316, -0.020605463534593582, -0.0002779187052510679, -0.05912763625383377, -0.06727048754692078, 0.18285079300403595, 0.0849122628569603, 0.0994248315691948, -0.03995193541049957, -0.042700041085481644, -0.050831280648708344, -0.015999358147382736, -0.028850458562374115, 0.09054771065711975, -0.023077435791492462, -0.16783851385116577, 0.06809525936841965, 0.06659355014562607, -0.047793205827474594, 0.13743296265602112, -0.017871588468551636, -0.054718222469091415, -0.05542363226413727, 0.024972807615995407, 0.028578216210007668, -0.011637288145720959, -0.10174912214279175, 0.01135315466672182, 0.04736499860882759, 0.027335889637470245, 0.030757347121834755, -0.05574197694659233, 0.045606523752212524, 0.03964102640748024, -0.0051095555536448956, 0.09455125778913498, 0.013052034191787243, 0.011371179483830929, 0.054487697780132294, 0.015018327161669731, 0.05509006604552269, -0.019673995673656464, -0.05075797438621521, -0.10895266383886337, 0.15361683070659637, -0.11960397660732269, -0.2024347484111786, -0.12362276017665863, -0.002160151954740286, -0.0707155242562294, 0.009432993829250336, 0.08783333003520966, -0.0576108954846859, -0.04639934003353119, -0.07054927200078964, 0.082401804625988, 0.06516814231872559, -0.06122073903679848, 0.022793015465140343, 0.044491324573755264, 0.019300153478980064, -0.13558506965637207, -0.01256467867642641, -0.0009698467911221087, -0.059992969036102295, -0.029468625783920288, -0.020502036437392235, 0.03529508784413338, 0.10439757257699966, 0.07838138192892075, 0.006268621422350407, -0.004847134463489056, 0.2311415821313858, -0.08132561296224594, 0.05793256685137749, 0.12925560772418976, -0.00492847291752696, 0.06785649806261063, 0.06823927164077759, 0.02921736054122448, -0.0604301393032074, 0.037793468683958054, 0.09207852184772491, -0.011134020984172821, -0.17220088839530945, -0.0825229212641716, -0.07316391915082932, -0.027987834066152573, 0.11363453418016434, 0.05443502217531204, 0.005788819398730993, 0.03511497378349304, -0.03509509935975075, -0.00786956213414669, 0.09723012149333954, 0.11181309074163437, 0.12088006734848022, -0.02245255932211876, 0.0961848646402359, -0.05983719229698181, -0.06555133312940598, 0.050372276455163956, -0.023450942710042, 0.1546611487865448, 0.029712120071053505, 0.16254034638404846, 0.07282111048698425, -0.027243897318840027, -0.010198643431067467, 0.07102791219949722, -0.038061968982219696, 0.029074180871248245, -0.03664639592170715, -0.09934395551681519, -0.0030994508415460587, 0.05371687188744545, 0.100886270403862, -0.049522656947374344, -0.03151976317167282, 0.02843596413731575, 0.14452403783798218, 0.148558109998703, 0.044620271772146225, -0.18480801582336426, -0.039886537939310074, 0.03783808648586273, -0.07628817856311798, -0.06614647805690765, -0.003292559180408716, 0.03291069343686104, -0.1307028979063034, 0.05845525115728378, -0.029032284393906593, 0.1035766750574112, -0.07195133715867996, 0.03236020356416702, -0.028543151915073395, 0.04899014160037041, 0.0006825145683251321, 0.07621081918478012, -0.21320325136184692, 0.0859026312828064, 0.020074283704161644, 0.06151001155376434, -0.059032514691352844, 0.03174557536840439, 0.06926750391721725, -0.0039482032880187035, 0.16935434937477112, -0.019252076745033264, -0.04226965084671974, 0.04488218575716019, -0.047553736716508865, 0.002045688219368458, 0.06305787712335587, -0.12111520022153854, 0.08598487824201584, -0.04973709210753441, -0.030283333733677864, 0.004278949927538633, 0.050496891140937805, -0.0806366354227066, -0.19231092929840088, 0.019068146124482155, -0.008364250883460045, -0.013499174267053604, -0.016947804018855095, -0.019649477675557137, 0.013293907046318054, 0.21630872786045074, -0.09595830738544464, -0.06680312752723694, -0.12551452219486237, -0.018326004967093468, 0.11170893162488937, -0.08121806383132935, 0.005361462943255901, -0.020493367686867714, 0.14915768802165985, -0.06271390616893768, -0.08328729122877121, 0.06980516016483307, -0.03865840658545494, -0.08975856751203537, -0.028537355363368988, 0.10778054594993591, 0.051850877702236176, 0.03245105966925621, 0.02609506994485855, 0.07413125038146973, -0.017917728051543236, -0.10499227792024612, -0.03785208612680435, 0.12753668427467346, -0.030980952084064484, 0.07510039210319519, -0.12638039886951447, -0.00975638534873724, -0.10269158333539963, 0.05047983676195145, 0.21886704862117767, 0.23353607952594757, -0.06733156740665436, 0.10750673711299896, 0.14714695513248444, -0.12030273675918579, -0.1968652904033661, -0.07495014369487762, 0.0115607725456357, 0.026041293516755104, 0.051818933337926865, -0.15990687906742096, 0.08031320571899414, 0.03865472972393036, -0.00502501567825675, -0.08410828560590744, -0.23540499806404114, -0.13724757730960846, 0.11072752624750137, 0.007635791320353746, -0.012949575670063496, -0.11198326200246811, -0.05573894456028938, -0.07305831462144852, -0.018547549843788147, 0.12014800310134888, -0.08825284242630005, 0.12330400198698044, 0.043592870235443115, -0.021762162446975708, 0.04751826077699661, -0.015998389571905136, 0.12174848467111588, 0.06361836194992065, 0.0344587005674839, -0.03886708244681358, -0.052533891052007675, 0.11618279665708542, -0.07856810837984085, 0.10408532619476318, -0.045437391847372055, 0.040324047207832336, -0.09231305122375488, -0.039485134184360504, -0.05779905989766121, 0.027527980506420135, -0.04963679611682892, -0.038333579897880554, -0.019586831331253052, 0.058486390858888626, 0.12144003063440323, -0.006944509223103523, 0.0823909193277359, -0.06708156317472458, 0.07591164857149124, 0.15254783630371094, 0.07095271348953247, 0.06936398893594742, -0.1942611038684845, -0.000014013337931828573, 0.005038081668317318, 0.04876699671149254, -0.10896433889865875, 0.08060649782419205, 0.06452307105064392, -0.000713262299541384, 0.1509888768196106, 0.0360390841960907, -0.099847711622715, -0.01586022973060608, 0.03472055122256279, -0.11073264479637146, -0.14526335895061493, -0.035220976918935776, -0.05206383764743805, -0.10470752418041229, -0.047716934233903885, 0.1626087874174118, -0.000772287487052381, -0.0021145150531083345, 0.0363813117146492, 0.0375530831515789, -0.0316913016140461, 0.08809469640254974, 0.011656288988888264, 0.04531177878379822, -0.05291866883635521, 0.11794206500053406, 0.08880818635225296, -0.08135505020618439, 0.03989305719733238, 0.1573043018579483, -0.07062137871980667, -0.07821036130189896, -0.036992091685533524, 0.15220214426517487, -0.05207199230790138, 0.04436153173446655, -0.06552889198064804, -0.06227628514170647, 0.025983653962612152, 0.05080271512269974, 0.039718177169561386, 0.06091129407286644, -0.09502368420362473, 0.0038243671879172325, -0.07755018770694733, 0.08589262515306473, 0.059400055557489395, 0.008105066604912281, -0.045791320502758026, 0.08680606633424759, 0.0026010945439338684, -0.004593072459101677, -0.02794632874429226, -0.04748992994427681, -0.10259758681058884, -0.0015770759200677276, -0.04721149802207947, -0.00333589781075716, -0.08921296149492264, -0.007841327227652073, 0.030791057273745537, 0.02957848832011223, 0.0043032169342041016, -0.0033071055077016354, -0.05066205933690071, -0.0801013708114624, -0.038670238107442856, 0.09294028580188751, -0.15240702033042908, 0.002621311694383621, 0.0319661945104599, -0.10741088539361954, 0.08716029673814774, 0.010922765359282494, -0.04566209763288498, 0.03329845517873764, -0.07390104234218597, -0.052819542586803436, -0.0012252030428498983, 0.016493530943989754, 0.03247801586985588, -0.11325902491807938, 0.018062002956867218, -0.051662079989910126, 0.03930958732962608, 0.013953128829598427, 0.05332906171679497, -0.10560635477304459, 0.034485701471567154, -0.005452655255794525, -0.030321255326271057, -0.08007221668958664, 0.012111720629036427, 0.025404486805200577, 0.030055446550250053, 0.13556228578090668, -0.0657658576965332, 0.09182772785425186, -0.13705681264400482, -0.0004509141726884991, 0.025467470288276672, -0.04395958036184311, 0.10668396204710007, -0.10147833824157715, 0.06326093524694443, -0.05107378959655762, 0.09807160496711731, -0.03885559365153313, 0.027904344722628593, 0.07051905244588852, 0.009445165283977985, -0.03731056675314903, 0.0317862369120121, 0.06640159338712692, 0.024821339175105095, -0.003514118492603302, -0.0341055691242218, -0.009305423125624657, 0.007855239324271679, -0.0023958205711096525, 0.0853574126958847, 0.1303388923406601, 0.04578126594424248, 0.07046723365783691, 0.09561236947774887, 0.005533057264983654, -0.0839286595582962, 0.061709947884082794, 0.007522348780184984, 0.05602838471531868, -0.06856687366962433, 0.007166938856244087, 0.1250058263540268, -0.13677969574928284, 0.11705687642097473, 0.009373586624860764, -0.062045786529779434, -0.09691495448350906, -0.11988254636526108, -0.06806539744138718, -0.033930521458387375, -0.00029943810659460723, -0.12326493114233017, 0.012530683539807796, 0.02643304504454136, 0.01839432492852211, -0.003293743822723627, 0.12045112252235413, -0.08425423502922058, -0.09150394052267075, 0.09220296889543533, -0.01622350513935089, 0.04016399011015892, 0.010102355852723122, 0.036132052540779114, 0.020340867340564728, 0.09055253118276596, 0.04071667790412903, 0.06058784946799278, 0.0406765379011631, 0.022403109818696976, -0.09553436189889908, -0.0846819132566452, 0.0032751220278441906, -0.003919276874512434, -0.04310031235218048, 0.09835168719291687, 0.03660186752676964, -0.08402950316667557, -0.015653973445296288, 0.222230926156044, -0.09502024948596954, -0.12386364489793777, -0.17473818361759186, 0.16873696446418762, 0.030060650780797005, 0.028306394815444946, -0.024526795372366905, -0.0880458876490593, -0.025812465697526932, 0.1529158055782318, 0.20833998918533325, -0.07100238651037216, 0.025658713653683662, 0.06401697546243668, 0.017550406977534294, 0.03837653249502182, 0.03021821565926075, 0.03615964204072952, 0.18278439342975616, -0.046115051954984665, 0.10104057192802429, -0.013751646503806114, -0.06887686252593994, -0.07846679538488388, 0.11736905574798584, 0.014623229391872883, 0.029069583863019943, -0.0247003473341465, 0.10637915879487991, -0.06499586254358292, -0.147087961435318, -0.033173102885484695, -0.09810777753591537, -0.12480176985263824, -0.03636280819773674, 0.044411756098270416, 0.026820342987775803, 0.08104401081800461, 0.037465889006853104, -0.04165663570165634, 0.1304418444633484, 0.001823189901188016, -0.053098585456609726, -0.024838844314217567, 0.0408797413110733, -0.06095685064792633, 0.1404743492603302, 0.006532035302370787, -0.01754666306078434, 0.12616582214832306, -0.004713025875389576, -0.05587528273463249, 0.059071216732263565, 0.039010189473629, -0.06444051861763, 0.09771361947059631, 0.09998604655265808, -0.021994443610310555, 0.10037614405155182, 0.07243965566158295, -0.16828422248363495, 0.058048613369464874, 0.005049819592386484, -0.04299210011959076, -0.06730222702026367, 0.0457376204431057, -0.09189700335264206, 0.10704787820577621, 0.18088111281394958, -0.02302515134215355, -0.010091161355376244, -0.008250445127487183, 0.007656346540898085, 0.0186519343405962, 0.039279066026210785, -0.06794095039367676, -0.12005452811717987, 0.00460387859493494, 0.018324417993426323, 0.037461042404174805, -0.26882800459861755, -0.12673038244247437, 0.03449464961886406, -0.011087811551988125, -0.04353330284357071, 0.12296539545059204, 0.08558213710784912, 0.011356789618730545, -0.03234255313873291, -0.20835435390472412, 0.016234485432505608, 0.10861578583717346, -0.12562894821166992, -0.07134871929883957 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bart-finetuned-pubmed This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.5363 - Rouge2 Precision: 0.3459 - Rouge2 Recall: 0.2455 - Rouge2 Fmeasure: 0.2731 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure | |:-------------:|:-----:|:-----:|:---------------:|:----------------:|:-------------:|:---------------:| | 1.652 | 1.0 | 1125 | 1.5087 | 0.3647 | 0.2425 | 0.2772 | | 1.4695 | 2.0 | 2250 | 1.5039 | 0.3448 | 0.2457 | 0.2732 | | 1.3714 | 3.0 | 3375 | 1.4842 | 0.3509 | 0.2474 | 0.277 | | 1.2734 | 4.0 | 4500 | 1.4901 | 0.3452 | 0.2426 | 0.2716 | | 1.1853 | 5.0 | 5625 | 1.5152 | 0.3658 | 0.2371 | 0.2744 | | 1.0975 | 6.0 | 6750 | 1.5133 | 0.3529 | 0.2417 | 0.2729 | | 1.0448 | 7.0 | 7875 | 1.5203 | 0.3485 | 0.2464 | 0.275 | | 0.9999 | 8.0 | 9000 | 1.5316 | 0.3437 | 0.2435 | 0.2719 | | 0.9732 | 9.0 | 10125 | 1.5338 | 0.3464 | 0.2446 | 0.2732 | | 0.954 | 10.0 | 11250 | 1.5363 | 0.3459 | 0.2455 | 0.2731 | ### Framework versions - Transformers 4.12.3 - Pytorch 1.9.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "bart-finetuned-pubmed", "results": []}]}
text2text-generation
gayanin/bart-finetuned-pubmed
[ "transformers", "pytorch", "tensorboard", "bart", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
bart-finetuned-pubmed ===================== This model is a fine-tuned version of facebook/bart-base on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 1.5363 * Rouge2 Precision: 0.3459 * Rouge2 Recall: 0.2455 * Rouge2 Fmeasure: 0.2731 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 10 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.12.3 * Pytorch 1.9.0+cu111 * Datasets 1.15.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ 57, 113, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ -0.08508943021297455, 0.059009723365306854, -0.0025067217648029327, 0.0979013592004776, 0.15000393986701965, 0.015550121665000916, 0.14306899905204773, 0.12910985946655273, -0.1104147657752037, 0.01497762743383646, 0.11377789825201035, 0.15263836085796356, 0.02282089740037918, 0.12985338270664215, -0.04324980080127716, -0.26661962270736694, -0.001931220875121653, 0.0425134040415287, -0.05081082880496979, 0.13756169378757477, 0.09351776540279388, -0.1283954679965973, 0.06209288537502289, 0.013259830884635448, -0.20614998042583466, 0.011243443004786968, 0.012654395774006844, -0.055894143879413605, 0.1563408225774765, 0.02879217453300953, 0.1248810738325119, 0.015392672270536423, 0.08919098973274231, -0.20269495248794556, 0.013288356363773346, 0.05712157115340233, 0.012704377993941307, 0.08693691343069077, 0.07019326090812683, 0.008502849377691746, 0.12195562571287155, -0.0550692044198513, 0.06569162756204605, 0.02171821892261505, -0.1268371343612671, -0.24371092021465302, -0.09617938101291656, 0.01730206236243248, 0.06793340295553207, 0.10327325761318207, -0.007038003299385309, 0.14202986657619476, -0.0907726138830185, 0.09685949236154556, 0.22535623610019684, -0.2912481427192688, -0.06453539431095123, 0.012027227319777012, 0.046310845762491226, 0.08498656749725342, -0.10936573892831802, -0.03121129423379898, 0.036769136786460876, 0.0533435083925724, 0.15099161863327026, -0.029004016891121864, -0.12851785123348236, 0.0067993816919624805, -0.13991206884384155, -0.04244416207075119, 0.11976934969425201, 0.033551059663295746, -0.03062145784497261, -0.049663301557302475, -0.06021909788250923, -0.15591871738433838, -0.04658692702651024, -0.015481983311474323, 0.04630829766392708, -0.020221740007400513, -0.07613331824541092, -0.0204966701567173, -0.10741250216960907, -0.06928207725286484, -0.06605354696512222, 0.11923778802156448, 0.04354887828230858, 0.0021922877058386803, -0.03597787395119667, 0.1066894605755806, 0.0031921761110424995, -0.13282306492328644, 0.029967913404107094, 0.03416052460670471, -0.01832577772438526, -0.04111199080944061, -0.06914754211902618, -0.08068716526031494, 0.0047184620052576065, 0.13986323773860931, -0.05247752368450165, 0.05505174398422241, -0.006124483421444893, 0.04676951840519905, -0.10856454819440842, 0.1862279772758484, -0.037313174456357956, -0.014556754380464554, 0.005099866539239883, 0.05809007212519646, 0.0073711019940674305, -0.018224989995360374, -0.10909736901521683, 0.0155162513256073, 0.11434587091207504, 0.01566154696047306, -0.051864951848983765, 0.06545158475637436, -0.04986627399921417, -0.019711673259735107, -0.005383384879678488, -0.09397685527801514, 0.036940839141607285, -0.00046732506598345935, -0.07648282498121262, -0.01392708346247673, 0.027037927880883217, 0.033059123903512955, -0.023060359060764313, 0.10372761636972427, -0.07326657325029373, 0.040119774639606476, -0.11034098267555237, -0.12985828518867493, 0.02379360795021057, -0.05093706399202347, 0.01257750578224659, -0.09680778533220291, -0.1690410077571869, -0.023198872804641724, 0.06315179914236069, -0.024221666157245636, -0.04872012883424759, -0.04764297977089882, -0.0591876357793808, 0.01857364922761917, -0.030084695667028427, 0.15341104567050934, -0.059667933732271194, 0.11093664914369583, 0.026811260730028152, 0.056811388581991196, -0.049244385212659836, 0.0640232115983963, -0.10023825615644455, 0.004804693628102541, -0.17909571528434753, 0.04144448786973953, -0.04886896163225174, 0.0691615492105484, -0.09584195911884308, -0.09476418048143387, 0.0070894756354391575, -0.0012252135202288628, 0.0922248363494873, 0.08148819953203201, -0.18312959372997284, -0.07345020771026611, 0.1803755760192871, -0.06637325882911682, -0.10603101551532745, 0.12413016706705093, -0.06125931069254875, 0.05605032294988632, 0.07485154271125793, 0.18447671830654144, 0.060495149344205856, -0.07972956448793411, 0.042815420776605606, -0.02310822531580925, 0.061461031436920166, -0.05061369016766548, 0.06144082173705101, -0.0015740091912448406, 0.008997824974358082, 0.027296042069792747, -0.018914513289928436, 0.0807005912065506, -0.09252570569515228, -0.08875570446252823, -0.03702326491475105, -0.08550652116537094, 0.0416337326169014, 0.06592554599046707, 0.07715988904237747, -0.09670800715684891, -0.08725818246603012, 0.07504115253686905, 0.07843797653913498, -0.07161097973585129, 0.042058903723955154, -0.05797753855586052, 0.053218938410282135, -0.031644925475120544, -0.010933670215308666, -0.18739967048168182, -0.011484996415674686, 0.01749764010310173, -0.01991698332130909, 0.035966865718364716, 0.01567731611430645, 0.07415371388196945, 0.060765426605939865, -0.04959912598133087, -0.026344681158661842, -0.03088304027915001, -0.005503952037543058, -0.13424921035766602, -0.20397503674030304, -0.028385931625962257, -0.021713485941290855, 0.13319361209869385, -0.20120537281036377, 0.036407094448804855, -0.021155470982193947, 0.07823749631643295, 0.012632254511117935, -0.009054061956703663, -0.04479849711060524, 0.09318310022354126, -0.03390050306916237, -0.04768531769514084, 0.07911723852157593, 0.015714123845100403, -0.09086770564317703, -0.007738233543932438, -0.13209213316440582, 0.15478350222110748, 0.13330721855163574, -0.10844607651233673, -0.07073138654232025, -0.021631086245179176, -0.05412955582141876, -0.042624473571777344, -0.03269109129905701, 0.028814762830734253, 0.18221105635166168, 0.005402253475040197, 0.15595926344394684, -0.06895509362220764, -0.04544135183095932, 0.014721529558300972, -0.030812514945864677, 0.03728782758116722, 0.1093631461262703, 0.10342521965503693, -0.06606509536504745, 0.13670805096626282, 0.1594262570142746, -0.09019361436367035, 0.1341284215450287, -0.04218955338001251, -0.0805460512638092, -0.018087977543473244, -0.015967579558491707, 0.003311882959678769, 0.08806391060352325, -0.1248554214835167, 0.0021502303425222635, 0.021898619830608368, 0.027150681242346764, 0.025007324293255806, -0.2304145097732544, -0.028616998344659805, 0.032168496400117874, -0.04988011717796326, -0.01193144079297781, -0.02748367004096508, 0.010204290971159935, 0.10510771721601486, -0.004440648481249809, -0.08200264722108841, 0.03024553507566452, 0.002979288110509515, -0.07817656546831131, 0.20116856694221497, -0.09601984173059464, -0.16362953186035156, -0.11990972608327866, -0.08466774225234985, -0.025407973676919937, 0.0006333276396617293, 0.07058113068342209, -0.0763314813375473, -0.02583201974630356, -0.06398262083530426, 0.022747932001948357, 0.005122330039739609, 0.016851583495736122, -0.005090611521154642, -0.006792239844799042, 0.07624347507953644, -0.11497985571622849, -0.005086334887892008, -0.04206455871462822, -0.06378201395273209, 0.056735120713710785, 0.04287910833954811, 0.119346484541893, 0.15807569026947021, -0.017323600128293037, 0.0033621115144342184, -0.03251056745648384, 0.21456749737262726, -0.07168623059988022, -0.02572282776236534, 0.13303974270820618, -0.012810795567929745, 0.05921174958348274, 0.11645162105560303, 0.06894069164991379, -0.08348559588193893, 0.01566815748810768, 0.030745603144168854, -0.028181428089737892, -0.2238590568304062, -0.03450171276926994, -0.05697457119822502, -0.021584440022706985, 0.09504575282335281, 0.022554855793714523, 0.05489036813378334, 0.05587419867515564, 0.03140489012002945, 0.0711350217461586, -0.0206539835780859, 0.06378024071455002, 0.13149724900722504, 0.0422884076833725, 0.13845385611057281, -0.039974238723516464, -0.06436894088983536, 0.033062320202589035, 0.0030130238737910986, 0.22307497262954712, 0.016209924593567848, 0.15206509828567505, 0.061359718441963196, 0.18071986734867096, 0.007526475470513105, 0.07722993940114975, 0.0016623445553705096, -0.03705189377069473, -0.018608173355460167, -0.04127879813313484, -0.03855584189295769, 0.017654506489634514, -0.04842754453420639, 0.03876636177301407, -0.10856906324625015, -0.043629225343465805, 0.04754512012004852, 0.2771961987018585, 0.023778056725859642, -0.31864532828330994, -0.07717172801494598, -0.00036911855568178, -0.055947743356227875, -0.019434863701462746, 0.01531248353421688, 0.08932564407587051, -0.10255590826272964, 0.030825791880488396, -0.07673387974500656, 0.10686583817005157, -0.0338410958647728, 0.0513697974383831, 0.05703167989850044, 0.09064888954162598, 0.0074427989311516285, 0.07733648270368576, -0.34322959184646606, 0.2757537066936493, 0.0015863314038142562, 0.0685863271355629, -0.07180868834257126, 0.004200204275548458, 0.03760520741343498, 0.030006008222699165, 0.03675540164113045, -0.023816727101802826, -0.05010373890399933, -0.1979699432849884, -0.06083349138498306, 0.03338848426938057, 0.08849044889211655, -0.015208135358989239, 0.10724756866693497, -0.03637489303946495, 0.013555774465203285, 0.07933959364891052, -0.01817752607166767, -0.08449028432369232, -0.10038159787654877, -0.0021290152799338102, 0.02349826693534851, -0.008148910477757454, -0.07504821568727493, -0.11203759163618088, -0.1086123138666153, 0.14376617968082428, -0.000505576201248914, -0.02043701335787773, -0.11413354426622391, 0.08418041467666626, 0.08054371178150177, -0.08268081396818161, 0.038226235657930374, 0.012076199986040592, 0.07253238558769226, 0.021928640082478523, -0.06869162619113922, 0.11743517965078354, -0.061591990292072296, -0.15685246884822845, -0.06077073514461517, 0.09617327898740768, 0.03495820239186287, 0.07084425538778305, -0.00955252442508936, 0.014181778766214848, -0.035478610545396805, -0.08395445346832275, 0.007323778234422207, -0.016649561002850533, 0.054046884179115295, 0.005613586865365505, -0.05917006731033325, 0.013119431212544441, -0.06673164665699005, -0.04861362278461456, 0.19685441255569458, 0.23643267154693604, -0.09236804395914078, 0.04123289883136749, 0.0566670298576355, -0.07564523071050644, -0.18556009232997894, 0.028464004397392273, 0.06030731275677681, 0.006424798630177975, 0.06219351291656494, -0.18720024824142456, 0.09042482823133469, 0.10375336557626724, -0.015704043209552765, 0.08750089257955551, -0.3450983464717865, -0.1287536919116974, 0.12261179834604263, 0.14835749566555023, 0.0773298367857933, -0.15730728209018707, -0.024423761293292046, -0.024487338960170746, -0.12253736704587936, 0.10799097269773483, -0.10343554615974426, 0.12569235265254974, -0.024922823533415794, 0.09015955030918121, 0.005423529539257288, -0.05952611565589905, 0.11246880888938904, -0.009608081541955471, 0.09860018640756607, -0.06457391381263733, 0.012750453315675259, 0.052541375160217285, -0.03626308590173721, 0.009744668379426003, -0.09385711699724197, 0.017950303852558136, -0.08402106165885925, -0.023100828751921654, -0.08023522794246674, 0.03166304528713226, -0.03804000839591026, -0.05968541279435158, -0.025253375992178917, 0.02366829104721546, 0.05212300270795822, -0.010717310942709446, 0.12041906267404556, 0.003172803670167923, 0.1666606068611145, 0.12051404267549515, 0.06679627299308777, -0.060455434024333954, -0.05864408612251282, -0.02040850557386875, -0.01916990801692009, 0.05407509580254555, -0.12644502520561218, 0.03148113936185837, 0.1482316106557846, 0.01112260390073061, 0.14705859124660492, 0.07279752939939499, -0.04138642176985741, 0.021508703008294106, 0.05808815732598305, -0.14601685106754303, -0.0978819876909256, 0.004262073896825314, -0.012549336068332195, -0.0993611142039299, 0.024093683809041977, 0.10582197457551956, -0.07025681436061859, -0.014245224185287952, -0.001300618750974536, 0.004427036270499229, -0.056685078889131546, 0.20954129099845886, 0.040825240314006805, 0.04488629102706909, -0.10133390873670578, 0.06986052542924881, 0.06670163571834564, -0.09199976176023483, 0.009679063223302364, 0.09358027577400208, -0.07091141492128372, -0.042494792491197586, 0.09825179725885391, 0.18157583475112915, -0.06053750962018967, -0.051968205720186234, -0.13865335285663605, -0.12631288170814514, 0.08258902281522751, 0.15249674022197723, 0.09179148823022842, 0.011042304337024689, -0.05701606348156929, 0.018072238191962242, -0.11137238144874573, 0.08557796478271484, 0.05146106705069542, 0.06654790043830872, -0.11732673645019531, 0.17865952849388123, 0.019580595195293427, 0.02943076565861702, -0.019377119839191437, 0.018659386783838272, -0.10073792934417725, 0.019098112359642982, -0.1542644202709198, -0.030792422592639923, -0.02281738445162773, 0.003935718443244696, -0.008415092714130878, -0.05548715218901634, -0.05324416235089302, 0.010335377417504787, -0.12480630725622177, -0.031239911913871765, 0.016602858901023865, 0.052919384092092514, -0.12200459092855453, -0.03997235372662544, 0.029674500226974487, -0.06168346107006073, 0.06309321522712708, 0.042134664952754974, 0.015569743700325489, 0.059671927243471146, -0.15463140606880188, -0.0007257514516822994, 0.05476606637239456, 0.014725166372954845, 0.053866829723119736, -0.08946948498487473, -0.014842146076261997, 0.009703661315143108, 0.06772691756486893, 0.014167545363307, 0.07901333272457123, -0.14037492871284485, -0.016553498804569244, -0.027473220601677895, -0.08790276944637299, -0.06567377597093582, 0.034813325852155685, 0.05885424092411995, 0.02534758858382702, 0.19230642914772034, -0.08663883060216904, 0.04423535615205765, -0.2236906737089157, 0.004885602742433548, -0.018386807292699814, -0.11208593100309372, -0.11076988279819489, -0.07442624121904373, 0.062069397419691086, -0.048831336200237274, 0.1306600421667099, 0.016975119709968567, 0.05512627959251404, 0.030559100210666656, -0.03522340580821037, 0.0068151578307151794, 0.017865292727947235, 0.20701195299625397, 0.035250384360551834, -0.02992667257785797, 0.0604645200073719, 0.05401676148176193, 0.08726005256175995, 0.11906958371400833, 0.1900363266468048, 0.16428078711032867, 0.013431509025394917, 0.08362184464931488, 0.036031950265169144, -0.057825151830911636, -0.1527434140443802, 0.05429701879620552, -0.03974258154630661, 0.10924030840396881, -0.03675759956240654, 0.24809126555919647, 0.08124599605798721, -0.16882838308811188, 0.05896889790892601, -0.053044259548187256, -0.08403248339891434, -0.09942585229873657, -0.05208740755915642, -0.08235858380794525, -0.15386471152305603, -0.010525263845920563, -0.10673623532056808, 0.044369589537382126, 0.1073075458407402, 0.012546357698738575, -0.021651610732078552, 0.1406983733177185, 0.04691668599843979, 0.006136175710707903, 0.05102555453777313, -0.00447231438010931, -0.019898148253560066, -0.11069278419017792, -0.07684430480003357, -0.0005416942294687033, 0.00044294176041148603, 0.0414026640355587, -0.04458095133304596, -0.057611048221588135, 0.03807045519351959, -0.03433486074209213, -0.10217992961406708, 0.020144127309322357, 0.015232115983963013, 0.07673296332359314, 0.06228181719779968, 0.016810225322842598, 0.00863565132021904, -0.011453939601778984, 0.23339354991912842, -0.07703941315412521, -0.10067247599363327, -0.10010910034179688, 0.2554469406604767, 0.03317548334598541, -0.01631402224302292, 0.02466154471039772, -0.0558704249560833, 0.0041338978335261345, 0.24604664742946625, 0.18470804393291473, -0.09337583184242249, -0.015010997653007507, 0.005467281676828861, -0.01287776604294777, -0.031064262613654137, 0.12513819336891174, 0.14502140879631042, 0.043014541268348694, -0.1038193628191948, -0.029462693259119987, -0.05220408737659454, -0.020500337705016136, -0.046871017664670944, 0.06598906219005585, 0.02909649722278118, 0.0028624404221773148, -0.031630344688892365, 0.06335922330617905, -0.04857010021805763, -0.08541370928287506, 0.012047038413584232, -0.19373968243598938, -0.16244837641716003, -0.017119649797677994, 0.12069724500179291, -0.0009714553598314524, 0.051065411418676376, -0.025881599634885788, 0.018005555495619774, 0.07385758310556412, -0.02856617048382759, -0.06588538736104965, -0.0843164250254631, 0.1033530905842781, -0.13006092607975006, 0.1929360181093216, -0.04420817643404007, 0.0557992085814476, 0.12757214903831482, 0.07224567979574203, -0.06711060553789139, 0.07589215785264969, 0.03937124088406563, -0.08052803575992584, 0.03466000780463219, 0.09806333482265472, -0.02974681183695793, 0.06524477899074554, 0.04520444571971893, -0.12981221079826355, 0.039959270507097244, -0.08914180845022202, -0.04995410516858101, -0.024910524487495422, -0.04553482308983803, -0.04854457080364227, 0.12769733369350433, 0.21056927740573883, -0.030881602317094803, 0.015373718924820423, -0.07948528230190277, 0.0021194927394390106, 0.04728483036160469, 0.046706512570381165, -0.0742158591747284, -0.23775643110275269, 0.008425078354775906, 0.0590129978954792, -0.017553891986608505, -0.24586652219295502, -0.10374633967876434, 0.0008912449120543897, -0.07367867231369019, -0.10168904811143875, 0.09770659357309341, 0.08680050075054169, 0.0485570915043354, -0.051154911518096924, -0.10804006457328796, -0.07781399041414261, 0.16249273717403412, -0.1471216082572937, -0.07565679401159286 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bart-mlm-pubmed-15 This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.4822 - Rouge2 Precision: 0.7578 - Rouge2 Recall: 0.5933 - Rouge2 Fmeasure: 0.6511 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure | |:-------------:|:-----:|:----:|:---------------:|:----------------:|:-------------:|:---------------:| | 0.7006 | 1.0 | 663 | 0.5062 | 0.7492 | 0.5855 | 0.6434 | | 0.5709 | 2.0 | 1326 | 0.4811 | 0.7487 | 0.5879 | 0.6447 | | 0.5011 | 3.0 | 1989 | 0.4734 | 0.7541 | 0.5906 | 0.6483 | | 0.4164 | 4.0 | 2652 | 0.4705 | 0.7515 | 0.5876 | 0.6452 | | 0.3888 | 5.0 | 3315 | 0.4703 | 0.7555 | 0.5946 | 0.6515 | | 0.3655 | 6.0 | 3978 | 0.4725 | 0.7572 | 0.5943 | 0.6516 | | 0.319 | 7.0 | 4641 | 0.4733 | 0.7557 | 0.5911 | 0.6491 | | 0.3089 | 8.0 | 5304 | 0.4792 | 0.7577 | 0.5936 | 0.6513 | | 0.2907 | 9.0 | 5967 | 0.4799 | 0.7577 | 0.5931 | 0.6509 | | 0.275 | 10.0 | 6630 | 0.4822 | 0.7578 | 0.5933 | 0.6511 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.10.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "bart-mlm-pubmed-15", "results": []}]}
text2text-generation
gayanin/bart-mlm-pubmed-15
[ "transformers", "pytorch", "tensorboard", "bart", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
bart-mlm-pubmed-15 ================== This model is a fine-tuned version of facebook/bart-base on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 0.4822 * Rouge2 Precision: 0.7578 * Rouge2 Recall: 0.5933 * Rouge2 Fmeasure: 0.6511 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 10 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.12.5 * Pytorch 1.10.0+cu111 * Datasets 1.15.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ 57, 113, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ -0.0867219790816307, 0.055462174117565155, -0.002274750731885433, 0.09676497429609299, 0.15085238218307495, 0.012637375853955746, 0.14072179794311523, 0.12473951280117035, -0.11230833828449249, 0.017588185146450996, 0.11828296631574631, 0.14802083373069763, 0.024210374802350998, 0.1252502202987671, -0.045356061309576035, -0.262150377035141, -0.00002328762820980046, 0.041856519877910614, -0.05285303294658661, 0.13920284807682037, 0.0906849279999733, -0.12862668931484222, 0.06566891074180603, 0.016234852373600006, -0.21103446185588837, 0.008428194560110569, 0.011844725348055363, -0.052661146968603134, 0.15579554438591003, 0.03129447251558304, 0.1261555701494217, 0.016640443354845047, 0.09018171578645706, -0.20001967251300812, 0.013460533693432808, 0.05625590682029724, 0.013338098302483559, 0.08852751553058624, 0.06908217072486877, 0.010756932199001312, 0.12144075334072113, -0.05637706071138382, 0.06317462772130966, 0.021377360448241234, -0.12813062965869904, -0.2361792027950287, -0.09788632392883301, 0.01742962934076786, 0.06630835682153702, 0.10380984842777252, -0.00795662496238947, 0.1452450305223465, -0.0946977362036705, 0.0953841581940651, 0.22252486646175385, -0.29312217235565186, -0.06660237908363342, 0.005923801567405462, 0.04689629748463631, 0.08137331157922745, -0.10814683139324188, -0.03077341988682747, 0.03911865875124931, 0.051733557134866714, 0.15108142793178558, -0.027031617239117622, -0.1255291998386383, 0.009782832115888596, -0.14126217365264893, -0.04250112920999527, 0.11205080151557922, 0.03479015454649925, -0.03083561733365059, -0.04649687185883522, -0.06090623512864113, -0.16255712509155273, -0.048318225890398026, -0.015000710263848305, 0.04703385382890701, -0.021401921287178993, -0.07937237620353699, -0.019585834816098213, -0.10484049469232559, -0.0716506764292717, -0.06771253049373627, 0.12509149312973022, 0.041989609599113464, 0.0011333179427310824, -0.03612881526350975, 0.10573788732290268, 0.0015543913468718529, -0.13245007395744324, 0.02922716923058033, 0.03680860623717308, -0.021345503628253937, -0.04338424280285835, -0.06522190570831299, -0.0797869861125946, 0.0042693219147622585, 0.14050699770450592, -0.05594496801495552, 0.05111481249332428, -0.0034115430898964405, 0.04727201908826828, -0.11071038246154785, 0.18794257938861847, -0.04299888759851456, -0.010948577895760536, 0.006456696894019842, 0.0576329380273819, 0.008746226318180561, -0.017962101846933365, -0.11141595244407654, 0.013874282129108906, 0.11075880378484726, 0.017528321593999863, -0.0515793077647686, 0.06489716470241547, -0.05094553157687187, -0.01589319296181202, 0.0007782398606650531, -0.09029772877693176, 0.03612196817994118, -0.0028094081208109856, -0.07799360901117325, -0.014370918273925781, 0.024872036650776863, 0.0323185995221138, -0.02158982865512371, 0.10786677896976471, -0.07814837992191315, 0.04184802621603012, -0.10921868681907654, -0.13197839260101318, 0.02444618009030819, -0.049501173198223114, 0.012133109383285046, -0.09642373770475388, -0.16728010773658752, -0.02236819453537464, 0.0593084841966629, -0.027130184695124626, -0.05025933310389519, -0.04999887943267822, -0.06405632197856903, 0.022886520251631737, -0.029339058324694633, 0.1537834256887436, -0.06181749328970909, 0.11160467565059662, 0.0309202428907156, 0.05685362219810486, -0.05136741325259209, 0.061943039298057556, -0.10100964456796646, 0.007471778895705938, -0.18199418485164642, 0.04249980300664902, -0.04797246307134628, 0.0761336013674736, -0.09424274414777756, -0.09534361958503723, 0.00478152884170413, -0.0008567497716285288, 0.09695705771446228, 0.08531241118907928, -0.1826077401638031, -0.07473284006118774, 0.18174810707569122, -0.06340796500444412, -0.10546950995922089, 0.12881304323673248, -0.059853445738554, 0.05344957113265991, 0.07425983995199203, 0.18069881200790405, 0.057555459439754486, -0.08471137285232544, 0.03798159584403038, -0.02432074397802353, 0.05915605276823044, -0.054047420620918274, 0.06302451342344284, 0.0009308145381510258, 0.011113142594695091, 0.024584030732512474, -0.01892632432281971, 0.08416716754436493, -0.09559695422649384, -0.08789679408073425, -0.03893548250198364, -0.08772057294845581, 0.039365991950035095, 0.06298413127660751, 0.07542391866445541, -0.09879676252603531, -0.09122679382562637, 0.07277120649814606, 0.07864084839820862, -0.06939317286014557, 0.041935428977012634, -0.060788344591856, 0.05782414972782135, -0.039004795253276825, -0.01181128527969122, -0.18654347956180573, -0.011404315009713173, 0.015132223255932331, -0.013601926155388355, 0.03996402397751808, 0.014380838721990585, 0.07381507754325867, 0.060963332653045654, -0.04813798889517784, -0.026154227554798126, -0.032566383481025696, -0.007290080189704895, -0.13407230377197266, -0.20358924567699432, -0.027620896697044373, -0.021122118458151817, 0.13551472127437592, -0.20138941705226898, 0.03468279913067818, -0.016711212694644928, 0.07772519439458847, 0.01212365459650755, -0.007773156277835369, -0.04515509679913521, 0.09454989433288574, -0.034737929701805115, -0.04874570295214653, 0.0757054015994072, 0.016684282571077347, -0.09071344882249832, -0.0038776614237576723, -0.12874078750610352, 0.1584293246269226, 0.13580897450447083, -0.10734445601701736, -0.07278770953416824, -0.018259037286043167, -0.054592207074165344, -0.04256129264831543, -0.033663660287857056, 0.03222687914967537, 0.18704986572265625, 0.006103041581809521, 0.15579484403133392, -0.07052551209926605, -0.047700610011816025, 0.015909289941191673, -0.02872025966644287, 0.038189150393009186, 0.11265859007835388, 0.10697872191667557, -0.07236873358488083, 0.13936497271060944, 0.15398815274238586, -0.0895383358001709, 0.13469843566417694, -0.041443824768066406, -0.083169125020504, -0.01970035210251808, -0.015816107392311096, 0.004152530804276466, 0.09045837819576263, -0.12357074022293091, -0.0009135896689258516, 0.020951200276613235, 0.0239117369055748, 0.024296946823596954, -0.23080766201019287, -0.0295211523771286, 0.03138891980051994, -0.04991867393255234, -0.01711665466427803, -0.028307540342211723, 0.009715849533677101, 0.10614278167486191, -0.00228509702719748, -0.08267837017774582, 0.027540652081370354, 0.0068334247916936874, -0.07535014301538467, 0.19980572164058685, -0.0934063196182251, -0.16080807149410248, -0.1213449239730835, -0.08109356462955475, -0.02593882940709591, -0.00087083870312199, 0.06687451153993607, -0.07908467203378677, -0.026373527944087982, -0.06183188781142235, 0.022726237773895264, 0.0034915136639028788, 0.01788340136408806, -0.005439021158963442, -0.008004619739949703, 0.0744732990860939, -0.11406894773244858, -0.004223077557981014, -0.04135558754205704, -0.059665754437446594, 0.05540837347507477, 0.04445236176252365, 0.12288699299097061, 0.15919733047485352, -0.019143497571349144, 0.002628567162901163, -0.03206859529018402, 0.21784807741641998, -0.07383282482624054, -0.027769632637500763, 0.12586188316345215, -0.009886739775538445, 0.059406962245702744, 0.11426973342895508, 0.06983331590890884, -0.08660981059074402, 0.015440749004483223, 0.029467498883605003, -0.02661033533513546, -0.22428545355796814, -0.0344984345138073, -0.05539081245660782, -0.025857549160718918, 0.09168455004692078, 0.02109621837735176, 0.05179550498723984, 0.054536715149879456, 0.02821865864098072, 0.07501064985990524, -0.02563261240720749, 0.0634828433394432, 0.12859152257442474, 0.044475167989730835, 0.13737396895885468, -0.042828015983104706, -0.06764766573905945, 0.03250397741794586, -0.0004876790044363588, 0.2203844040632248, 0.017161650583148003, 0.1652754843235016, 0.056817904114723206, 0.18229453265666962, 0.011886241845786572, 0.07815506309270859, 0.004910087678581476, -0.03957361355423927, -0.016728149726986885, -0.04107232019305229, -0.04007142037153244, 0.01632532849907875, -0.04514501616358757, 0.0391554981470108, -0.10711868852376938, -0.03853718563914299, 0.04627954959869385, 0.27932265400886536, 0.021928509697318077, -0.32240667939186096, -0.07944931834936142, -0.003353038104251027, -0.05579080432653427, -0.021180815994739532, 0.014041367918252945, 0.09300810843706131, -0.09714332222938538, 0.0287052933126688, -0.0773022472858429, 0.10984886437654495, -0.036719538271427155, 0.049680717289447784, 0.05390894040465355, 0.09313826262950897, 0.004795422777533531, 0.07552178204059601, -0.34024909138679504, 0.2803536653518677, 0.00493947509676218, 0.07066044211387634, -0.07081294059753418, 0.0007122082752175629, 0.036577872931957245, 0.03317338600754738, 0.0364554263651371, -0.022193681448698044, -0.04955914989113808, -0.19878827035427094, -0.06574621051549911, 0.030401309952139854, 0.09008432924747467, -0.01220324169844389, 0.11066174507141113, -0.03505552560091019, 0.011278372257947922, 0.07857942581176758, -0.021211452782154083, -0.08122111856937408, -0.09851276129484177, -0.003853484522551298, 0.02492848038673401, -0.01119220070540905, -0.07381158322095871, -0.11267706006765366, -0.10793549567461014, 0.14149123430252075, 0.0014742484781891108, -0.017179269343614578, -0.11443088203668594, 0.08256468921899796, 0.0818248838186264, -0.08289701491594315, 0.03618624061346054, 0.01137845404446125, 0.07406225055456161, 0.020498428493738174, -0.06446254998445511, 0.11837790161371231, -0.059787798672914505, -0.1557784527540207, -0.05858810245990753, 0.09455057233572006, 0.03461139276623726, 0.07255010306835175, -0.010767288506031036, 0.014234802685678005, -0.03665432706475258, -0.08398229628801346, 0.00625862181186676, -0.022087380290031433, 0.05600825324654579, 0.004411609843373299, -0.05165262892842293, 0.020836126059293747, -0.06702107191085815, -0.047911353409290314, 0.19497588276863098, 0.2395961731672287, -0.09146451205015182, 0.03606722876429558, 0.05541786924004555, -0.07244808971881866, -0.1859205812215805, 0.03345244377851486, 0.058628231287002563, 0.01092927809804678, 0.05417318269610405, -0.18450801074504852, 0.08502987027168274, 0.10271044820547104, -0.015260961838066578, 0.08886101096868515, -0.33973658084869385, -0.12911880016326904, 0.11879928410053253, 0.14507518708705902, 0.07987871766090393, -0.15548162162303925, -0.024761725217103958, -0.029475124552845955, -0.12206004559993744, 0.10974925011396408, -0.10247453302145004, 0.12515240907669067, -0.024495430290699005, 0.0931176021695137, 0.005357546731829643, -0.06057319790124893, 0.11408938467502594, -0.012519733980298042, 0.10009486228227615, -0.06385384500026703, 0.013631567358970642, 0.054003387689590454, -0.03547247126698494, 0.01032607164233923, -0.09419726580381393, 0.016805468127131462, -0.0844309851527214, -0.022668074816465378, -0.07763423770666122, 0.03207119181752205, -0.037714798003435135, -0.05581261217594147, -0.02734486758708954, 0.02279684692621231, 0.04741416871547699, -0.0103211160749197, 0.11845803260803223, 0.002253623679280281, 0.16185960173606873, 0.11441992223262787, 0.07241763174533844, -0.06480717658996582, -0.053468264639377594, -0.014491192065179348, -0.018878111615777016, 0.05418642237782478, -0.12668992578983307, 0.027285650372505188, 0.14959746599197388, 0.013112148270010948, 0.14402221143245697, 0.07148712873458862, -0.04400071129202843, 0.020384516566991806, 0.058407384902238846, -0.14605429768562317, -0.10016372054815292, 0.0015120514435693622, -0.014121695421636105, -0.09633656591176987, 0.02186219021677971, 0.104742631316185, -0.07350589334964752, -0.015269096940755844, -0.00009402789874002337, 0.0009875482646748424, -0.05830801650881767, 0.20820464193820953, 0.041482239961624146, 0.04539477452635765, -0.10268230736255646, 0.06860265135765076, 0.06770068407058716, -0.09284508973360062, 0.009055975824594498, 0.09270795434713364, -0.06811615824699402, -0.041182056069374084, 0.09842722862958908, 0.1891988068819046, -0.061397816985845566, -0.05311920866370201, -0.13982443511486053, -0.12608177959918976, 0.08418476581573486, 0.16326147317886353, 0.09047462791204453, 0.010690012946724892, -0.05424687638878822, 0.01794775016605854, -0.11534462124109268, 0.08347894251346588, 0.05206974968314171, 0.06622795015573502, -0.1187308132648468, 0.18090631067752838, 0.017406543716788292, 0.030131004750728607, -0.018872877582907677, 0.020791858434677124, -0.1013152003288269, 0.020750652998685837, -0.1570514291524887, -0.029977548867464066, -0.02095569297671318, 0.00531708775088191, -0.00911243911832571, -0.05623884126543999, -0.055717792361974716, 0.007587141823023558, -0.12510471045970917, -0.029829248785972595, 0.0198475681245327, 0.0492250993847847, -0.12176875025033951, -0.03792301192879677, 0.03158491849899292, -0.061427321285009384, 0.06153474003076553, 0.044302329421043396, 0.015935782343149185, 0.06060129031538963, -0.1510305404663086, 0.0015090981032699347, 0.054345134645700455, 0.014039680361747742, 0.05875590071082115, -0.08778750151395798, -0.01577935926616192, 0.008487395010888577, 0.07210314273834229, 0.01493874005973339, 0.07730013132095337, -0.1365908682346344, -0.02042602002620697, -0.024301405996084213, -0.08779192715883255, -0.06440678983926773, 0.036969974637031555, 0.0639105960726738, 0.02465238608419895, 0.1896926462650299, -0.0847133919596672, 0.041419923305511475, -0.22318339347839355, 0.005688908044248819, -0.018722617998719215, -0.11305885016918182, -0.10584550350904465, -0.0761919617652893, 0.06391319632530212, -0.04928651079535484, 0.12931008636951447, 0.016942420974373817, 0.05590570345520973, 0.03130513057112694, -0.03239591047167778, 0.000992286833934486, 0.01864616759121418, 0.20299074053764343, 0.03815624862909317, -0.029569365084171295, 0.05931583791971207, 0.05377772077918053, 0.09008264541625977, 0.12118542194366455, 0.19257573783397675, 0.16044528782367706, 0.01828458532691002, 0.08911080658435822, 0.03772636875510216, -0.05856596678495407, -0.14712370932102203, 0.056585051119327545, -0.04469308257102966, 0.10746489465236664, -0.03561236709356308, 0.25183579325675964, 0.08441885560750961, -0.16680240631103516, 0.05733546242117882, -0.054255396127700806, -0.08562584221363068, -0.09988097846508026, -0.05033312365412712, -0.08291010558605194, -0.15420004725456238, -0.008836298249661922, -0.10521621257066727, 0.04416395351290703, 0.1034187451004982, 0.011722159571945667, -0.022144833579659462, 0.14252611994743347, 0.049787841737270355, 0.009630057029426098, 0.054459646344184875, -0.005084357690066099, -0.019953832030296326, -0.11368576437234879, -0.07848115265369415, 0.003735193982720375, -0.002830169629305601, 0.04411040619015694, -0.04335101321339607, -0.060959719121456146, 0.04084630310535431, -0.029876695945858955, -0.1036081612110138, 0.01943010650575161, 0.017292972654104233, 0.07820700854063034, 0.06008705869317055, 0.014094533398747444, 0.01108044944703579, -0.012546200305223465, 0.22888067364692688, -0.07803371548652649, -0.09721923619508743, -0.0996541753411293, 0.25048452615737915, 0.028660345822572708, -0.015824580565094948, 0.02607366256415844, -0.05443009361624718, 0.004403687082231045, 0.2448265254497528, 0.1809670329093933, -0.09583774954080582, -0.015260185115039349, 0.0016676436644047499, -0.01293659396469593, -0.03642828390002251, 0.12183144688606262, 0.145652636885643, 0.03883199393749237, -0.10613173246383667, -0.036546703428030014, -0.0550968237221241, -0.02225862815976143, -0.04098888486623764, 0.06264974176883698, 0.027477219700813293, 0.003945797681808472, -0.034828055649995804, 0.06270168721675873, -0.05248544365167618, -0.08482006192207336, 0.013557428494095802, -0.19405968487262726, -0.16283732652664185, -0.017171639949083328, 0.11503660678863525, -0.0020884971600025892, 0.0517377033829689, -0.026106225326657295, 0.016429852694272995, 0.07086073607206345, -0.029722098261117935, -0.06770484149456024, -0.08544480055570602, 0.10278977453708649, -0.1238647997379303, 0.18897047638893127, -0.04384840279817581, 0.056988347321748734, 0.12976819276809692, 0.07265783846378326, -0.06945961713790894, 0.07593805342912674, 0.039460960775613785, -0.08218730241060257, 0.03488754853606224, 0.10043741017580032, -0.030200093984603882, 0.0640149712562561, 0.042659394443035126, -0.12833768129348755, 0.042023394256830215, -0.0923188328742981, -0.046105559915304184, -0.027127021923661232, -0.046274203807115555, -0.049073364585638046, 0.1290901005268097, 0.20984455943107605, -0.02893080934882164, 0.017928577959537506, -0.07876405864953995, 0.0013245618902146816, 0.046588193625211716, 0.04883263632655144, -0.0731220617890358, -0.24251648783683777, 0.006194929126650095, 0.06343667954206467, -0.01754559576511383, -0.25067535042762756, -0.09885697066783905, 0.0006914784898981452, -0.07237914204597473, -0.10031825304031372, 0.09725553542375565, 0.08772692829370499, 0.05043494328856468, -0.05213455483317375, -0.10574068874120712, -0.07714276760816574, 0.16188697516918182, -0.1455913782119751, -0.07354407012462616 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bart-mlm-pubmed-35 This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.9359 - Rouge2 Precision: 0.5451 - Rouge2 Recall: 0.4232 - Rouge2 Fmeasure: 0.4666 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure | |:-------------:|:-----:|:----:|:---------------:|:----------------:|:-------------:|:---------------:| | 1.4156 | 1.0 | 663 | 1.0366 | 0.5165 | 0.3967 | 0.4394 | | 1.1773 | 2.0 | 1326 | 0.9841 | 0.5354 | 0.4168 | 0.4589 | | 1.0894 | 3.0 | 1989 | 0.9554 | 0.5346 | 0.4133 | 0.4563 | | 0.9359 | 4.0 | 2652 | 0.9440 | 0.5357 | 0.4163 | 0.4587 | | 0.8758 | 5.0 | 3315 | 0.9340 | 0.5428 | 0.4226 | 0.465 | | 0.8549 | 6.0 | 3978 | 0.9337 | 0.5385 | 0.422 | 0.4634 | | 0.7743 | 7.0 | 4641 | 0.9330 | 0.542 | 0.422 | 0.4647 | | 0.7465 | 8.0 | 5304 | 0.9315 | 0.5428 | 0.4231 | 0.4654 | | 0.7348 | 9.0 | 5967 | 0.9344 | 0.5462 | 0.4244 | 0.4674 | | 0.7062 | 10.0 | 6630 | 0.9359 | 0.5451 | 0.4232 | 0.4666 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.10.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "bart-mlm-pubmed-35", "results": []}]}
text2text-generation
gayanin/bart-mlm-pubmed-35
[ "transformers", "pytorch", "tensorboard", "bart", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
bart-mlm-pubmed-35 ================== This model is a fine-tuned version of facebook/bart-base on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 0.9359 * Rouge2 Precision: 0.5451 * Rouge2 Recall: 0.4232 * Rouge2 Fmeasure: 0.4666 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 10 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.12.5 * Pytorch 1.10.0+cu111 * Datasets 1.15.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ 57, 113, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ -0.0867219790816307, 0.055462174117565155, -0.002274750731885433, 0.09676497429609299, 0.15085238218307495, 0.012637375853955746, 0.14072179794311523, 0.12473951280117035, -0.11230833828449249, 0.017588185146450996, 0.11828296631574631, 0.14802083373069763, 0.024210374802350998, 0.1252502202987671, -0.045356061309576035, -0.262150377035141, -0.00002328762820980046, 0.041856519877910614, -0.05285303294658661, 0.13920284807682037, 0.0906849279999733, -0.12862668931484222, 0.06566891074180603, 0.016234852373600006, -0.21103446185588837, 0.008428194560110569, 0.011844725348055363, -0.052661146968603134, 0.15579554438591003, 0.03129447251558304, 0.1261555701494217, 0.016640443354845047, 0.09018171578645706, -0.20001967251300812, 0.013460533693432808, 0.05625590682029724, 0.013338098302483559, 0.08852751553058624, 0.06908217072486877, 0.010756932199001312, 0.12144075334072113, -0.05637706071138382, 0.06317462772130966, 0.021377360448241234, -0.12813062965869904, -0.2361792027950287, -0.09788632392883301, 0.01742962934076786, 0.06630835682153702, 0.10380984842777252, -0.00795662496238947, 0.1452450305223465, -0.0946977362036705, 0.0953841581940651, 0.22252486646175385, -0.29312217235565186, -0.06660237908363342, 0.005923801567405462, 0.04689629748463631, 0.08137331157922745, -0.10814683139324188, -0.03077341988682747, 0.03911865875124931, 0.051733557134866714, 0.15108142793178558, -0.027031617239117622, -0.1255291998386383, 0.009782832115888596, -0.14126217365264893, -0.04250112920999527, 0.11205080151557922, 0.03479015454649925, -0.03083561733365059, -0.04649687185883522, -0.06090623512864113, -0.16255712509155273, -0.048318225890398026, -0.015000710263848305, 0.04703385382890701, -0.021401921287178993, -0.07937237620353699, -0.019585834816098213, -0.10484049469232559, -0.0716506764292717, -0.06771253049373627, 0.12509149312973022, 0.041989609599113464, 0.0011333179427310824, -0.03612881526350975, 0.10573788732290268, 0.0015543913468718529, -0.13245007395744324, 0.02922716923058033, 0.03680860623717308, -0.021345503628253937, -0.04338424280285835, -0.06522190570831299, -0.0797869861125946, 0.0042693219147622585, 0.14050699770450592, -0.05594496801495552, 0.05111481249332428, -0.0034115430898964405, 0.04727201908826828, -0.11071038246154785, 0.18794257938861847, -0.04299888759851456, -0.010948577895760536, 0.006456696894019842, 0.0576329380273819, 0.008746226318180561, -0.017962101846933365, -0.11141595244407654, 0.013874282129108906, 0.11075880378484726, 0.017528321593999863, -0.0515793077647686, 0.06489716470241547, -0.05094553157687187, -0.01589319296181202, 0.0007782398606650531, -0.09029772877693176, 0.03612196817994118, -0.0028094081208109856, -0.07799360901117325, -0.014370918273925781, 0.024872036650776863, 0.0323185995221138, -0.02158982865512371, 0.10786677896976471, -0.07814837992191315, 0.04184802621603012, -0.10921868681907654, -0.13197839260101318, 0.02444618009030819, -0.049501173198223114, 0.012133109383285046, -0.09642373770475388, -0.16728010773658752, -0.02236819453537464, 0.0593084841966629, -0.027130184695124626, -0.05025933310389519, -0.04999887943267822, -0.06405632197856903, 0.022886520251631737, -0.029339058324694633, 0.1537834256887436, -0.06181749328970909, 0.11160467565059662, 0.0309202428907156, 0.05685362219810486, -0.05136741325259209, 0.061943039298057556, -0.10100964456796646, 0.007471778895705938, -0.18199418485164642, 0.04249980300664902, -0.04797246307134628, 0.0761336013674736, -0.09424274414777756, -0.09534361958503723, 0.00478152884170413, -0.0008567497716285288, 0.09695705771446228, 0.08531241118907928, -0.1826077401638031, -0.07473284006118774, 0.18174810707569122, -0.06340796500444412, -0.10546950995922089, 0.12881304323673248, -0.059853445738554, 0.05344957113265991, 0.07425983995199203, 0.18069881200790405, 0.057555459439754486, -0.08471137285232544, 0.03798159584403038, -0.02432074397802353, 0.05915605276823044, -0.054047420620918274, 0.06302451342344284, 0.0009308145381510258, 0.011113142594695091, 0.024584030732512474, -0.01892632432281971, 0.08416716754436493, -0.09559695422649384, -0.08789679408073425, -0.03893548250198364, -0.08772057294845581, 0.039365991950035095, 0.06298413127660751, 0.07542391866445541, -0.09879676252603531, -0.09122679382562637, 0.07277120649814606, 0.07864084839820862, -0.06939317286014557, 0.041935428977012634, -0.060788344591856, 0.05782414972782135, -0.039004795253276825, -0.01181128527969122, -0.18654347956180573, -0.011404315009713173, 0.015132223255932331, -0.013601926155388355, 0.03996402397751808, 0.014380838721990585, 0.07381507754325867, 0.060963332653045654, -0.04813798889517784, -0.026154227554798126, -0.032566383481025696, -0.007290080189704895, -0.13407230377197266, -0.20358924567699432, -0.027620896697044373, -0.021122118458151817, 0.13551472127437592, -0.20138941705226898, 0.03468279913067818, -0.016711212694644928, 0.07772519439458847, 0.01212365459650755, -0.007773156277835369, -0.04515509679913521, 0.09454989433288574, -0.034737929701805115, -0.04874570295214653, 0.0757054015994072, 0.016684282571077347, -0.09071344882249832, -0.0038776614237576723, -0.12874078750610352, 0.1584293246269226, 0.13580897450447083, -0.10734445601701736, -0.07278770953416824, -0.018259037286043167, -0.054592207074165344, -0.04256129264831543, -0.033663660287857056, 0.03222687914967537, 0.18704986572265625, 0.006103041581809521, 0.15579484403133392, -0.07052551209926605, -0.047700610011816025, 0.015909289941191673, -0.02872025966644287, 0.038189150393009186, 0.11265859007835388, 0.10697872191667557, -0.07236873358488083, 0.13936497271060944, 0.15398815274238586, -0.0895383358001709, 0.13469843566417694, -0.041443824768066406, -0.083169125020504, -0.01970035210251808, -0.015816107392311096, 0.004152530804276466, 0.09045837819576263, -0.12357074022293091, -0.0009135896689258516, 0.020951200276613235, 0.0239117369055748, 0.024296946823596954, -0.23080766201019287, -0.0295211523771286, 0.03138891980051994, -0.04991867393255234, -0.01711665466427803, -0.028307540342211723, 0.009715849533677101, 0.10614278167486191, -0.00228509702719748, -0.08267837017774582, 0.027540652081370354, 0.0068334247916936874, -0.07535014301538467, 0.19980572164058685, -0.0934063196182251, -0.16080807149410248, -0.1213449239730835, -0.08109356462955475, -0.02593882940709591, -0.00087083870312199, 0.06687451153993607, -0.07908467203378677, -0.026373527944087982, -0.06183188781142235, 0.022726237773895264, 0.0034915136639028788, 0.01788340136408806, -0.005439021158963442, -0.008004619739949703, 0.0744732990860939, -0.11406894773244858, -0.004223077557981014, -0.04135558754205704, -0.059665754437446594, 0.05540837347507477, 0.04445236176252365, 0.12288699299097061, 0.15919733047485352, -0.019143497571349144, 0.002628567162901163, -0.03206859529018402, 0.21784807741641998, -0.07383282482624054, -0.027769632637500763, 0.12586188316345215, -0.009886739775538445, 0.059406962245702744, 0.11426973342895508, 0.06983331590890884, -0.08660981059074402, 0.015440749004483223, 0.029467498883605003, -0.02661033533513546, -0.22428545355796814, -0.0344984345138073, -0.05539081245660782, -0.025857549160718918, 0.09168455004692078, 0.02109621837735176, 0.05179550498723984, 0.054536715149879456, 0.02821865864098072, 0.07501064985990524, -0.02563261240720749, 0.0634828433394432, 0.12859152257442474, 0.044475167989730835, 0.13737396895885468, -0.042828015983104706, -0.06764766573905945, 0.03250397741794586, -0.0004876790044363588, 0.2203844040632248, 0.017161650583148003, 0.1652754843235016, 0.056817904114723206, 0.18229453265666962, 0.011886241845786572, 0.07815506309270859, 0.004910087678581476, -0.03957361355423927, -0.016728149726986885, -0.04107232019305229, -0.04007142037153244, 0.01632532849907875, -0.04514501616358757, 0.0391554981470108, -0.10711868852376938, -0.03853718563914299, 0.04627954959869385, 0.27932265400886536, 0.021928509697318077, -0.32240667939186096, -0.07944931834936142, -0.003353038104251027, -0.05579080432653427, -0.021180815994739532, 0.014041367918252945, 0.09300810843706131, -0.09714332222938538, 0.0287052933126688, -0.0773022472858429, 0.10984886437654495, -0.036719538271427155, 0.049680717289447784, 0.05390894040465355, 0.09313826262950897, 0.004795422777533531, 0.07552178204059601, -0.34024909138679504, 0.2803536653518677, 0.00493947509676218, 0.07066044211387634, -0.07081294059753418, 0.0007122082752175629, 0.036577872931957245, 0.03317338600754738, 0.0364554263651371, -0.022193681448698044, -0.04955914989113808, -0.19878827035427094, -0.06574621051549911, 0.030401309952139854, 0.09008432924747467, -0.01220324169844389, 0.11066174507141113, -0.03505552560091019, 0.011278372257947922, 0.07857942581176758, -0.021211452782154083, -0.08122111856937408, -0.09851276129484177, -0.003853484522551298, 0.02492848038673401, -0.01119220070540905, -0.07381158322095871, -0.11267706006765366, -0.10793549567461014, 0.14149123430252075, 0.0014742484781891108, -0.017179269343614578, -0.11443088203668594, 0.08256468921899796, 0.0818248838186264, -0.08289701491594315, 0.03618624061346054, 0.01137845404446125, 0.07406225055456161, 0.020498428493738174, -0.06446254998445511, 0.11837790161371231, -0.059787798672914505, -0.1557784527540207, -0.05858810245990753, 0.09455057233572006, 0.03461139276623726, 0.07255010306835175, -0.010767288506031036, 0.014234802685678005, -0.03665432706475258, -0.08398229628801346, 0.00625862181186676, -0.022087380290031433, 0.05600825324654579, 0.004411609843373299, -0.05165262892842293, 0.020836126059293747, -0.06702107191085815, -0.047911353409290314, 0.19497588276863098, 0.2395961731672287, -0.09146451205015182, 0.03606722876429558, 0.05541786924004555, -0.07244808971881866, -0.1859205812215805, 0.03345244377851486, 0.058628231287002563, 0.01092927809804678, 0.05417318269610405, -0.18450801074504852, 0.08502987027168274, 0.10271044820547104, -0.015260961838066578, 0.08886101096868515, -0.33973658084869385, -0.12911880016326904, 0.11879928410053253, 0.14507518708705902, 0.07987871766090393, -0.15548162162303925, -0.024761725217103958, -0.029475124552845955, -0.12206004559993744, 0.10974925011396408, -0.10247453302145004, 0.12515240907669067, -0.024495430290699005, 0.0931176021695137, 0.005357546731829643, -0.06057319790124893, 0.11408938467502594, -0.012519733980298042, 0.10009486228227615, -0.06385384500026703, 0.013631567358970642, 0.054003387689590454, -0.03547247126698494, 0.01032607164233923, -0.09419726580381393, 0.016805468127131462, -0.0844309851527214, -0.022668074816465378, -0.07763423770666122, 0.03207119181752205, -0.037714798003435135, -0.05581261217594147, -0.02734486758708954, 0.02279684692621231, 0.04741416871547699, -0.0103211160749197, 0.11845803260803223, 0.002253623679280281, 0.16185960173606873, 0.11441992223262787, 0.07241763174533844, -0.06480717658996582, -0.053468264639377594, -0.014491192065179348, -0.018878111615777016, 0.05418642237782478, -0.12668992578983307, 0.027285650372505188, 0.14959746599197388, 0.013112148270010948, 0.14402221143245697, 0.07148712873458862, -0.04400071129202843, 0.020384516566991806, 0.058407384902238846, -0.14605429768562317, -0.10016372054815292, 0.0015120514435693622, -0.014121695421636105, -0.09633656591176987, 0.02186219021677971, 0.104742631316185, -0.07350589334964752, -0.015269096940755844, -0.00009402789874002337, 0.0009875482646748424, -0.05830801650881767, 0.20820464193820953, 0.041482239961624146, 0.04539477452635765, -0.10268230736255646, 0.06860265135765076, 0.06770068407058716, -0.09284508973360062, 0.009055975824594498, 0.09270795434713364, -0.06811615824699402, -0.041182056069374084, 0.09842722862958908, 0.1891988068819046, -0.061397816985845566, -0.05311920866370201, -0.13982443511486053, -0.12608177959918976, 0.08418476581573486, 0.16326147317886353, 0.09047462791204453, 0.010690012946724892, -0.05424687638878822, 0.01794775016605854, -0.11534462124109268, 0.08347894251346588, 0.05206974968314171, 0.06622795015573502, -0.1187308132648468, 0.18090631067752838, 0.017406543716788292, 0.030131004750728607, -0.018872877582907677, 0.020791858434677124, -0.1013152003288269, 0.020750652998685837, -0.1570514291524887, -0.029977548867464066, -0.02095569297671318, 0.00531708775088191, -0.00911243911832571, -0.05623884126543999, -0.055717792361974716, 0.007587141823023558, -0.12510471045970917, -0.029829248785972595, 0.0198475681245327, 0.0492250993847847, -0.12176875025033951, -0.03792301192879677, 0.03158491849899292, -0.061427321285009384, 0.06153474003076553, 0.044302329421043396, 0.015935782343149185, 0.06060129031538963, -0.1510305404663086, 0.0015090981032699347, 0.054345134645700455, 0.014039680361747742, 0.05875590071082115, -0.08778750151395798, -0.01577935926616192, 0.008487395010888577, 0.07210314273834229, 0.01493874005973339, 0.07730013132095337, -0.1365908682346344, -0.02042602002620697, -0.024301405996084213, -0.08779192715883255, -0.06440678983926773, 0.036969974637031555, 0.0639105960726738, 0.02465238608419895, 0.1896926462650299, -0.0847133919596672, 0.041419923305511475, -0.22318339347839355, 0.005688908044248819, -0.018722617998719215, -0.11305885016918182, -0.10584550350904465, -0.0761919617652893, 0.06391319632530212, -0.04928651079535484, 0.12931008636951447, 0.016942420974373817, 0.05590570345520973, 0.03130513057112694, -0.03239591047167778, 0.000992286833934486, 0.01864616759121418, 0.20299074053764343, 0.03815624862909317, -0.029569365084171295, 0.05931583791971207, 0.05377772077918053, 0.09008264541625977, 0.12118542194366455, 0.19257573783397675, 0.16044528782367706, 0.01828458532691002, 0.08911080658435822, 0.03772636875510216, -0.05856596678495407, -0.14712370932102203, 0.056585051119327545, -0.04469308257102966, 0.10746489465236664, -0.03561236709356308, 0.25183579325675964, 0.08441885560750961, -0.16680240631103516, 0.05733546242117882, -0.054255396127700806, -0.08562584221363068, -0.09988097846508026, -0.05033312365412712, -0.08291010558605194, -0.15420004725456238, -0.008836298249661922, -0.10521621257066727, 0.04416395351290703, 0.1034187451004982, 0.011722159571945667, -0.022144833579659462, 0.14252611994743347, 0.049787841737270355, 0.009630057029426098, 0.054459646344184875, -0.005084357690066099, -0.019953832030296326, -0.11368576437234879, -0.07848115265369415, 0.003735193982720375, -0.002830169629305601, 0.04411040619015694, -0.04335101321339607, -0.060959719121456146, 0.04084630310535431, -0.029876695945858955, -0.1036081612110138, 0.01943010650575161, 0.017292972654104233, 0.07820700854063034, 0.06008705869317055, 0.014094533398747444, 0.01108044944703579, -0.012546200305223465, 0.22888067364692688, -0.07803371548652649, -0.09721923619508743, -0.0996541753411293, 0.25048452615737915, 0.028660345822572708, -0.015824580565094948, 0.02607366256415844, -0.05443009361624718, 0.004403687082231045, 0.2448265254497528, 0.1809670329093933, -0.09583774954080582, -0.015260185115039349, 0.0016676436644047499, -0.01293659396469593, -0.03642828390002251, 0.12183144688606262, 0.145652636885643, 0.03883199393749237, -0.10613173246383667, -0.036546703428030014, -0.0550968237221241, -0.02225862815976143, -0.04098888486623764, 0.06264974176883698, 0.027477219700813293, 0.003945797681808472, -0.034828055649995804, 0.06270168721675873, -0.05248544365167618, -0.08482006192207336, 0.013557428494095802, -0.19405968487262726, -0.16283732652664185, -0.017171639949083328, 0.11503660678863525, -0.0020884971600025892, 0.0517377033829689, -0.026106225326657295, 0.016429852694272995, 0.07086073607206345, -0.029722098261117935, -0.06770484149456024, -0.08544480055570602, 0.10278977453708649, -0.1238647997379303, 0.18897047638893127, -0.04384840279817581, 0.056988347321748734, 0.12976819276809692, 0.07265783846378326, -0.06945961713790894, 0.07593805342912674, 0.039460960775613785, -0.08218730241060257, 0.03488754853606224, 0.10043741017580032, -0.030200093984603882, 0.0640149712562561, 0.042659394443035126, -0.12833768129348755, 0.042023394256830215, -0.0923188328742981, -0.046105559915304184, -0.027127021923661232, -0.046274203807115555, -0.049073364585638046, 0.1290901005268097, 0.20984455943107605, -0.02893080934882164, 0.017928577959537506, -0.07876405864953995, 0.0013245618902146816, 0.046588193625211716, 0.04883263632655144, -0.0731220617890358, -0.24251648783683777, 0.006194929126650095, 0.06343667954206467, -0.01754559576511383, -0.25067535042762756, -0.09885697066783905, 0.0006914784898981452, -0.07237914204597473, -0.10031825304031372, 0.09725553542375565, 0.08772692829370499, 0.05043494328856468, -0.05213455483317375, -0.10574068874120712, -0.07714276760816574, 0.16188697516918182, -0.1455913782119751, -0.07354407012462616 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bart-mlm-pubmed-45 This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.1797 - Rouge2 Precision: 0.4333 - Rouge2 Recall: 0.3331 - Rouge2 Fmeasure: 0.3684 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure | |:-------------:|:-----:|:----:|:---------------:|:----------------:|:-------------:|:---------------:| | 1.7989 | 1.0 | 663 | 1.3385 | 0.4097 | 0.3086 | 0.3444 | | 1.5072 | 2.0 | 1326 | 1.2582 | 0.4218 | 0.3213 | 0.3569 | | 1.4023 | 3.0 | 1989 | 1.2236 | 0.4207 | 0.3211 | 0.3562 | | 1.2205 | 4.0 | 2652 | 1.2025 | 0.4359 | 0.3331 | 0.3696 | | 1.1584 | 5.0 | 3315 | 1.1910 | 0.4304 | 0.3307 | 0.3658 | | 1.1239 | 6.0 | 3978 | 1.1830 | 0.4247 | 0.3279 | 0.3618 | | 1.0384 | 7.0 | 4641 | 1.1761 | 0.4308 | 0.3325 | 0.367 | | 1.0168 | 8.0 | 5304 | 1.1762 | 0.4314 | 0.3336 | 0.368 | | 0.9966 | 9.0 | 5967 | 1.1773 | 0.4335 | 0.3341 | 0.369 | | 0.961 | 10.0 | 6630 | 1.1797 | 0.4333 | 0.3331 | 0.3684 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.10.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "bart-mlm-pubmed-45", "results": []}]}
text2text-generation
gayanin/bart-mlm-pubmed-45
[ "transformers", "pytorch", "tensorboard", "bart", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
bart-mlm-pubmed-45 ================== This model is a fine-tuned version of facebook/bart-base on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 1.1797 * Rouge2 Precision: 0.4333 * Rouge2 Recall: 0.3331 * Rouge2 Fmeasure: 0.3684 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 10 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.12.5 * Pytorch 1.10.0+cu111 * Datasets 1.15.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ 57, 113, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ -0.0867219790816307, 0.055462174117565155, -0.002274750731885433, 0.09676497429609299, 0.15085238218307495, 0.012637375853955746, 0.14072179794311523, 0.12473951280117035, -0.11230833828449249, 0.017588185146450996, 0.11828296631574631, 0.14802083373069763, 0.024210374802350998, 0.1252502202987671, -0.045356061309576035, -0.262150377035141, -0.00002328762820980046, 0.041856519877910614, -0.05285303294658661, 0.13920284807682037, 0.0906849279999733, -0.12862668931484222, 0.06566891074180603, 0.016234852373600006, -0.21103446185588837, 0.008428194560110569, 0.011844725348055363, -0.052661146968603134, 0.15579554438591003, 0.03129447251558304, 0.1261555701494217, 0.016640443354845047, 0.09018171578645706, -0.20001967251300812, 0.013460533693432808, 0.05625590682029724, 0.013338098302483559, 0.08852751553058624, 0.06908217072486877, 0.010756932199001312, 0.12144075334072113, -0.05637706071138382, 0.06317462772130966, 0.021377360448241234, -0.12813062965869904, -0.2361792027950287, -0.09788632392883301, 0.01742962934076786, 0.06630835682153702, 0.10380984842777252, -0.00795662496238947, 0.1452450305223465, -0.0946977362036705, 0.0953841581940651, 0.22252486646175385, -0.29312217235565186, -0.06660237908363342, 0.005923801567405462, 0.04689629748463631, 0.08137331157922745, -0.10814683139324188, -0.03077341988682747, 0.03911865875124931, 0.051733557134866714, 0.15108142793178558, -0.027031617239117622, -0.1255291998386383, 0.009782832115888596, -0.14126217365264893, -0.04250112920999527, 0.11205080151557922, 0.03479015454649925, -0.03083561733365059, -0.04649687185883522, -0.06090623512864113, -0.16255712509155273, -0.048318225890398026, -0.015000710263848305, 0.04703385382890701, -0.021401921287178993, -0.07937237620353699, -0.019585834816098213, -0.10484049469232559, -0.0716506764292717, -0.06771253049373627, 0.12509149312973022, 0.041989609599113464, 0.0011333179427310824, -0.03612881526350975, 0.10573788732290268, 0.0015543913468718529, -0.13245007395744324, 0.02922716923058033, 0.03680860623717308, -0.021345503628253937, -0.04338424280285835, -0.06522190570831299, -0.0797869861125946, 0.0042693219147622585, 0.14050699770450592, -0.05594496801495552, 0.05111481249332428, -0.0034115430898964405, 0.04727201908826828, -0.11071038246154785, 0.18794257938861847, -0.04299888759851456, -0.010948577895760536, 0.006456696894019842, 0.0576329380273819, 0.008746226318180561, -0.017962101846933365, -0.11141595244407654, 0.013874282129108906, 0.11075880378484726, 0.017528321593999863, -0.0515793077647686, 0.06489716470241547, -0.05094553157687187, -0.01589319296181202, 0.0007782398606650531, -0.09029772877693176, 0.03612196817994118, -0.0028094081208109856, -0.07799360901117325, -0.014370918273925781, 0.024872036650776863, 0.0323185995221138, -0.02158982865512371, 0.10786677896976471, -0.07814837992191315, 0.04184802621603012, -0.10921868681907654, -0.13197839260101318, 0.02444618009030819, -0.049501173198223114, 0.012133109383285046, -0.09642373770475388, -0.16728010773658752, -0.02236819453537464, 0.0593084841966629, -0.027130184695124626, -0.05025933310389519, -0.04999887943267822, -0.06405632197856903, 0.022886520251631737, -0.029339058324694633, 0.1537834256887436, -0.06181749328970909, 0.11160467565059662, 0.0309202428907156, 0.05685362219810486, -0.05136741325259209, 0.061943039298057556, -0.10100964456796646, 0.007471778895705938, -0.18199418485164642, 0.04249980300664902, -0.04797246307134628, 0.0761336013674736, -0.09424274414777756, -0.09534361958503723, 0.00478152884170413, -0.0008567497716285288, 0.09695705771446228, 0.08531241118907928, -0.1826077401638031, -0.07473284006118774, 0.18174810707569122, -0.06340796500444412, -0.10546950995922089, 0.12881304323673248, -0.059853445738554, 0.05344957113265991, 0.07425983995199203, 0.18069881200790405, 0.057555459439754486, -0.08471137285232544, 0.03798159584403038, -0.02432074397802353, 0.05915605276823044, -0.054047420620918274, 0.06302451342344284, 0.0009308145381510258, 0.011113142594695091, 0.024584030732512474, -0.01892632432281971, 0.08416716754436493, -0.09559695422649384, -0.08789679408073425, -0.03893548250198364, -0.08772057294845581, 0.039365991950035095, 0.06298413127660751, 0.07542391866445541, -0.09879676252603531, -0.09122679382562637, 0.07277120649814606, 0.07864084839820862, -0.06939317286014557, 0.041935428977012634, -0.060788344591856, 0.05782414972782135, -0.039004795253276825, -0.01181128527969122, -0.18654347956180573, -0.011404315009713173, 0.015132223255932331, -0.013601926155388355, 0.03996402397751808, 0.014380838721990585, 0.07381507754325867, 0.060963332653045654, -0.04813798889517784, -0.026154227554798126, -0.032566383481025696, -0.007290080189704895, -0.13407230377197266, -0.20358924567699432, -0.027620896697044373, -0.021122118458151817, 0.13551472127437592, -0.20138941705226898, 0.03468279913067818, -0.016711212694644928, 0.07772519439458847, 0.01212365459650755, -0.007773156277835369, -0.04515509679913521, 0.09454989433288574, -0.034737929701805115, -0.04874570295214653, 0.0757054015994072, 0.016684282571077347, -0.09071344882249832, -0.0038776614237576723, -0.12874078750610352, 0.1584293246269226, 0.13580897450447083, -0.10734445601701736, -0.07278770953416824, -0.018259037286043167, -0.054592207074165344, -0.04256129264831543, -0.033663660287857056, 0.03222687914967537, 0.18704986572265625, 0.006103041581809521, 0.15579484403133392, -0.07052551209926605, -0.047700610011816025, 0.015909289941191673, -0.02872025966644287, 0.038189150393009186, 0.11265859007835388, 0.10697872191667557, -0.07236873358488083, 0.13936497271060944, 0.15398815274238586, -0.0895383358001709, 0.13469843566417694, -0.041443824768066406, -0.083169125020504, -0.01970035210251808, -0.015816107392311096, 0.004152530804276466, 0.09045837819576263, -0.12357074022293091, -0.0009135896689258516, 0.020951200276613235, 0.0239117369055748, 0.024296946823596954, -0.23080766201019287, -0.0295211523771286, 0.03138891980051994, -0.04991867393255234, -0.01711665466427803, -0.028307540342211723, 0.009715849533677101, 0.10614278167486191, -0.00228509702719748, -0.08267837017774582, 0.027540652081370354, 0.0068334247916936874, -0.07535014301538467, 0.19980572164058685, -0.0934063196182251, -0.16080807149410248, -0.1213449239730835, -0.08109356462955475, -0.02593882940709591, -0.00087083870312199, 0.06687451153993607, -0.07908467203378677, -0.026373527944087982, -0.06183188781142235, 0.022726237773895264, 0.0034915136639028788, 0.01788340136408806, -0.005439021158963442, -0.008004619739949703, 0.0744732990860939, -0.11406894773244858, -0.004223077557981014, -0.04135558754205704, -0.059665754437446594, 0.05540837347507477, 0.04445236176252365, 0.12288699299097061, 0.15919733047485352, -0.019143497571349144, 0.002628567162901163, -0.03206859529018402, 0.21784807741641998, -0.07383282482624054, -0.027769632637500763, 0.12586188316345215, -0.009886739775538445, 0.059406962245702744, 0.11426973342895508, 0.06983331590890884, -0.08660981059074402, 0.015440749004483223, 0.029467498883605003, -0.02661033533513546, -0.22428545355796814, -0.0344984345138073, -0.05539081245660782, -0.025857549160718918, 0.09168455004692078, 0.02109621837735176, 0.05179550498723984, 0.054536715149879456, 0.02821865864098072, 0.07501064985990524, -0.02563261240720749, 0.0634828433394432, 0.12859152257442474, 0.044475167989730835, 0.13737396895885468, -0.042828015983104706, -0.06764766573905945, 0.03250397741794586, -0.0004876790044363588, 0.2203844040632248, 0.017161650583148003, 0.1652754843235016, 0.056817904114723206, 0.18229453265666962, 0.011886241845786572, 0.07815506309270859, 0.004910087678581476, -0.03957361355423927, -0.016728149726986885, -0.04107232019305229, -0.04007142037153244, 0.01632532849907875, -0.04514501616358757, 0.0391554981470108, -0.10711868852376938, -0.03853718563914299, 0.04627954959869385, 0.27932265400886536, 0.021928509697318077, -0.32240667939186096, -0.07944931834936142, -0.003353038104251027, -0.05579080432653427, -0.021180815994739532, 0.014041367918252945, 0.09300810843706131, -0.09714332222938538, 0.0287052933126688, -0.0773022472858429, 0.10984886437654495, -0.036719538271427155, 0.049680717289447784, 0.05390894040465355, 0.09313826262950897, 0.004795422777533531, 0.07552178204059601, -0.34024909138679504, 0.2803536653518677, 0.00493947509676218, 0.07066044211387634, -0.07081294059753418, 0.0007122082752175629, 0.036577872931957245, 0.03317338600754738, 0.0364554263651371, -0.022193681448698044, -0.04955914989113808, -0.19878827035427094, -0.06574621051549911, 0.030401309952139854, 0.09008432924747467, -0.01220324169844389, 0.11066174507141113, -0.03505552560091019, 0.011278372257947922, 0.07857942581176758, -0.021211452782154083, -0.08122111856937408, -0.09851276129484177, -0.003853484522551298, 0.02492848038673401, -0.01119220070540905, -0.07381158322095871, -0.11267706006765366, -0.10793549567461014, 0.14149123430252075, 0.0014742484781891108, -0.017179269343614578, -0.11443088203668594, 0.08256468921899796, 0.0818248838186264, -0.08289701491594315, 0.03618624061346054, 0.01137845404446125, 0.07406225055456161, 0.020498428493738174, -0.06446254998445511, 0.11837790161371231, -0.059787798672914505, -0.1557784527540207, -0.05858810245990753, 0.09455057233572006, 0.03461139276623726, 0.07255010306835175, -0.010767288506031036, 0.014234802685678005, -0.03665432706475258, -0.08398229628801346, 0.00625862181186676, -0.022087380290031433, 0.05600825324654579, 0.004411609843373299, -0.05165262892842293, 0.020836126059293747, -0.06702107191085815, -0.047911353409290314, 0.19497588276863098, 0.2395961731672287, -0.09146451205015182, 0.03606722876429558, 0.05541786924004555, -0.07244808971881866, -0.1859205812215805, 0.03345244377851486, 0.058628231287002563, 0.01092927809804678, 0.05417318269610405, -0.18450801074504852, 0.08502987027168274, 0.10271044820547104, -0.015260961838066578, 0.08886101096868515, -0.33973658084869385, -0.12911880016326904, 0.11879928410053253, 0.14507518708705902, 0.07987871766090393, -0.15548162162303925, -0.024761725217103958, -0.029475124552845955, -0.12206004559993744, 0.10974925011396408, -0.10247453302145004, 0.12515240907669067, -0.024495430290699005, 0.0931176021695137, 0.005357546731829643, -0.06057319790124893, 0.11408938467502594, -0.012519733980298042, 0.10009486228227615, -0.06385384500026703, 0.013631567358970642, 0.054003387689590454, -0.03547247126698494, 0.01032607164233923, -0.09419726580381393, 0.016805468127131462, -0.0844309851527214, -0.022668074816465378, -0.07763423770666122, 0.03207119181752205, -0.037714798003435135, -0.05581261217594147, -0.02734486758708954, 0.02279684692621231, 0.04741416871547699, -0.0103211160749197, 0.11845803260803223, 0.002253623679280281, 0.16185960173606873, 0.11441992223262787, 0.07241763174533844, -0.06480717658996582, -0.053468264639377594, -0.014491192065179348, -0.018878111615777016, 0.05418642237782478, -0.12668992578983307, 0.027285650372505188, 0.14959746599197388, 0.013112148270010948, 0.14402221143245697, 0.07148712873458862, -0.04400071129202843, 0.020384516566991806, 0.058407384902238846, -0.14605429768562317, -0.10016372054815292, 0.0015120514435693622, -0.014121695421636105, -0.09633656591176987, 0.02186219021677971, 0.104742631316185, -0.07350589334964752, -0.015269096940755844, -0.00009402789874002337, 0.0009875482646748424, -0.05830801650881767, 0.20820464193820953, 0.041482239961624146, 0.04539477452635765, -0.10268230736255646, 0.06860265135765076, 0.06770068407058716, -0.09284508973360062, 0.009055975824594498, 0.09270795434713364, -0.06811615824699402, -0.041182056069374084, 0.09842722862958908, 0.1891988068819046, -0.061397816985845566, -0.05311920866370201, -0.13982443511486053, -0.12608177959918976, 0.08418476581573486, 0.16326147317886353, 0.09047462791204453, 0.010690012946724892, -0.05424687638878822, 0.01794775016605854, -0.11534462124109268, 0.08347894251346588, 0.05206974968314171, 0.06622795015573502, -0.1187308132648468, 0.18090631067752838, 0.017406543716788292, 0.030131004750728607, -0.018872877582907677, 0.020791858434677124, -0.1013152003288269, 0.020750652998685837, -0.1570514291524887, -0.029977548867464066, -0.02095569297671318, 0.00531708775088191, -0.00911243911832571, -0.05623884126543999, -0.055717792361974716, 0.007587141823023558, -0.12510471045970917, -0.029829248785972595, 0.0198475681245327, 0.0492250993847847, -0.12176875025033951, -0.03792301192879677, 0.03158491849899292, -0.061427321285009384, 0.06153474003076553, 0.044302329421043396, 0.015935782343149185, 0.06060129031538963, -0.1510305404663086, 0.0015090981032699347, 0.054345134645700455, 0.014039680361747742, 0.05875590071082115, -0.08778750151395798, -0.01577935926616192, 0.008487395010888577, 0.07210314273834229, 0.01493874005973339, 0.07730013132095337, -0.1365908682346344, -0.02042602002620697, -0.024301405996084213, -0.08779192715883255, -0.06440678983926773, 0.036969974637031555, 0.0639105960726738, 0.02465238608419895, 0.1896926462650299, -0.0847133919596672, 0.041419923305511475, -0.22318339347839355, 0.005688908044248819, -0.018722617998719215, -0.11305885016918182, -0.10584550350904465, -0.0761919617652893, 0.06391319632530212, -0.04928651079535484, 0.12931008636951447, 0.016942420974373817, 0.05590570345520973, 0.03130513057112694, -0.03239591047167778, 0.000992286833934486, 0.01864616759121418, 0.20299074053764343, 0.03815624862909317, -0.029569365084171295, 0.05931583791971207, 0.05377772077918053, 0.09008264541625977, 0.12118542194366455, 0.19257573783397675, 0.16044528782367706, 0.01828458532691002, 0.08911080658435822, 0.03772636875510216, -0.05856596678495407, -0.14712370932102203, 0.056585051119327545, -0.04469308257102966, 0.10746489465236664, -0.03561236709356308, 0.25183579325675964, 0.08441885560750961, -0.16680240631103516, 0.05733546242117882, -0.054255396127700806, -0.08562584221363068, -0.09988097846508026, -0.05033312365412712, -0.08291010558605194, -0.15420004725456238, -0.008836298249661922, -0.10521621257066727, 0.04416395351290703, 0.1034187451004982, 0.011722159571945667, -0.022144833579659462, 0.14252611994743347, 0.049787841737270355, 0.009630057029426098, 0.054459646344184875, -0.005084357690066099, -0.019953832030296326, -0.11368576437234879, -0.07848115265369415, 0.003735193982720375, -0.002830169629305601, 0.04411040619015694, -0.04335101321339607, -0.060959719121456146, 0.04084630310535431, -0.029876695945858955, -0.1036081612110138, 0.01943010650575161, 0.017292972654104233, 0.07820700854063034, 0.06008705869317055, 0.014094533398747444, 0.01108044944703579, -0.012546200305223465, 0.22888067364692688, -0.07803371548652649, -0.09721923619508743, -0.0996541753411293, 0.25048452615737915, 0.028660345822572708, -0.015824580565094948, 0.02607366256415844, -0.05443009361624718, 0.004403687082231045, 0.2448265254497528, 0.1809670329093933, -0.09583774954080582, -0.015260185115039349, 0.0016676436644047499, -0.01293659396469593, -0.03642828390002251, 0.12183144688606262, 0.145652636885643, 0.03883199393749237, -0.10613173246383667, -0.036546703428030014, -0.0550968237221241, -0.02225862815976143, -0.04098888486623764, 0.06264974176883698, 0.027477219700813293, 0.003945797681808472, -0.034828055649995804, 0.06270168721675873, -0.05248544365167618, -0.08482006192207336, 0.013557428494095802, -0.19405968487262726, -0.16283732652664185, -0.017171639949083328, 0.11503660678863525, -0.0020884971600025892, 0.0517377033829689, -0.026106225326657295, 0.016429852694272995, 0.07086073607206345, -0.029722098261117935, -0.06770484149456024, -0.08544480055570602, 0.10278977453708649, -0.1238647997379303, 0.18897047638893127, -0.04384840279817581, 0.056988347321748734, 0.12976819276809692, 0.07265783846378326, -0.06945961713790894, 0.07593805342912674, 0.039460960775613785, -0.08218730241060257, 0.03488754853606224, 0.10043741017580032, -0.030200093984603882, 0.0640149712562561, 0.042659394443035126, -0.12833768129348755, 0.042023394256830215, -0.0923188328742981, -0.046105559915304184, -0.027127021923661232, -0.046274203807115555, -0.049073364585638046, 0.1290901005268097, 0.20984455943107605, -0.02893080934882164, 0.017928577959537506, -0.07876405864953995, 0.0013245618902146816, 0.046588193625211716, 0.04883263632655144, -0.0731220617890358, -0.24251648783683777, 0.006194929126650095, 0.06343667954206467, -0.01754559576511383, -0.25067535042762756, -0.09885697066783905, 0.0006914784898981452, -0.07237914204597473, -0.10031825304031372, 0.09725553542375565, 0.08772692829370499, 0.05043494328856468, -0.05213455483317375, -0.10574068874120712, -0.07714276760816574, 0.16188697516918182, -0.1455913782119751, -0.07354407012462616 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bart-mlm-pubmed-medterm This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.0000 - Rouge2 Precision: 0.985 - Rouge2 Recall: 0.7208 - Rouge2 Fmeasure: 0.8088 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure | |:-------------:|:-----:|:------:|:---------------:|:----------------:|:-------------:|:---------------:| | 0.0018 | 1.0 | 13833 | 0.0003 | 0.985 | 0.7208 | 0.8088 | | 0.0014 | 2.0 | 27666 | 0.0006 | 0.9848 | 0.7207 | 0.8086 | | 0.0009 | 3.0 | 41499 | 0.0002 | 0.9848 | 0.7207 | 0.8086 | | 0.0007 | 4.0 | 55332 | 0.0002 | 0.985 | 0.7208 | 0.8088 | | 0.0006 | 5.0 | 69165 | 0.0001 | 0.9848 | 0.7207 | 0.8087 | | 0.0001 | 6.0 | 82998 | 0.0002 | 0.9846 | 0.7206 | 0.8086 | | 0.0009 | 7.0 | 96831 | 0.0001 | 0.9848 | 0.7208 | 0.8087 | | 0.0 | 8.0 | 110664 | 0.0000 | 0.9848 | 0.7207 | 0.8087 | | 0.0001 | 9.0 | 124497 | 0.0000 | 0.985 | 0.7208 | 0.8088 | | 0.0 | 10.0 | 138330 | 0.0000 | 0.985 | 0.7208 | 0.8088 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "bart-mlm-pubmed-medterm", "results": []}]}
text2text-generation
gayanin/bart-mlm-pubmed-medterm
[ "transformers", "pytorch", "tensorboard", "bart", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
bart-mlm-pubmed-medterm ======================= This model is a fine-tuned version of facebook/bart-base on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 0.0000 * Rouge2 Precision: 0.985 * Rouge2 Recall: 0.7208 * Rouge2 Fmeasure: 0.8088 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 10 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.12.5 * Pytorch 1.10.0+cu111 * Datasets 1.16.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ 57, 113, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ -0.08575264364480972, 0.05633566528558731, -0.0022802120074629784, 0.09782730787992477, 0.15030108392238617, 0.013678183779120445, 0.141473650932312, 0.12577228248119354, -0.10942047834396362, 0.017514344304800034, 0.11677248030900955, 0.14822839200496674, 0.023955507203936577, 0.12322043627500534, -0.04460148513317108, -0.26430022716522217, 0.0007802951731719077, 0.04253416508436203, -0.04874259978532791, 0.13914833962917328, 0.09036122262477875, -0.12921038269996643, 0.06532150506973267, 0.013905899599194527, -0.21144849061965942, 0.008773364126682281, 0.0109938345849514, -0.05301390588283539, 0.15549762547016144, 0.029926272109150887, 0.12534675002098083, 0.017150504514575005, 0.09051694720983505, -0.2026723176240921, 0.013532525859773159, 0.05570119991898537, 0.012264726683497429, 0.08733553439378738, 0.06785966455936432, 0.012129535898566246, 0.12319324910640717, -0.05600752308964729, 0.06364559382200241, 0.020987169817090034, -0.12775586545467377, -0.23507221043109894, -0.09750444442033768, 0.018745429813861847, 0.06682736426591873, 0.10478084534406662, -0.008177259936928749, 0.14650700986385345, -0.09424923360347748, 0.09521288424730301, 0.22506146132946014, -0.29395800828933716, -0.06649722158908844, 0.00632872898131609, 0.04742015525698662, 0.0841280072927475, -0.10912875086069107, -0.030779220163822174, 0.03865108639001846, 0.0519852414727211, 0.14812932908535004, -0.027532117441296577, -0.12487629801034927, 0.009007291868329048, -0.1418655961751938, -0.042740195989608765, 0.11268030852079391, 0.03594076633453369, -0.030479609966278076, -0.04767574369907379, -0.06057160347700119, -0.1626424491405487, -0.048348866403102875, -0.01552506908774376, 0.04619044065475464, -0.0218792874366045, -0.07858514785766602, -0.01798616163432598, -0.10579685121774673, -0.07064461708068848, -0.06668353080749512, 0.12267826497554779, 0.04238506406545639, 0.0007529190042987466, -0.03509489446878433, 0.10703020542860031, 0.0012143761850893497, -0.13320787250995636, 0.0286741703748703, 0.03672683611512184, -0.020675139501690865, -0.04306180030107498, -0.06466236710548401, -0.07850472629070282, 0.0049675097689032555, 0.1392647624015808, -0.05250490456819534, 0.05253743380308151, -0.0039702826179564, 0.04716503620147705, -0.11051322519779205, 0.18810424208641052, -0.039532989263534546, -0.012605647556483746, 0.004277979489415884, 0.05894504860043526, 0.008444076403975487, -0.01897946000099182, -0.1124170646071434, 0.01385465543717146, 0.11295322328805923, 0.016912929713726044, -0.05155456066131592, 0.0643906518816948, -0.05068904533982277, -0.017133908346295357, 0.0006101011531427503, -0.09007594734430313, 0.036902159452438354, -0.0020986811723560095, -0.07644974440336227, -0.012982060201466084, 0.023349689319729805, 0.03351760655641556, -0.022193046286702156, 0.10540727525949478, -0.07631063461303711, 0.04231596738100052, -0.10969653725624084, -0.1309623122215271, 0.025154022499918938, -0.05063905566930771, 0.01235120464116335, -0.09696058183908463, -0.16413970291614532, -0.02303752675652504, 0.06007036194205284, -0.026015454903244972, -0.04996287822723389, -0.050214983522892, -0.06328963488340378, 0.022223953157663345, -0.02934548445045948, 0.15295207500457764, -0.06233030930161476, 0.11128061264753342, 0.029938897117972374, 0.056255340576171875, -0.05393235757946968, 0.06150523200631142, -0.09904954582452774, 0.006346821319311857, -0.18083801865577698, 0.040659643709659576, -0.04777887463569641, 0.07561630755662918, -0.09412787109613419, -0.09539339691400528, 0.0063524250872433186, -0.0009580380283296108, 0.09529048949480057, 0.08460646867752075, -0.18373724818229675, -0.07355949282646179, 0.18068185448646545, -0.0649222582578659, -0.10637593269348145, 0.1288325935602188, -0.05928051471710205, 0.05420795828104019, 0.07415416836738586, 0.18164043128490448, 0.060068126767873764, -0.08332160115242004, 0.03723680227994919, -0.025397533550858498, 0.06186477467417717, -0.053701866418123245, 0.06198422983288765, 0.0011611221125349402, 0.014060075394809246, 0.025520555675029755, -0.01815463975071907, 0.0824170783162117, -0.09457416832447052, -0.08770126104354858, -0.03838174045085907, -0.0859527438879013, 0.0381641760468483, 0.06410212069749832, 0.07577045261859894, -0.098390132188797, -0.0898209735751152, 0.07167989015579224, 0.07796735316514969, -0.07007282227277756, 0.0425422377884388, -0.06090964376926422, 0.05713639780879021, -0.03761081397533417, -0.012134892866015434, -0.18692655861377716, -0.01238585077226162, 0.016539976000785828, -0.015093738213181496, 0.038710419088602066, 0.015645556151866913, 0.07298159599304199, 0.05917958915233612, -0.048086028546094894, -0.025832030922174454, -0.03214012086391449, -0.006841087248176336, -0.13551123440265656, -0.20390565693378448, -0.028174996376037598, -0.021327819675207138, 0.13421887159347534, -0.20119331777095795, 0.035115212202072144, -0.014079409651458263, 0.07783723622560501, 0.011992991901934147, -0.008237873204052448, -0.0467047318816185, 0.09556680917739868, -0.03450560197234154, -0.048291854560375214, 0.07716714590787888, 0.016040265560150146, -0.09176047146320343, -0.006321043707430363, -0.13036394119262695, 0.15792037546634674, 0.13538609445095062, -0.1059211865067482, -0.0717877745628357, -0.020348945632576942, -0.05446043610572815, -0.04287105053663254, -0.03260893374681473, 0.031612202525138855, 0.18769261240959167, 0.007168282754719257, 0.1538822054862976, -0.06949309259653091, -0.04615984484553337, 0.016867468133568764, -0.0279021505266428, 0.03791814669966698, 0.1121511235833168, 0.10752896219491959, -0.07208601385354996, 0.1379559189081192, 0.15636184811592102, -0.08756851404905319, 0.1336057037115097, -0.04171377047896385, -0.08164983242750168, -0.01841726526618004, -0.016257163137197495, 0.004034469369798899, 0.09123393148183823, -0.12616704404354095, -0.0021091168746352196, 0.020060596987605095, 0.025283752009272575, 0.025186793878674507, -0.23158307373523712, -0.029813088476657867, 0.030598992481827736, -0.0493461936712265, -0.016993029043078423, -0.028274988755583763, 0.008319595828652382, 0.10563178360462189, -0.002137316856533289, -0.08418715000152588, 0.027631979435682297, 0.005347121972590685, -0.07632587850093842, 0.19925782084465027, -0.09273555874824524, -0.15955707430839539, -0.12044167518615723, -0.08297817409038544, -0.025361139327287674, -0.0014082134002819657, 0.06905682384967804, -0.07837706804275513, -0.02517872303724289, -0.06351431459188461, 0.018899017944931984, 0.005058370530605316, 0.018207650631666183, -0.005107344593852758, -0.007524035405367613, 0.07329044491052628, -0.11334706097841263, -0.005163710564374924, -0.042103588581085205, -0.060004785656929016, 0.05661643669009209, 0.04208831116557121, 0.1230076402425766, 0.16047611832618713, -0.016655107960104942, 0.0035277714487165213, -0.03200680390000343, 0.2148171216249466, -0.07441481202840805, -0.027610111981630325, 0.1259373128414154, -0.011454383842647076, 0.0596701055765152, 0.11659708619117737, 0.0697733536362648, -0.08605074882507324, 0.01577095501124859, 0.028945576399564743, -0.027355307713150978, -0.22245140373706818, -0.033566128462553024, -0.05571353808045387, -0.025065600872039795, 0.0926528200507164, 0.022606508806347847, 0.049877554178237915, 0.05440289527177811, 0.030406907200813293, 0.07166926562786102, -0.024742992594838142, 0.0636567547917366, 0.12718959152698517, 0.04409758001565933, 0.1369646042585373, -0.042125411331653595, -0.06682644784450531, 0.0313715860247612, -0.0003490304807201028, 0.2218756228685379, 0.01742665469646454, 0.16291430592536926, 0.05752427875995636, 0.18306316435337067, 0.011591986753046513, 0.07704518735408783, 0.004655129741877317, -0.039861634373664856, -0.017220260575413704, -0.040385473519563675, -0.03864952549338341, 0.01716863550245762, -0.04468127712607384, 0.038970354944467545, -0.10800634324550629, -0.03966984525322914, 0.04582959786057472, 0.2805701494216919, 0.0234502162784338, -0.32086268067359924, -0.07784395664930344, -0.002933875424787402, -0.05667600780725479, -0.02009502239525318, 0.014272820204496384, 0.09295147657394409, -0.09741602092981339, 0.028531944379210472, -0.07672535628080368, 0.10915838927030563, -0.035857461392879486, 0.05112425610423088, 0.05466713011264801, 0.09014900773763657, 0.005634683184325695, 0.07596607506275177, -0.3418184220790863, 0.27913907170295715, 0.0035512058530002832, 0.07026582211256027, -0.07185917347669601, 0.0010848584352061152, 0.03570804372429848, 0.03300431743264198, 0.036197688430547714, -0.022898167371749878, -0.05096342787146568, -0.19794617593288422, -0.06559854000806808, 0.03320569545030594, 0.09086280316114426, -0.012960069812834263, 0.11116974800825119, -0.03480583801865578, 0.01181504875421524, 0.07831013202667236, -0.018684467300772667, -0.08071529865264893, -0.10003743320703506, -0.002356293611228466, 0.023923400789499283, -0.010330325923860073, -0.07409828901290894, -0.11259674280881882, -0.10809888690710068, 0.14173465967178345, 0.0020893430337309837, -0.016284501180052757, -0.11290579289197922, 0.08080129325389862, 0.08113890141248703, -0.08332797139883041, 0.03528117761015892, 0.012755313888192177, 0.07323883473873138, 0.021402893587946892, -0.06344722956418991, 0.1190379187464714, -0.05855443328619003, -0.15667855739593506, -0.058737073093652725, 0.09405089914798737, 0.03514012694358826, 0.07228415459394455, -0.00902622751891613, 0.014607502147555351, -0.03630761429667473, -0.084701307117939, 0.00601168954744935, -0.019750747829675674, 0.05543022230267525, 0.0034648599103093147, -0.05275597795844078, 0.020979274064302444, -0.0680873766541481, -0.04976948723196983, 0.1957470327615738, 0.23959538340568542, -0.09119350463151932, 0.03613875433802605, 0.05567459762096405, -0.07291441410779953, -0.18549086153507233, 0.03218400850892067, 0.05840104818344116, 0.009620487689971924, 0.05564259737730026, -0.1827802211046219, 0.08506029844284058, 0.10315603762865067, -0.014640217646956444, 0.08903633058071136, -0.34148305654525757, -0.1284351497888565, 0.11879125237464905, 0.14526502788066864, 0.08004280924797058, -0.15655362606048584, -0.02457815781235695, -0.028370102867484093, -0.12450671941041946, 0.10885964334011078, -0.10318321734666824, 0.1254090815782547, -0.02534843049943447, 0.09177672117948532, 0.005116436630487442, -0.06131298094987869, 0.11363709717988968, -0.009998580440878868, 0.09883057326078415, -0.0647587925195694, 0.010600718669593334, 0.05156908929347992, -0.03684339299798012, 0.0082807382568717, -0.09471861273050308, 0.01480564009398222, -0.08488670736551285, -0.022304046899080276, -0.07831832021474838, 0.030786538496613503, -0.037675611674785614, -0.05542261525988579, -0.02595551498234272, 0.022361744195222855, 0.04858902096748352, -0.010251673869788647, 0.12104091048240662, 0.0005702523048967123, 0.16362980008125305, 0.11582574248313904, 0.07169033586978912, -0.06722074747085571, -0.05341736599802971, -0.01676936447620392, -0.018892155960202217, 0.05318767577409744, -0.1259271502494812, 0.027742162346839905, 0.15015973150730133, 0.013143335469067097, 0.14531812071800232, 0.07239607721567154, -0.04277273640036583, 0.022233575582504272, 0.0584661029279232, -0.14510472118854523, -0.09924758225679398, 0.0029859417118132114, -0.014042996801435947, -0.09831659495830536, 0.022247619926929474, 0.10482482612133026, -0.07261839509010315, -0.014422910287976265, 0.00021522586757782847, 0.001090998761355877, -0.058643948286771774, 0.20791734755039215, 0.04050423204898834, 0.045383431017398834, -0.10209398716688156, 0.06729616969823837, 0.06620173156261444, -0.09533673524856567, 0.00877863634377718, 0.09268695116043091, -0.06964728981256485, -0.04220130667090416, 0.0984969511628151, 0.1865660697221756, -0.06108783930540085, -0.05217863991856575, -0.14057166874408722, -0.12643536925315857, 0.0827636644244194, 0.16121089458465576, 0.0911259725689888, 0.011288322508335114, -0.054632388055324554, 0.017814412713050842, -0.1148185208439827, 0.08209650963544846, 0.05194268375635147, 0.06719308346509933, -0.11875301599502563, 0.17898906767368317, 0.017479006201028824, 0.030278317630290985, -0.01837179809808731, 0.020641343668103218, -0.10173510015010834, 0.020659759640693665, -0.15812957286834717, -0.030440930277109146, -0.021631648764014244, 0.00470460532233119, -0.008570555597543716, -0.0555935874581337, -0.05527716130018234, 0.007208582945168018, -0.12551534175872803, -0.030424470081925392, 0.019766151905059814, 0.049807652831077576, -0.11980216205120087, -0.03763768449425697, 0.030968077480793, -0.06193746626377106, 0.06235473230481148, 0.043445587158203125, 0.015511097386479378, 0.06104012951254845, -0.15085268020629883, 0.0010267047910019755, 0.054892826825380325, 0.014522165060043335, 0.05781368911266327, -0.08622811734676361, -0.016390033066272736, 0.008856775239109993, 0.0718768835067749, 0.016361279413104057, 0.07785818725824356, -0.1370134800672531, -0.01947687193751335, -0.025037327781319618, -0.08905724436044693, -0.06457877904176712, 0.03672981634736061, 0.06321347504854202, 0.025027794763445854, 0.19022849202156067, -0.08564849197864532, 0.04235750809311867, -0.22190020978450775, 0.006087371613830328, -0.017453210428357124, -0.11171382665634155, -0.10376555472612381, -0.07694901525974274, 0.0636405497789383, -0.049289241433143616, 0.12881110608577728, 0.013897447846829891, 0.05562526732683182, 0.030905907973647118, -0.03500235453248024, 0.0013203182024881244, 0.018607862293720245, 0.20426422357559204, 0.03814717382192612, -0.029759522527456284, 0.05985923856496811, 0.053494859486818314, 0.0884547010064125, 0.12186294794082642, 0.1928272694349289, 0.16242268681526184, 0.017916886135935783, 0.08802594989538193, 0.03757762163877487, -0.0573461540043354, -0.1466362178325653, 0.05849537253379822, -0.04488181695342064, 0.10887710005044937, -0.03640427812933922, 0.2512339651584625, 0.08298500627279282, -0.1661251038312912, 0.055606063455343246, -0.05510346591472626, -0.08544636517763138, -0.09842987358570099, -0.04847333952784538, -0.08379113674163818, -0.15463006496429443, -0.008621136657893658, -0.10481017827987671, 0.04454461857676506, 0.10537915676832199, 0.012801490724086761, -0.02081133797764778, 0.14149221777915955, 0.050684064626693726, 0.00928298570215702, 0.05558963492512703, -0.004826862830668688, -0.020937755703926086, -0.11282166838645935, -0.07637485861778259, 0.002765788696706295, -0.0031779573764652014, 0.0432286411523819, -0.04307671636343002, -0.06078961491584778, 0.039571139961481094, -0.03079097531735897, -0.10382705181837082, 0.019429145380854607, 0.016280125826597214, 0.07814209163188934, 0.05941585823893547, 0.015224304981529713, 0.009962396696209908, -0.011860898695886135, 0.22947782278060913, -0.07859647274017334, -0.09914393723011017, -0.0993993729352951, 0.25183388590812683, 0.029122240841388702, -0.016079038381576538, 0.02508092299103737, -0.05449232831597328, 0.005650653969496489, 0.24216462671756744, 0.1803433746099472, -0.09621535241603851, -0.013776007108390331, 0.0034953949507325888, -0.012509147636592388, -0.036544010043144226, 0.12241394817829132, 0.14573757350444794, 0.03968300297856331, -0.1052585318684578, -0.03567322716116905, -0.054358579218387604, -0.021771596744656563, -0.042321834713220596, 0.0629144161939621, 0.029354361817240715, 0.004280323162674904, -0.035623565316200256, 0.06325401365756989, -0.05200126767158508, -0.08271918445825577, 0.013550087809562683, -0.1951812356710434, -0.16296735405921936, -0.015713701024651527, 0.11731792986392975, -0.0023019297514110804, 0.0517619326710701, -0.026834743097424507, 0.015933360904455185, 0.07142571359872818, -0.02918364480137825, -0.06707494705915451, -0.08409074693918228, 0.10277403146028519, -0.12542040646076202, 0.18858781456947327, -0.04445721209049225, 0.057031869888305664, 0.12994222342967987, 0.07283755391836166, -0.06822879612445831, 0.07568690925836563, 0.039116792380809784, -0.08222810924053192, 0.03369520977139473, 0.09890315681695938, -0.029990965500473976, 0.06380590796470642, 0.043939776718616486, -0.12770088016986847, 0.04210788011550903, -0.09335237741470337, -0.045684244483709335, -0.026865167543292046, -0.046466339379549026, -0.04868560656905174, 0.12961025536060333, 0.20932228863239288, -0.029295748099684715, 0.017757877707481384, -0.07764005661010742, 0.00124599679838866, 0.045705512166023254, 0.049732059240341187, -0.07432383298873901, -0.24187178909778595, 0.00669033033773303, 0.062164995819330215, -0.01819278672337532, -0.2504955530166626, -0.10075169801712036, 0.0019417153671383858, -0.07289865612983704, -0.10078714787960052, 0.09705903381109238, 0.08812442421913147, 0.049639929085969925, -0.05273124575614929, -0.1075003445148468, -0.07764784246683121, 0.16339652240276337, -0.14682643115520477, -0.07510910928249359 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bart-mlm-pubmed This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.7223 - Rouge2 Precision: 0.6572 - Rouge2 Recall: 0.5164 - Rouge2 Fmeasure: 0.5662 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure | |:-------------:|:-----:|:----:|:---------------:|:----------------:|:-------------:|:---------------:| | 1.0322 | 1.0 | 663 | 0.7891 | 0.639 | 0.4989 | 0.5491 | | 0.8545 | 2.0 | 1326 | 0.7433 | 0.6461 | 0.5057 | 0.5556 | | 0.758 | 3.0 | 1989 | 0.7299 | 0.647 | 0.5033 | 0.5547 | | 0.6431 | 4.0 | 2652 | 0.7185 | 0.6556 | 0.5101 | 0.5616 | | 0.6058 | 5.0 | 3315 | 0.7126 | 0.6537 | 0.5144 | 0.5638 | | 0.5726 | 6.0 | 3978 | 0.7117 | 0.6567 | 0.5169 | 0.5666 | | 0.5168 | 7.0 | 4641 | 0.7150 | 0.6585 | 0.5154 | 0.566 | | 0.5011 | 8.0 | 5304 | 0.7220 | 0.6568 | 0.5164 | 0.5664 | | 0.4803 | 9.0 | 5967 | 0.7208 | 0.6573 | 0.5161 | 0.5662 | | 0.4577 | 10.0 | 6630 | 0.7223 | 0.6572 | 0.5164 | 0.5662 | ### Framework versions - Transformers 4.12.3 - Pytorch 1.9.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "bart-mlm-pubmed", "results": []}]}
text2text-generation
gayanin/bart-mlm-pubmed
[ "transformers", "pytorch", "tensorboard", "bart", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
bart-mlm-pubmed =============== This model is a fine-tuned version of facebook/bart-base on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 0.7223 * Rouge2 Precision: 0.6572 * Rouge2 Recall: 0.5164 * Rouge2 Fmeasure: 0.5662 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 10 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.12.3 * Pytorch 1.9.0+cu111 * Datasets 1.15.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ 57, 113, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ -0.08557509630918503, 0.05797784775495529, -0.0024650858249515295, 0.0973440408706665, 0.15082214772701263, 0.014361673034727573, 0.14202666282653809, 0.1284749060869217, -0.1130785346031189, 0.015352019108831882, 0.11496320366859436, 0.1527942717075348, 0.02294573374092579, 0.13084737956523895, -0.04365701228380203, -0.2648164629936218, -0.0026389164850115776, 0.04189163073897362, -0.05457921326160431, 0.13746899366378784, 0.0935695543885231, -0.12811213731765747, 0.06262031197547913, 0.015224087983369827, -0.20650480687618256, 0.011299044825136662, 0.013253763318061829, -0.054878585040569305, 0.1566205471754074, 0.029767753556370735, 0.12552888691425323, 0.014870206825435162, 0.08888030052185059, -0.1990465372800827, 0.013165589421987534, 0.057609137147665024, 0.013492707163095474, 0.08827997744083405, 0.07116074115037918, 0.0077233994379639626, 0.12105370312929153, -0.055542074143886566, 0.06459270417690277, 0.022096144035458565, -0.12724025547504425, -0.24369101226329803, -0.09698525071144104, 0.014968804083764553, 0.06725514680147171, 0.1028880700469017, -0.006567833013832569, 0.14157108962535858, -0.09084468334913254, 0.09689459949731827, 0.22284746170043945, -0.2906811237335205, -0.06498780101537704, 0.011874345131218433, 0.045909855514764786, 0.08240236341953278, -0.10873782634735107, -0.03148496150970459, 0.03735921159386635, 0.0532090961933136, 0.15261325240135193, -0.02854987047612667, -0.13056209683418274, 0.0071847327053546906, -0.13974504172801971, -0.041684433817863464, 0.11841420829296112, 0.03234250470995903, -0.03081713244318962, -0.04771360382437706, -0.061295535415410995, -0.15475089848041534, -0.046411070972681046, -0.013999939896166325, 0.047232769429683685, -0.020092923194169998, -0.0768132358789444, -0.022629056125879288, -0.10648367553949356, -0.07063766568899155, -0.0673685073852539, 0.12165956199169159, 0.042844004929065704, 0.0031615335028618574, -0.03630446642637253, 0.10548682510852814, 0.0035586943849921227, -0.13221335411071777, 0.030069373548030853, 0.03492699936032295, -0.01865115389227867, -0.04105285555124283, -0.06947466731071472, -0.08198700845241547, 0.003853237023577094, 0.14006449282169342, -0.05646827444434166, 0.053593363612890244, -0.0061585004441440105, 0.04709697887301445, -0.10937396436929703, 0.1860518455505371, -0.03982982039451599, -0.011398683302104473, 0.006969108246266842, 0.05643102154135704, 0.007961134426295757, -0.017133241519331932, -0.1084834411740303, 0.015050229616463184, 0.11153305321931839, 0.016323400661349297, -0.05195974186062813, 0.06597777456045151, -0.049927856773138046, -0.018759995698928833, -0.003833961207419634, -0.09406312555074692, 0.03616093471646309, -0.0017729229293763638, -0.07796341180801392, -0.015656862407922745, 0.02844221331179142, 0.031641390174627304, -0.02240990847349167, 0.10548911988735199, -0.07533113658428192, 0.03979979455471039, -0.11055824905633926, -0.13053302466869354, 0.02338314801454544, -0.0502239391207695, 0.012467453256249428, -0.09571076929569244, -0.17265865206718445, -0.02260986529290676, 0.06242571026086807, -0.025519538670778275, -0.048985522240400314, -0.04684624448418617, -0.06004047021269798, 0.019441721960902214, -0.02986130118370056, 0.15474113821983337, -0.059180356562137604, 0.11162208020687103, 0.027848316356539726, 0.057262178510427475, -0.04753875732421875, 0.06424582004547119, -0.10222961008548737, 0.006006013136357069, -0.18065530061721802, 0.0426718108355999, -0.04955210164189339, 0.07003756612539291, -0.09592458605766296, -0.09480317682027817, 0.005579913966357708, -0.0015064466279000044, 0.09391278028488159, 0.08172187209129333, -0.1822434812784195, -0.07377175986766815, 0.18064740300178528, -0.06511407345533371, -0.10493358969688416, 0.1239113137125969, -0.06153952702879906, 0.05544251203536987, 0.07404625415802002, 0.18374469876289368, 0.05792616307735443, -0.08046869188547134, 0.04303304851055145, -0.021938232704997063, 0.05866285040974617, -0.05184716731309891, 0.06214688718318939, -0.0013054247247055173, 0.006874904036521912, 0.0265116598457098, -0.019144419580698013, 0.08176027983427048, -0.09347759187221527, -0.08844824135303497, -0.03787373751401901, -0.08717095851898193, 0.04232139512896538, 0.06494518369436264, 0.07661247253417969, -0.09733280539512634, -0.08833606541156769, 0.07622445374727249, 0.07923863083124161, -0.07062407582998276, 0.04146646335721016, -0.05760251730680466, 0.053641967475414276, -0.03358220309019089, -0.011056359857320786, -0.18741513788700104, -0.011289747431874275, 0.01638415828347206, -0.017600003629922867, 0.037580668926239014, 0.014952936209738255, 0.07492576539516449, 0.06219782680273056, -0.04979800805449486, -0.026544280350208282, -0.03119697794318199, -0.006423572544008493, -0.13293619453907013, -0.20306040346622467, -0.028028221800923347, -0.021641265600919724, 0.13343122601509094, -0.2009158879518509, 0.03583341836929321, -0.023619288578629494, 0.07827150821685791, 0.01246605347841978, -0.008513009175658226, -0.04345930367708206, 0.09174907952547073, -0.03419872373342514, -0.048438847064971924, 0.0778726115822792, 0.016632501035928726, -0.08936432749032974, -0.005213967990130186, -0.13059812784194946, 0.154335618019104, 0.13398605585098267, -0.109267458319664, -0.07169313728809357, -0.01927037537097931, -0.05461978167295456, -0.042211223393678665, -0.03387223929166794, 0.029679985716938972, 0.18176357448101044, 0.004334183409810066, 0.1574278622865677, -0.07010001689195633, -0.04696783050894737, 0.014187801629304886, -0.031674742698669434, 0.03749563544988632, 0.10982924699783325, 0.10342871397733688, -0.06769175082445145, 0.13820020854473114, 0.15716181695461273, -0.09168849885463715, 0.13502740859985352, -0.04192240908741951, -0.08164186775684357, -0.018649788573384285, -0.01536570768803358, 0.003787263063713908, 0.0873529314994812, -0.12143122404813766, 0.003933616913855076, 0.02258227951824665, 0.02594759501516819, 0.024460578337311745, -0.22989404201507568, -0.028422435745596886, 0.03205085173249245, -0.05031414330005646, -0.011806132271885872, -0.02739398553967476, 0.011359723284840584, 0.10586884617805481, -0.004531248938292265, -0.08050142228603363, 0.02993703819811344, 0.004699133802205324, -0.0769304484128952, 0.20169587433338165, -0.09706494212150574, -0.1637854427099228, -0.12091536819934845, -0.08293202519416809, -0.026052171364426613, 0.0014434658223763108, 0.06812053173780441, -0.07721155136823654, -0.027105147019028664, -0.0622563362121582, 0.025878239423036575, 0.00303762243129313, 0.016288071870803833, -0.005396385211497545, -0.006983431056141853, 0.07750478386878967, -0.1158980205655098, -0.004352658521384001, -0.04156622290611267, -0.06408201903104782, 0.055429182946681976, 0.045250020921230316, 0.11900973320007324, 0.1562872976064682, -0.019941333681344986, 0.0020934613421559334, -0.032178498804569244, 0.21723729372024536, -0.0711238756775856, -0.025527210906147957, 0.13363201916217804, -0.010772963054478168, 0.05935594066977501, 0.11434641480445862, 0.06896944344043732, -0.08439167588949203, 0.015641551464796066, 0.03146744519472122, -0.02775386907160282, -0.22607465088367462, -0.035509444773197174, -0.05682922527194023, -0.022771256044507027, 0.09378010034561157, 0.021831363439559937, 0.055628132075071335, 0.05593869462609291, 0.028792228549718857, 0.07442972809076309, -0.022520264610648155, 0.06372927874326706, 0.1331590712070465, 0.042940836399793625, 0.1387576460838318, -0.04075496643781662, -0.06493838876485825, 0.03414444252848625, 0.0032412968575954437, 0.22154109179973602, 0.015358719043433666, 0.15438289940357208, 0.060494668781757355, 0.17960532009601593, 0.007917347364127636, 0.07860751450061798, 0.0015619045589119196, -0.036762550473213196, -0.01835392601788044, -0.041727956384420395, -0.03988157957792282, 0.01668858714401722, -0.04960950091481209, 0.03889947757124901, -0.10701512545347214, -0.04220839962363243, 0.047638606280088425, 0.27584484219551086, 0.022613098844885826, -0.32008102536201477, -0.07924925535917282, -0.000800048466771841, -0.05493105575442314, -0.02113347128033638, 0.015032713301479816, 0.0895565003156662, -0.10183075815439224, 0.030688468366861343, -0.07749426364898682, 0.10763554275035858, -0.03400316834449768, 0.05004250630736351, 0.05611645430326462, 0.09402332454919815, 0.006991265341639519, 0.07751007378101349, -0.34100282192230225, 0.2773095667362213, 0.003207253059372306, 0.0686740130186081, -0.0708940401673317, 0.003884126665070653, 0.03813540190458298, 0.029243171215057373, 0.03716603294014931, -0.022894810885190964, -0.04986367002129555, -0.19813089072704315, -0.060870710760354996, 0.030555516481399536, 0.087855763733387, -0.013954204507172108, 0.10670018941164017, -0.03637736290693283, 0.012944468297064304, 0.07936744391918182, -0.021180756390094757, -0.08449053019285202, -0.09916507452726364, -0.003497461089864373, 0.024422915652394295, -0.01013598870486021, -0.07431060820817947, -0.11199549585580826, -0.10714338719844818, 0.14454849064350128, 0.00030904225423000753, -0.021153245121240616, -0.11560752242803574, 0.08575820922851562, 0.08170140534639359, -0.08235042542219162, 0.03866954520344734, 0.010728503577411175, 0.0735451802611351, 0.020851613953709602, -0.06992997974157333, 0.11719205230474472, -0.06279655545949936, -0.15596836805343628, -0.06059526279568672, 0.09620831906795502, 0.03372067213058472, 0.07123098522424698, -0.011486504226922989, 0.014384274370968342, -0.03605905547738075, -0.08338286727666855, 0.008236547000706196, -0.01939668133854866, 0.05514785274863243, 0.006125152111053467, -0.058195680379867554, 0.014784802682697773, -0.06521721929311752, -0.04701821878552437, 0.19605928659439087, 0.23573876917362213, -0.09226791560649872, 0.04106112942099571, 0.05637664347887039, -0.07482227683067322, -0.18666554987430573, 0.02983306162059307, 0.06072240695357323, 0.007387315854430199, 0.060703735798597336, -0.1887173056602478, 0.09051591902971268, 0.1033984050154686, -0.016317160800099373, 0.08762941509485245, -0.3422962725162506, -0.12914228439331055, 0.12140638381242752, 0.1478712409734726, 0.07760870456695557, -0.15588858723640442, -0.024483660236001015, -0.025420840829610825, -0.11983033269643784, 0.10904578864574432, -0.10273055732250214, 0.12601962685585022, -0.024580011144280434, 0.09276609122753143, 0.005451680161058903, -0.05881782993674278, 0.11245585978031158, -0.01189667172729969, 0.09958674013614655, -0.06396038830280304, 0.015464957803487778, 0.05506577715277672, -0.03509014844894409, 0.011536392383277416, -0.09352053701877594, 0.020168304443359375, -0.08376681804656982, -0.023155800998210907, -0.07963050901889801, 0.032956406474113464, -0.03793349862098694, -0.05986899882555008, -0.026036793366074562, 0.02429935336112976, 0.050531789660453796, -0.010982504114508629, 0.11777891218662262, 0.005092812702059746, 0.16532002389431, 0.11997238546609879, 0.06758877635002136, -0.05856377258896828, -0.057965926826000214, -0.018287530168890953, -0.018600642681121826, 0.055018987506628036, -0.12774129211902618, 0.030484061688184738, 0.14785519242286682, 0.011324583552777767, 0.14517802000045776, 0.07210396230220795, -0.043099530041217804, 0.020173687487840652, 0.05760704725980759, -0.14726482331752777, -0.09851435571908951, 0.0032092025503516197, -0.011646303348243237, -0.0974501222372055, 0.023828880861401558, 0.10530957579612732, -0.07094137370586395, -0.015073563903570175, -0.001354108564555645, 0.00421906728297472, -0.056278858333826065, 0.20935192704200745, 0.04181188717484474, 0.045024573802948, -0.10191436111927032, 0.0707210823893547, 0.06832195073366165, -0.08964808285236359, 0.00993502140045166, 0.09424503892660141, -0.06908190995454788, -0.041838258504867554, 0.09895788133144379, 0.18326790630817413, -0.06115535646677017, -0.052227284759283066, -0.13775545358657837, -0.1257747858762741, 0.0836097002029419, 0.15467534959316254, 0.09101999551057816, 0.01022416539490223, -0.056607432663440704, 0.01793774962425232, -0.11219318956136703, 0.08645049482584, 0.05046258121728897, 0.06604874134063721, -0.11693543195724487, 0.17879492044448853, 0.019278259947896004, 0.03014407679438591, -0.01975470408797264, 0.01953020691871643, -0.0996992364525795, 0.018736697733402252, -0.15295851230621338, -0.03050967864692211, -0.021909181028604507, 0.004243950359523296, -0.009330005384981632, -0.056228574365377426, -0.05361832678318024, 0.01074991561472416, -0.1244160607457161, -0.03095104545354843, 0.016803549602627754, 0.051998887211084366, -0.12358442693948746, -0.040309298783540726, 0.030399562790989876, -0.060959529131650925, 0.06211674585938454, 0.042984794825315475, 0.01629852131009102, 0.05957023426890373, -0.1543388068675995, -0.0008564500021748245, 0.054273493587970734, 0.014685038477182388, 0.05535905808210373, -0.09014321863651276, -0.01455523818731308, 0.00920642726123333, 0.06838034093379974, 0.012648557312786579, 0.07839054614305496, -0.1401091367006302, -0.017275916412472725, -0.026878677308559418, -0.08726391196250916, -0.06537812203168869, 0.0349012091755867, 0.05977591872215271, 0.025268997997045517, 0.19140265882015228, -0.08591519296169281, 0.04349004477262497, -0.2249266803264618, 0.00443578464910388, -0.01949417218565941, -0.11361955851316452, -0.11297199875116348, -0.07334589958190918, 0.06251771748065948, -0.04872793331742287, 0.1303475946187973, 0.01994432508945465, 0.055724598467350006, 0.030796509236097336, -0.03247934207320213, 0.006079598795622587, 0.017334716394543648, 0.20509415864944458, 0.035108782351017, -0.029343655332922935, 0.06059243902564049, 0.053900592029094696, 0.08851487934589386, 0.11867611110210419, 0.19023163616657257, 0.16243740916252136, 0.014272973872721195, 0.08425895124673843, 0.036830462515354156, -0.05881085619330406, -0.15314625203609467, 0.05231079086661339, -0.03973090648651123, 0.10766033083200455, -0.035850416868925095, 0.24789296090602875, 0.08230912685394287, -0.16906382143497467, 0.06066376343369484, -0.05281450226902962, -0.08429483324289322, -0.10045257210731506, -0.05304477736353874, -0.08157717436552048, -0.1533229649066925, -0.010376306250691414, -0.10730087012052536, 0.04410395398736, 0.10569559037685394, 0.011313079856336117, -0.022711845114827156, 0.14224213361740112, 0.04678516834974289, 0.006921378429979086, 0.049485716968774796, -0.004396454896777868, -0.018589219078421593, -0.11076974123716354, -0.07876267284154892, 0.0007373534608632326, 0.0008972617797553539, 0.04222412034869194, -0.044756557792425156, -0.05759548395872116, 0.039335522800683975, -0.03332393243908882, -0.10205753147602081, 0.01991400122642517, 0.01644257828593254, 0.07677613943815231, 0.06256134063005447, 0.015582529827952385, 0.010071598924696445, -0.012238532304763794, 0.23328664898872375, -0.07652296125888824, -0.09849634766578674, -0.09997638314962387, 0.25519120693206787, 0.03242529183626175, -0.016361214220523834, 0.02551874704658985, -0.05580533668398857, 0.0024922143202275038, 0.24792322516441345, 0.18503901362419128, -0.09330018609762192, -0.016223885118961334, 0.003951533697545528, -0.013180235400795937, -0.030911341309547424, 0.12488389760255814, 0.14532248675823212, 0.04233239218592644, -0.10485904663801193, -0.03146449103951454, -0.05317186191678047, -0.020958123728632927, -0.04599064961075783, 0.06540154665708542, 0.02744894102215767, 0.002543752547353506, -0.030882662162184715, 0.06242403760552406, -0.049079589545726776, -0.08729781955480576, 0.011982234194874763, -0.1934814304113388, -0.16230864822864532, -0.018154799938201904, 0.11844339221715927, -0.0005745822563767433, 0.051182594150304794, -0.02575855329632759, 0.0185923520475626, 0.07263938337564468, -0.02951509691774845, -0.06663591414690018, -0.08516581356525421, 0.10318087786436081, -0.12906059622764587, 0.19309887290000916, -0.04397737234830856, 0.05572250112891197, 0.12750227749347687, 0.07198160886764526, -0.06880002468824387, 0.07642154395580292, 0.039752088487148285, -0.08129467070102692, 0.03581864759325981, 0.0997769683599472, -0.03008105792105198, 0.06486620008945465, 0.04390885308384895, -0.13144659996032715, 0.0396687313914299, -0.08762539178133011, -0.04982801154255867, -0.025038480758666992, -0.04512716829776764, -0.049123141914606094, 0.127562016248703, 0.21153198182582855, -0.03048686683177948, 0.015169437043368816, -0.08045081794261932, 0.0019922079518437386, 0.04781306907534599, 0.04550519585609436, -0.0732780173420906, -0.23838646709918976, 0.007762589957565069, 0.06057143956422806, -0.01739509217441082, -0.24684923887252808, -0.10164336115121841, -0.00023782998323440552, -0.07349865883588791, -0.1015484556555748, 0.09750201553106308, 0.08572869747877121, 0.04867485910654068, -0.05064929276704788, -0.10593841969966888, -0.0775187537074089, 0.1609395295381546, -0.14609578251838684, -0.07436643540859222 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bart-paraphrase-pubmed-1.1 This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.4236 - Rouge2 Precision: 0.8482 - Rouge2 Recall: 0.673 - Rouge2 Fmeasure: 0.7347 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure | |:-------------:|:-----:|:----:|:---------------:|:----------------:|:-------------:|:---------------:| | 0.6534 | 1.0 | 663 | 0.4641 | 0.8448 | 0.6691 | 0.7313 | | 0.5078 | 2.0 | 1326 | 0.4398 | 0.8457 | 0.6719 | 0.7333 | | 0.4367 | 3.0 | 1989 | 0.4274 | 0.847 | 0.6717 | 0.7335 | | 0.3575 | 4.0 | 2652 | 0.4149 | 0.8481 | 0.6733 | 0.735 | | 0.3319 | 5.0 | 3315 | 0.4170 | 0.8481 | 0.6724 | 0.7343 | | 0.3179 | 6.0 | 3978 | 0.4264 | 0.8484 | 0.6733 | 0.735 | | 0.2702 | 7.0 | 4641 | 0.4207 | 0.8489 | 0.6732 | 0.7353 | | 0.2606 | 8.0 | 5304 | 0.4205 | 0.8487 | 0.6725 | 0.7347 | | 0.2496 | 9.0 | 5967 | 0.4247 | 0.8466 | 0.6717 | 0.7334 | | 0.2353 | 10.0 | 6630 | 0.4236 | 0.8482 | 0.673 | 0.7347 | ### Framework versions - Transformers 4.12.3 - Pytorch 1.9.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "bart-paraphrase-pubmed-1.1", "results": []}]}
text2text-generation
gayanin/bart-paraphrase-pubmed-1.1
[ "transformers", "pytorch", "tensorboard", "bart", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
bart-paraphrase-pubmed-1.1 ========================== This model is a fine-tuned version of facebook/bart-base on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 0.4236 * Rouge2 Precision: 0.8482 * Rouge2 Recall: 0.673 * Rouge2 Fmeasure: 0.7347 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 10 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.12.3 * Pytorch 1.9.0+cu111 * Datasets 1.15.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ 57, 113, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ -0.08557509630918503, 0.05797784775495529, -0.0024650858249515295, 0.0973440408706665, 0.15082214772701263, 0.014361673034727573, 0.14202666282653809, 0.1284749060869217, -0.1130785346031189, 0.015352019108831882, 0.11496320366859436, 0.1527942717075348, 0.02294573374092579, 0.13084737956523895, -0.04365701228380203, -0.2648164629936218, -0.0026389164850115776, 0.04189163073897362, -0.05457921326160431, 0.13746899366378784, 0.0935695543885231, -0.12811213731765747, 0.06262031197547913, 0.015224087983369827, -0.20650480687618256, 0.011299044825136662, 0.013253763318061829, -0.054878585040569305, 0.1566205471754074, 0.029767753556370735, 0.12552888691425323, 0.014870206825435162, 0.08888030052185059, -0.1990465372800827, 0.013165589421987534, 0.057609137147665024, 0.013492707163095474, 0.08827997744083405, 0.07116074115037918, 0.0077233994379639626, 0.12105370312929153, -0.055542074143886566, 0.06459270417690277, 0.022096144035458565, -0.12724025547504425, -0.24369101226329803, -0.09698525071144104, 0.014968804083764553, 0.06725514680147171, 0.1028880700469017, -0.006567833013832569, 0.14157108962535858, -0.09084468334913254, 0.09689459949731827, 0.22284746170043945, -0.2906811237335205, -0.06498780101537704, 0.011874345131218433, 0.045909855514764786, 0.08240236341953278, -0.10873782634735107, -0.03148496150970459, 0.03735921159386635, 0.0532090961933136, 0.15261325240135193, -0.02854987047612667, -0.13056209683418274, 0.0071847327053546906, -0.13974504172801971, -0.041684433817863464, 0.11841420829296112, 0.03234250470995903, -0.03081713244318962, -0.04771360382437706, -0.061295535415410995, -0.15475089848041534, -0.046411070972681046, -0.013999939896166325, 0.047232769429683685, -0.020092923194169998, -0.0768132358789444, -0.022629056125879288, -0.10648367553949356, -0.07063766568899155, -0.0673685073852539, 0.12165956199169159, 0.042844004929065704, 0.0031615335028618574, -0.03630446642637253, 0.10548682510852814, 0.0035586943849921227, -0.13221335411071777, 0.030069373548030853, 0.03492699936032295, -0.01865115389227867, -0.04105285555124283, -0.06947466731071472, -0.08198700845241547, 0.003853237023577094, 0.14006449282169342, -0.05646827444434166, 0.053593363612890244, -0.0061585004441440105, 0.04709697887301445, -0.10937396436929703, 0.1860518455505371, -0.03982982039451599, -0.011398683302104473, 0.006969108246266842, 0.05643102154135704, 0.007961134426295757, -0.017133241519331932, -0.1084834411740303, 0.015050229616463184, 0.11153305321931839, 0.016323400661349297, -0.05195974186062813, 0.06597777456045151, -0.049927856773138046, -0.018759995698928833, -0.003833961207419634, -0.09406312555074692, 0.03616093471646309, -0.0017729229293763638, -0.07796341180801392, -0.015656862407922745, 0.02844221331179142, 0.031641390174627304, -0.02240990847349167, 0.10548911988735199, -0.07533113658428192, 0.03979979455471039, -0.11055824905633926, -0.13053302466869354, 0.02338314801454544, -0.0502239391207695, 0.012467453256249428, -0.09571076929569244, -0.17265865206718445, -0.02260986529290676, 0.06242571026086807, -0.025519538670778275, -0.048985522240400314, -0.04684624448418617, -0.06004047021269798, 0.019441721960902214, -0.02986130118370056, 0.15474113821983337, -0.059180356562137604, 0.11162208020687103, 0.027848316356539726, 0.057262178510427475, -0.04753875732421875, 0.06424582004547119, -0.10222961008548737, 0.006006013136357069, -0.18065530061721802, 0.0426718108355999, -0.04955210164189339, 0.07003756612539291, -0.09592458605766296, -0.09480317682027817, 0.005579913966357708, -0.0015064466279000044, 0.09391278028488159, 0.08172187209129333, -0.1822434812784195, -0.07377175986766815, 0.18064740300178528, -0.06511407345533371, -0.10493358969688416, 0.1239113137125969, -0.06153952702879906, 0.05544251203536987, 0.07404625415802002, 0.18374469876289368, 0.05792616307735443, -0.08046869188547134, 0.04303304851055145, -0.021938232704997063, 0.05866285040974617, -0.05184716731309891, 0.06214688718318939, -0.0013054247247055173, 0.006874904036521912, 0.0265116598457098, -0.019144419580698013, 0.08176027983427048, -0.09347759187221527, -0.08844824135303497, -0.03787373751401901, -0.08717095851898193, 0.04232139512896538, 0.06494518369436264, 0.07661247253417969, -0.09733280539512634, -0.08833606541156769, 0.07622445374727249, 0.07923863083124161, -0.07062407582998276, 0.04146646335721016, -0.05760251730680466, 0.053641967475414276, -0.03358220309019089, -0.011056359857320786, -0.18741513788700104, -0.011289747431874275, 0.01638415828347206, -0.017600003629922867, 0.037580668926239014, 0.014952936209738255, 0.07492576539516449, 0.06219782680273056, -0.04979800805449486, -0.026544280350208282, -0.03119697794318199, -0.006423572544008493, -0.13293619453907013, -0.20306040346622467, -0.028028221800923347, -0.021641265600919724, 0.13343122601509094, -0.2009158879518509, 0.03583341836929321, -0.023619288578629494, 0.07827150821685791, 0.01246605347841978, -0.008513009175658226, -0.04345930367708206, 0.09174907952547073, -0.03419872373342514, -0.048438847064971924, 0.0778726115822792, 0.016632501035928726, -0.08936432749032974, -0.005213967990130186, -0.13059812784194946, 0.154335618019104, 0.13398605585098267, -0.109267458319664, -0.07169313728809357, -0.01927037537097931, -0.05461978167295456, -0.042211223393678665, -0.03387223929166794, 0.029679985716938972, 0.18176357448101044, 0.004334183409810066, 0.1574278622865677, -0.07010001689195633, -0.04696783050894737, 0.014187801629304886, -0.031674742698669434, 0.03749563544988632, 0.10982924699783325, 0.10342871397733688, -0.06769175082445145, 0.13820020854473114, 0.15716181695461273, -0.09168849885463715, 0.13502740859985352, -0.04192240908741951, -0.08164186775684357, -0.018649788573384285, -0.01536570768803358, 0.003787263063713908, 0.0873529314994812, -0.12143122404813766, 0.003933616913855076, 0.02258227951824665, 0.02594759501516819, 0.024460578337311745, -0.22989404201507568, -0.028422435745596886, 0.03205085173249245, -0.05031414330005646, -0.011806132271885872, -0.02739398553967476, 0.011359723284840584, 0.10586884617805481, -0.004531248938292265, -0.08050142228603363, 0.02993703819811344, 0.004699133802205324, -0.0769304484128952, 0.20169587433338165, -0.09706494212150574, -0.1637854427099228, -0.12091536819934845, -0.08293202519416809, -0.026052171364426613, 0.0014434658223763108, 0.06812053173780441, -0.07721155136823654, -0.027105147019028664, -0.0622563362121582, 0.025878239423036575, 0.00303762243129313, 0.016288071870803833, -0.005396385211497545, -0.006983431056141853, 0.07750478386878967, -0.1158980205655098, -0.004352658521384001, -0.04156622290611267, -0.06408201903104782, 0.055429182946681976, 0.045250020921230316, 0.11900973320007324, 0.1562872976064682, -0.019941333681344986, 0.0020934613421559334, -0.032178498804569244, 0.21723729372024536, -0.0711238756775856, -0.025527210906147957, 0.13363201916217804, -0.010772963054478168, 0.05935594066977501, 0.11434641480445862, 0.06896944344043732, -0.08439167588949203, 0.015641551464796066, 0.03146744519472122, -0.02775386907160282, -0.22607465088367462, -0.035509444773197174, -0.05682922527194023, -0.022771256044507027, 0.09378010034561157, 0.021831363439559937, 0.055628132075071335, 0.05593869462609291, 0.028792228549718857, 0.07442972809076309, -0.022520264610648155, 0.06372927874326706, 0.1331590712070465, 0.042940836399793625, 0.1387576460838318, -0.04075496643781662, -0.06493838876485825, 0.03414444252848625, 0.0032412968575954437, 0.22154109179973602, 0.015358719043433666, 0.15438289940357208, 0.060494668781757355, 0.17960532009601593, 0.007917347364127636, 0.07860751450061798, 0.0015619045589119196, -0.036762550473213196, -0.01835392601788044, -0.041727956384420395, -0.03988157957792282, 0.01668858714401722, -0.04960950091481209, 0.03889947757124901, -0.10701512545347214, -0.04220839962363243, 0.047638606280088425, 0.27584484219551086, 0.022613098844885826, -0.32008102536201477, -0.07924925535917282, -0.000800048466771841, -0.05493105575442314, -0.02113347128033638, 0.015032713301479816, 0.0895565003156662, -0.10183075815439224, 0.030688468366861343, -0.07749426364898682, 0.10763554275035858, -0.03400316834449768, 0.05004250630736351, 0.05611645430326462, 0.09402332454919815, 0.006991265341639519, 0.07751007378101349, -0.34100282192230225, 0.2773095667362213, 0.003207253059372306, 0.0686740130186081, -0.0708940401673317, 0.003884126665070653, 0.03813540190458298, 0.029243171215057373, 0.03716603294014931, -0.022894810885190964, -0.04986367002129555, -0.19813089072704315, -0.060870710760354996, 0.030555516481399536, 0.087855763733387, -0.013954204507172108, 0.10670018941164017, -0.03637736290693283, 0.012944468297064304, 0.07936744391918182, -0.021180756390094757, -0.08449053019285202, -0.09916507452726364, -0.003497461089864373, 0.024422915652394295, -0.01013598870486021, -0.07431060820817947, -0.11199549585580826, -0.10714338719844818, 0.14454849064350128, 0.00030904225423000753, -0.021153245121240616, -0.11560752242803574, 0.08575820922851562, 0.08170140534639359, -0.08235042542219162, 0.03866954520344734, 0.010728503577411175, 0.0735451802611351, 0.020851613953709602, -0.06992997974157333, 0.11719205230474472, -0.06279655545949936, -0.15596836805343628, -0.06059526279568672, 0.09620831906795502, 0.03372067213058472, 0.07123098522424698, -0.011486504226922989, 0.014384274370968342, -0.03605905547738075, -0.08338286727666855, 0.008236547000706196, -0.01939668133854866, 0.05514785274863243, 0.006125152111053467, -0.058195680379867554, 0.014784802682697773, -0.06521721929311752, -0.04701821878552437, 0.19605928659439087, 0.23573876917362213, -0.09226791560649872, 0.04106112942099571, 0.05637664347887039, -0.07482227683067322, -0.18666554987430573, 0.02983306162059307, 0.06072240695357323, 0.007387315854430199, 0.060703735798597336, -0.1887173056602478, 0.09051591902971268, 0.1033984050154686, -0.016317160800099373, 0.08762941509485245, -0.3422962725162506, -0.12914228439331055, 0.12140638381242752, 0.1478712409734726, 0.07760870456695557, -0.15588858723640442, -0.024483660236001015, -0.025420840829610825, -0.11983033269643784, 0.10904578864574432, -0.10273055732250214, 0.12601962685585022, -0.024580011144280434, 0.09276609122753143, 0.005451680161058903, -0.05881782993674278, 0.11245585978031158, -0.01189667172729969, 0.09958674013614655, -0.06396038830280304, 0.015464957803487778, 0.05506577715277672, -0.03509014844894409, 0.011536392383277416, -0.09352053701877594, 0.020168304443359375, -0.08376681804656982, -0.023155800998210907, -0.07963050901889801, 0.032956406474113464, -0.03793349862098694, -0.05986899882555008, -0.026036793366074562, 0.02429935336112976, 0.050531789660453796, -0.010982504114508629, 0.11777891218662262, 0.005092812702059746, 0.16532002389431, 0.11997238546609879, 0.06758877635002136, -0.05856377258896828, -0.057965926826000214, -0.018287530168890953, -0.018600642681121826, 0.055018987506628036, -0.12774129211902618, 0.030484061688184738, 0.14785519242286682, 0.011324583552777767, 0.14517802000045776, 0.07210396230220795, -0.043099530041217804, 0.020173687487840652, 0.05760704725980759, -0.14726482331752777, -0.09851435571908951, 0.0032092025503516197, -0.011646303348243237, -0.0974501222372055, 0.023828880861401558, 0.10530957579612732, -0.07094137370586395, -0.015073563903570175, -0.001354108564555645, 0.00421906728297472, -0.056278858333826065, 0.20935192704200745, 0.04181188717484474, 0.045024573802948, -0.10191436111927032, 0.0707210823893547, 0.06832195073366165, -0.08964808285236359, 0.00993502140045166, 0.09424503892660141, -0.06908190995454788, -0.041838258504867554, 0.09895788133144379, 0.18326790630817413, -0.06115535646677017, -0.052227284759283066, -0.13775545358657837, -0.1257747858762741, 0.0836097002029419, 0.15467534959316254, 0.09101999551057816, 0.01022416539490223, -0.056607432663440704, 0.01793774962425232, -0.11219318956136703, 0.08645049482584, 0.05046258121728897, 0.06604874134063721, -0.11693543195724487, 0.17879492044448853, 0.019278259947896004, 0.03014407679438591, -0.01975470408797264, 0.01953020691871643, -0.0996992364525795, 0.018736697733402252, -0.15295851230621338, -0.03050967864692211, -0.021909181028604507, 0.004243950359523296, -0.009330005384981632, -0.056228574365377426, -0.05361832678318024, 0.01074991561472416, -0.1244160607457161, -0.03095104545354843, 0.016803549602627754, 0.051998887211084366, -0.12358442693948746, -0.040309298783540726, 0.030399562790989876, -0.060959529131650925, 0.06211674585938454, 0.042984794825315475, 0.01629852131009102, 0.05957023426890373, -0.1543388068675995, -0.0008564500021748245, 0.054273493587970734, 0.014685038477182388, 0.05535905808210373, -0.09014321863651276, -0.01455523818731308, 0.00920642726123333, 0.06838034093379974, 0.012648557312786579, 0.07839054614305496, -0.1401091367006302, -0.017275916412472725, -0.026878677308559418, -0.08726391196250916, -0.06537812203168869, 0.0349012091755867, 0.05977591872215271, 0.025268997997045517, 0.19140265882015228, -0.08591519296169281, 0.04349004477262497, -0.2249266803264618, 0.00443578464910388, -0.01949417218565941, -0.11361955851316452, -0.11297199875116348, -0.07334589958190918, 0.06251771748065948, -0.04872793331742287, 0.1303475946187973, 0.01994432508945465, 0.055724598467350006, 0.030796509236097336, -0.03247934207320213, 0.006079598795622587, 0.017334716394543648, 0.20509415864944458, 0.035108782351017, -0.029343655332922935, 0.06059243902564049, 0.053900592029094696, 0.08851487934589386, 0.11867611110210419, 0.19023163616657257, 0.16243740916252136, 0.014272973872721195, 0.08425895124673843, 0.036830462515354156, -0.05881085619330406, -0.15314625203609467, 0.05231079086661339, -0.03973090648651123, 0.10766033083200455, -0.035850416868925095, 0.24789296090602875, 0.08230912685394287, -0.16906382143497467, 0.06066376343369484, -0.05281450226902962, -0.08429483324289322, -0.10045257210731506, -0.05304477736353874, -0.08157717436552048, -0.1533229649066925, -0.010376306250691414, -0.10730087012052536, 0.04410395398736, 0.10569559037685394, 0.011313079856336117, -0.022711845114827156, 0.14224213361740112, 0.04678516834974289, 0.006921378429979086, 0.049485716968774796, -0.004396454896777868, -0.018589219078421593, -0.11076974123716354, -0.07876267284154892, 0.0007373534608632326, 0.0008972617797553539, 0.04222412034869194, -0.044756557792425156, -0.05759548395872116, 0.039335522800683975, -0.03332393243908882, -0.10205753147602081, 0.01991400122642517, 0.01644257828593254, 0.07677613943815231, 0.06256134063005447, 0.015582529827952385, 0.010071598924696445, -0.012238532304763794, 0.23328664898872375, -0.07652296125888824, -0.09849634766578674, -0.09997638314962387, 0.25519120693206787, 0.03242529183626175, -0.016361214220523834, 0.02551874704658985, -0.05580533668398857, 0.0024922143202275038, 0.24792322516441345, 0.18503901362419128, -0.09330018609762192, -0.016223885118961334, 0.003951533697545528, -0.013180235400795937, -0.030911341309547424, 0.12488389760255814, 0.14532248675823212, 0.04233239218592644, -0.10485904663801193, -0.03146449103951454, -0.05317186191678047, -0.020958123728632927, -0.04599064961075783, 0.06540154665708542, 0.02744894102215767, 0.002543752547353506, -0.030882662162184715, 0.06242403760552406, -0.049079589545726776, -0.08729781955480576, 0.011982234194874763, -0.1934814304113388, -0.16230864822864532, -0.018154799938201904, 0.11844339221715927, -0.0005745822563767433, 0.051182594150304794, -0.02575855329632759, 0.0185923520475626, 0.07263938337564468, -0.02951509691774845, -0.06663591414690018, -0.08516581356525421, 0.10318087786436081, -0.12906059622764587, 0.19309887290000916, -0.04397737234830856, 0.05572250112891197, 0.12750227749347687, 0.07198160886764526, -0.06880002468824387, 0.07642154395580292, 0.039752088487148285, -0.08129467070102692, 0.03581864759325981, 0.0997769683599472, -0.03008105792105198, 0.06486620008945465, 0.04390885308384895, -0.13144659996032715, 0.0396687313914299, -0.08762539178133011, -0.04982801154255867, -0.025038480758666992, -0.04512716829776764, -0.049123141914606094, 0.127562016248703, 0.21153198182582855, -0.03048686683177948, 0.015169437043368816, -0.08045081794261932, 0.0019922079518437386, 0.04781306907534599, 0.04550519585609436, -0.0732780173420906, -0.23838646709918976, 0.007762589957565069, 0.06057143956422806, -0.01739509217441082, -0.24684923887252808, -0.10164336115121841, -0.00023782998323440552, -0.07349865883588791, -0.1015484556555748, 0.09750201553106308, 0.08572869747877121, 0.04867485910654068, -0.05064929276704788, -0.10593841969966888, -0.0775187537074089, 0.1609395295381546, -0.14609578251838684, -0.07436643540859222 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bart-paraphrase-pubmed This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.6340 - Rouge2 Precision: 0.83 - Rouge2 Recall: 0.6526 - Rouge2 Fmeasure: 0.7144 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 40 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure | |:-------------:|:-----:|:-----:|:---------------:|:----------------:|:-------------:|:---------------:| | 0.6613 | 1.0 | 663 | 0.4750 | 0.8321 | 0.6552 | 0.7167 | | 0.4993 | 2.0 | 1326 | 0.4404 | 0.8366 | 0.6583 | 0.7203 | | 0.443 | 3.0 | 1989 | 0.4261 | 0.8319 | 0.6562 | 0.7176 | | 0.3482 | 4.0 | 2652 | 0.4198 | 0.8348 | 0.6571 | 0.7191 | | 0.3206 | 5.0 | 3315 | 0.4233 | 0.8344 | 0.656 | 0.7183 | | 0.294 | 6.0 | 3978 | 0.4334 | 0.835 | 0.657 | 0.719 | | 0.2404 | 7.0 | 4641 | 0.4437 | 0.8334 | 0.6559 | 0.7178 | | 0.2228 | 8.0 | 5304 | 0.4438 | 0.8348 | 0.6565 | 0.7187 | | 0.211 | 9.0 | 5967 | 0.4516 | 0.8329 | 0.6549 | 0.717 | | 0.1713 | 10.0 | 6630 | 0.4535 | 0.8332 | 0.6547 | 0.7169 | | 0.1591 | 11.0 | 7293 | 0.4763 | 0.8349 | 0.6561 | 0.7184 | | 0.1555 | 12.0 | 7956 | 0.4824 | 0.8311 | 0.6534 | 0.7153 | | 0.1262 | 13.0 | 8619 | 0.4883 | 0.8322 | 0.655 | 0.7167 | | 0.1164 | 14.0 | 9282 | 0.5025 | 0.8312 | 0.6539 | 0.7158 | | 0.1108 | 15.0 | 9945 | 0.5149 | 0.8321 | 0.6535 | 0.7157 | | 0.0926 | 16.0 | 10608 | 0.5340 | 0.8315 | 0.6544 | 0.7159 | | 0.0856 | 17.0 | 11271 | 0.5322 | 0.8306 | 0.6518 | 0.7142 | | 0.0785 | 18.0 | 11934 | 0.5346 | 0.8324 | 0.6549 | 0.7167 | | 0.071 | 19.0 | 12597 | 0.5488 | 0.8311 | 0.652 | 0.714 | | 0.0635 | 20.0 | 13260 | 0.5624 | 0.8287 | 0.6517 | 0.7132 | | 0.0608 | 21.0 | 13923 | 0.5612 | 0.8299 | 0.6527 | 0.7141 | | 0.0531 | 22.0 | 14586 | 0.5764 | 0.8283 | 0.6498 | 0.7119 | | 0.0486 | 23.0 | 15249 | 0.5832 | 0.8298 | 0.6532 | 0.7148 | | 0.0465 | 24.0 | 15912 | 0.5866 | 0.83 | 0.6522 | 0.7142 | | 0.0418 | 25.0 | 16575 | 0.5825 | 0.83 | 0.6523 | 0.7141 | | 0.0391 | 26.0 | 17238 | 0.5997 | 0.8306 | 0.6545 | 0.716 | | 0.0376 | 27.0 | 17901 | 0.5894 | 0.8315 | 0.6546 | 0.7164 | | 0.035 | 28.0 | 18564 | 0.6045 | 0.8306 | 0.6529 | 0.7149 | | 0.0316 | 29.0 | 19227 | 0.6168 | 0.8311 | 0.6546 | 0.7162 | | 0.0314 | 30.0 | 19890 | 0.6203 | 0.8311 | 0.6552 | 0.7164 | | 0.0292 | 31.0 | 20553 | 0.6173 | 0.8315 | 0.6548 | 0.7163 | | 0.0265 | 32.0 | 21216 | 0.6226 | 0.832 | 0.6548 | 0.7166 | | 0.0274 | 33.0 | 21879 | 0.6264 | 0.8314 | 0.6538 | 0.7155 | | 0.0247 | 34.0 | 22542 | 0.6254 | 0.8289 | 0.6515 | 0.7132 | | 0.0238 | 35.0 | 23205 | 0.6254 | 0.8307 | 0.6519 | 0.7142 | | 0.0232 | 36.0 | 23868 | 0.6295 | 0.8287 | 0.6515 | 0.7133 | | 0.0215 | 37.0 | 24531 | 0.6326 | 0.8293 | 0.6523 | 0.7138 | | 0.0212 | 38.0 | 25194 | 0.6332 | 0.8295 | 0.6522 | 0.714 | | 0.0221 | 39.0 | 25857 | 0.6335 | 0.8305 | 0.6528 | 0.7147 | | 0.0202 | 40.0 | 26520 | 0.6340 | 0.83 | 0.6526 | 0.7144 | ### Framework versions - Transformers 4.12.3 - Pytorch 1.9.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "bart-paraphrase-pubmed", "results": []}]}
text2text-generation
gayanin/bart-paraphrase-pubmed
[ "transformers", "pytorch", "tensorboard", "bart", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
bart-paraphrase-pubmed ====================== This model is a fine-tuned version of facebook/bart-base on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 0.6340 * Rouge2 Precision: 0.83 * Rouge2 Recall: 0.6526 * Rouge2 Fmeasure: 0.7144 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 40 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.12.3 * Pytorch 1.9.0+cu111 * Datasets 1.15.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ 57, 113, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bart #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ -0.08599434047937393, 0.05834865942597389, -0.002454254077747464, 0.09755201637744904, 0.1502911001443863, 0.01487046293914318, 0.1404426544904709, 0.12856683135032654, -0.11431876569986343, 0.015327106229960918, 0.11476201564073563, 0.15231545269489288, 0.0231939610093832, 0.13046801090240479, -0.04386463761329651, -0.2653008699417114, -0.003989884164184332, 0.04227236658334732, -0.05577414855360985, 0.13822516798973083, 0.09378889948129654, -0.1278133988380432, 0.06302109360694885, 0.015066556632518768, -0.2076697200536728, 0.011048037558794022, 0.014454306103289127, -0.054308392107486725, 0.15650887787342072, 0.030611077323555946, 0.12592890858650208, 0.015065697021782398, 0.08929910510778427, -0.19830070436000824, 0.013111257925629616, 0.05774792656302452, 0.013414233922958374, 0.08879932761192322, 0.07067517191171646, 0.008151953108608723, 0.12202291935682297, -0.05537032335996628, 0.06277098506689072, 0.021911723539233208, -0.12743841111660004, -0.2447482943534851, -0.09741463512182236, 0.014734630472958088, 0.06658800691366196, 0.10341746360063553, -0.00609555933624506, 0.1419001966714859, -0.09010140597820282, 0.09710690379142761, 0.2209998518228531, -0.2926889955997467, -0.06498176604509354, 0.011604208499193192, 0.04376288875937462, 0.08313845098018646, -0.10857803374528885, -0.030330512672662735, 0.036809053272008896, 0.05362755432724953, 0.15197400748729706, -0.02883593738079071, -0.1294025480747223, 0.007756366394460201, -0.13953395187854767, -0.041564859449863434, 0.11867208033800125, 0.03181876614689827, -0.030901506543159485, -0.04666568338871002, -0.06212157383561134, -0.15422138571739197, -0.046199627220630646, -0.013807124458253384, 0.04773496091365814, -0.0194022785872221, -0.07690723240375519, -0.02159295603632927, -0.10685940086841583, -0.07100570946931839, -0.06920818984508514, 0.12076534330844879, 0.04215540736913681, 0.003820213722065091, -0.036157261580228806, 0.10596241801977158, 0.0032408111728727818, -0.13212591409683228, 0.030532224103808403, 0.0355200320482254, -0.016898075118660927, -0.04100484028458595, -0.06971294432878494, -0.08249813318252563, 0.0032743695192039013, 0.13864581286907196, -0.059629589319229126, 0.05299123749136925, -0.005841272417455912, 0.048070140182971954, -0.10928124189376831, 0.18581390380859375, -0.03982224687933922, -0.012938288040459156, 0.006947798188775778, 0.05558207258582115, 0.007906414568424225, -0.01702549308538437, -0.10830985754728317, 0.013855022378265858, 0.11134301871061325, 0.015975503250956535, -0.05256016179919243, 0.06535071134567261, -0.05007942020893097, -0.018320439383387566, -0.0037524248473346233, -0.09331618994474411, 0.0360235758125782, -0.002973729046061635, -0.07712061703205109, -0.01437020767480135, 0.029453426599502563, 0.030974099412560463, -0.022710375487804413, 0.10471224039793015, -0.07534614205360413, 0.0398501455783844, -0.11078204214572906, -0.12888851761817932, 0.02241513319313526, -0.046284183859825134, 0.01215242501348257, -0.09574145078659058, -0.17133113741874695, -0.022494420409202576, 0.06273683905601501, -0.026282090693712234, -0.04963444173336029, -0.04529763013124466, -0.05876407027244568, 0.02018505521118641, -0.030140971764922142, 0.15476298332214355, -0.05932031199336052, 0.11273343116044998, 0.028262639418244362, 0.05743509158492088, -0.047183748334646225, 0.0643644779920578, -0.10184387862682343, 0.0058274162001907825, -0.18359054625034332, 0.042391981929540634, -0.04856310412287712, 0.06811285763978958, -0.09599843621253967, -0.09564989060163498, 0.0063094086945056915, -0.0011966119054704905, 0.09327573329210281, 0.08143791556358337, -0.18195630609989166, -0.07405339926481247, 0.18069756031036377, -0.06527987122535706, -0.10461360216140747, 0.12433458864688873, -0.0609377920627594, 0.055302560329437256, 0.0735134556889534, 0.18300841748714447, 0.05626527965068817, -0.08017414808273315, 0.04448419809341431, -0.020772967487573624, 0.05846629664301872, -0.05272767320275307, 0.061861976981163025, -0.0023110059555619955, 0.0053677186369895935, 0.025919923558831215, -0.02039193920791149, 0.08219429850578308, -0.09373018145561218, -0.08836131542921066, -0.03760780021548271, -0.0870809406042099, 0.04120631143450737, 0.06481700390577316, 0.07725662738084793, -0.09664175659418106, -0.08841022104024887, 0.07498371601104736, 0.07948129624128342, -0.06956784427165985, 0.04176681488752365, -0.05800706893205643, 0.0531475692987442, -0.03338198363780975, -0.011284689418971539, -0.18806269764900208, -0.010880351066589355, 0.015555355697870255, -0.01668579690158367, 0.0375010184943676, 0.015380295924842358, 0.07520195841789246, 0.06256834417581558, -0.050752658396959305, -0.026338422670960426, -0.030818190425634384, -0.006728185806423426, -0.13332432508468628, -0.20328551530838013, -0.027543889358639717, -0.020419932901859283, 0.13216787576675415, -0.19993245601654053, 0.035051651298999786, -0.02316482551395893, 0.07827211171388626, 0.012307415716350079, -0.00761375529691577, -0.04414622485637665, 0.09227070212364197, -0.03376131132245064, -0.04809686169028282, 0.07805051654577255, 0.01736290380358696, -0.08876808732748032, -0.004819985944777727, -0.1285131424665451, 0.1565212607383728, 0.1337907761335373, -0.11123388260602951, -0.07292395830154419, -0.020189132541418076, -0.054163336753845215, -0.0424845814704895, -0.03429996967315674, 0.030658941715955734, 0.1824934333562851, 0.004374104086309671, 0.15783318877220154, -0.06951592862606049, -0.04679398238658905, 0.013309096917510033, -0.032099347561597824, 0.03841041773557663, 0.10916014015674591, 0.10253839939832687, -0.0677894651889801, 0.13854530453681946, 0.15660111606121063, -0.09240374714136124, 0.1350812315940857, -0.042517371475696564, -0.0818207785487175, -0.01916096732020378, -0.015511785633862019, 0.003808314446359873, 0.08775798976421356, -0.1198357343673706, 0.004381198436021805, 0.022737978026270866, 0.025767171755433083, 0.024198438972234726, -0.22922059893608093, -0.02885458618402481, 0.031522348523139954, -0.05068572983145714, -0.011481457389891148, -0.02703692950308323, 0.011496715247631073, 0.10633011907339096, -0.003929548431187868, -0.08029170334339142, 0.030034147202968597, 0.004846271593123674, -0.07664937525987625, 0.2015686184167862, -0.09671233594417572, -0.16333068907260895, -0.12181244045495987, -0.08247335255146027, -0.02614292874932289, 0.0011053421767428517, 0.06721315532922745, -0.07785338163375854, -0.027016695588827133, -0.06105837598443031, 0.02719043381512165, 0.003346031066030264, 0.01755821891129017, -0.006226200610399246, -0.007615254260599613, 0.07894682884216309, -0.114702969789505, -0.00468524731695652, -0.041900474578142166, -0.06408455967903137, 0.05618305504322052, 0.04559933394193649, 0.11963074654340744, 0.156169131398201, -0.019627146422863007, 0.002127718413248658, -0.030812988057732582, 0.21843256056308746, -0.07088525593280792, -0.024929802864789963, 0.13419407606124878, -0.009526311419904232, 0.05983005836606026, 0.1130843460559845, 0.06967785954475403, -0.08454716950654984, 0.015465336851775646, 0.03113757260143757, -0.027322424575686455, -0.2247588336467743, -0.03611665591597557, -0.05704684928059578, -0.024400845170021057, 0.09338723868131638, 0.021422995254397392, 0.05451149865984917, 0.05544328689575195, 0.029544414952397346, 0.07530797272920609, -0.023491373285651207, 0.06308522820472717, 0.13335305452346802, 0.0431525893509388, 0.13874323666095734, -0.04044165462255478, -0.06549319624900818, 0.03553732857108116, 0.003377513261511922, 0.22062459588050842, 0.014951895922422409, 0.154981791973114, 0.06006443873047829, 0.18112115561962128, 0.009083702228963375, 0.0782301276922226, 0.0005739140324294567, -0.036878205835819244, -0.018276695162057877, -0.04196077212691307, -0.04090369492769241, 0.01679270528256893, -0.04979055002331734, 0.03841731324791908, -0.10672920197248459, -0.04182201996445656, 0.04762829467654228, 0.27746331691741943, 0.02176586166024208, -0.3201792538166046, -0.07984248548746109, -0.0007871415582485497, -0.05361107364296913, -0.021867292001843452, 0.01515277475118637, 0.08859307318925858, -0.10185211896896362, 0.030164752155542374, -0.07742176949977875, 0.10732197761535645, -0.03574932739138603, 0.04951346293091774, 0.05603883042931557, 0.09383442252874374, 0.007012224290519953, 0.07679852843284607, -0.3394135534763336, 0.27763456106185913, 0.003764447523280978, 0.06777896732091904, -0.0701209232211113, 0.0036184967029839754, 0.038402654230594635, 0.032647181302309036, 0.037591949105262756, -0.02252223528921604, -0.05036517605185509, -0.1983892321586609, -0.060426272451877594, 0.029836470261216164, 0.08702293038368225, -0.014342278242111206, 0.10627799481153488, -0.03698701038956642, 0.01315427664667368, 0.07902888208627701, -0.023033667355775833, -0.08282702416181564, -0.09866967797279358, -0.004151536151766777, 0.025075281038880348, -0.00956557597965002, -0.07447735220193863, -0.11201302707195282, -0.10496702790260315, 0.14582470059394836, -0.00004441500277607702, -0.021528594195842743, -0.11622561514377594, 0.085303895175457, 0.08086128532886505, -0.08262412995100021, 0.03746214881539345, 0.010665342211723328, 0.0743183046579361, 0.021477313712239265, -0.06991749256849289, 0.1176801472902298, -0.06282133609056473, -0.1546827256679535, -0.06094122678041458, 0.09567979723215103, 0.03419576212763786, 0.0712725892663002, -0.012237746268510818, 0.014417000114917755, -0.037038590759038925, -0.083699531853199, 0.00878449622541666, -0.020739970728754997, 0.05666366219520569, 0.005531974136829376, -0.05859399959445, 0.016763746738433838, -0.06393236666917801, -0.04729767143726349, 0.19484016299247742, 0.23572388291358948, -0.09256384521722794, 0.04066557064652443, 0.056808553636074066, -0.0740208625793457, -0.18730060756206512, 0.02930448018014431, 0.06022385507822037, 0.0074267820455133915, 0.060453787446022034, -0.1894194632768631, 0.09141681343317032, 0.10287632793188095, -0.016828540712594986, 0.08792013674974442, -0.3403591811656952, -0.12867721915245056, 0.12182667851448059, 0.14719249308109283, 0.08007444441318512, -0.15642644464969635, -0.02459191530942917, -0.02505306527018547, -0.11938125640153885, 0.10730750113725662, -0.10354623943567276, 0.12561848759651184, -0.02535521797835827, 0.09497386962175369, 0.0052956705912947655, -0.05858417972922325, 0.11266792565584183, -0.012096613645553589, 0.09939070791006088, -0.0640457421541214, 0.017021771520376205, 0.05313762649893761, -0.03517783805727959, 0.010760773904621601, -0.09323219954967499, 0.02104518935084343, -0.08497314155101776, -0.023238597437739372, -0.07976443320512772, 0.03336747735738754, -0.037948109209537506, -0.06015245243906975, -0.02621028572320938, 0.024976614862680435, 0.04956984892487526, -0.011363592930138111, 0.11780847609043121, 0.004889215342700481, 0.16515231132507324, 0.12180675566196442, 0.06730155646800995, -0.05830175802111626, -0.059517696499824524, -0.018047139048576355, -0.018978359177708626, 0.05564846470952034, -0.12694543600082397, 0.02999177575111389, 0.14766855537891388, 0.011121339164674282, 0.14453694224357605, 0.07168641686439514, -0.042829953134059906, 0.019224146381020546, 0.05798894166946411, -0.14816665649414062, -0.09743458777666092, 0.0027818072121590376, -0.011850985698401928, -0.09649297595024109, 0.02246815338730812, 0.10547858476638794, -0.07011464238166809, -0.014487535692751408, -0.0006218290072865784, 0.0038345586508512497, -0.05655380338430405, 0.20849324762821198, 0.04210299253463745, 0.044809740036726, -0.10197317600250244, 0.07018331438302994, 0.06750201433897018, -0.08765801042318344, 0.009321319870650768, 0.09485206007957458, -0.0698072612285614, -0.04174066707491875, 0.09983614087104797, 0.18448463082313538, -0.06503373384475708, -0.052133236080408096, -0.13725246489048004, -0.12592288851737976, 0.08328177034854889, 0.15304912626743317, 0.09171387553215027, 0.009489074349403381, -0.0571175292134285, 0.017233725637197495, -0.11167974025011063, 0.0855131521821022, 0.049648843705654144, 0.06554700434207916, -0.116292804479599, 0.17906299233436584, 0.019488023594021797, 0.030350396409630775, -0.02017747424542904, 0.02003367245197296, -0.0991997942328453, 0.01900547370314598, -0.15307964384555817, -0.030345972627401352, -0.020811056718230247, 0.0043624332174658775, -0.010301414877176285, -0.057195622473955154, -0.053986404091119766, 0.011322357691824436, -0.12456579506397247, -0.03048091195523739, 0.01719452440738678, 0.051429715007543564, -0.1229899451136589, -0.04011981561779976, 0.030934035778045654, -0.06069336086511612, 0.06172921136021614, 0.04295166954398155, 0.017236728221178055, 0.060201868414878845, -0.15163962543010712, -0.0006529937381856143, 0.05337587371468544, 0.014920718036592007, 0.05591290816664696, -0.09013031423091888, -0.014849647879600525, 0.009182192385196686, 0.06785354763269424, 0.01277496013790369, 0.07918918877840042, -0.14033469557762146, -0.01634557545185089, -0.02647225372493267, -0.08701763302087784, -0.06568551063537598, 0.034911442548036575, 0.06179738789796829, 0.02639692835509777, 0.1906956285238266, -0.08661317080259323, 0.04268795996904373, -0.22533157467842102, 0.004841512069106102, -0.019352147355675697, -0.11312513798475266, -0.11331541836261749, -0.07353511452674866, 0.06226040795445442, -0.04896393418312073, 0.13002046942710876, 0.020305665209889412, 0.054751187562942505, 0.03133288770914078, -0.03366745263338089, 0.005750593263655901, 0.01608583889901638, 0.205226331949234, 0.035753216594457626, -0.029129182919859886, 0.06155847758054733, 0.05452892184257507, 0.08767583221197128, 0.11909259855747223, 0.19100654125213623, 0.16124440729618073, 0.015551093965768814, 0.08405792713165283, 0.03592834621667862, -0.05907253548502922, -0.1542709469795227, 0.05198408663272858, -0.03987092524766922, 0.10657256841659546, -0.035640012472867966, 0.24772721529006958, 0.08283884078264236, -0.1692878007888794, 0.06053163483738899, -0.05334022268652916, -0.08406694233417511, -0.10109049826860428, -0.051476649940013885, -0.08149106800556183, -0.15377075970172882, -0.011047353968024254, -0.10708335041999817, 0.044735558331012726, 0.105055071413517, 0.010855378583073616, -0.022852279245853424, 0.14219476282596588, 0.046534523367881775, 0.007016164716333151, 0.04950641468167305, -0.004040314815938473, -0.0184505432844162, -0.10999295115470886, -0.07957183569669724, 0.00043241571984253824, 0.0004928381531499326, 0.041723351925611496, -0.04442113637924194, -0.05812237784266472, 0.03885599598288536, -0.03259952366352081, -0.10228612273931503, 0.01896536722779274, 0.0171419195830822, 0.07640837877988815, 0.06222086399793625, 0.015102550387382507, 0.010839882306754589, -0.01276362594217062, 0.23270754516124725, -0.07615658640861511, -0.09687668085098267, -0.09961318224668503, 0.2568187117576599, 0.03216251730918884, -0.016101347282528877, 0.02539229579269886, -0.055350761860609055, 0.0017620661528781056, 0.24776576459407806, 0.1838623583316803, -0.09586544334888458, -0.016557447612285614, 0.0039223902858793736, -0.013034794479608536, -0.029956461861729622, 0.1252327263355255, 0.1458398848772049, 0.0435309074819088, -0.10537716746330261, -0.03171052783727646, -0.054796136915683746, -0.020922789350152016, -0.04549646005034447, 0.06623286008834839, 0.027551455423235893, 0.003005504608154297, -0.03157292678952217, 0.06206898018717766, -0.0487452857196331, -0.08868007361888885, 0.01347963884472847, -0.19344396889209747, -0.16272121667861938, -0.017151057720184326, 0.11802516877651215, -0.0016480647027492523, 0.05056711286306381, -0.02561979927122593, 0.018958674743771553, 0.07432592660188675, -0.02965911477804184, -0.06678115576505661, -0.08528096228837967, 0.10371428728103638, -0.12842990458011627, 0.1926184445619583, -0.04361995682120323, 0.055147428065538406, 0.127189040184021, 0.07203391939401627, -0.06863263994455338, 0.07754376530647278, 0.03895392641425133, -0.08267904818058014, 0.035562075674533844, 0.10019126534461975, -0.030049525201320648, 0.06459307670593262, 0.04369943588972092, -0.13094839453697205, 0.03887425363063812, -0.08860503882169724, -0.050679948180913925, -0.024899795651435852, -0.04618152603507042, -0.049721308052539825, 0.1278197169303894, 0.21208956837654114, -0.030007870867848396, 0.015435700304806232, -0.0805194154381752, 0.0027713386807590723, 0.048479996621608734, 0.0453515462577343, -0.07323454320430756, -0.23722012341022491, 0.007858792319893837, 0.061782654374837875, -0.01760774664580822, -0.24717341363430023, -0.10177654772996902, 0.00009661721560405567, -0.0726672112941742, -0.1017979234457016, 0.09821505844593048, 0.08440291881561279, 0.04887804761528969, -0.050239745527505875, -0.10361301898956299, -0.07760649919509888, 0.15979816019535065, -0.14623968303203583, -0.07455562055110931 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-small-finetuned-pubmed This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.6131 - Rouge2 Precision: 0.3 - Rouge2 Recall: 0.2152 - Rouge2 Fmeasure: 0.2379 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure | |:-------------:|:-----:|:----:|:---------------:|:----------------:|:-------------:|:---------------:| | 2.1335 | 1.0 | 563 | 1.7632 | 0.2716 | 0.1936 | 0.2135 | | 1.9373 | 2.0 | 1126 | 1.7037 | 0.2839 | 0.2068 | 0.2265 | | 1.8827 | 3.0 | 1689 | 1.6723 | 0.2901 | 0.2118 | 0.2316 | | 1.8257 | 4.0 | 2252 | 1.6503 | 0.2938 | 0.2115 | 0.2332 | | 1.8152 | 5.0 | 2815 | 1.6386 | 0.2962 | 0.2139 | 0.2357 | | 1.7939 | 6.0 | 3378 | 1.6284 | 0.2976 | 0.212 | 0.2354 | | 1.7845 | 7.0 | 3941 | 1.6211 | 0.2991 | 0.2155 | 0.2383 | | 1.7468 | 8.0 | 4504 | 1.6167 | 0.2994 | 0.217 | 0.239 | | 1.7464 | 9.0 | 5067 | 1.6137 | 0.3007 | 0.2154 | 0.2382 | | 1.744 | 10.0 | 5630 | 1.6131 | 0.3 | 0.2152 | 0.2379 | ### Framework versions - Transformers 4.12.3 - Pytorch 1.9.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "t5-small-finetuned-pubmed", "results": []}]}
text2text-generation
gayanin/t5-small-finetuned-pubmed
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
t5-small-finetuned-pubmed ========================= This model is a fine-tuned version of t5-small on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 1.6131 * Rouge2 Precision: 0.3 * Rouge2 Recall: 0.2152 * Rouge2 Fmeasure: 0.2379 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 10 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.12.3 * Pytorch 1.9.0+cu111 * Datasets 1.15.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ 67, 113, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ -0.07949404418468475, 0.0564143992960453, -0.003547149244695902, 0.09548448026180267, 0.134885773062706, 0.015752414241433144, 0.13106422126293182, 0.14192859828472137, -0.11324204504489899, 0.04366486147046089, 0.11466746032238007, 0.14547954499721527, 0.03547332435846329, 0.12752412259578705, -0.05534541979432106, -0.27006328105926514, 0.006268559489399195, 0.04371552914381027, -0.04095698148012161, 0.13530462980270386, 0.08237200975418091, -0.11423757672309875, 0.07748546451330185, 0.010784623213112354, -0.16411936283111572, 0.01406471710652113, 0.003399496665224433, -0.06483443826436996, 0.1398109495639801, 0.04237164929509163, 0.10752451419830322, 0.018908314406871796, 0.06679096817970276, -0.19446660578250885, 0.0109678590670228, 0.06959249824285507, -0.0018445359310135245, 0.08894182741641998, 0.06829437613487244, 0.003577023046091199, 0.16443464159965515, -0.07851828634738922, 0.05713105946779251, 0.026997538283467293, -0.11338674277067184, -0.2119010090827942, -0.08342599868774414, 0.04453987255692482, 0.07187700271606445, 0.10639701038599014, -0.011891402304172516, 0.12583711743354797, -0.05546113848686218, 0.11122795194387436, 0.25110960006713867, -0.2940060794353485, -0.06238178908824921, -0.010530545376241207, 0.03900611028075218, 0.07814677804708481, -0.08047302067279816, -0.028269629925489426, 0.03524890914559364, 0.05202828347682953, 0.1354394406080246, -0.019140755757689476, -0.09542679041624069, -0.0005180437583476305, -0.1471855491399765, -0.052294254302978516, 0.1346442550420761, 0.0307223629206419, -0.026553509756922722, -0.06383319944143295, -0.08567296713590622, -0.18611612915992737, -0.037392325699329376, -0.015058370307087898, 0.0420556515455246, -0.022022457793354988, -0.056070391088724136, -0.03214196860790253, -0.10048644244670868, -0.055914439260959625, -0.06864920258522034, 0.11180728673934937, 0.05407477170228958, 0.0006952894036658108, -0.045621998608112335, 0.09688213467597961, -0.0035035377368330956, -0.13610920310020447, 0.013030585832893848, 0.02864496409893036, 0.004643809050321579, -0.028874831274151802, -0.05769568309187889, -0.09659367799758911, 0.010559597052633762, 0.13373468816280365, -0.07725301384925842, 0.05615345761179924, -0.015284072607755661, 0.04096401855349541, -0.10524129122495651, 0.1655028611421585, -0.036691274493932724, -0.005090360064059496, 0.0152931809425354, 0.056691497564315796, 0.03659380227327347, -0.024689797312021255, -0.1106666848063469, 0.009564433246850967, 0.1051572859287262, 0.02166169136762619, -0.04937496781349182, 0.07525641471147537, -0.037675052881240845, -0.021132027730345726, -0.00619869539514184, -0.10757511854171753, 0.023151380941271782, 0.0007577617070637643, -0.06451638787984848, 0.005180902313441038, 0.04195364937186241, -0.0016261404380202293, -0.05820774286985397, 0.10298141092061996, -0.07695844769477844, 0.017937934026122093, -0.09426338225603104, -0.131855309009552, 0.03295193985104561, -0.07928472012281418, -0.0003243583778385073, -0.09992272406816483, -0.1592864692211151, -0.012299121357500553, 0.050725724548101425, -0.03181161731481552, -0.05502237007021904, -0.053982075303792953, -0.0851273164153099, 0.032034195959568024, -0.02382550947368145, 0.1275702714920044, -0.06002156063914299, 0.09820174425840378, 0.02618243917822838, 0.05881549045443535, -0.031452473253011703, 0.0585092157125473, -0.08886926621198654, 0.017082594335079193, -0.16854308545589447, 0.04778769612312317, -0.038849614560604095, 0.05416354537010193, -0.09745929390192032, -0.10490196198225021, -0.008448571898043156, -0.0028444514609873295, 0.09182807058095932, 0.08716139942407608, -0.16208472847938538, -0.08082988113164902, 0.1869829148054123, -0.07983431220054626, -0.11703549325466156, 0.1313040405511856, -0.049363113939762115, 0.016233744099736214, 0.051688142120838165, 0.18089652061462402, 0.06451244652271271, -0.09654229879379272, 0.012901064939796925, -0.013497622683644295, 0.05176795274019241, -0.038879767060279846, 0.06196818873286247, -0.00548528553918004, 0.028259962797164917, 0.015097291208803654, -0.0006526846555061638, 0.05387027934193611, -0.08495680242776871, -0.08202605694532394, -0.052987437695264816, -0.06848649680614471, 0.021879900246858597, 0.0572112612426281, 0.06443130224943161, -0.10825537890195847, -0.1086103618144989, 0.0562802329659462, 0.0727447047829628, -0.08792860060930252, 0.05227958783507347, -0.067689910531044, 0.07822953164577484, -0.029315488412976265, -0.00040971985436044633, -0.18319328129291534, -0.026695091277360916, 0.02026013284921646, -0.019780587404966354, 0.029061948880553246, 0.0013278945116326213, 0.06571736186742783, 0.06184765696525574, -0.0491892546415329, -0.025172991678118706, -0.04403429850935936, -0.009468089789152145, -0.11788172274827957, -0.2014685720205307, -0.025572270154953003, -0.015570782124996185, 0.09911118447780609, -0.19183824956417084, 0.043172407895326614, 0.006429973524063826, 0.08800740540027618, 0.01662595197558403, -0.006448050960898399, -0.03452456369996071, 0.08168385177850723, -0.05559932813048363, -0.04861126095056534, 0.07598654180765152, 0.014693357981741428, -0.09227608144283295, -0.005764368921518326, -0.15114983916282654, 0.13481079041957855, 0.1349031925201416, -0.09844910353422165, -0.07131806761026382, 0.0019263345748186111, -0.06290996819734573, -0.038050465285778046, -0.03042401559650898, 0.00847144890576601, 0.1867779642343521, 0.0008264269563369453, 0.16217701137065887, -0.08554030954837799, -0.057506218552589417, 0.031518734991550446, -0.021643029525876045, 0.01932627148926258, 0.12811386585235596, 0.10083401203155518, -0.07101471722126007, 0.13798166811466217, 0.13670098781585693, -0.08034153282642365, 0.1501668393611908, -0.04592064768075943, -0.09172110259532928, -0.013009236194193363, -0.0003499274607747793, 0.010255217552185059, 0.07129430025815964, -0.1597914844751358, -0.0024991624522954226, 0.027955835685133934, 0.024384986609220505, 0.02751421555876732, -0.21544723212718964, -0.009807469323277473, 0.040275637060403824, -0.06126367300748825, -0.01219665352255106, -0.0038576938677579165, 0.01719529740512371, 0.11636614054441452, 0.004349178168922663, -0.06725382059812546, 0.024087026715278625, -0.0023639334831386805, -0.08714976906776428, 0.19712290167808533, -0.09088673442602158, -0.17832201719284058, -0.11985688656568527, -0.0727715715765953, -0.044163819402456284, -0.001775681390427053, 0.07676776498556137, -0.07832981646060944, -0.034010306000709534, -0.09042573720216751, 0.032673101872205734, -0.01596715860068798, 0.02971854992210865, 0.011373902671039104, 0.00005765316018369049, 0.06619273871183395, -0.11374592036008835, -0.015900399535894394, -0.040923312306404114, -0.050963252782821655, 0.04550565406680107, 0.03193211928009987, 0.11296466737985611, 0.15573342144489288, -0.02694494090974331, 0.018585816025733948, -0.0413634292781353, 0.19837069511413574, -0.06112223118543625, -0.01109871082007885, 0.14711245894432068, -0.011359009891748428, 0.0646899938583374, 0.11733540892601013, 0.05341801047325134, -0.07728442549705505, 0.014905186370015144, 0.0409904308617115, -0.03399444371461868, -0.24462155997753143, -0.03691887855529785, -0.06698370724916458, 0.015516904182732105, 0.09392675012350082, 0.029529696330428123, 0.050052352249622345, 0.048251982778310776, 0.015280768275260925, 0.07179639488458633, -0.013814661651849747, 0.08132308721542358, 0.15918441116809845, 0.04225151613354683, 0.13460592925548553, -0.051051780581474304, -0.052015382796525955, 0.049338337033987045, -0.009570916183292866, 0.21982136368751526, -0.002154654124751687, 0.16932719945907593, 0.05807902663946152, 0.15438787639141083, 0.007616914343088865, 0.0743485614657402, -0.012367313727736473, -0.019787879660725594, -0.01157250627875328, -0.051077958196401596, -0.030345242470502853, 0.022071612998843193, -0.07789615541696548, 0.04886065050959587, -0.12414181977510452, 0.0043195029720664024, 0.04672384634613991, 0.2811325192451477, 0.033064089715480804, -0.31928494572639465, -0.09526072442531586, -0.0019406348001211882, -0.06411085277795792, -0.022756416350603104, 0.03210729360580444, 0.0992027074098587, -0.07551776617765427, 0.054377347230911255, -0.07991065084934235, 0.10294454544782639, -0.03725973144173622, 0.044787317514419556, 0.06894634664058685, 0.10142703354358673, 0.00565936416387558, 0.07388850301504135, -0.317076712846756, 0.2721509337425232, 0.003970673307776451, 0.06336425989866257, -0.07463841140270233, 0.017451444640755653, 0.027469024062156677, 0.030897585675120354, 0.06747712939977646, -0.024185657501220703, -0.06442418694496155, -0.15417131781578064, -0.0708034411072731, 0.02198013663291931, 0.10074534267187119, -0.012309446930885315, 0.11437331885099411, -0.04243314638733864, 0.0015587025554850698, 0.06967709213495255, -0.001816388452425599, -0.06905435025691986, -0.107118621468544, 0.012800514698028564, 0.03152332827448845, -0.04090585559606552, -0.06599149107933044, -0.1115439161658287, -0.09727595001459122, 0.16968639194965363, -0.034095387905836105, -0.04554363712668419, -0.11179470270872116, 0.07620351761579514, 0.07603275775909424, -0.08771482855081558, 0.041208416223526, 0.0011836019111797214, 0.0867062658071518, 0.02028629556298256, -0.0919828936457634, 0.11604481935501099, -0.06488057225942612, -0.178857684135437, -0.052733030170202255, 0.12812520563602448, 0.015450065024197102, 0.06481190770864487, -0.023927276954054832, 0.014400248415768147, -0.03586616739630699, -0.08066627383232117, 0.01280206348747015, 0.008862454444169998, 0.07081472128629684, 0.0009371654014103115, -0.06937971711158752, 0.002123666927218437, -0.06266848742961884, -0.0390707328915596, 0.19336336851119995, 0.2217472493648529, -0.08751484006643295, 0.046471863985061646, 0.031407617032527924, -0.07661525905132294, -0.1804649829864502, 0.011302393861114979, 0.06078340485692024, 0.004391801543533802, 0.0186516884714365, -0.18716906011104584, 0.07209649682044983, 0.0976984053850174, -0.003250669687986374, 0.09224601089954376, -0.34940195083618164, -0.13862144947052002, 0.10465477406978607, 0.1389959752559662, 0.0879070982336998, -0.15785366296768188, -0.02520362287759781, -0.01454087719321251, -0.1104147881269455, 0.13120298087596893, -0.09409182518720627, 0.12222310155630112, -0.028405671939253807, 0.09469716250896454, 0.010953881777822971, -0.05430290848016739, 0.09253909438848495, -0.01591566577553749, 0.07530958205461502, -0.0686631128191948, 0.015330145135521889, 0.05687771737575531, -0.047380607575178146, 0.03163624182343483, -0.09741169959306717, 0.031077176332473755, -0.0888667106628418, -0.028745243325829506, -0.07290823012590408, 0.026976455003023148, -0.0375811830163002, -0.05688906088471413, -0.03807510808110237, 0.003683975897729397, 0.07402101904153824, -0.015751710161566734, 0.15194736421108246, 0.00024299607321154326, 0.1600034087896347, 0.11890453845262527, 0.09521564841270447, -0.05958735570311546, -0.05743212625384331, -0.006683378480374813, -0.02039499208331108, 0.05214744061231613, -0.1445227414369583, 0.02554505318403244, 0.15122447907924652, 0.014207675121724606, 0.13785220682621002, 0.08036656677722931, -0.0441175140440464, 0.008824958465993404, 0.05637658014893532, -0.1657281368970871, -0.1166725605726242, -0.013719539158046246, -0.0265200175344944, -0.10901714861392975, 0.04098965600132942, 0.11308068037033081, -0.07383134216070175, -0.006775813177227974, 0.0048089041374623775, 0.013856743462383747, -0.037588685750961304, 0.17984139919281006, 0.029517831280827522, 0.04664522036910057, -0.0938263013958931, 0.08207742869853973, 0.04805797338485718, -0.12336578220129013, 0.03551991283893585, 0.12367785722017288, -0.07805367559194565, -0.0412609688937664, 0.06370855122804642, 0.1596069186925888, -0.053466495126485825, -0.05010019242763519, -0.13781219720840454, -0.1259523630142212, 0.10536480695009232, 0.1849544495344162, 0.0730019137263298, 0.012445244938135147, -0.05698233097791672, 0.018529828637838364, -0.1192287802696228, 0.10663890838623047, 0.042442355304956436, 0.06186125800013542, -0.12253224104642868, 0.15938889980316162, 0.011982965283095837, 0.03284047171473503, -0.020286934450268745, 0.028225207701325417, -0.09365732222795486, 0.004719262011349201, -0.13101078569889069, -0.011329809203743935, -0.029805008322000504, -0.002207574201747775, -0.008025893941521645, -0.039233602583408356, -0.06596103310585022, 0.018577856943011284, -0.10683992505073547, -0.0332418754696846, 0.008546729572117329, 0.052161481231451035, -0.11427126824855804, -0.0236129742115736, 0.013924947939813137, -0.07580752670764923, 0.08239991962909698, 0.04676820710301399, 0.0007834918214939535, 0.050252243876457214, -0.12124282866716385, 0.0160363782197237, 0.060466598719358444, 0.023202048614621162, 0.04757588729262352, -0.0938132181763649, -0.0038980136159807444, 0.001960037974640727, 0.030710404738783836, 0.012896386906504631, 0.07152638584375381, -0.1342019885778427, -0.0021948199719190598, -0.015553135424852371, -0.07139328867197037, -0.06982050091028214, 0.039244409650564194, 0.05780696123838425, 0.019756343215703964, 0.17778824269771576, -0.09133607894182205, 0.04732578992843628, -0.22043386101722717, 0.007376316003501415, 0.003592558903619647, -0.10848762094974518, -0.08558480441570282, -0.06354396790266037, 0.06657565385103226, -0.06249623000621796, 0.1238265410065651, 0.012564717791974545, 0.04685128480195999, 0.03990457206964493, -0.03994515910744667, -0.004111339338123798, 0.016321422532200813, 0.21000239253044128, 0.02498401515185833, -0.04485245794057846, 0.04260806739330292, 0.02145540714263916, 0.0970880538225174, 0.1239950880408287, 0.21669401228427887, 0.1441294103860855, 0.013526647351682186, 0.09898082911968231, 0.03593965992331505, -0.06104731559753418, -0.16846713423728943, 0.05571479722857475, -0.036268044263124466, 0.13944850862026215, -0.024710601195693016, 0.23139667510986328, 0.10478629916906357, -0.15444132685661316, 0.05611685290932655, -0.03912433236837387, -0.0757886990904808, -0.12081985175609589, -0.0795322135090828, -0.08269605040550232, -0.14490585029125214, -0.005862410645931959, -0.12498612701892853, 0.050618261098861694, 0.07064805179834366, 0.025601714849472046, -0.018688833341002464, 0.141510471701622, 0.033637676388025284, -0.001853355672210455, 0.06893128156661987, 0.012806639075279236, -0.015005754306912422, -0.09326738864183426, -0.07781705260276794, 0.009869667701423168, -0.0145048713311553, 0.03749828413128853, -0.029682129621505737, -0.033939871937036514, 0.04095948114991188, -0.02899901196360588, -0.09470562636852264, 0.018235784024000168, 0.019794195890426636, 0.06964737176895142, 0.06957338750362396, 0.012003181502223015, -0.005165791604667902, -0.012233116663992405, 0.21462875604629517, -0.08277639001607895, -0.07864833623170853, -0.0962117612361908, 0.24701710045337677, 0.03354884311556816, -0.023462366312742233, 0.03413587808609009, -0.05914740264415741, -0.016909543424844742, 0.24458587169647217, 0.18621590733528137, -0.05043504759669304, -0.009927776642143726, 0.010951543226838112, -0.004921457264572382, -0.017238618806004524, 0.11052000522613525, 0.14688393473625183, 0.07711103558540344, -0.07585299760103226, -0.043082062155008316, -0.04625839740037918, 0.0014172104420140386, -0.050019312649965286, 0.08392122387886047, 0.03463425114750862, -0.008173596113920212, -0.017196059226989746, 0.04761171340942383, -0.06158741936087608, -0.08485296368598938, 0.013713723048567772, -0.2120974063873291, -0.15565037727355957, -0.013451878912746906, 0.1100158840417862, -0.0028883814811706543, 0.052325475960969925, -0.014684191904962063, -0.00044790736865252256, 0.09290171414613724, -0.018471060320734978, -0.0878685787320137, -0.05778178572654724, 0.08202111721038818, -0.136373832821846, 0.1854904294013977, -0.0358562134206295, 0.04023417830467224, 0.13094095885753632, 0.0689321979880333, -0.0757412537932396, 0.07597552984952927, 0.048465341329574585, -0.06701251119375229, 0.0337127149105072, 0.11274240911006927, -0.031843896955251694, 0.06079224497079849, 0.050436656922101974, -0.13376568257808685, 0.01459560077637434, -0.07657342404127121, -0.05887550115585327, -0.028058242052793503, -0.03225749731063843, -0.05919809266924858, 0.12936878204345703, 0.22054564952850342, -0.03515980765223503, 0.0019767635967582464, -0.0786546990275383, -0.004331674892455339, 0.04922102019190788, 0.050458721816539764, -0.04007605090737343, -0.2270025610923767, 0.008689581416547298, 0.056300751864910126, -0.005901097785681486, -0.26316550374031067, -0.0928284302353859, 0.00967988558113575, -0.05751662328839302, -0.13092462718486786, 0.08820579946041107, 0.09286501258611679, 0.04819212108850479, -0.04950564727187157, -0.057574011385440826, -0.060785893350839615, 0.16556264460086823, -0.14724871516227722, -0.07210414856672287 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-small-mlm-pubmed-15 This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.5389 - Rouge2 Precision: 0.7165 - Rouge2 Recall: 0.5375 - Rouge2 Fmeasure: 0.5981 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 40 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure | |:-------------:|:-----:|:-----:|:---------------:|:----------------:|:-------------:|:---------------:| | 1.1024 | 0.75 | 500 | 0.7890 | 0.6854 | 0.4813 | 0.5502 | | 0.8788 | 1.51 | 1000 | 0.7176 | 0.6906 | 0.4989 | 0.5638 | | 0.8086 | 2.26 | 1500 | 0.6830 | 0.6872 | 0.5052 | 0.5663 | | 0.7818 | 3.02 | 2000 | 0.6650 | 0.6912 | 0.5104 | 0.5711 | | 0.7466 | 3.77 | 2500 | 0.6458 | 0.6965 | 0.5167 | 0.5774 | | 0.731 | 4.52 | 3000 | 0.6355 | 0.6955 | 0.5161 | 0.5763 | | 0.7126 | 5.28 | 3500 | 0.6249 | 0.6924 | 0.517 | 0.576 | | 0.6998 | 6.03 | 4000 | 0.6166 | 0.6995 | 0.5207 | 0.5809 | | 0.6855 | 6.79 | 4500 | 0.6076 | 0.6981 | 0.5215 | 0.5813 | | 0.676 | 7.54 | 5000 | 0.6015 | 0.7003 | 0.5242 | 0.5836 | | 0.6688 | 8.3 | 5500 | 0.5962 | 0.7004 | 0.5235 | 0.583 | | 0.6569 | 9.05 | 6000 | 0.5900 | 0.6997 | 0.5234 | 0.5827 | | 0.6503 | 9.8 | 6500 | 0.5880 | 0.703 | 0.5257 | 0.5856 | | 0.6455 | 10.56 | 7000 | 0.5818 | 0.7008 | 0.5259 | 0.5849 | | 0.635 | 11.31 | 7500 | 0.5796 | 0.7017 | 0.5271 | 0.5861 | | 0.6323 | 12.07 | 8000 | 0.5769 | 0.7053 | 0.5276 | 0.5877 | | 0.6241 | 12.82 | 8500 | 0.5730 | 0.7011 | 0.5243 | 0.5838 | | 0.6224 | 13.57 | 9000 | 0.5696 | 0.7046 | 0.5286 | 0.5879 | | 0.6139 | 14.33 | 9500 | 0.5685 | 0.7047 | 0.5295 | 0.5886 | | 0.6118 | 15.08 | 10000 | 0.5653 | 0.704 | 0.5297 | 0.5886 | | 0.6089 | 15.84 | 10500 | 0.5633 | 0.703 | 0.5272 | 0.5865 | | 0.598 | 16.59 | 11000 | 0.5613 | 0.7059 | 0.5293 | 0.5889 | | 0.6003 | 17.35 | 11500 | 0.5602 | 0.7085 | 0.532 | 0.5918 | | 0.5981 | 18.1 | 12000 | 0.5587 | 0.7106 | 0.5339 | 0.5938 | | 0.5919 | 18.85 | 12500 | 0.5556 | 0.708 | 0.5319 | 0.5914 | | 0.5897 | 19.61 | 13000 | 0.5556 | 0.7106 | 0.5327 | 0.5931 | | 0.5899 | 20.36 | 13500 | 0.5526 | 0.7114 | 0.534 | 0.5939 | | 0.5804 | 21.12 | 14000 | 0.5521 | 0.7105 | 0.5328 | 0.5928 | | 0.5764 | 21.87 | 14500 | 0.5520 | 0.715 | 0.537 | 0.5976 | | 0.5793 | 22.62 | 15000 | 0.5506 | 0.713 | 0.5346 | 0.5951 | | 0.5796 | 23.38 | 15500 | 0.5492 | 0.7124 | 0.5352 | 0.5952 | | 0.5672 | 24.13 | 16000 | 0.5482 | 0.7124 | 0.5346 | 0.5948 | | 0.5737 | 24.89 | 16500 | 0.5470 | 0.7134 | 0.5352 | 0.5956 | | 0.5685 | 25.64 | 17000 | 0.5463 | 0.7117 | 0.5346 | 0.5946 | | 0.5658 | 26.4 | 17500 | 0.5457 | 0.7145 | 0.5359 | 0.5965 | | 0.5657 | 27.15 | 18000 | 0.5447 | 0.7145 | 0.5367 | 0.597 | | 0.5645 | 27.9 | 18500 | 0.5441 | 0.7141 | 0.5362 | 0.5964 | | 0.565 | 28.66 | 19000 | 0.5436 | 0.7151 | 0.5367 | 0.5972 | | 0.5579 | 29.41 | 19500 | 0.5426 | 0.7162 | 0.5378 | 0.5982 | | 0.563 | 30.17 | 20000 | 0.5424 | 0.7155 | 0.5373 | 0.5977 | | 0.556 | 30.92 | 20500 | 0.5418 | 0.7148 | 0.536 | 0.5966 | | 0.5576 | 31.67 | 21000 | 0.5411 | 0.7141 | 0.5356 | 0.5961 | | 0.5546 | 32.43 | 21500 | 0.5409 | 0.7149 | 0.5364 | 0.5967 | | 0.556 | 33.18 | 22000 | 0.5405 | 0.7143 | 0.5356 | 0.596 | | 0.5536 | 33.94 | 22500 | 0.5401 | 0.7165 | 0.5377 | 0.5982 | | 0.5527 | 34.69 | 23000 | 0.5397 | 0.7188 | 0.5389 | 0.5999 | | 0.5531 | 35.44 | 23500 | 0.5395 | 0.7172 | 0.538 | 0.5989 | | 0.5508 | 36.2 | 24000 | 0.5392 | 0.7166 | 0.538 | 0.5985 | | 0.5495 | 36.95 | 24500 | 0.5391 | 0.7176 | 0.5387 | 0.5993 | | 0.5539 | 37.71 | 25000 | 0.5391 | 0.7169 | 0.5372 | 0.598 | | 0.5452 | 38.46 | 25500 | 0.5390 | 0.7179 | 0.5384 | 0.5991 | | 0.5513 | 39.22 | 26000 | 0.5390 | 0.717 | 0.5377 | 0.5984 | | 0.5506 | 39.97 | 26500 | 0.5389 | 0.7165 | 0.5375 | 0.5981 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.10.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "t5-small-mlm-pubmed-15", "results": []}]}
text2text-generation
gayanin/t5-small-mlm-pubmed-15
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
t5-small-mlm-pubmed-15 ====================== This model is a fine-tuned version of t5-small on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 0.5389 * Rouge2 Precision: 0.7165 * Rouge2 Recall: 0.5375 * Rouge2 Fmeasure: 0.5981 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 40 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.12.5 * Pytorch 1.10.0+cu111 * Datasets 1.15.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ 67, 113, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ -0.08290655165910721, 0.0561562180519104, -0.0033341292291879654, 0.0974818542599678, 0.13760358095169067, 0.01766536757349968, 0.13184024393558502, 0.1418495625257492, -0.10419638454914093, 0.04303500056266785, 0.11713067442178726, 0.13870346546173096, 0.03724949434399605, 0.11987947672605515, -0.0533820204436779, -0.2704218029975891, 0.004514544736593962, 0.042675621807575226, -0.04484877735376358, 0.138296976685524, 0.0836346298456192, -0.11288135498762131, 0.08066939562559128, 0.008555017411708832, -0.16280622780323029, 0.012109514325857162, 0.004350124858319759, -0.059513889253139496, 0.14368391036987305, 0.0429895743727684, 0.11031325161457062, 0.017344066873192787, 0.0726812407374382, -0.19599047303199768, 0.0101163974031806, 0.06806007027626038, 0.00038959956145845354, 0.0930684432387352, 0.06934904307126999, 0.007495777681469917, 0.15162503719329834, -0.07378671318292618, 0.05114918202161789, 0.026230188086628914, -0.11362802982330322, -0.206829234957695, -0.08438733965158463, 0.04544178768992424, 0.0728008821606636, 0.1082300916314125, -0.012888996861875057, 0.11570046842098236, -0.05872626230120659, 0.1074431985616684, 0.24895934760570526, -0.2966533601284027, -0.0611410066485405, -0.0022558283526450396, 0.037833523005247116, 0.07777697592973709, -0.08357521146535873, -0.027721943333745003, 0.03477223217487335, 0.051343031227588654, 0.13402600586414337, -0.01978500932455063, -0.09307581931352615, 0.003447385737672448, -0.14764487743377686, -0.05139296129345894, 0.13220742344856262, 0.03533041104674339, -0.0252921674400568, -0.05920717865228653, -0.08439136296510696, -0.1922561228275299, -0.03614458814263344, -0.013815004378557205, 0.04315968602895737, -0.01912953145802021, -0.05469510704278946, -0.0314909927546978, -0.10494361072778702, -0.06342916190624237, -0.07069374620914459, 0.11002932488918304, 0.05088857561349869, 0.0005525643937289715, -0.039969056844711304, 0.10632190108299255, -0.008618105202913284, -0.13256314396858215, 0.015247875824570656, 0.031929828226566315, -0.0001388030214002356, -0.02761031873524189, -0.05723525583744049, -0.09524934738874435, 0.00600000936537981, 0.1342083364725113, -0.07112326472997665, 0.05284997075796127, -0.006939806509763002, 0.045952942222356796, -0.1070401668548584, 0.16755802929401398, -0.0486505962908268, -0.010803737677633762, 0.01158976461738348, 0.05981434881687164, 0.03288842737674713, -0.020392410457134247, -0.11220861226320267, 0.00010222693526884541, 0.10959841310977936, 0.0202386025339365, -0.04950649291276932, 0.0738593116402626, -0.04075304791331291, -0.021159959957003593, -0.005638741888105869, -0.10210371762514114, 0.021969862282276154, 0.0007769876392558217, -0.06896360963582993, 0.005692490842193365, 0.04732730612158775, -0.000047709978389320895, -0.052339911460876465, 0.10835372656583786, -0.08030639588832855, 0.020474359393119812, -0.09411584585905075, -0.12684805691242218, 0.031222136691212654, -0.06927908211946487, -0.001445225439965725, -0.09956303238868713, -0.16846881806850433, -0.014985563233494759, 0.050374485552310944, -0.0367213599383831, -0.05347895249724388, -0.053169094026088715, -0.08290962129831314, 0.03254782408475876, -0.026594385504722595, 0.1353824883699417, -0.060081757605075836, 0.10388627648353577, 0.02719009481370449, 0.05812229588627815, -0.035124484449625015, 0.06163717433810234, -0.08869641274213791, 0.017949244007468224, -0.16754627227783203, 0.04582368582487106, -0.037347063422203064, 0.05102720111608505, -0.10030282288789749, -0.10429289937019348, -0.014025570824742317, -0.0014420845545828342, 0.09350264817476273, 0.09072556346654892, -0.1596563160419464, -0.08030329644680023, 0.1847245842218399, -0.075993113219738, -0.11475680023431778, 0.1324533075094223, -0.04795621708035469, 0.023776903748512268, 0.050072357058525085, 0.17126426100730896, 0.061164312064647675, -0.09448344260454178, 0.00980286207050085, -0.011430212296545506, 0.05237822234630585, -0.043501004576683044, 0.06699581444263458, -0.005940420087426901, 0.024207424372434616, 0.015904854983091354, -0.00988923478871584, 0.0590413399040699, -0.08703697472810745, -0.08166613429784775, -0.05124528706073761, -0.07000484317541122, 0.02541288360953331, 0.057920899242162704, 0.06738544255495071, -0.10146501660346985, -0.10995489358901978, 0.055052727460861206, 0.07795330137014389, -0.08591156452894211, 0.049616727977991104, -0.06362432986497879, 0.07748861610889435, -0.03717665746808052, -0.004858267959207296, -0.1842920184135437, -0.02507300302386284, 0.016461417078971863, -0.01565955951809883, 0.03205730393528938, 0.010646531358361244, 0.06852484494447708, 0.05945669859647751, -0.0539558045566082, -0.02887731045484543, -0.04653764143586159, -0.011250624433159828, -0.12204498797655106, -0.19815650582313538, -0.028808830305933952, -0.01213879231363535, 0.10921090841293335, -0.19918854534626007, 0.040859926491975784, -0.0005712375859729946, 0.08669878542423248, 0.015414310619235039, -0.005056044086813927, -0.03622424229979515, 0.07871338725090027, -0.054688118398189545, -0.046889789402484894, 0.07630103826522827, 0.011968146078288555, -0.09326254576444626, -0.0067756446078419685, -0.14590005576610565, 0.13550059497356415, 0.13269974291324615, -0.10939951986074448, -0.07498525828123093, -0.00014988928160164505, -0.06308727711439133, -0.044364847242832184, -0.03433700278401375, 0.002084161154925823, 0.1847485452890396, 0.004152368754148483, 0.1612563580274582, -0.08164374530315399, -0.05449668690562248, 0.027899647131562233, -0.022480769082903862, 0.02206249348819256, 0.1271718144416809, 0.11015284061431885, -0.07089956104755402, 0.14137734472751617, 0.13799802958965302, -0.08419525623321533, 0.1481219232082367, -0.041925687342882156, -0.0960870310664177, -0.015579030849039555, 0.0015935116680338979, 0.006605233997106552, 0.0730024203658104, -0.15685798227787018, -0.0012449485948309302, 0.028459347784519196, 0.025288591161370277, 0.030444685369729996, -0.21850360929965973, -0.01255981158465147, 0.037460025399923325, -0.06330183893442154, -0.00524491211399436, -0.0028231998439878225, 0.016079260036349297, 0.11657550185918808, 0.006073521915823221, -0.07516241073608398, 0.02481776848435402, -0.002794104628264904, -0.08657848089933395, 0.19870367646217346, -0.0845867171883583, -0.18074966967105865, -0.12891380488872528, -0.06726356595754623, -0.043366700410842896, -0.002374825067818165, 0.06971656531095505, -0.07397567480802536, -0.031976085156202316, -0.08561195433139801, 0.03829355537891388, -0.014810682274401188, 0.031260158866643906, 0.002424344653263688, 0.0011621731100603938, 0.06859128177165985, -0.10968838632106781, -0.013017759658396244, -0.04131608456373215, -0.05467443913221359, 0.044275809079408646, 0.032206129282712936, 0.10946238040924072, 0.15918144583702087, -0.02134254388511181, 0.014209355227649212, -0.03856508433818817, 0.19614090025424957, -0.0616222582757473, -0.010480225086212158, 0.14962492883205414, -0.010375459678471088, 0.06336773186922073, 0.11880208551883698, 0.05551348254084587, -0.07656800001859665, 0.015353626571595669, 0.04298843443393707, -0.02995165064930916, -0.24456535279750824, -0.03745952621102333, -0.06430353224277496, 0.01859467849135399, 0.09515346586704254, 0.028852233663201332, 0.049921270459890366, 0.05071214586496353, 0.01629592664539814, 0.0726606696844101, -0.01879890076816082, 0.07679726928472519, 0.15386727452278137, 0.03886941820383072, 0.1346341073513031, -0.04746706783771515, -0.05455458164215088, 0.05045757070183754, -0.004078770987689495, 0.21942612528800964, -0.0005766464746557176, 0.17156188189983368, 0.05442199483513832, 0.15693192183971405, 0.004887423012405634, 0.07499101012945175, -0.009609416127204895, -0.022819530218839645, -0.01337562594562769, -0.05020713061094284, -0.03291146084666252, 0.02542036585509777, -0.07104518264532089, 0.04833095148205757, -0.12325827032327652, -0.00188556092325598, 0.04320943355560303, 0.2802351415157318, 0.0405685119330883, -0.3180699050426483, -0.09586703777313232, -0.001863937359303236, -0.06601335108280182, -0.025169864296913147, 0.03086915798485279, 0.10135548561811447, -0.07850915938615799, 0.04553009569644928, -0.08417942374944687, 0.1017647311091423, -0.04550463706254959, 0.04261774942278862, 0.06964059174060822, 0.0973525196313858, 0.006177486386150122, 0.07683391869068146, -0.3115845322608948, 0.2770364582538605, 0.003649738384410739, 0.05838273838162422, -0.07231909036636353, 0.015402166172862053, 0.029973959550261497, 0.03283792361617088, 0.06709308922290802, -0.0236468818038702, -0.053502075374126434, -0.15808293223381042, -0.0766480416059494, 0.017629850655794144, 0.09788084030151367, -0.022138727828860283, 0.11446476727724075, -0.0450059249997139, 0.0014125618617981672, 0.07177630811929703, -0.0017192758386954665, -0.0657576322555542, -0.09894618391990662, 0.01636740006506443, 0.03453396260738373, -0.03535209968686104, -0.06613287329673767, -0.11288671940565109, -0.099558025598526, 0.16747832298278809, -0.039388880133628845, -0.041787441819906235, -0.11062997579574585, 0.08566928654909134, 0.08272646367549896, -0.08951807022094727, 0.039972852915525436, 0.0027922180015593767, 0.08221258223056793, 0.024498604238033295, -0.08637084066867828, 0.11773484200239182, -0.06372465193271637, -0.18094107508659363, -0.05394289642572403, 0.1287798285484314, 0.01562205795198679, 0.06453610211610794, -0.027446415275335312, 0.011150774545967579, -0.03938901796936989, -0.08023043721914291, 0.013192030601203442, 0.003363831667229533, 0.0681382343173027, 0.012272682040929794, -0.06686056405305862, 0.005658120382577181, -0.05888858437538147, -0.04186416044831276, 0.19300760328769684, 0.22520455718040466, -0.08679668605327606, 0.035767678171396255, 0.033091478049755096, -0.07429971545934677, -0.1849990040063858, 0.012744343839585781, 0.06097197160124779, 0.007211062591522932, 0.024243300780653954, -0.18836943805217743, 0.07346469908952713, 0.10004008561372757, -0.00837626587599516, 0.09914879500865936, -0.34950441122055054, -0.1374751478433609, 0.10743189603090286, 0.13604730367660522, 0.09628299623727798, -0.15808777511119843, -0.028050824999809265, -0.012906467542052269, -0.1121537983417511, 0.12720459699630737, -0.08926551789045334, 0.1276235580444336, -0.030113140121102333, 0.10204938799142838, 0.011549760587513447, -0.05622809752821922, 0.0967482402920723, -0.014391757547855377, 0.07529748231172562, -0.06632299721240997, 0.010276787914335728, 0.04653660207986832, -0.04602847620844841, 0.033551398664712906, -0.09545015543699265, 0.028441082686185837, -0.09289832413196564, -0.028064562007784843, -0.07354013621807098, 0.02864394150674343, -0.03696542978286743, -0.05775268003344536, -0.03842325136065483, 0.010819452814757824, 0.06380131840705872, -0.012068573385477066, 0.14870263636112213, -0.0005498268292285502, 0.14971645176410675, 0.12274178862571716, 0.08925654739141464, -0.049326732754707336, -0.06268030405044556, -0.011371415108442307, -0.020137067884206772, 0.05212213844060898, -0.14427445828914642, 0.023769235238432884, 0.15216124057769775, 0.012204814702272415, 0.1438562124967575, 0.07859069108963013, -0.04671945050358772, 0.0067559233866631985, 0.059334784746170044, -0.16533371806144714, -0.11403258144855499, -0.016219502314925194, -0.030757775530219078, -0.11229962855577469, 0.039819300174713135, 0.11831185221672058, -0.0757538452744484, -0.005467086099088192, 0.007986614480614662, 0.016132740303874016, -0.043064091354608536, 0.18508249521255493, 0.03280956670641899, 0.04500225931406021, -0.09153584390878677, 0.08438543975353241, 0.04874846339225769, -0.11189202964305878, 0.02673557586967945, 0.11879381537437439, -0.07513748109340668, -0.04230424761772156, 0.056355223059654236, 0.15814174711704254, -0.06083619222044945, -0.051546331495046616, -0.13888682425022125, -0.12150862067937851, 0.1011362224817276, 0.17383639514446259, 0.07477602362632751, 0.015202626585960388, -0.05597279593348503, 0.016755463555455208, -0.11632344126701355, 0.10590853542089462, 0.04650687798857689, 0.06333526223897934, -0.12795105576515198, 0.1621592491865158, 0.009975633583962917, 0.0341961532831192, -0.020248830318450928, 0.02993115596473217, -0.09240244328975677, 0.00516399834305048, -0.1327214539051056, -0.015830542892217636, -0.025055525824427605, -0.005159457679837942, -0.007224248256534338, -0.04342233017086983, -0.06602351367473602, 0.01573561504483223, -0.1079404205083847, -0.03205976262688637, 0.010696139186620712, 0.0527566596865654, -0.11295486986637115, -0.02667165920138359, 0.018598835915327072, -0.07231608033180237, 0.08585590124130249, 0.047582969069480896, 0.008423063904047012, 0.050962433218955994, -0.12281996011734009, 0.021969828754663467, 0.05819125473499298, 0.023421231657266617, 0.045176632702350616, -0.09510499984025955, -0.0053449166007339954, -0.00416087731719017, 0.036273691803216934, 0.012846813537180424, 0.0688384398818016, -0.13683339953422546, -0.005447857081890106, -0.02121860533952713, -0.06594370305538177, -0.06882797181606293, 0.03822512924671173, 0.05455278977751732, 0.025747045874595642, 0.1743641197681427, -0.09025660902261734, 0.04786162078380585, -0.22445723414421082, 0.013717317953705788, 0.0007456443272531033, -0.10603298246860504, -0.0822920948266983, -0.06746797263622284, 0.06580676883459091, -0.06137210875749588, 0.1194010004401207, 0.011731178499758244, 0.04491567984223366, 0.037451643496751785, -0.03913581371307373, 0.0003387695178389549, 0.015274430625140667, 0.20671822130680084, 0.028274817392230034, -0.04267042130231857, 0.03994293138384819, 0.027011286467313766, 0.09897870570421219, 0.1251721829175949, 0.21746008098125458, 0.1443902999162674, 0.00819371361285448, 0.10330741852521896, 0.03774707764387131, -0.06433349847793579, -0.16620855033397675, 0.05777188017964363, -0.038903187960386276, 0.13317857682704926, -0.027176300063729286, 0.23982073366641998, 0.09933306276798248, -0.15459823608398438, 0.053737152367830276, -0.040833480656147, -0.07642215490341187, -0.11874419450759888, -0.07482520490884781, -0.08116868138313293, -0.1470830738544464, -0.007038090378046036, -0.12125971913337708, 0.04796804487705231, 0.07687056809663773, 0.024951323866844177, -0.022664962336421013, 0.14206930994987488, 0.03099041059613228, -0.0015508835203945637, 0.0718890056014061, 0.01312553696334362, -0.010794972069561481, -0.1007872074842453, -0.07965445518493652, 0.008685879409313202, -0.009174277074635029, 0.03900023177266121, -0.03424133360385895, -0.038705188781023026, 0.04055759683251381, -0.022366680204868317, -0.09681612253189087, 0.012983053922653198, 0.020274659618735313, 0.07080096751451492, 0.07250792533159256, 0.0088870320469141, 0.0009708370780572295, -0.013526457361876965, 0.2187499850988388, -0.08535981923341751, -0.07158172130584717, -0.09586427360773087, 0.23818866908550262, 0.027675556018948555, -0.02125565893948078, 0.03135984390974045, -0.062381561845541, -0.015156809240579605, 0.248570516705513, 0.1799585074186325, -0.06191599741578102, -0.011147504672408104, 0.016132691875100136, -0.005703626200556755, -0.01867654174566269, 0.10559917241334915, 0.14206047356128693, 0.07876604050397873, -0.08216157555580139, -0.04418375715613365, -0.043920811265707016, -0.0011924505233764648, -0.04161557927727699, 0.08116697520017624, 0.03127100318670273, -0.0031100884079933167, -0.022759346291422844, 0.05014672502875328, -0.062130700796842575, -0.09346071630716324, 0.01612163707613945, -0.21261119842529297, -0.15939190983772278, -0.014445329084992409, 0.10830047726631165, -0.0029632048681378365, 0.052350714802742004, -0.02021626941859722, 0.0011673958506435156, 0.09261893481016159, -0.020598411560058594, -0.08719086647033691, -0.061430081725120544, 0.0800943523645401, -0.1197943463921547, 0.18815921247005463, -0.03740065544843674, 0.03992569446563721, 0.13078317046165466, 0.06729324907064438, -0.07798749953508377, 0.07431641966104507, 0.048254434019327164, -0.06731877475976944, 0.03658498823642731, 0.11344766616821289, -0.03134822100400925, 0.06376076489686966, 0.04725462943315506, -0.1335436850786209, 0.014687446877360344, -0.07266610860824585, -0.06185290217399597, -0.03238183259963989, -0.035577576607465744, -0.058154065161943436, 0.13258427381515503, 0.21921570599079132, -0.03687327727675438, 0.00048057673848234117, -0.07764729857444763, -0.004444161430001259, 0.048580095171928406, 0.055419228971004486, -0.038020554929971695, -0.22537297010421753, 0.009807797148823738, 0.0571817122399807, -0.0041002691723406315, -0.26522406935691833, -0.09097064286470413, 0.012601674534380436, -0.058848533779382706, -0.1294889599084854, 0.08956053853034973, 0.08972277492284775, 0.048941295593976974, -0.046956077218055725, -0.059232089668512344, -0.06484432518482208, 0.1634431779384613, -0.15104427933692932, -0.07681909203529358 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-small-mlm-pubmed-35 This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.1101 - Rouge2 Precision: 0.4758 - Rouge2 Recall: 0.3498 - Rouge2 Fmeasure: 0.3927 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 40 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure | |:-------------:|:-----:|:-----:|:---------------:|:----------------:|:-------------:|:---------------:| | 1.8404 | 0.75 | 500 | 1.5005 | 0.4265 | 0.2786 | 0.3273 | | 1.6858 | 1.51 | 1000 | 1.4216 | 0.4318 | 0.2946 | 0.3404 | | 1.6071 | 2.26 | 1500 | 1.3777 | 0.4472 | 0.3148 | 0.3598 | | 1.5551 | 3.02 | 2000 | 1.3360 | 0.4406 | 0.3168 | 0.3586 | | 1.5116 | 3.77 | 2500 | 1.3128 | 0.4523 | 0.3234 | 0.3671 | | 1.4837 | 4.52 | 3000 | 1.2937 | 0.4477 | 0.3215 | 0.3645 | | 1.4513 | 5.28 | 3500 | 1.2766 | 0.4511 | 0.3262 | 0.3689 | | 1.4336 | 6.03 | 4000 | 1.2626 | 0.4548 | 0.3283 | 0.3718 | | 1.4149 | 6.79 | 4500 | 1.2449 | 0.4495 | 0.3274 | 0.3687 | | 1.3977 | 7.54 | 5000 | 1.2349 | 0.4507 | 0.3305 | 0.3712 | | 1.3763 | 8.3 | 5500 | 1.2239 | 0.4519 | 0.3266 | 0.3688 | | 1.371 | 9.05 | 6000 | 1.2171 | 0.4546 | 0.3305 | 0.3727 | | 1.3501 | 9.8 | 6500 | 1.2080 | 0.4575 | 0.3329 | 0.3755 | | 1.3443 | 10.56 | 7000 | 1.2017 | 0.4576 | 0.3314 | 0.3742 | | 1.326 | 11.31 | 7500 | 1.1926 | 0.4578 | 0.333 | 0.3757 | | 1.3231 | 12.07 | 8000 | 1.1866 | 0.4606 | 0.3357 | 0.3782 | | 1.3089 | 12.82 | 8500 | 1.1816 | 0.4591 | 0.3338 | 0.3765 | | 1.3007 | 13.57 | 9000 | 1.1764 | 0.4589 | 0.3361 | 0.3777 | | 1.2943 | 14.33 | 9500 | 1.1717 | 0.4641 | 0.3382 | 0.3811 | | 1.2854 | 15.08 | 10000 | 1.1655 | 0.4617 | 0.3378 | 0.38 | | 1.2777 | 15.84 | 10500 | 1.1612 | 0.464 | 0.3401 | 0.3823 | | 1.2684 | 16.59 | 11000 | 1.1581 | 0.4608 | 0.3367 | 0.3789 | | 1.2612 | 17.35 | 11500 | 1.1554 | 0.4623 | 0.3402 | 0.3818 | | 1.2625 | 18.1 | 12000 | 1.1497 | 0.4613 | 0.3381 | 0.3802 | | 1.2529 | 18.85 | 12500 | 1.1465 | 0.4671 | 0.3419 | 0.3848 | | 1.2461 | 19.61 | 13000 | 1.1431 | 0.4646 | 0.3399 | 0.3824 | | 1.2415 | 20.36 | 13500 | 1.1419 | 0.4659 | 0.341 | 0.3835 | | 1.2375 | 21.12 | 14000 | 1.1377 | 0.4693 | 0.3447 | 0.3873 | | 1.2315 | 21.87 | 14500 | 1.1353 | 0.4672 | 0.3433 | 0.3855 | | 1.2263 | 22.62 | 15000 | 1.1333 | 0.467 | 0.3433 | 0.3854 | | 1.2214 | 23.38 | 15500 | 1.1305 | 0.4682 | 0.3446 | 0.3869 | | 1.2202 | 24.13 | 16000 | 1.1291 | 0.4703 | 0.3465 | 0.3888 | | 1.2155 | 24.89 | 16500 | 1.1270 | 0.472 | 0.348 | 0.3903 | | 1.2064 | 25.64 | 17000 | 1.1261 | 0.4724 | 0.3479 | 0.3905 | | 1.2173 | 26.4 | 17500 | 1.1236 | 0.4734 | 0.3485 | 0.3912 | | 1.1994 | 27.15 | 18000 | 1.1220 | 0.4739 | 0.3486 | 0.3915 | | 1.2018 | 27.9 | 18500 | 1.1217 | 0.4747 | 0.3489 | 0.3921 | | 1.2045 | 28.66 | 19000 | 1.1194 | 0.4735 | 0.3488 | 0.3916 | | 1.1949 | 29.41 | 19500 | 1.1182 | 0.4732 | 0.3484 | 0.3911 | | 1.19 | 30.17 | 20000 | 1.1166 | 0.4724 | 0.3479 | 0.3904 | | 1.1932 | 30.92 | 20500 | 1.1164 | 0.4753 | 0.3494 | 0.3924 | | 1.1952 | 31.67 | 21000 | 1.1147 | 0.4733 | 0.3485 | 0.3911 | | 1.1922 | 32.43 | 21500 | 1.1146 | 0.475 | 0.3494 | 0.3923 | | 1.1889 | 33.18 | 22000 | 1.1132 | 0.4765 | 0.3499 | 0.3933 | | 1.1836 | 33.94 | 22500 | 1.1131 | 0.4768 | 0.351 | 0.3939 | | 1.191 | 34.69 | 23000 | 1.1127 | 0.4755 | 0.3495 | 0.3926 | | 1.1811 | 35.44 | 23500 | 1.1113 | 0.4748 | 0.349 | 0.3919 | | 1.1864 | 36.2 | 24000 | 1.1107 | 0.4751 | 0.3494 | 0.3921 | | 1.1789 | 36.95 | 24500 | 1.1103 | 0.4756 | 0.3499 | 0.3927 | | 1.1819 | 37.71 | 25000 | 1.1101 | 0.4758 | 0.35 | 0.3932 | | 1.1862 | 38.46 | 25500 | 1.1099 | 0.4755 | 0.3497 | 0.3926 | | 1.1764 | 39.22 | 26000 | 1.1101 | 0.4759 | 0.3498 | 0.3928 | | 1.1819 | 39.97 | 26500 | 1.1101 | 0.4758 | 0.3498 | 0.3927 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.10.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "t5-small-mlm-pubmed-35", "results": []}]}
text2text-generation
gayanin/t5-small-mlm-pubmed-35
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
t5-small-mlm-pubmed-35 ====================== This model is a fine-tuned version of t5-small on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 1.1101 * Rouge2 Precision: 0.4758 * Rouge2 Recall: 0.3498 * Rouge2 Fmeasure: 0.3927 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 40 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.12.5 * Pytorch 1.10.0+cu111 * Datasets 1.15.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ 67, 113, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ -0.08290655165910721, 0.0561562180519104, -0.0033341292291879654, 0.0974818542599678, 0.13760358095169067, 0.01766536757349968, 0.13184024393558502, 0.1418495625257492, -0.10419638454914093, 0.04303500056266785, 0.11713067442178726, 0.13870346546173096, 0.03724949434399605, 0.11987947672605515, -0.0533820204436779, -0.2704218029975891, 0.004514544736593962, 0.042675621807575226, -0.04484877735376358, 0.138296976685524, 0.0836346298456192, -0.11288135498762131, 0.08066939562559128, 0.008555017411708832, -0.16280622780323029, 0.012109514325857162, 0.004350124858319759, -0.059513889253139496, 0.14368391036987305, 0.0429895743727684, 0.11031325161457062, 0.017344066873192787, 0.0726812407374382, -0.19599047303199768, 0.0101163974031806, 0.06806007027626038, 0.00038959956145845354, 0.0930684432387352, 0.06934904307126999, 0.007495777681469917, 0.15162503719329834, -0.07378671318292618, 0.05114918202161789, 0.026230188086628914, -0.11362802982330322, -0.206829234957695, -0.08438733965158463, 0.04544178768992424, 0.0728008821606636, 0.1082300916314125, -0.012888996861875057, 0.11570046842098236, -0.05872626230120659, 0.1074431985616684, 0.24895934760570526, -0.2966533601284027, -0.0611410066485405, -0.0022558283526450396, 0.037833523005247116, 0.07777697592973709, -0.08357521146535873, -0.027721943333745003, 0.03477223217487335, 0.051343031227588654, 0.13402600586414337, -0.01978500932455063, -0.09307581931352615, 0.003447385737672448, -0.14764487743377686, -0.05139296129345894, 0.13220742344856262, 0.03533041104674339, -0.0252921674400568, -0.05920717865228653, -0.08439136296510696, -0.1922561228275299, -0.03614458814263344, -0.013815004378557205, 0.04315968602895737, -0.01912953145802021, -0.05469510704278946, -0.0314909927546978, -0.10494361072778702, -0.06342916190624237, -0.07069374620914459, 0.11002932488918304, 0.05088857561349869, 0.0005525643937289715, -0.039969056844711304, 0.10632190108299255, -0.008618105202913284, -0.13256314396858215, 0.015247875824570656, 0.031929828226566315, -0.0001388030214002356, -0.02761031873524189, -0.05723525583744049, -0.09524934738874435, 0.00600000936537981, 0.1342083364725113, -0.07112326472997665, 0.05284997075796127, -0.006939806509763002, 0.045952942222356796, -0.1070401668548584, 0.16755802929401398, -0.0486505962908268, -0.010803737677633762, 0.01158976461738348, 0.05981434881687164, 0.03288842737674713, -0.020392410457134247, -0.11220861226320267, 0.00010222693526884541, 0.10959841310977936, 0.0202386025339365, -0.04950649291276932, 0.0738593116402626, -0.04075304791331291, -0.021159959957003593, -0.005638741888105869, -0.10210371762514114, 0.021969862282276154, 0.0007769876392558217, -0.06896360963582993, 0.005692490842193365, 0.04732730612158775, -0.000047709978389320895, -0.052339911460876465, 0.10835372656583786, -0.08030639588832855, 0.020474359393119812, -0.09411584585905075, -0.12684805691242218, 0.031222136691212654, -0.06927908211946487, -0.001445225439965725, -0.09956303238868713, -0.16846881806850433, -0.014985563233494759, 0.050374485552310944, -0.0367213599383831, -0.05347895249724388, -0.053169094026088715, -0.08290962129831314, 0.03254782408475876, -0.026594385504722595, 0.1353824883699417, -0.060081757605075836, 0.10388627648353577, 0.02719009481370449, 0.05812229588627815, -0.035124484449625015, 0.06163717433810234, -0.08869641274213791, 0.017949244007468224, -0.16754627227783203, 0.04582368582487106, -0.037347063422203064, 0.05102720111608505, -0.10030282288789749, -0.10429289937019348, -0.014025570824742317, -0.0014420845545828342, 0.09350264817476273, 0.09072556346654892, -0.1596563160419464, -0.08030329644680023, 0.1847245842218399, -0.075993113219738, -0.11475680023431778, 0.1324533075094223, -0.04795621708035469, 0.023776903748512268, 0.050072357058525085, 0.17126426100730896, 0.061164312064647675, -0.09448344260454178, 0.00980286207050085, -0.011430212296545506, 0.05237822234630585, -0.043501004576683044, 0.06699581444263458, -0.005940420087426901, 0.024207424372434616, 0.015904854983091354, -0.00988923478871584, 0.0590413399040699, -0.08703697472810745, -0.08166613429784775, -0.05124528706073761, -0.07000484317541122, 0.02541288360953331, 0.057920899242162704, 0.06738544255495071, -0.10146501660346985, -0.10995489358901978, 0.055052727460861206, 0.07795330137014389, -0.08591156452894211, 0.049616727977991104, -0.06362432986497879, 0.07748861610889435, -0.03717665746808052, -0.004858267959207296, -0.1842920184135437, -0.02507300302386284, 0.016461417078971863, -0.01565955951809883, 0.03205730393528938, 0.010646531358361244, 0.06852484494447708, 0.05945669859647751, -0.0539558045566082, -0.02887731045484543, -0.04653764143586159, -0.011250624433159828, -0.12204498797655106, -0.19815650582313538, -0.028808830305933952, -0.01213879231363535, 0.10921090841293335, -0.19918854534626007, 0.040859926491975784, -0.0005712375859729946, 0.08669878542423248, 0.015414310619235039, -0.005056044086813927, -0.03622424229979515, 0.07871338725090027, -0.054688118398189545, -0.046889789402484894, 0.07630103826522827, 0.011968146078288555, -0.09326254576444626, -0.0067756446078419685, -0.14590005576610565, 0.13550059497356415, 0.13269974291324615, -0.10939951986074448, -0.07498525828123093, -0.00014988928160164505, -0.06308727711439133, -0.044364847242832184, -0.03433700278401375, 0.002084161154925823, 0.1847485452890396, 0.004152368754148483, 0.1612563580274582, -0.08164374530315399, -0.05449668690562248, 0.027899647131562233, -0.022480769082903862, 0.02206249348819256, 0.1271718144416809, 0.11015284061431885, -0.07089956104755402, 0.14137734472751617, 0.13799802958965302, -0.08419525623321533, 0.1481219232082367, -0.041925687342882156, -0.0960870310664177, -0.015579030849039555, 0.0015935116680338979, 0.006605233997106552, 0.0730024203658104, -0.15685798227787018, -0.0012449485948309302, 0.028459347784519196, 0.025288591161370277, 0.030444685369729996, -0.21850360929965973, -0.01255981158465147, 0.037460025399923325, -0.06330183893442154, -0.00524491211399436, -0.0028231998439878225, 0.016079260036349297, 0.11657550185918808, 0.006073521915823221, -0.07516241073608398, 0.02481776848435402, -0.002794104628264904, -0.08657848089933395, 0.19870367646217346, -0.0845867171883583, -0.18074966967105865, -0.12891380488872528, -0.06726356595754623, -0.043366700410842896, -0.002374825067818165, 0.06971656531095505, -0.07397567480802536, -0.031976085156202316, -0.08561195433139801, 0.03829355537891388, -0.014810682274401188, 0.031260158866643906, 0.002424344653263688, 0.0011621731100603938, 0.06859128177165985, -0.10968838632106781, -0.013017759658396244, -0.04131608456373215, -0.05467443913221359, 0.044275809079408646, 0.032206129282712936, 0.10946238040924072, 0.15918144583702087, -0.02134254388511181, 0.014209355227649212, -0.03856508433818817, 0.19614090025424957, -0.0616222582757473, -0.010480225086212158, 0.14962492883205414, -0.010375459678471088, 0.06336773186922073, 0.11880208551883698, 0.05551348254084587, -0.07656800001859665, 0.015353626571595669, 0.04298843443393707, -0.02995165064930916, -0.24456535279750824, -0.03745952621102333, -0.06430353224277496, 0.01859467849135399, 0.09515346586704254, 0.028852233663201332, 0.049921270459890366, 0.05071214586496353, 0.01629592664539814, 0.0726606696844101, -0.01879890076816082, 0.07679726928472519, 0.15386727452278137, 0.03886941820383072, 0.1346341073513031, -0.04746706783771515, -0.05455458164215088, 0.05045757070183754, -0.004078770987689495, 0.21942612528800964, -0.0005766464746557176, 0.17156188189983368, 0.05442199483513832, 0.15693192183971405, 0.004887423012405634, 0.07499101012945175, -0.009609416127204895, -0.022819530218839645, -0.01337562594562769, -0.05020713061094284, -0.03291146084666252, 0.02542036585509777, -0.07104518264532089, 0.04833095148205757, -0.12325827032327652, -0.00188556092325598, 0.04320943355560303, 0.2802351415157318, 0.0405685119330883, -0.3180699050426483, -0.09586703777313232, -0.001863937359303236, -0.06601335108280182, -0.025169864296913147, 0.03086915798485279, 0.10135548561811447, -0.07850915938615799, 0.04553009569644928, -0.08417942374944687, 0.1017647311091423, -0.04550463706254959, 0.04261774942278862, 0.06964059174060822, 0.0973525196313858, 0.006177486386150122, 0.07683391869068146, -0.3115845322608948, 0.2770364582538605, 0.003649738384410739, 0.05838273838162422, -0.07231909036636353, 0.015402166172862053, 0.029973959550261497, 0.03283792361617088, 0.06709308922290802, -0.0236468818038702, -0.053502075374126434, -0.15808293223381042, -0.0766480416059494, 0.017629850655794144, 0.09788084030151367, -0.022138727828860283, 0.11446476727724075, -0.0450059249997139, 0.0014125618617981672, 0.07177630811929703, -0.0017192758386954665, -0.0657576322555542, -0.09894618391990662, 0.01636740006506443, 0.03453396260738373, -0.03535209968686104, -0.06613287329673767, -0.11288671940565109, -0.099558025598526, 0.16747832298278809, -0.039388880133628845, -0.041787441819906235, -0.11062997579574585, 0.08566928654909134, 0.08272646367549896, -0.08951807022094727, 0.039972852915525436, 0.0027922180015593767, 0.08221258223056793, 0.024498604238033295, -0.08637084066867828, 0.11773484200239182, -0.06372465193271637, -0.18094107508659363, -0.05394289642572403, 0.1287798285484314, 0.01562205795198679, 0.06453610211610794, -0.027446415275335312, 0.011150774545967579, -0.03938901796936989, -0.08023043721914291, 0.013192030601203442, 0.003363831667229533, 0.0681382343173027, 0.012272682040929794, -0.06686056405305862, 0.005658120382577181, -0.05888858437538147, -0.04186416044831276, 0.19300760328769684, 0.22520455718040466, -0.08679668605327606, 0.035767678171396255, 0.033091478049755096, -0.07429971545934677, -0.1849990040063858, 0.012744343839585781, 0.06097197160124779, 0.007211062591522932, 0.024243300780653954, -0.18836943805217743, 0.07346469908952713, 0.10004008561372757, -0.00837626587599516, 0.09914879500865936, -0.34950441122055054, -0.1374751478433609, 0.10743189603090286, 0.13604730367660522, 0.09628299623727798, -0.15808777511119843, -0.028050824999809265, -0.012906467542052269, -0.1121537983417511, 0.12720459699630737, -0.08926551789045334, 0.1276235580444336, -0.030113140121102333, 0.10204938799142838, 0.011549760587513447, -0.05622809752821922, 0.0967482402920723, -0.014391757547855377, 0.07529748231172562, -0.06632299721240997, 0.010276787914335728, 0.04653660207986832, -0.04602847620844841, 0.033551398664712906, -0.09545015543699265, 0.028441082686185837, -0.09289832413196564, -0.028064562007784843, -0.07354013621807098, 0.02864394150674343, -0.03696542978286743, -0.05775268003344536, -0.03842325136065483, 0.010819452814757824, 0.06380131840705872, -0.012068573385477066, 0.14870263636112213, -0.0005498268292285502, 0.14971645176410675, 0.12274178862571716, 0.08925654739141464, -0.049326732754707336, -0.06268030405044556, -0.011371415108442307, -0.020137067884206772, 0.05212213844060898, -0.14427445828914642, 0.023769235238432884, 0.15216124057769775, 0.012204814702272415, 0.1438562124967575, 0.07859069108963013, -0.04671945050358772, 0.0067559233866631985, 0.059334784746170044, -0.16533371806144714, -0.11403258144855499, -0.016219502314925194, -0.030757775530219078, -0.11229962855577469, 0.039819300174713135, 0.11831185221672058, -0.0757538452744484, -0.005467086099088192, 0.007986614480614662, 0.016132740303874016, -0.043064091354608536, 0.18508249521255493, 0.03280956670641899, 0.04500225931406021, -0.09153584390878677, 0.08438543975353241, 0.04874846339225769, -0.11189202964305878, 0.02673557586967945, 0.11879381537437439, -0.07513748109340668, -0.04230424761772156, 0.056355223059654236, 0.15814174711704254, -0.06083619222044945, -0.051546331495046616, -0.13888682425022125, -0.12150862067937851, 0.1011362224817276, 0.17383639514446259, 0.07477602362632751, 0.015202626585960388, -0.05597279593348503, 0.016755463555455208, -0.11632344126701355, 0.10590853542089462, 0.04650687798857689, 0.06333526223897934, -0.12795105576515198, 0.1621592491865158, 0.009975633583962917, 0.0341961532831192, -0.020248830318450928, 0.02993115596473217, -0.09240244328975677, 0.00516399834305048, -0.1327214539051056, -0.015830542892217636, -0.025055525824427605, -0.005159457679837942, -0.007224248256534338, -0.04342233017086983, -0.06602351367473602, 0.01573561504483223, -0.1079404205083847, -0.03205976262688637, 0.010696139186620712, 0.0527566596865654, -0.11295486986637115, -0.02667165920138359, 0.018598835915327072, -0.07231608033180237, 0.08585590124130249, 0.047582969069480896, 0.008423063904047012, 0.050962433218955994, -0.12281996011734009, 0.021969828754663467, 0.05819125473499298, 0.023421231657266617, 0.045176632702350616, -0.09510499984025955, -0.0053449166007339954, -0.00416087731719017, 0.036273691803216934, 0.012846813537180424, 0.0688384398818016, -0.13683339953422546, -0.005447857081890106, -0.02121860533952713, -0.06594370305538177, -0.06882797181606293, 0.03822512924671173, 0.05455278977751732, 0.025747045874595642, 0.1743641197681427, -0.09025660902261734, 0.04786162078380585, -0.22445723414421082, 0.013717317953705788, 0.0007456443272531033, -0.10603298246860504, -0.0822920948266983, -0.06746797263622284, 0.06580676883459091, -0.06137210875749588, 0.1194010004401207, 0.011731178499758244, 0.04491567984223366, 0.037451643496751785, -0.03913581371307373, 0.0003387695178389549, 0.015274430625140667, 0.20671822130680084, 0.028274817392230034, -0.04267042130231857, 0.03994293138384819, 0.027011286467313766, 0.09897870570421219, 0.1251721829175949, 0.21746008098125458, 0.1443902999162674, 0.00819371361285448, 0.10330741852521896, 0.03774707764387131, -0.06433349847793579, -0.16620855033397675, 0.05777188017964363, -0.038903187960386276, 0.13317857682704926, -0.027176300063729286, 0.23982073366641998, 0.09933306276798248, -0.15459823608398438, 0.053737152367830276, -0.040833480656147, -0.07642215490341187, -0.11874419450759888, -0.07482520490884781, -0.08116868138313293, -0.1470830738544464, -0.007038090378046036, -0.12125971913337708, 0.04796804487705231, 0.07687056809663773, 0.024951323866844177, -0.022664962336421013, 0.14206930994987488, 0.03099041059613228, -0.0015508835203945637, 0.0718890056014061, 0.01312553696334362, -0.010794972069561481, -0.1007872074842453, -0.07965445518493652, 0.008685879409313202, -0.009174277074635029, 0.03900023177266121, -0.03424133360385895, -0.038705188781023026, 0.04055759683251381, -0.022366680204868317, -0.09681612253189087, 0.012983053922653198, 0.020274659618735313, 0.07080096751451492, 0.07250792533159256, 0.0088870320469141, 0.0009708370780572295, -0.013526457361876965, 0.2187499850988388, -0.08535981923341751, -0.07158172130584717, -0.09586427360773087, 0.23818866908550262, 0.027675556018948555, -0.02125565893948078, 0.03135984390974045, -0.062381561845541, -0.015156809240579605, 0.248570516705513, 0.1799585074186325, -0.06191599741578102, -0.011147504672408104, 0.016132691875100136, -0.005703626200556755, -0.01867654174566269, 0.10559917241334915, 0.14206047356128693, 0.07876604050397873, -0.08216157555580139, -0.04418375715613365, -0.043920811265707016, -0.0011924505233764648, -0.04161557927727699, 0.08116697520017624, 0.03127100318670273, -0.0031100884079933167, -0.022759346291422844, 0.05014672502875328, -0.062130700796842575, -0.09346071630716324, 0.01612163707613945, -0.21261119842529297, -0.15939190983772278, -0.014445329084992409, 0.10830047726631165, -0.0029632048681378365, 0.052350714802742004, -0.02021626941859722, 0.0011673958506435156, 0.09261893481016159, -0.020598411560058594, -0.08719086647033691, -0.061430081725120544, 0.0800943523645401, -0.1197943463921547, 0.18815921247005463, -0.03740065544843674, 0.03992569446563721, 0.13078317046165466, 0.06729324907064438, -0.07798749953508377, 0.07431641966104507, 0.048254434019327164, -0.06731877475976944, 0.03658498823642731, 0.11344766616821289, -0.03134822100400925, 0.06376076489686966, 0.04725462943315506, -0.1335436850786209, 0.014687446877360344, -0.07266610860824585, -0.06185290217399597, -0.03238183259963989, -0.035577576607465744, -0.058154065161943436, 0.13258427381515503, 0.21921570599079132, -0.03687327727675438, 0.00048057673848234117, -0.07764729857444763, -0.004444161430001259, 0.048580095171928406, 0.055419228971004486, -0.038020554929971695, -0.22537297010421753, 0.009807797148823738, 0.0571817122399807, -0.0041002691723406315, -0.26522406935691833, -0.09097064286470413, 0.012601674534380436, -0.058848533779382706, -0.1294889599084854, 0.08956053853034973, 0.08972277492284775, 0.048941295593976974, -0.046956077218055725, -0.059232089668512344, -0.06484432518482208, 0.1634431779384613, -0.15104427933692932, -0.07681909203529358 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-small-mlm-pubmed-45 This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.6395 - Rouge2 Precision: 0.3383 - Rouge2 Recall: 0.2424 - Rouge2 Fmeasure: 0.2753 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure | |:-------------:|:-----:|:----:|:---------------:|:----------------:|:-------------:|:---------------:| | 2.519 | 0.75 | 500 | 1.9659 | 0.3178 | 0.1888 | 0.2299 | | 2.169 | 1.51 | 1000 | 1.8450 | 0.3256 | 0.2138 | 0.25 | | 2.0796 | 2.26 | 1500 | 1.7900 | 0.3368 | 0.2265 | 0.2636 | | 1.9978 | 3.02 | 2000 | 1.7553 | 0.3427 | 0.234 | 0.2709 | | 1.9686 | 3.77 | 2500 | 1.7172 | 0.3356 | 0.2347 | 0.2692 | | 1.9142 | 4.52 | 3000 | 1.6986 | 0.3358 | 0.238 | 0.2715 | | 1.921 | 5.28 | 3500 | 1.6770 | 0.3349 | 0.2379 | 0.2709 | | 1.8848 | 6.03 | 4000 | 1.6683 | 0.3346 | 0.2379 | 0.2708 | | 1.8674 | 6.79 | 4500 | 1.6606 | 0.3388 | 0.2419 | 0.2752 | | 1.8606 | 7.54 | 5000 | 1.6514 | 0.3379 | 0.2409 | 0.274 | | 1.8515 | 8.3 | 5500 | 1.6438 | 0.3356 | 0.2407 | 0.2731 | | 1.8403 | 9.05 | 6000 | 1.6401 | 0.3367 | 0.2421 | 0.2744 | | 1.8411 | 9.8 | 6500 | 1.6395 | 0.3383 | 0.2424 | 0.2753 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.10.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "t5-small-mlm-pubmed-45", "results": []}]}
text2text-generation
gayanin/t5-small-mlm-pubmed-45
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
t5-small-mlm-pubmed-45 ====================== This model is a fine-tuned version of t5-small on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 1.6395 * Rouge2 Precision: 0.3383 * Rouge2 Recall: 0.2424 * Rouge2 Fmeasure: 0.2753 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 10 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.12.5 * Pytorch 1.10.0+cu111 * Datasets 1.15.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ 67, 113, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ -0.08269525319337845, 0.056378915905952454, -0.0033615895081311464, 0.09726845473051071, 0.1380135864019394, 0.017358558252453804, 0.13250045478343964, 0.1417226940393448, -0.10345281660556793, 0.04318553954362869, 0.11734073609113693, 0.13928459584712982, 0.03708208352327347, 0.11994913220405579, -0.053264033049345016, -0.27000662684440613, 0.005596803035587072, 0.042485322803258896, -0.044491712003946304, 0.13760170340538025, 0.08332482725381851, -0.11321394145488739, 0.08034632354974747, 0.008888037875294685, -0.16212846338748932, 0.012178964912891388, 0.0033482930157333612, -0.059928275644779205, 0.14373448491096497, 0.04238886386156082, 0.11004441231489182, 0.017295870929956436, 0.07233617454767227, -0.19673123955726624, 0.010151540860533714, 0.06811850517988205, 0.0005693810526281595, 0.09264732897281647, 0.06996092200279236, 0.0072451322339475155, 0.15095973014831543, -0.07351623475551605, 0.052414678037166595, 0.02655074931681156, -0.1135457381606102, -0.20676463842391968, -0.08416391164064407, 0.04574306309223175, 0.07331861555576324, 0.10789787024259567, -0.013276481069624424, 0.11553036421537399, -0.059303686022758484, 0.10711460560560226, 0.25034427642822266, -0.29492226243019104, -0.06128635257482529, -0.0022695267107337713, 0.03915403410792351, 0.07694810628890991, -0.08384211361408234, -0.028664587065577507, 0.034901462495326996, 0.0509931817650795, 0.13463430106639862, -0.019712962210178375, -0.09382731467485428, 0.002993146190419793, -0.14766260981559753, -0.05127188563346863, 0.13165341317653656, 0.03568609803915024, -0.024856511503458023, -0.05997200682759285, -0.08379334956407547, -0.19245201349258423, -0.03627685829997063, -0.01434552762657404, 0.04285723716020584, -0.01977437548339367, -0.054337866604328156, -0.03188811242580414, -0.10448281466960907, -0.06290357559919357, -0.06915207207202911, 0.11096615344285965, 0.051424119621515274, 0.00011151981743751094, -0.0397210493683815, 0.10572453588247299, -0.007833397015929222, -0.13282829523086548, 0.014494688250124454, 0.03131069242954254, -0.0016167850699275732, -0.02790338173508644, -0.05696355924010277, -0.09389249980449677, 0.006581631954759359, 0.13552556931972504, -0.06912436336278915, 0.05344173684716225, -0.006928566377609968, 0.04504135996103287, -0.10696142911911011, 0.16744956374168396, -0.04871731624007225, -0.009845289401710033, 0.011549752205610275, 0.06054366007447243, 0.0328734926879406, -0.02049337327480316, -0.11201692372560501, 0.0010903767542913556, 0.10973610728979111, 0.02047094888985157, -0.049063894897699356, 0.07413005083799362, -0.04057883471250534, -0.021568577736616135, -0.005884911864995956, -0.10268031060695648, 0.0221256036311388, 0.001785362372174859, -0.06933064758777618, 0.004972131457179785, 0.04628436639904976, 0.0005319282063283026, -0.05206027254462242, 0.10894733667373657, -0.08053117990493774, 0.020323770120739937, -0.09386667609214783, -0.12771724164485931, 0.03203904628753662, -0.07229843735694885, -0.0008894350612536073, -0.09936651587486267, -0.16893015801906586, -0.015167842619121075, 0.05025872215628624, -0.03634479269385338, -0.05311787128448486, -0.05400995537638664, -0.08402786403894424, 0.031788334250450134, -0.026317749172449112, 0.13538812100887299, -0.05995629355311394, 0.10286474227905273, 0.026920361444354057, 0.0578741580247879, -0.035238347947597504, 0.0616101436316967, -0.0889873206615448, 0.01812722720205784, -0.16571000218391418, 0.04600019380450249, -0.038386911153793335, 0.052468717098236084, -0.10034013539552689, -0.10366327315568924, -0.01440427266061306, -0.0016803444596007466, 0.09378878027200699, 0.09093115478754044, -0.15961578488349915, -0.08025752007961273, 0.18463566899299622, -0.07589662820100784, -0.11519395560026169, 0.1318536102771759, -0.048300016671419144, 0.023699544370174408, 0.05046416074037552, 0.17146648466587067, 0.06272238492965698, -0.09445425868034363, 0.008414136245846748, -0.012561875395476818, 0.05291658267378807, -0.04295893758535385, 0.06724854558706284, -0.005161365959793329, 0.025250060483813286, 0.016542961820960045, -0.00924696959555149, 0.05842602625489235, -0.08680278807878494, -0.08183000981807709, -0.05125518515706062, -0.07013542205095291, 0.025736374780535698, 0.05821191892027855, 0.06687890738248825, -0.10170835256576538, -0.10978864878416061, 0.05612782761454582, 0.0777045488357544, -0.08659040182828903, 0.0491458885371685, -0.06330514699220657, 0.07790636271238327, -0.037381015717983246, -0.004637535661458969, -0.18389855325222015, -0.025765148922801018, 0.01693103089928627, -0.01687784492969513, 0.03195309266448021, 0.00992311630398035, 0.06839224696159363, 0.05935089662671089, -0.05325128510594368, -0.02889467217028141, -0.046754296869039536, -0.01083369366824627, -0.12166513502597809, -0.19820936024188995, -0.029339097440242767, -0.012906184419989586, 0.10996276885271072, -0.19980613887310028, 0.04117908328771591, -0.0008601741283200681, 0.08701878041028976, 0.015517298132181168, -0.0059236790984869, -0.035857487469911575, 0.0784507766366005, -0.05494728684425354, -0.047110334038734436, 0.07603180408477783, 0.011445947922766209, -0.09326332807540894, -0.007221799343824387, -0.147250697016716, 0.13424332439899445, 0.13277801871299744, -0.10779416561126709, -0.07405058294534683, 0.000604522880166769, -0.06356871873140335, -0.04413803666830063, -0.0340212807059288, 0.0016274807276204228, 0.18422681093215942, 0.004109219182282686, 0.16078519821166992, -0.0821070447564125, -0.05481322482228279, 0.02870483696460724, -0.021909138187766075, 0.021282700821757317, 0.12766748666763306, 0.11134835332632065, -0.07082132250070572, 0.14148126542568207, 0.13859781622886658, -0.08368471264839172, 0.147953063249588, -0.04156362637877464, -0.09589220583438873, -0.01521033700555563, 0.0014992242213338614, 0.0065142326056957245, 0.07284099608659744, -0.15796105563640594, -0.0015062465099617839, 0.02833535522222519, 0.025335397571325302, 0.030536135658621788, -0.2187902182340622, -0.012304075062274933, 0.037896398454904556, -0.06272267550230026, -0.005430673249065876, -0.003159643616527319, 0.015953944995999336, 0.11623569577932358, 0.005600794218480587, -0.07519179582595825, 0.02453729137778282, -0.0028576848562806845, -0.08647949248552322, 0.19883769750595093, -0.0849945917725563, -0.1808682233095169, -0.12791162729263306, -0.0677480399608612, -0.04339669644832611, -0.002173554850742221, 0.0702272281050682, -0.07362733781337738, -0.032201532274484634, -0.08636964857578278, 0.03719964623451233, -0.015163491480052471, 0.030274158343672752, 0.0032852974254637957, 0.0017474624328315258, 0.06742336601018906, -0.11047136038541794, -0.012818917632102966, -0.04099436476826668, -0.05448686331510544, 0.043724220246076584, 0.03221651911735535, 0.10896811634302139, 0.15915575623512268, -0.021351315081119537, 0.014196757227182388, -0.03945118561387062, 0.19537794589996338, -0.0620412677526474, -0.01090088952332735, 0.1490119844675064, -0.011384861543774605, 0.06295274943113327, 0.11998633295297623, 0.05519726872444153, -0.07660055160522461, 0.01516230683773756, 0.04347888380289078, -0.030182501301169395, -0.2452249377965927, -0.036999884992837906, -0.06423996388912201, 0.019275818020105362, 0.09551515430212021, 0.029116734862327576, 0.05069415643811226, 0.05116363242268562, 0.015689419582486153, 0.0717720165848732, -0.018054937943816185, 0.07714737206697464, 0.15393786132335663, 0.038777049630880356, 0.13492351770401, -0.04749971255660057, -0.0541992112994194, 0.04947451502084732, -0.003922663629055023, 0.2202150821685791, -0.00026676373090595007, 0.17084971070289612, 0.054769616574048996, 0.15596789121627808, 0.004101146012544632, 0.07521362602710724, -0.008631471544504166, -0.022696271538734436, -0.01335220504552126, -0.05006207525730133, -0.032055653631687164, 0.025346677750349045, -0.07101451605558395, 0.0486719124019146, -0.1233769953250885, -0.0020515350624918938, 0.04311871528625488, 0.27941274642944336, 0.04106856882572174, -0.3180472254753113, -0.09526669979095459, -0.0019105826504528522, -0.0668366402387619, -0.02458605356514454, 0.03074367344379425, 0.10163965076208115, -0.07827594876289368, 0.04606161639094353, -0.08417335897684097, 0.10201042145490646, -0.043985240161418915, 0.04298200458288193, 0.07026321440935135, 0.09741707891225815, 0.006220438051968813, 0.07725043594837189, -0.31219181418418884, 0.27684107422828674, 0.00319865089841187, 0.05907602235674858, -0.07293906807899475, 0.015789980068802834, 0.029849328100681305, 0.030358120799064636, 0.06695136427879333, -0.023928217589855194, -0.053665246814489365, -0.158283069729805, -0.07674533128738403, 0.01836743764579296, 0.09846096485853195, -0.022080333903431892, 0.11480162292718887, -0.044587600976228714, 0.0011874983320012689, 0.07196878641843796, -0.0006222384981811047, -0.06674335151910782, -0.09944204241037369, 0.01700431853532791, 0.03379006311297417, -0.035738371312618256, -0.06590922176837921, -0.1128842756152153, -0.1011509895324707, 0.1666242480278015, -0.03917703405022621, -0.041605401784181595, -0.11047035455703735, 0.08587080985307693, 0.08362783491611481, -0.08941028267145157, 0.04098344221711159, 0.0026268598157912493, 0.08178117871284485, 0.024251021444797516, -0.08665957301855087, 0.11725389957427979, -0.0638246163725853, -0.18192006647586823, -0.053622905164957047, 0.12911221385002136, 0.015225201845169067, 0.06438282877206802, -0.0267353393137455, 0.011391929350793362, -0.038634780794382095, -0.08037734031677246, 0.013032876886427402, 0.0043837870471179485, 0.06723761558532715, 0.012349408119916916, -0.06642144173383713, 0.004431111738085747, -0.059756308794021606, -0.041329238563776016, 0.19353334605693817, 0.22535426914691925, -0.08669967949390411, 0.03599875047802925, 0.033000413328409195, -0.07494721561670303, -0.18455864489078522, 0.012883306480944157, 0.061455342918634415, 0.007004934828728437, 0.024494977667927742, -0.18805061280727386, 0.0737152099609375, 0.10081442445516586, -0.007979331538081169, 0.09884322434663773, -0.3505803942680359, -0.1376723349094391, 0.10717718303203583, 0.13668929040431976, 0.09429292380809784, -0.1577870100736618, -0.027832427993416786, -0.013011494651436806, -0.11239242553710938, 0.1287219077348709, -0.08839474618434906, 0.12759807705879211, -0.0295058973133564, 0.10005053877830505, 0.011605529114603996, -0.056386034935712814, 0.0966484397649765, -0.014278127811849117, 0.07524306327104568, -0.06615935266017914, 0.00922487210482359, 0.04807693511247635, -0.04585577920079231, 0.03357836604118347, -0.09540718048810959, 0.027833018451929092, -0.09177707880735397, -0.02811633236706257, -0.07351407408714294, 0.028258604928851128, -0.037035729736089706, -0.057568978518247604, -0.03813588619232178, 0.010525871999561787, 0.06474247574806213, -0.011947987601161003, 0.14867018163204193, -0.00030866681481711566, 0.1502327024936676, 0.12151180952787399, 0.08947128057479858, -0.05011026933789253, -0.06166475638747215, -0.011435549706220627, -0.019902849569916725, 0.05156496912240982, -0.14429137110710144, 0.024305559694767, 0.15239167213439941, 0.012361976318061352, 0.14391875267028809, 0.07913386821746826, -0.04669825732707977, 0.007309648208320141, 0.05904959887266159, -0.16460445523262024, -0.11493266373872757, -0.015880193561315536, -0.030602402985095978, -0.11311869323253632, 0.040661171078681946, 0.11809003353118896, -0.07640529423952103, -0.005917554255574942, 0.007487940602004528, 0.016412269324064255, -0.04268287494778633, 0.18580596148967743, 0.03240061178803444, 0.045250631868839264, -0.0915227010846138, 0.08461984246969223, 0.04938613250851631, -0.11333908885717392, 0.02700362354516983, 0.11827096343040466, -0.07460956275463104, -0.04249544069170952, 0.056030500680208206, 0.15699324011802673, -0.058109868317842484, -0.0514342226088047, -0.13905341923236847, -0.12185586988925934, 0.10150935500860214, 0.17514614760875702, 0.07429982721805573, 0.0158491600304842, -0.05547500401735306, 0.01750873774290085, -0.1167503073811531, 0.10663007199764252, 0.04698600992560387, 0.0637093186378479, -0.12853676080703735, 0.1613733172416687, 0.009821878746151924, 0.03393711522221565, -0.0199320986866951, 0.029620300978422165, -0.09307966381311417, 0.004899723455309868, -0.13255731761455536, -0.016058633103966713, -0.02612416446208954, -0.005226100329309702, -0.006490012630820274, -0.04287366941571236, -0.06589032709598541, 0.015216692350804806, -0.10787488520145416, -0.03235069289803505, 0.010425429791212082, 0.05320208892226219, -0.11354006826877594, -0.026639046147465706, 0.018161142244935036, -0.07261420786380768, 0.08592170476913452, 0.047300346195697784, 0.007859506644308567, 0.050588447600603104, -0.1244892105460167, 0.021579725667834282, 0.05894581601023674, 0.023242393508553505, 0.044990621507167816, -0.09535739570856094, -0.005086883436888456, -0.003991192672401667, 0.036462265998125076, 0.012634573504328728, 0.06807177513837814, -0.13644780218601227, -0.006225551012903452, -0.021476207301020622, -0.06657961755990982, -0.06849218159914017, 0.037936046719551086, 0.053487151861190796, 0.02460109442472458, 0.17492060363292694, -0.08943071216344833, 0.04843921586871147, -0.22437000274658203, 0.013288824819028378, 0.0006127831293269992, -0.1065606102347374, -0.08215301483869553, -0.0672241598367691, 0.06605275720357895, -0.06108418479561806, 0.11998360604047775, 0.011219877749681473, 0.04583999887108803, 0.03713931888341904, -0.03837547078728676, 0.0004274279053788632, 0.01642737165093422, 0.20671513676643372, 0.028101854026317596, -0.04290441423654556, 0.03951560705900192, 0.026594560593366623, 0.099412702023983, 0.12426470965147018, 0.21719783544540405, 0.14552831649780273, 0.007416215725243092, 0.10298871994018555, 0.03850937262177467, -0.06438405811786652, -0.1655460000038147, 0.05801280960440636, -0.039102133363485336, 0.13383296132087708, -0.027338379994034767, 0.2398751676082611, 0.09891008585691452, -0.15449953079223633, 0.05394202098250389, -0.04069485142827034, -0.07641230523586273, -0.11851092427968979, -0.07599102705717087, -0.08121068775653839, -0.14669637382030487, -0.006592465098947287, -0.12147950381040573, 0.04737407714128494, 0.07723213732242584, 0.025379927828907967, -0.0224272720515728, 0.14215174317359924, 0.03127937763929367, -0.001587958075106144, 0.07157371193170547, 0.01308489590883255, -0.011080355383455753, -0.10127421468496323, -0.07883825898170471, 0.008688746951520443, -0.008731166832149029, 0.03941936045885086, -0.03465891256928444, -0.0387042798101902, 0.04094085842370987, -0.023018335923552513, -0.09658120572566986, 0.013782097958028316, 0.019648026674985886, 0.07085202634334564, 0.07227272540330887, 0.008937516249716282, 0.00041281149606220424, -0.0131387859582901, 0.21901138126850128, -0.08560163527727127, -0.07266795635223389, -0.09617272764444351, 0.2374449074268341, 0.027929166331887245, -0.021551519632339478, 0.0316794216632843, -0.06247735396027565, -0.014416265301406384, 0.2484855353832245, 0.18090471625328064, -0.059918977320194244, -0.010903073474764824, 0.01627921871840954, -0.005722172558307648, -0.019365373998880386, 0.10553712397813797, 0.14187146723270416, 0.07826556265354156, -0.08175606280565262, -0.0438702367246151, -0.0426865816116333, -0.0014341978821903467, -0.04207398742437363, 0.08042506873607635, 0.031545039266347885, -0.0033897224348038435, -0.022281965240836143, 0.05040707066655159, -0.06268028914928436, -0.09229660779237747, 0.015594256110489368, -0.21274876594543457, -0.15923573076725006, -0.01536442618817091, 0.10842075943946838, -0.002407025545835495, 0.05266395956277847, -0.020109323784708977, 0.0004890338168479502, 0.09164372831583023, -0.020243996754288673, -0.08709855377674103, -0.06154970824718475, 0.07987271249294281, -0.11974356323480606, 0.18853157758712769, -0.037729211151599884, 0.03990333154797554, 0.13098415732383728, 0.06742548197507858, -0.07831884920597076, 0.07312920689582825, 0.04901792109012604, -0.06688553839921951, 0.0366656631231308, 0.11311990022659302, -0.03137825056910515, 0.06388898193836212, 0.047431036829948425, -0.13387876749038696, 0.015139265917241573, -0.07251381874084473, -0.0612214021384716, -0.03221260756254196, -0.034659262746572495, -0.05775070935487747, 0.1323966532945633, 0.2190065085887909, -0.03710855543613434, 0.00018983685004059225, -0.07764235883951187, -0.004742419347167015, 0.04807257652282715, 0.055883441120386124, -0.038031067699193954, -0.2264697402715683, 0.00991019792854786, 0.056176088750362396, -0.004065029788762331, -0.264145165681839, -0.09081718325614929, 0.012506717815995216, -0.05934730917215347, -0.12927012145519257, 0.0889926478266716, 0.0906132236123085, 0.04876545071601868, -0.04732590168714523, -0.06102409213781357, -0.06475313752889633, 0.16419334709644318, -0.15079237520694733, -0.07652173936367035 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-small-mlm-pubmed This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.8008 - Rouge2 Precision: 0.6071 - Rouge2 Recall: 0.4566 - Rouge2 Fmeasure: 0.5079 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 40 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure | |:-------------:|:-----:|:-----:|:---------------:|:----------------:|:-------------:|:---------------:| | 0.914 | 0.75 | 500 | 0.8691 | 0.5901 | 0.4357 | 0.4879 | | 0.9093 | 1.51 | 1000 | 0.8646 | 0.5867 | 0.4372 | 0.488 | | 0.895 | 2.26 | 1500 | 0.8618 | 0.5891 | 0.4387 | 0.49 | | 0.8842 | 3.02 | 2000 | 0.8571 | 0.5899 | 0.4374 | 0.4891 | | 0.8796 | 3.77 | 2500 | 0.8544 | 0.5903 | 0.4406 | 0.4916 | | 0.8759 | 4.52 | 3000 | 0.8513 | 0.5921 | 0.4395 | 0.4912 | | 0.8621 | 5.28 | 3500 | 0.8485 | 0.5934 | 0.4413 | 0.493 | | 0.8613 | 6.03 | 4000 | 0.8442 | 0.5944 | 0.4428 | 0.4944 | | 0.8537 | 6.79 | 4500 | 0.8406 | 0.594 | 0.4414 | 0.4932 | | 0.8518 | 7.54 | 5000 | 0.8399 | 0.5956 | 0.4424 | 0.4945 | | 0.8438 | 8.3 | 5500 | 0.8365 | 0.5953 | 0.4452 | 0.4964 | | 0.8339 | 9.05 | 6000 | 0.8353 | 0.5983 | 0.4468 | 0.4983 | | 0.8307 | 9.8 | 6500 | 0.8331 | 0.5979 | 0.4461 | 0.4976 | | 0.8328 | 10.56 | 7000 | 0.8304 | 0.5975 | 0.4465 | 0.4979 | | 0.8263 | 11.31 | 7500 | 0.8283 | 0.5977 | 0.4467 | 0.4981 | | 0.8168 | 12.07 | 8000 | 0.8267 | 0.5971 | 0.4463 | 0.4976 | | 0.8165 | 12.82 | 8500 | 0.8248 | 0.5969 | 0.4462 | 0.4976 | | 0.8084 | 13.57 | 9000 | 0.8245 | 0.6018 | 0.4527 | 0.5035 | | 0.8136 | 14.33 | 9500 | 0.8219 | 0.6023 | 0.4509 | 0.5023 | | 0.8073 | 15.08 | 10000 | 0.8206 | 0.6002 | 0.4486 | 0.5001 | | 0.808 | 15.84 | 10500 | 0.8185 | 0.6009 | 0.4506 | 0.5019 | | 0.8027 | 16.59 | 11000 | 0.8173 | 0.5978 | 0.4478 | 0.4989 | | 0.8061 | 17.35 | 11500 | 0.8169 | 0.6022 | 0.4513 | 0.5026 | | 0.7922 | 18.1 | 12000 | 0.8152 | 0.6016 | 0.4501 | 0.5016 | | 0.7928 | 18.85 | 12500 | 0.8141 | 0.6009 | 0.45 | 0.5012 | | 0.7909 | 19.61 | 13000 | 0.8143 | 0.6019 | 0.4521 | 0.5028 | | 0.7909 | 20.36 | 13500 | 0.8115 | 0.5997 | 0.4505 | 0.5011 | | 0.7949 | 21.12 | 14000 | 0.8115 | 0.6043 | 0.4536 | 0.5048 | | 0.7853 | 21.87 | 14500 | 0.8095 | 0.6033 | 0.4527 | 0.5038 | | 0.7819 | 22.62 | 15000 | 0.8095 | 0.6054 | 0.4541 | 0.5056 | | 0.7828 | 23.38 | 15500 | 0.8075 | 0.6036 | 0.453 | 0.5042 | | 0.787 | 24.13 | 16000 | 0.8068 | 0.6031 | 0.4528 | 0.504 | | 0.7739 | 24.89 | 16500 | 0.8072 | 0.6043 | 0.4529 | 0.5045 | | 0.7782 | 25.64 | 17000 | 0.8073 | 0.606 | 0.4551 | 0.5063 | | 0.7772 | 26.4 | 17500 | 0.8063 | 0.6055 | 0.4549 | 0.5062 | | 0.7718 | 27.15 | 18000 | 0.8057 | 0.606 | 0.4546 | 0.5059 | | 0.7747 | 27.9 | 18500 | 0.8045 | 0.6046 | 0.4543 | 0.5054 | | 0.7738 | 28.66 | 19000 | 0.8035 | 0.6059 | 0.4549 | 0.506 | | 0.7642 | 29.41 | 19500 | 0.8041 | 0.6053 | 0.4545 | 0.5058 | | 0.7666 | 30.17 | 20000 | 0.8039 | 0.6066 | 0.457 | 0.508 | | 0.7686 | 30.92 | 20500 | 0.8027 | 0.6075 | 0.4571 | 0.5081 | | 0.7664 | 31.67 | 21000 | 0.8026 | 0.6062 | 0.4566 | 0.5076 | | 0.77 | 32.43 | 21500 | 0.8022 | 0.6068 | 0.4571 | 0.5081 | | 0.7618 | 33.18 | 22000 | 0.8015 | 0.6065 | 0.4563 | 0.5072 | | 0.7615 | 33.94 | 22500 | 0.8013 | 0.6064 | 0.4565 | 0.5074 | | 0.7611 | 34.69 | 23000 | 0.8017 | 0.607 | 0.4567 | 0.5078 | | 0.7611 | 35.44 | 23500 | 0.8013 | 0.608 | 0.4565 | 0.5082 | | 0.7604 | 36.2 | 24000 | 0.8012 | 0.6069 | 0.4561 | 0.5072 | | 0.7599 | 36.95 | 24500 | 0.8013 | 0.6078 | 0.4571 | 0.5085 | | 0.7542 | 37.71 | 25000 | 0.8016 | 0.6083 | 0.4579 | 0.5091 | | 0.7637 | 38.46 | 25500 | 0.8009 | 0.6072 | 0.4569 | 0.5081 | | 0.7596 | 39.22 | 26000 | 0.8008 | 0.6069 | 0.4566 | 0.5078 | | 0.7604 | 39.97 | 26500 | 0.8008 | 0.6071 | 0.4566 | 0.5079 | ### Framework versions - Transformers 4.12.3 - Pytorch 1.9.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "t5-small-mlm-pubmed", "results": []}]}
text2text-generation
gayanin/t5-small-mlm-pubmed
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
t5-small-mlm-pubmed =================== This model is a fine-tuned version of t5-small on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 0.8008 * Rouge2 Precision: 0.6071 * Rouge2 Recall: 0.4566 * Rouge2 Fmeasure: 0.5079 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 40 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.12.3 * Pytorch 1.9.0+cu111 * Datasets 1.15.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ 67, 113, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ -0.07970647513866425, 0.05618209391832352, -0.0035211448557674885, 0.09573755413293839, 0.13439251482486725, 0.01604393869638443, 0.13040132820606232, 0.14207349717617035, -0.11404426395893097, 0.04347899556159973, 0.11448727548122406, 0.14493705332279205, 0.03565381094813347, 0.12750902771949768, -0.05548149347305298, -0.27041757106781006, 0.005191842094063759, 0.04392910376191139, -0.041267987340688705, 0.13601571321487427, 0.0827043280005455, -0.11394234746694565, 0.07784605771303177, 0.010510270483791828, -0.1648477017879486, 0.014006838202476501, 0.004427968058735132, -0.06444551795721054, 0.1397617608308792, 0.042953427881002426, 0.10778803378343582, 0.019004065543413162, 0.0670994445681572, -0.19370298087596893, 0.010933644138276577, 0.0695246234536171, -0.002024224493652582, 0.0893910676240921, 0.06763961911201477, 0.0037759514525532722, 0.16508980095386505, -0.07879544049501419, 0.055875882506370544, 0.026623399928212166, -0.11347684264183044, -0.21202772855758667, -0.08364398777484894, 0.044286902993917465, 0.07125360518693924, 0.10669179260730743, -0.011496015824377537, 0.1261020451784134, -0.054858338087797165, 0.11157506704330444, 0.24977892637252808, -0.29577717185020447, -0.06224033609032631, -0.010642322711646557, 0.037630289793014526, 0.07900764048099518, -0.08015309274196625, -0.027335336431860924, 0.035151831805706024, 0.0523827001452446, 0.13492220640182495, -0.019226429983973503, -0.0948009118437767, -0.00007149617886170745, -0.1471523642539978, -0.05248448997735977, 0.13527248799800873, 0.030327139422297478, -0.027001528069376945, -0.06306556612253189, -0.086290143430233, -0.18589968979358673, -0.03719177097082138, -0.014552822336554527, 0.042396750301122665, -0.02139066345989704, -0.0564255490899086, -0.03179588541388512, -0.10097865015268326, -0.05643310025334358, -0.07024245709180832, 0.11095584183931351, 0.05354992300271988, 0.0011953155044466257, -0.04592759534716606, 0.09748810529708862, -0.004272218327969313, -0.1359105408191681, 0.013752578757703304, 0.02925274148583412, 0.006220732815563679, -0.028572140261530876, -0.05794379860162735, -0.09798003733158112, 0.010030772536993027, 0.13240289688110352, -0.07927518337965012, 0.05555834248661995, -0.015303871594369411, 0.04185829311609268, -0.1052953228354454, 0.16563715040683746, -0.03657083958387375, -0.005941031035035849, 0.015368702821433544, 0.05598761513829231, 0.036626555025577545, -0.02462504431605339, -0.11085943132638931, 0.008543760515749454, 0.10500930994749069, 0.02141900733113289, -0.04978107661008835, 0.0750170350074768, -0.03790786862373352, -0.020713964477181435, -0.005967091768980026, -0.10705368965864182, 0.023009024560451508, -0.0002794674946926534, -0.06417135149240494, 0.005814921110868454, 0.04299601539969444, -0.0022012984845787287, -0.058519333600997925, 0.102350614964962, -0.07670111954212189, 0.018065376207232475, -0.09453786164522171, -0.1309826523065567, 0.03209294006228447, -0.076261006295681, -0.0008669015951454639, -0.10015710443258286, -0.1587812304496765, -0.012086953036487103, 0.05086781829595566, -0.03213105350732803, -0.05539444088935852, -0.05316904932260513, -0.084006667137146, 0.03278766945004463, -0.024140965193510056, 0.12758025527000427, -0.06016784533858299, 0.09923254698514938, 0.0264546237885952, 0.05907397344708443, -0.03135816007852554, 0.058520298451185226, -0.08861006051301956, 0.01691306009888649, -0.17035819590091705, 0.04763250797986984, -0.03778949752449989, 0.05268533527851105, -0.09737878292798996, -0.10552085191011429, -0.00809414591640234, -0.0026318728923797607, 0.09155330806970596, 0.08696578443050385, -0.1621638834476471, -0.08089955896139145, 0.187085822224617, -0.07997633516788483, -0.11667120456695557, 0.13191409409046173, -0.04904620721936226, 0.016333499923348427, 0.05132472142577171, 0.18084654211997986, 0.06298808753490448, -0.09667132049798965, 0.014284934848546982, -0.01233536098152399, 0.0511188767850399, -0.03936023637652397, 0.061658069491386414, -0.006255249958485365, 0.027179894968867302, 0.014478320255875587, -0.0012712603202089667, 0.054477039724588394, -0.08519433438777924, -0.08185864984989166, -0.053023673593997955, -0.06830500811338425, 0.021541286259889603, 0.05691271647810936, 0.0649256780743599, -0.10807033628225327, -0.10874702036380768, 0.05520942807197571, 0.07296762615442276, -0.08724057674407959, 0.05275844782590866, -0.06808307021856308, 0.07790038734674454, -0.029076624661684036, -0.0006249735015444458, -0.1835891306400299, -0.025952648371458054, 0.019814496859908104, -0.01861627958714962, 0.029163185507059097, 0.001983806723728776, 0.06586001813411713, 0.06195047125220299, -0.049851518124341965, -0.025172851979732513, -0.043830569833517075, -0.009892961010336876, -0.11822175979614258, -0.2014486938714981, -0.025029929354786873, -0.014790442772209644, 0.09835125505924225, -0.191099151968956, 0.04286830127239227, 0.006814300082623959, 0.08768846094608307, 0.016575083136558533, -0.00556919677183032, -0.03490392118692398, 0.08196140825748444, -0.05533399060368538, -0.048375941812992096, 0.07622221112251282, 0.015259234234690666, -0.09228192269802094, -0.005303113721311092, -0.1497860848903656, 0.13611219823360443, 0.13487353920936584, -0.10002174228429794, -0.07222650200128555, 0.0011263459455221891, -0.062444865703582764, -0.038231171667575836, -0.030713841319084167, 0.008991996757686138, 0.18739847838878632, 0.0008081691921688616, 0.162650004029274, -0.08515044301748276, -0.0572257936000824, 0.030708005651831627, -0.022270403802394867, 0.020106498152017593, 0.12760965526103973, 0.09955139458179474, -0.07106159627437592, 0.1378687024116516, 0.13609221577644348, -0.0808141902089119, 0.1503780335187912, -0.04630235210061073, -0.09189415723085403, -0.013379528187215328, -0.00024326455604750663, 0.010356050916016102, 0.0715009868144989, -0.15866462886333466, -0.0022577564232051373, 0.028105996549129486, 0.024323290213942528, 0.02742689661681652, -0.2151990681886673, -0.010028054937720299, 0.03984561562538147, -0.06185489147901535, -0.012049520388245583, -0.003524145809933543, 0.017359228804707527, 0.11673406511545181, 0.00482036080211401, -0.0671842023730278, 0.02437196858227253, -0.0022935455199331045, -0.08724077045917511, 0.19693607091903687, -0.09051427990198135, -0.17822737991809845, -0.12083493918180466, -0.07224565744400024, -0.04414281249046326, -0.00197895267046988, 0.07629862427711487, -0.078654445707798, -0.03377571702003479, -0.08970330655574799, 0.03379515931010246, -0.015554885379970074, 0.03069380111992359, 0.010514451190829277, -0.0005087167373858392, 0.06738945841789246, -0.11300215125083923, -0.016138961538672447, -0.041255123913288116, -0.05108978599309921, 0.046114541590213776, 0.031934503465890884, 0.11350259184837341, 0.1557045429944992, -0.026999326422810555, 0.018642611801624298, -0.04047708585858345, 0.19916129112243652, -0.060722049325704575, -0.010691525414586067, 0.14770998060703278, -0.010340921580791473, 0.06512372940778732, 0.11612339317798615, 0.053692761808633804, -0.07726564258337021, 0.015086568892002106, 0.040422044694423676, -0.03380247578024864, -0.24393326044082642, -0.037388723343610764, -0.06705563515424728, 0.01479910034686327, 0.09357575327157974, 0.029263248667120934, 0.0492623969912529, 0.047790899872779846, 0.015922728925943375, 0.07264577597379684, -0.014511949382722378, 0.08097585290670395, 0.1591137945652008, 0.04234624654054642, 0.13429881632328033, -0.051052577793598175, -0.05240398272871971, 0.05033881962299347, -0.009724665433168411, 0.2190238982439041, -0.0025243207346647978, 0.17009945213794708, 0.05772107467055321, 0.15547960996627808, 0.008451160974800587, 0.07413522899150848, -0.013356735929846764, -0.019881488755345345, -0.011598740704357624, -0.05124952271580696, -0.031233662739396095, 0.0221539493650198, -0.07792928069829941, 0.048571985214948654, -0.12400656938552856, 0.004553573671728373, 0.04684997349977493, 0.2820889353752136, 0.032490409910678864, -0.3193526566028595, -0.09583453834056854, -0.001912693027406931, -0.06323714554309845, -0.023387599736452103, 0.032233938574790955, 0.09896223992109299, -0.07572834193706512, 0.0538797453045845, -0.07986144721508026, 0.10270402580499649, -0.03876457363367081, 0.044424109160900116, 0.06827439367771149, 0.10139644145965576, 0.005614049732685089, 0.07345929741859436, -0.3165425658226013, 0.2723497152328491, 0.004449883941560984, 0.06270064413547516, -0.07406271249055862, 0.01707313023507595, 0.02754279039800167, 0.03340395912528038, 0.067631796002388, -0.023887569084763527, -0.06422402709722519, -0.15395256876945496, -0.07072997838258743, 0.02128000743687153, 0.10018029063940048, -0.012308824807405472, 0.11400365829467773, -0.04282495751976967, 0.0017974338261410594, 0.0694587305188179, -0.0029336728621274233, -0.06805136054754257, -0.10664937645196915, 0.012098431587219238, 0.03227362036705017, -0.04054863750934601, -0.06624250113964081, -0.11151982843875885, -0.0956583246588707, 0.1705619990825653, -0.03427789360284805, -0.04580521211028099, -0.11193536221981049, 0.07590723782777786, 0.0750650092959404, -0.08781548589468002, 0.040160343050956726, 0.0013525610556825995, 0.08718224614858627, 0.02052129991352558, -0.09173212200403214, 0.11652883142232895, -0.06473710387945175, -0.17786680161952972, -0.05303249508142471, 0.12779755890369415, 0.015867331996560097, 0.0649731382727623, -0.0246312003582716, 0.014183612540364265, -0.03666013479232788, -0.08050292730331421, 0.012996288016438484, 0.007864437066018581, 0.07169833034276962, 0.0007786322967149317, -0.06979358941316605, 0.003303206292912364, -0.06181299313902855, -0.03957941010594368, 0.19281525909900665, 0.22161921858787537, -0.08764229714870453, 0.04631498083472252, 0.03150693327188492, -0.07596632093191147, -0.1809237003326416, 0.011169633828103542, 0.06028871610760689, 0.004642752464860678, 0.018349921330809593, -0.18748262524604797, 0.07177360355854034, 0.09691809117794037, -0.003659474663436413, 0.09257175773382187, -0.3483312427997589, -0.13846953213214874, 0.10489019006490707, 0.1383417844772339, 0.08988142013549805, -0.1581147462129593, -0.025414153933525085, -0.014400533400475979, -0.11019179970026016, 0.12966777384281158, -0.09501716494560242, 0.12220399081707001, -0.02902134135365486, 0.09671389311552048, 0.010922052897512913, -0.05411289259791374, 0.09262816607952118, -0.016036048531532288, 0.0754116028547287, -0.06883502751588821, 0.016398794949054718, 0.055385489016771317, -0.04756981506943703, 0.031636178493499756, -0.09747293591499329, 0.03171762824058533, -0.08999159187078476, -0.028688043355941772, -0.07292381674051285, 0.027399513870477676, -0.03748006001114845, -0.05711767077445984, -0.0383770577609539, 0.0039001954719424248, 0.07311125844717026, -0.015877295285463333, 0.15198276937007904, 0.00000899718543223571, 0.15947996079921722, 0.12020239233970642, 0.09504806995391846, -0.05893948674201965, -0.05834357440471649, -0.006569803226739168, -0.020636118948459625, 0.05274699255824089, -0.14446239173412323, 0.025033049285411835, 0.15100868046283722, 0.01405982207506895, 0.13777358829975128, 0.07981224358081818, -0.04413124918937683, 0.00827434379607439, 0.05664626136422157, -0.16655807197093964, -0.11572705954313278, -0.014058628119528294, -0.026674563065171242, -0.10813114792108536, 0.040123067796230316, 0.11327245831489563, -0.073194719851017, -0.006334878038614988, 0.005280733574181795, 0.013525931164622307, -0.03798902779817581, 0.17906631529331207, 0.029950255528092384, 0.04638643190264702, -0.0938672348856926, 0.08184632658958435, 0.04741538316011429, -0.12192073464393616, 0.035292934626340866, 0.12413190305233002, -0.0786174014210701, -0.04104587435722351, 0.06406554579734802, 0.16084440052509308, -0.0561397559940815, -0.05026200786232948, -0.13764561712741852, -0.12562040984630585, 0.1049945205450058, 0.18362648785114288, 0.07349322736263275, 0.011823346838355064, -0.057436514645814896, 0.017766080796718597, -0.11880122870206833, 0.10590046644210815, 0.04193080961704254, 0.061493344604969025, -0.12190298736095428, 0.16019462049007416, 0.012135082855820656, 0.033096931874752045, -0.020632507279515266, 0.02850898541510105, -0.09297958761453629, 0.00500874686986208, -0.13116006553173065, -0.011053605936467648, -0.02873915806412697, -0.0021040525753051043, -0.008743512444198132, -0.039798080921173096, -0.06608514487743378, 0.019103234633803368, -0.106943778693676, -0.03293268382549286, 0.008849622681736946, 0.051718760281801224, -0.11371468007564545, -0.02366635948419571, 0.014356747269630432, -0.07551728934049606, 0.08233707398176193, 0.047079745680093765, 0.001286335987970233, 0.05061236768960953, -0.11951661854982376, 0.016410095617175102, 0.05972582474350929, 0.023433901369571686, 0.04779763147234917, -0.09356691688299179, -0.0041692377999424934, 0.001841875957325101, 0.030438091605901718, 0.013103047385811806, 0.07235664129257202, -0.13459599018096924, -0.0013596918433904648, -0.015245144255459309, -0.07079464197158813, -0.07017932087182999, 0.03954507037997246, 0.058870453387498856, 0.02090909704566002, 0.17723441123962402, -0.09214934706687927, 0.04669949412345886, -0.2205420434474945, 0.007798021659255028, 0.003780747065320611, -0.10795415192842484, -0.08578450232744217, -0.06377839297056198, 0.06636075675487518, -0.06282336264848709, 0.12327748537063599, 0.0131275849416852, 0.04582061618566513, 0.040229663252830505, -0.040760818868875504, -0.004250772297382355, 0.015156992711126804, 0.20994704961776733, 0.02512585185468197, -0.04463261738419533, 0.04308757930994034, 0.02179460972547531, 0.09661579877138138, 0.12489418685436249, 0.21694600582122803, 0.14298763871192932, 0.014387080445885658, 0.09928768873214722, 0.035198990255594254, -0.06094584986567497, -0.16915419697761536, 0.05543579161167145, -0.036076463758945465, 0.13880451023578644, -0.024507401511073112, 0.23132513463497162, 0.1052137017250061, -0.1545604020357132, 0.055927906185388565, -0.03922995924949646, -0.07576817274093628, -0.12108166515827179, -0.07842722535133362, -0.08267813920974731, -0.14525106549263, -0.006304868031293154, -0.12476762384176254, 0.05123329907655716, 0.07021632045507431, 0.02515234984457493, -0.018900301307439804, 0.1414327621459961, 0.03335326910018921, -0.0018002375727519393, 0.06924034655094147, 0.012822285294532776, -0.01474764198064804, -0.09282629191875458, -0.0786755234003067, 0.009894825518131256, -0.014927028678357601, 0.03704913333058357, -0.029248319566249847, -0.033917296677827835, 0.040545471012592316, -0.02836235612630844, -0.094931460916996, 0.017457181587815285, 0.02041895128786564, 0.06959208846092224, 0.06971155107021332, 0.012002513743937016, -0.0046219537034630775, -0.012609101831912994, 0.21433262526988983, -0.08251576870679855, -0.07761818915605545, -0.0959196537733078, 0.24777209758758545, 0.03326538950204849, -0.023167623206973076, 0.03386855870485306, -0.05904306098818779, -0.017711732536554337, 0.24465857446193695, 0.18535126745700836, -0.05242941901087761, -0.01013960875570774, 0.010733258910477161, -0.004880309570580721, -0.016577109694480896, 0.11061552166938782, 0.1471039056777954, 0.07755723595619202, -0.07625018060207367, -0.04338677227497101, -0.04754672572016716, 0.001727356924675405, -0.04957426339387894, 0.08472298085689545, 0.03432685136795044, -0.007935767993330956, -0.017704032361507416, 0.04727589711546898, -0.06101274490356445, -0.08601145446300507, 0.014239969663321972, -0.21198400855064392, -0.15579576790332794, -0.012501013465225697, 0.1099143922328949, -0.0034667314030230045, 0.05202975124120712, -0.014795253053307533, 0.0002383115643169731, 0.09384431689977646, -0.01882326230406761, -0.08798610419034958, -0.05763716623187065, 0.0822770744562149, -0.13656635582447052, 0.185106560587883, -0.03552374243736267, 0.040308497846126556, 0.13072547316551208, 0.06880321353673935, -0.07539548724889755, 0.0771835520863533, 0.047684431076049805, -0.06742265075445175, 0.033548370003700256, 0.11302043497562408, -0.031768228858709335, 0.060636065900325775, 0.05028706416487694, -0.13336050510406494, 0.014125366695225239, -0.07677320390939713, -0.05943436920642853, -0.028233028948307037, -0.03321053460240364, -0.05958905816078186, 0.12950733304023743, 0.2207465022802353, -0.03490869700908661, 0.0022883478086441755, -0.07871167361736298, -0.004025994334369898, 0.04979107901453972, 0.049898549914360046, -0.04005827009677887, -0.22587619721889496, 0.008576999418437481, 0.057376209646463394, -0.005984125193208456, -0.264218807220459, -0.09301101416349411, 0.009716426022350788, -0.05701451748609543, -0.13112474977970123, 0.08875549584627151, 0.09198290854692459, 0.04841488227248192, -0.04912622645497322, -0.05570525676012039, -0.06084301695227623, 0.16480278968811035, -0.14747744798660278, -0.07238201797008514 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-small-paraphrase-pubmed This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.4032 - Rouge2 Precision: 0.8281 - Rouge2 Recall: 0.6346 - Rouge2 Fmeasure: 0.6996 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 40 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure | |:-------------:|:-----:|:-----:|:---------------:|:----------------:|:-------------:|:---------------:| | 0.5253 | 1.0 | 663 | 0.4895 | 0.8217 | 0.6309 | 0.695 | | 0.5385 | 2.0 | 1326 | 0.4719 | 0.822 | 0.6307 | 0.6953 | | 0.5255 | 3.0 | 1989 | 0.4579 | 0.8225 | 0.631 | 0.6954 | | 0.4927 | 4.0 | 2652 | 0.4510 | 0.824 | 0.6315 | 0.6965 | | 0.484 | 5.0 | 3315 | 0.4426 | 0.8254 | 0.6323 | 0.6974 | | 0.4691 | 6.0 | 3978 | 0.4383 | 0.8241 | 0.6311 | 0.6962 | | 0.4546 | 7.0 | 4641 | 0.4319 | 0.8248 | 0.6322 | 0.6969 | | 0.4431 | 8.0 | 5304 | 0.4270 | 0.8254 | 0.633 | 0.6977 | | 0.4548 | 9.0 | 5967 | 0.4257 | 0.8257 | 0.6322 | 0.6976 | | 0.4335 | 10.0 | 6630 | 0.4241 | 0.8271 | 0.6333 | 0.6986 | | 0.4234 | 11.0 | 7293 | 0.4203 | 0.827 | 0.6341 | 0.6992 | | 0.433 | 12.0 | 7956 | 0.4185 | 0.8279 | 0.6347 | 0.6998 | | 0.4108 | 13.0 | 8619 | 0.4161 | 0.8285 | 0.6352 | 0.7004 | | 0.4101 | 14.0 | 9282 | 0.4133 | 0.8289 | 0.6356 | 0.7008 | | 0.4155 | 15.0 | 9945 | 0.4149 | 0.8279 | 0.635 | 0.6998 | | 0.3991 | 16.0 | 10608 | 0.4124 | 0.8289 | 0.6353 | 0.7005 | | 0.3962 | 17.0 | 11271 | 0.4113 | 0.829 | 0.6353 | 0.7006 | | 0.3968 | 18.0 | 11934 | 0.4114 | 0.8285 | 0.6352 | 0.7002 | | 0.3962 | 19.0 | 12597 | 0.4100 | 0.8282 | 0.6346 | 0.6998 | | 0.3771 | 20.0 | 13260 | 0.4078 | 0.829 | 0.6352 | 0.7005 | | 0.3902 | 21.0 | 13923 | 0.4083 | 0.8295 | 0.6351 | 0.7006 | | 0.3811 | 22.0 | 14586 | 0.4077 | 0.8276 | 0.6346 | 0.6995 | | 0.38 | 23.0 | 15249 | 0.4076 | 0.8281 | 0.6346 | 0.6997 | | 0.3695 | 24.0 | 15912 | 0.4059 | 0.8277 | 0.6344 | 0.6993 | | 0.3665 | 25.0 | 16575 | 0.4043 | 0.8278 | 0.6343 | 0.6992 | | 0.3728 | 26.0 | 17238 | 0.4059 | 0.8279 | 0.6346 | 0.6994 | | 0.3669 | 27.0 | 17901 | 0.4048 | 0.8271 | 0.6342 | 0.6991 | | 0.3702 | 28.0 | 18564 | 0.4058 | 0.8265 | 0.6338 | 0.6985 | | 0.3674 | 29.0 | 19227 | 0.4049 | 0.8277 | 0.6345 | 0.6993 | | 0.364 | 30.0 | 19890 | 0.4048 | 0.8273 | 0.6341 | 0.699 | | 0.3618 | 31.0 | 20553 | 0.4041 | 0.828 | 0.6349 | 0.6997 | | 0.3609 | 32.0 | 21216 | 0.4040 | 0.8275 | 0.6346 | 0.6994 | | 0.357 | 33.0 | 21879 | 0.4037 | 0.8278 | 0.6348 | 0.6996 | | 0.3638 | 34.0 | 22542 | 0.4038 | 0.8275 | 0.634 | 0.6989 | | 0.3551 | 35.0 | 23205 | 0.4035 | 0.8275 | 0.6344 | 0.6992 | | 0.358 | 36.0 | 23868 | 0.4035 | 0.8279 | 0.6347 | 0.6995 | | 0.3519 | 37.0 | 24531 | 0.4034 | 0.8277 | 0.6343 | 0.6992 | | 0.359 | 38.0 | 25194 | 0.4035 | 0.8281 | 0.6346 | 0.6996 | | 0.3542 | 39.0 | 25857 | 0.4033 | 0.8281 | 0.6346 | 0.6996 | | 0.3592 | 40.0 | 26520 | 0.4032 | 0.8281 | 0.6346 | 0.6996 | ### Framework versions - Transformers 4.12.3 - Pytorch 1.9.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "t5-small-paraphrase-pubmed", "results": []}]}
text2text-generation
gayanin/t5-small-paraphrase-pubmed
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
t5-small-paraphrase-pubmed ========================== This model is a fine-tuned version of t5-small on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 0.4032 * Rouge2 Precision: 0.8281 * Rouge2 Recall: 0.6346 * Rouge2 Fmeasure: 0.6996 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 40 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.12.3 * Pytorch 1.9.0+cu111 * Datasets 1.15.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ 67, 113, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ -0.07970647513866425, 0.05618209391832352, -0.0035211448557674885, 0.09573755413293839, 0.13439251482486725, 0.01604393869638443, 0.13040132820606232, 0.14207349717617035, -0.11404426395893097, 0.04347899556159973, 0.11448727548122406, 0.14493705332279205, 0.03565381094813347, 0.12750902771949768, -0.05548149347305298, -0.27041757106781006, 0.005191842094063759, 0.04392910376191139, -0.041267987340688705, 0.13601571321487427, 0.0827043280005455, -0.11394234746694565, 0.07784605771303177, 0.010510270483791828, -0.1648477017879486, 0.014006838202476501, 0.004427968058735132, -0.06444551795721054, 0.1397617608308792, 0.042953427881002426, 0.10778803378343582, 0.019004065543413162, 0.0670994445681572, -0.19370298087596893, 0.010933644138276577, 0.0695246234536171, -0.002024224493652582, 0.0893910676240921, 0.06763961911201477, 0.0037759514525532722, 0.16508980095386505, -0.07879544049501419, 0.055875882506370544, 0.026623399928212166, -0.11347684264183044, -0.21202772855758667, -0.08364398777484894, 0.044286902993917465, 0.07125360518693924, 0.10669179260730743, -0.011496015824377537, 0.1261020451784134, -0.054858338087797165, 0.11157506704330444, 0.24977892637252808, -0.29577717185020447, -0.06224033609032631, -0.010642322711646557, 0.037630289793014526, 0.07900764048099518, -0.08015309274196625, -0.027335336431860924, 0.035151831805706024, 0.0523827001452446, 0.13492220640182495, -0.019226429983973503, -0.0948009118437767, -0.00007149617886170745, -0.1471523642539978, -0.05248448997735977, 0.13527248799800873, 0.030327139422297478, -0.027001528069376945, -0.06306556612253189, -0.086290143430233, -0.18589968979358673, -0.03719177097082138, -0.014552822336554527, 0.042396750301122665, -0.02139066345989704, -0.0564255490899086, -0.03179588541388512, -0.10097865015268326, -0.05643310025334358, -0.07024245709180832, 0.11095584183931351, 0.05354992300271988, 0.0011953155044466257, -0.04592759534716606, 0.09748810529708862, -0.004272218327969313, -0.1359105408191681, 0.013752578757703304, 0.02925274148583412, 0.006220732815563679, -0.028572140261530876, -0.05794379860162735, -0.09798003733158112, 0.010030772536993027, 0.13240289688110352, -0.07927518337965012, 0.05555834248661995, -0.015303871594369411, 0.04185829311609268, -0.1052953228354454, 0.16563715040683746, -0.03657083958387375, -0.005941031035035849, 0.015368702821433544, 0.05598761513829231, 0.036626555025577545, -0.02462504431605339, -0.11085943132638931, 0.008543760515749454, 0.10500930994749069, 0.02141900733113289, -0.04978107661008835, 0.0750170350074768, -0.03790786862373352, -0.020713964477181435, -0.005967091768980026, -0.10705368965864182, 0.023009024560451508, -0.0002794674946926534, -0.06417135149240494, 0.005814921110868454, 0.04299601539969444, -0.0022012984845787287, -0.058519333600997925, 0.102350614964962, -0.07670111954212189, 0.018065376207232475, -0.09453786164522171, -0.1309826523065567, 0.03209294006228447, -0.076261006295681, -0.0008669015951454639, -0.10015710443258286, -0.1587812304496765, -0.012086953036487103, 0.05086781829595566, -0.03213105350732803, -0.05539444088935852, -0.05316904932260513, -0.084006667137146, 0.03278766945004463, -0.024140965193510056, 0.12758025527000427, -0.06016784533858299, 0.09923254698514938, 0.0264546237885952, 0.05907397344708443, -0.03135816007852554, 0.058520298451185226, -0.08861006051301956, 0.01691306009888649, -0.17035819590091705, 0.04763250797986984, -0.03778949752449989, 0.05268533527851105, -0.09737878292798996, -0.10552085191011429, -0.00809414591640234, -0.0026318728923797607, 0.09155330806970596, 0.08696578443050385, -0.1621638834476471, -0.08089955896139145, 0.187085822224617, -0.07997633516788483, -0.11667120456695557, 0.13191409409046173, -0.04904620721936226, 0.016333499923348427, 0.05132472142577171, 0.18084654211997986, 0.06298808753490448, -0.09667132049798965, 0.014284934848546982, -0.01233536098152399, 0.0511188767850399, -0.03936023637652397, 0.061658069491386414, -0.006255249958485365, 0.027179894968867302, 0.014478320255875587, -0.0012712603202089667, 0.054477039724588394, -0.08519433438777924, -0.08185864984989166, -0.053023673593997955, -0.06830500811338425, 0.021541286259889603, 0.05691271647810936, 0.0649256780743599, -0.10807033628225327, -0.10874702036380768, 0.05520942807197571, 0.07296762615442276, -0.08724057674407959, 0.05275844782590866, -0.06808307021856308, 0.07790038734674454, -0.029076624661684036, -0.0006249735015444458, -0.1835891306400299, -0.025952648371458054, 0.019814496859908104, -0.01861627958714962, 0.029163185507059097, 0.001983806723728776, 0.06586001813411713, 0.06195047125220299, -0.049851518124341965, -0.025172851979732513, -0.043830569833517075, -0.009892961010336876, -0.11822175979614258, -0.2014486938714981, -0.025029929354786873, -0.014790442772209644, 0.09835125505924225, -0.191099151968956, 0.04286830127239227, 0.006814300082623959, 0.08768846094608307, 0.016575083136558533, -0.00556919677183032, -0.03490392118692398, 0.08196140825748444, -0.05533399060368538, -0.048375941812992096, 0.07622221112251282, 0.015259234234690666, -0.09228192269802094, -0.005303113721311092, -0.1497860848903656, 0.13611219823360443, 0.13487353920936584, -0.10002174228429794, -0.07222650200128555, 0.0011263459455221891, -0.062444865703582764, -0.038231171667575836, -0.030713841319084167, 0.008991996757686138, 0.18739847838878632, 0.0008081691921688616, 0.162650004029274, -0.08515044301748276, -0.0572257936000824, 0.030708005651831627, -0.022270403802394867, 0.020106498152017593, 0.12760965526103973, 0.09955139458179474, -0.07106159627437592, 0.1378687024116516, 0.13609221577644348, -0.0808141902089119, 0.1503780335187912, -0.04630235210061073, -0.09189415723085403, -0.013379528187215328, -0.00024326455604750663, 0.010356050916016102, 0.0715009868144989, -0.15866462886333466, -0.0022577564232051373, 0.028105996549129486, 0.024323290213942528, 0.02742689661681652, -0.2151990681886673, -0.010028054937720299, 0.03984561562538147, -0.06185489147901535, -0.012049520388245583, -0.003524145809933543, 0.017359228804707527, 0.11673406511545181, 0.00482036080211401, -0.0671842023730278, 0.02437196858227253, -0.0022935455199331045, -0.08724077045917511, 0.19693607091903687, -0.09051427990198135, -0.17822737991809845, -0.12083493918180466, -0.07224565744400024, -0.04414281249046326, -0.00197895267046988, 0.07629862427711487, -0.078654445707798, -0.03377571702003479, -0.08970330655574799, 0.03379515931010246, -0.015554885379970074, 0.03069380111992359, 0.010514451190829277, -0.0005087167373858392, 0.06738945841789246, -0.11300215125083923, -0.016138961538672447, -0.041255123913288116, -0.05108978599309921, 0.046114541590213776, 0.031934503465890884, 0.11350259184837341, 0.1557045429944992, -0.026999326422810555, 0.018642611801624298, -0.04047708585858345, 0.19916129112243652, -0.060722049325704575, -0.010691525414586067, 0.14770998060703278, -0.010340921580791473, 0.06512372940778732, 0.11612339317798615, 0.053692761808633804, -0.07726564258337021, 0.015086568892002106, 0.040422044694423676, -0.03380247578024864, -0.24393326044082642, -0.037388723343610764, -0.06705563515424728, 0.01479910034686327, 0.09357575327157974, 0.029263248667120934, 0.0492623969912529, 0.047790899872779846, 0.015922728925943375, 0.07264577597379684, -0.014511949382722378, 0.08097585290670395, 0.1591137945652008, 0.04234624654054642, 0.13429881632328033, -0.051052577793598175, -0.05240398272871971, 0.05033881962299347, -0.009724665433168411, 0.2190238982439041, -0.0025243207346647978, 0.17009945213794708, 0.05772107467055321, 0.15547960996627808, 0.008451160974800587, 0.07413522899150848, -0.013356735929846764, -0.019881488755345345, -0.011598740704357624, -0.05124952271580696, -0.031233662739396095, 0.0221539493650198, -0.07792928069829941, 0.048571985214948654, -0.12400656938552856, 0.004553573671728373, 0.04684997349977493, 0.2820889353752136, 0.032490409910678864, -0.3193526566028595, -0.09583453834056854, -0.001912693027406931, -0.06323714554309845, -0.023387599736452103, 0.032233938574790955, 0.09896223992109299, -0.07572834193706512, 0.0538797453045845, -0.07986144721508026, 0.10270402580499649, -0.03876457363367081, 0.044424109160900116, 0.06827439367771149, 0.10139644145965576, 0.005614049732685089, 0.07345929741859436, -0.3165425658226013, 0.2723497152328491, 0.004449883941560984, 0.06270064413547516, -0.07406271249055862, 0.01707313023507595, 0.02754279039800167, 0.03340395912528038, 0.067631796002388, -0.023887569084763527, -0.06422402709722519, -0.15395256876945496, -0.07072997838258743, 0.02128000743687153, 0.10018029063940048, -0.012308824807405472, 0.11400365829467773, -0.04282495751976967, 0.0017974338261410594, 0.0694587305188179, -0.0029336728621274233, -0.06805136054754257, -0.10664937645196915, 0.012098431587219238, 0.03227362036705017, -0.04054863750934601, -0.06624250113964081, -0.11151982843875885, -0.0956583246588707, 0.1705619990825653, -0.03427789360284805, -0.04580521211028099, -0.11193536221981049, 0.07590723782777786, 0.0750650092959404, -0.08781548589468002, 0.040160343050956726, 0.0013525610556825995, 0.08718224614858627, 0.02052129991352558, -0.09173212200403214, 0.11652883142232895, -0.06473710387945175, -0.17786680161952972, -0.05303249508142471, 0.12779755890369415, 0.015867331996560097, 0.0649731382727623, -0.0246312003582716, 0.014183612540364265, -0.03666013479232788, -0.08050292730331421, 0.012996288016438484, 0.007864437066018581, 0.07169833034276962, 0.0007786322967149317, -0.06979358941316605, 0.003303206292912364, -0.06181299313902855, -0.03957941010594368, 0.19281525909900665, 0.22161921858787537, -0.08764229714870453, 0.04631498083472252, 0.03150693327188492, -0.07596632093191147, -0.1809237003326416, 0.011169633828103542, 0.06028871610760689, 0.004642752464860678, 0.018349921330809593, -0.18748262524604797, 0.07177360355854034, 0.09691809117794037, -0.003659474663436413, 0.09257175773382187, -0.3483312427997589, -0.13846953213214874, 0.10489019006490707, 0.1383417844772339, 0.08988142013549805, -0.1581147462129593, -0.025414153933525085, -0.014400533400475979, -0.11019179970026016, 0.12966777384281158, -0.09501716494560242, 0.12220399081707001, -0.02902134135365486, 0.09671389311552048, 0.010922052897512913, -0.05411289259791374, 0.09262816607952118, -0.016036048531532288, 0.0754116028547287, -0.06883502751588821, 0.016398794949054718, 0.055385489016771317, -0.04756981506943703, 0.031636178493499756, -0.09747293591499329, 0.03171762824058533, -0.08999159187078476, -0.028688043355941772, -0.07292381674051285, 0.027399513870477676, -0.03748006001114845, -0.05711767077445984, -0.0383770577609539, 0.0039001954719424248, 0.07311125844717026, -0.015877295285463333, 0.15198276937007904, 0.00000899718543223571, 0.15947996079921722, 0.12020239233970642, 0.09504806995391846, -0.05893948674201965, -0.05834357440471649, -0.006569803226739168, -0.020636118948459625, 0.05274699255824089, -0.14446239173412323, 0.025033049285411835, 0.15100868046283722, 0.01405982207506895, 0.13777358829975128, 0.07981224358081818, -0.04413124918937683, 0.00827434379607439, 0.05664626136422157, -0.16655807197093964, -0.11572705954313278, -0.014058628119528294, -0.026674563065171242, -0.10813114792108536, 0.040123067796230316, 0.11327245831489563, -0.073194719851017, -0.006334878038614988, 0.005280733574181795, 0.013525931164622307, -0.03798902779817581, 0.17906631529331207, 0.029950255528092384, 0.04638643190264702, -0.0938672348856926, 0.08184632658958435, 0.04741538316011429, -0.12192073464393616, 0.035292934626340866, 0.12413190305233002, -0.0786174014210701, -0.04104587435722351, 0.06406554579734802, 0.16084440052509308, -0.0561397559940815, -0.05026200786232948, -0.13764561712741852, -0.12562040984630585, 0.1049945205450058, 0.18362648785114288, 0.07349322736263275, 0.011823346838355064, -0.057436514645814896, 0.017766080796718597, -0.11880122870206833, 0.10590046644210815, 0.04193080961704254, 0.061493344604969025, -0.12190298736095428, 0.16019462049007416, 0.012135082855820656, 0.033096931874752045, -0.020632507279515266, 0.02850898541510105, -0.09297958761453629, 0.00500874686986208, -0.13116006553173065, -0.011053605936467648, -0.02873915806412697, -0.0021040525753051043, -0.008743512444198132, -0.039798080921173096, -0.06608514487743378, 0.019103234633803368, -0.106943778693676, -0.03293268382549286, 0.008849622681736946, 0.051718760281801224, -0.11371468007564545, -0.02366635948419571, 0.014356747269630432, -0.07551728934049606, 0.08233707398176193, 0.047079745680093765, 0.001286335987970233, 0.05061236768960953, -0.11951661854982376, 0.016410095617175102, 0.05972582474350929, 0.023433901369571686, 0.04779763147234917, -0.09356691688299179, -0.0041692377999424934, 0.001841875957325101, 0.030438091605901718, 0.013103047385811806, 0.07235664129257202, -0.13459599018096924, -0.0013596918433904648, -0.015245144255459309, -0.07079464197158813, -0.07017932087182999, 0.03954507037997246, 0.058870453387498856, 0.02090909704566002, 0.17723441123962402, -0.09214934706687927, 0.04669949412345886, -0.2205420434474945, 0.007798021659255028, 0.003780747065320611, -0.10795415192842484, -0.08578450232744217, -0.06377839297056198, 0.06636075675487518, -0.06282336264848709, 0.12327748537063599, 0.0131275849416852, 0.04582061618566513, 0.040229663252830505, -0.040760818868875504, -0.004250772297382355, 0.015156992711126804, 0.20994704961776733, 0.02512585185468197, -0.04463261738419533, 0.04308757930994034, 0.02179460972547531, 0.09661579877138138, 0.12489418685436249, 0.21694600582122803, 0.14298763871192932, 0.014387080445885658, 0.09928768873214722, 0.035198990255594254, -0.06094584986567497, -0.16915419697761536, 0.05543579161167145, -0.036076463758945465, 0.13880451023578644, -0.024507401511073112, 0.23132513463497162, 0.1052137017250061, -0.1545604020357132, 0.055927906185388565, -0.03922995924949646, -0.07576817274093628, -0.12108166515827179, -0.07842722535133362, -0.08267813920974731, -0.14525106549263, -0.006304868031293154, -0.12476762384176254, 0.05123329907655716, 0.07021632045507431, 0.02515234984457493, -0.018900301307439804, 0.1414327621459961, 0.03335326910018921, -0.0018002375727519393, 0.06924034655094147, 0.012822285294532776, -0.01474764198064804, -0.09282629191875458, -0.0786755234003067, 0.009894825518131256, -0.014927028678357601, 0.03704913333058357, -0.029248319566249847, -0.033917296677827835, 0.040545471012592316, -0.02836235612630844, -0.094931460916996, 0.017457181587815285, 0.02041895128786564, 0.06959208846092224, 0.06971155107021332, 0.012002513743937016, -0.0046219537034630775, -0.012609101831912994, 0.21433262526988983, -0.08251576870679855, -0.07761818915605545, -0.0959196537733078, 0.24777209758758545, 0.03326538950204849, -0.023167623206973076, 0.03386855870485306, -0.05904306098818779, -0.017711732536554337, 0.24465857446193695, 0.18535126745700836, -0.05242941901087761, -0.01013960875570774, 0.010733258910477161, -0.004880309570580721, -0.016577109694480896, 0.11061552166938782, 0.1471039056777954, 0.07755723595619202, -0.07625018060207367, -0.04338677227497101, -0.04754672572016716, 0.001727356924675405, -0.04957426339387894, 0.08472298085689545, 0.03432685136795044, -0.007935767993330956, -0.017704032361507416, 0.04727589711546898, -0.06101274490356445, -0.08601145446300507, 0.014239969663321972, -0.21198400855064392, -0.15579576790332794, -0.012501013465225697, 0.1099143922328949, -0.0034667314030230045, 0.05202975124120712, -0.014795253053307533, 0.0002383115643169731, 0.09384431689977646, -0.01882326230406761, -0.08798610419034958, -0.05763716623187065, 0.0822770744562149, -0.13656635582447052, 0.185106560587883, -0.03552374243736267, 0.040308497846126556, 0.13072547316551208, 0.06880321353673935, -0.07539548724889755, 0.0771835520863533, 0.047684431076049805, -0.06742265075445175, 0.033548370003700256, 0.11302043497562408, -0.031768228858709335, 0.060636065900325775, 0.05028706416487694, -0.13336050510406494, 0.014125366695225239, -0.07677320390939713, -0.05943436920642853, -0.028233028948307037, -0.03321053460240364, -0.05958905816078186, 0.12950733304023743, 0.2207465022802353, -0.03490869700908661, 0.0022883478086441755, -0.07871167361736298, -0.004025994334369898, 0.04979107901453972, 0.049898549914360046, -0.04005827009677887, -0.22587619721889496, 0.008576999418437481, 0.057376209646463394, -0.005984125193208456, -0.264218807220459, -0.09301101416349411, 0.009716426022350788, -0.05701451748609543, -0.13112474977970123, 0.08875549584627151, 0.09198290854692459, 0.04841488227248192, -0.04912622645497322, -0.05570525676012039, -0.06084301695227623, 0.16480278968811035, -0.14747744798660278, -0.07238201797008514 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-emotion This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset. It achieves the following results on the evaluation set: - Loss: 0.2180 - Accuracy: 0.923 - F1: 0.9233 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.8217 | 1.0 | 250 | 0.3137 | 0.903 | 0.8999 | | 0.2484 | 2.0 | 500 | 0.2180 | 0.923 | 0.9233 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["emotion"], "metrics": ["accuracy", "f1"], "model-index": [{"name": "distilbert-base-uncased-finetuned-emotion", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "emotion", "type": "emotion", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.923, "name": "Accuracy"}, {"type": "f1", "value": 0.9233262687967644, "name": "F1"}]}]}]}
text-classification
gbade786/distilbert-base-uncased-finetuned-emotion
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "dataset:emotion", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-emotion ========================================= This model is a fine-tuned version of distilbert-base-uncased on the emotion dataset. It achieves the following results on the evaluation set: * Loss: 0.2180 * Accuracy: 0.923 * F1: 0.9233 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 64 * eval\_batch\_size: 64 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 2 ### Training results ### Framework versions * Transformers 4.15.0 * Pytorch 1.10.0+cu111 * Datasets 1.17.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2", "### Training results", "### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.17.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2", "### Training results", "### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.17.0\n* Tokenizers 0.10.3" ]
[ 67, 98, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.17.0\n* Tokenizers 0.10.3" ]
[ -0.10559017211198807, 0.11264306306838989, -0.0026052736211568117, 0.1316038817167282, 0.16395239531993866, 0.04522356390953064, 0.11307826638221741, 0.12593752145767212, -0.08274517953395844, 0.03218365088105202, 0.10883709043264389, 0.1616138219833374, 0.023710818961262703, 0.09626796096563339, -0.05841570347547531, -0.2749282121658325, -0.013689970597624779, 0.051934823393821716, -0.015424271114170551, 0.13255959749221802, 0.09328529983758926, -0.1244332566857338, 0.09686582535505295, 0.004273436032235622, -0.17551159858703613, 0.003513350849971175, 0.003246294567361474, -0.04379364848136902, 0.1468280851840973, 0.020350197330117226, 0.1062086969614029, 0.0077477674931287766, 0.08351071923971176, -0.22485455870628357, 0.018106499686837196, 0.03976118192076683, 0.000556860351935029, 0.08888021856546402, 0.03686147555708885, -0.01465404499322176, 0.15519283711910248, -0.065745510160923, 0.05322560667991638, 0.02166222780942917, -0.11221104115247726, -0.2214886099100113, -0.0819297507405281, 0.04374919831752777, 0.061586588621139526, 0.11929657310247421, -0.019513610750436783, 0.12983739376068115, -0.09577435255050659, 0.09564735740423203, 0.23557506501674652, -0.2459212839603424, -0.06816665828227997, 0.020716998726129532, 0.015153581276535988, 0.04325677081942558, -0.12044794112443924, -0.038527995347976685, 0.050822459161281586, 0.04983973875641823, 0.11995408684015274, -0.032514531165361404, -0.10042685270309448, 0.00965651124715805, -0.1285530924797058, -0.04680818319320679, 0.16843664646148682, 0.06076216325163841, -0.025980258360505104, -0.055126845836639404, -0.05622677877545357, -0.1681162565946579, -0.0311527531594038, -0.01656068116426468, 0.0550268217921257, -0.015687160193920135, -0.06279774010181427, 0.009829756803810596, -0.11961845308542252, -0.04593752324581146, -0.06522849202156067, 0.1062057688832283, 0.022034084424376488, 0.007314522285014391, -0.017791900783777237, 0.10297752171754837, -0.0011767797404900193, -0.12136783450841904, 0.02087540552020073, 0.020303498953580856, 0.02800610288977623, -0.02963714301586151, -0.0692429468035698, -0.055642709136009216, -0.0034981591161340475, 0.1002635583281517, -0.06712748110294342, 0.04515835642814636, 0.04488805681467056, 0.03783594071865082, -0.0700572058558464, 0.19676558673381805, -0.034064095467329025, -0.03367667645215988, -0.01071123592555523, 0.06047074496746063, 0.02233787067234516, -0.005910118576139212, -0.12084519863128662, 0.02086716517806053, 0.08938821405172348, -0.0011196117848157883, -0.09251783043146133, 0.0812760517001152, -0.0756930410861969, -0.019283419474959373, -0.01990571618080139, -0.07603959739208221, 0.028990641236305237, 0.02057429403066635, -0.07318393141031265, 0.004666510969400406, 0.031070781871676445, 0.00842572282999754, -0.015963401645421982, 0.09091777354478836, -0.07897259294986725, 0.025106968358159065, -0.09586040675640106, -0.10381422936916351, 0.02860204130411148, -0.09549640864133835, 0.03398827463388443, -0.0923539325594902, -0.1936139166355133, -0.023878617212176323, 0.06723587960004807, -0.020858589559793472, -0.0472467802464962, -0.07276830077171326, -0.06319067627191544, 0.0186002217233181, -0.002766136545687914, 0.09782975167036057, -0.06582771986722946, 0.09141599386930466, 0.027836086228489876, 0.08264293521642685, -0.03408736363053322, 0.055807944387197495, -0.11438050121068954, 0.0034559001214802265, -0.1353979855775833, 0.05005458742380142, -0.046664875000715256, 0.07236746698617935, -0.06320130825042725, -0.11174054443836212, 0.013232680968940258, -0.007041184231638908, 0.06484793871641159, 0.11016754806041718, -0.1939847767353058, -0.09443826228380203, 0.16742154955863953, -0.06945346295833588, -0.1077503114938736, 0.12909770011901855, -0.06702858954668045, 0.06736191362142563, 0.07169003784656525, 0.17843753099441528, 0.058913156390190125, -0.07334905862808228, -0.017036985605955124, 0.01299421675503254, 0.05008311942219734, -0.03252929821610451, 0.05339537560939789, 0.02528526447713375, 0.03297878056764603, 0.03895782306790352, -0.012813988141715527, 0.07122855633497238, -0.09237135946750641, -0.1004953607916832, -0.03618174046278, -0.09172490239143372, 0.04879666864871979, 0.09336817264556885, 0.062292490154504776, -0.10711373388767242, -0.07389577478170395, 0.029286310076713562, 0.09498550742864609, -0.06431128829717636, 0.028741229325532913, -0.05952073633670807, 0.06417015194892883, 0.0028456461150199175, -0.015565698966383934, -0.1757330298423767, 0.01494549959897995, 0.006148091051727533, 0.02566537633538246, 0.005958637688308954, 0.03546323627233505, 0.06599592417478561, 0.04299850016832352, -0.05588868260383606, -0.024363212287425995, -0.046865615993738174, -0.0033683376386761665, -0.11250913143157959, -0.22324204444885254, -0.01647019572556019, -0.023582227528095245, 0.1781032830476761, -0.2108297199010849, 0.04474042356014252, -0.005936608649790287, 0.05654168128967285, 0.014749058522284031, -0.018868330866098404, -0.03266744688153267, 0.06624644994735718, -0.05379496514797211, -0.041723620146512985, 0.07800924777984619, 0.01095687784254551, -0.08746950328350067, -0.037527114152908325, -0.10513173788785934, 0.14405284821987152, 0.13149382174015045, -0.11232934147119522, -0.06894700974225998, -0.017026443034410477, -0.06565356999635696, -0.0189244132488966, -0.03795301541686058, 0.0394805446267128, 0.1949654370546341, -0.007500956300646067, 0.1385962963104248, -0.06390349566936493, -0.024583594873547554, 0.023477846756577492, -0.04374828562140465, 0.00653071328997612, 0.13479819893836975, 0.12033825367689133, -0.06005723029375076, 0.15039968490600586, 0.13567441701889038, -0.09016793221235275, 0.16381137073040009, -0.03634655103087425, -0.05900914594531059, -0.024822773411870003, -0.048643484711647034, -0.019447380676865578, 0.10666008293628693, -0.18583443760871887, -0.01165559608489275, 0.02392948791384697, 0.0005404149414971471, 0.006637858226895332, -0.2256203293800354, -0.05095876380801201, 0.048096779733896255, -0.04394644498825073, -0.006285902112722397, -0.008653471246361732, 0.005744840484112501, 0.10518727451562881, -0.0039021559059619904, -0.08448732644319534, 0.030458370223641396, -0.001102047972381115, -0.08532645553350449, 0.20413623750209808, -0.09191438555717468, -0.17333106696605682, -0.11084707826375961, -0.0697750374674797, -0.04891122505068779, 0.005643229465931654, 0.0712379738688469, -0.11663885414600372, -0.018724054098129272, -0.07814854383468628, 0.025077644735574722, 0.01236546877771616, 0.01991431601345539, 0.027560051530599594, -0.0023026540875434875, 0.04822791740298271, -0.10861843079328537, -0.019350044429302216, -0.06478270143270493, -0.048040971159935, 0.05265131592750549, 0.01994260400533676, 0.11923782527446747, 0.16831080615520477, -0.006132509093731642, 0.012047170661389828, -0.038554199039936066, 0.22607320547103882, -0.07233158499002457, -0.020518459379673004, 0.13886161148548126, -0.01101786457002163, 0.05214766785502434, 0.11381560564041138, 0.06801068037748337, -0.09092135727405548, 0.012833732180297375, 0.045043062418699265, -0.03651098534464836, -0.22116237878799438, -0.04171822592616081, -0.04950914531946182, 0.024271536618471146, 0.06821717321872711, 0.021066254004836082, 0.0479927733540535, 0.07647651433944702, 0.04087156802415848, 0.050012923777103424, -0.049643442034721375, 0.05142870172858238, 0.1310501992702484, 0.018975134938955307, 0.10166387259960175, -0.037927038967609406, -0.05215701088309288, 0.0577872209250927, -0.021185241639614105, 0.2116309553384781, 0.0013306924374774098, 0.14567583799362183, 0.05787299945950508, 0.1722598671913147, -0.030557777732610703, 0.0716283917427063, -0.015625491738319397, -0.04045099392533302, -0.031199650838971138, -0.029640333727002144, -0.06493087112903595, 0.0335986465215683, -0.05755437910556793, 0.08579656481742859, -0.13957542181015015, 0.011322961188852787, 0.06333821266889572, 0.27820372581481934, 0.02674824185669422, -0.3197600841522217, -0.11224586516618729, 0.004311186261475086, -0.036882564425468445, -0.0069759804755449295, 0.02162455953657627, 0.09255705773830414, -0.09507271647453308, 0.03591146692633629, -0.06060466915369034, 0.08438726514577866, -0.07138898223638535, 0.0637422427535057, 0.04541429504752159, 0.07373111695051193, 0.010996204800903797, 0.08788669854402542, -0.2866887152194977, 0.2671714425086975, -0.009589733555912971, 0.06033339723944664, -0.08664324879646301, -0.0006567959790118039, 0.06249820068478584, 0.06929805129766464, 0.0664716586470604, -0.007572469301521778, 0.004489549435675144, -0.1798592507839203, -0.04003551974892616, 0.03019147925078869, 0.06256838887929916, -0.035383641719818115, 0.08538605272769928, -0.02528219111263752, 0.008161421865224838, 0.07721453160047531, 0.033974796533584595, -0.04932934790849686, -0.10141760110855103, -0.011030618101358414, 0.03499726578593254, -0.05170112103223801, -0.05305099859833717, -0.12744249403476715, -0.10991018265485764, 0.14330187439918518, -0.005647995509207249, -0.02335965819656849, -0.10269445180892944, 0.08304596692323685, 0.040489356964826584, -0.08795482665300369, 0.028254425153136253, 0.008109893649816513, 0.08079451322555542, 0.021728457883000374, -0.06958382576704025, 0.1057387962937355, -0.07534609735012054, -0.17269738018512726, -0.06878910958766937, 0.09481937438249588, 0.05280325561761856, 0.07659976184368134, -0.005562508013099432, -0.010335355065762997, -0.04971720278263092, -0.08487866818904877, 0.037139248102903366, 0.032902251929044724, 0.060758426785469055, 0.015200793743133545, -0.050298888236284256, 0.004957201424986124, -0.0699288621544838, -0.03662749007344246, 0.19893009960651398, 0.23122532665729523, -0.08881437033414841, 0.03023918718099594, 0.03237885236740112, -0.07464572787284851, -0.1931084841489792, 0.05057067051529884, 0.05882105976343155, 0.010720467194914818, 0.039678867906332016, -0.1932009905576706, 0.12213222682476044, 0.08543569594621658, -0.011459185741841793, 0.09868625551462173, -0.30179840326309204, -0.1130232885479927, 0.13859358429908752, 0.1429702788591385, 0.12109575420618057, -0.14072492718696594, -0.0029403718654066324, -0.03228345885872841, -0.12321203202009201, 0.11565089970827103, -0.08540214598178864, 0.12271682173013687, -0.025011513382196426, 0.11743595451116562, 0.009247012436389923, -0.04667966067790985, 0.11675525456666946, 0.019977673888206482, 0.09897599369287491, -0.07062285393476486, -0.028732264414429665, 0.023122841492295265, -0.04019331932067871, 0.02985725924372673, -0.10084102302789688, 0.0192367322742939, -0.11847047507762909, -0.032050833106040955, -0.08900802582502365, 0.03642430156469345, -0.039857007563114166, -0.07250478863716125, -0.05153486132621765, 0.02710239589214325, 0.07554112374782562, -0.003674109233543277, 0.08479476720094681, 0.019898628816008568, 0.11462193727493286, 0.09720813482999802, 0.09680981934070587, -0.05649472400546074, -0.07261354476213455, -0.022060200572013855, -0.009932723827660084, 0.04927678033709526, -0.14903327822685242, 0.015952538698911667, 0.13865165412425995, 0.020093567669391632, 0.1662343591451645, 0.08495385199785233, -0.03860313445329666, 0.016921745613217354, 0.06068721413612366, -0.15193581581115723, -0.08758597075939178, -0.02065819315612316, -0.07010462135076523, -0.12239024043083191, 0.03171529993414879, 0.08232713490724564, -0.07343604415655136, -0.00006629877316299826, -0.01560384500771761, 0.01660032942891121, -0.04140070825815201, 0.16355851292610168, 0.04825963079929352, 0.028392326086759567, -0.10368098318576813, 0.07722143828868866, 0.020419621840119362, -0.11000621318817139, 0.02880868688225746, 0.07140245288610458, -0.07745356857776642, -0.056092482060194016, 0.06686534732580185, 0.21415922045707703, -0.060029905289411545, -0.05179377272725105, -0.14825160801410675, -0.12671586871147156, 0.0851065069437027, 0.14789752662181854, 0.11383510380983353, 0.008939139544963837, -0.08564038574695587, 0.023657387122511864, -0.11646664142608643, 0.08940546214580536, 0.0573774091899395, 0.04157112166285515, -0.13376374542713165, 0.12169450521469116, 0.01104568038135767, 0.03941497206687927, -0.02057637833058834, 0.011740822345018387, -0.09073038399219513, 0.007765982765704393, -0.11749641597270966, -0.02614375203847885, -0.039267461746931076, 0.012166643515229225, 0.001341469120234251, -0.04397457838058472, -0.04555152729153633, 0.0038994071073830128, -0.11604350060224533, -0.0161400455981493, 0.03639943152666092, 0.07709427177906036, -0.11367599666118622, -0.03741530328989029, 0.028428995981812477, -0.0646262839436531, 0.09122604131698608, 0.06303120404481888, 0.013874838128685951, 0.05793087184429169, -0.16439029574394226, 0.026330994442105293, 0.09115877002477646, 0.015774326398968697, 0.0535723976790905, -0.08209728449583054, -0.010306304320693016, -0.008861285634338856, 0.039777353405952454, 0.01680448278784752, 0.08040016144514084, -0.12696930766105652, 0.017301060259342194, 0.005590124521404505, -0.08521867543458939, -0.06823625415563583, 0.03237314894795418, 0.08137436211109161, 0.010186384432017803, 0.19773492217063904, -0.07728539407253265, 0.044470664113759995, -0.21847271919250488, 0.008044037967920303, 0.00012132310075685382, -0.10208303481340408, -0.1305316686630249, -0.07412595301866531, 0.057601526379585266, -0.05616994947195053, 0.13297711312770844, 0.0463821142911911, 0.010264740325510502, 0.010886321775615215, -0.008825544267892838, 0.02307543158531189, 0.0027402520645409822, 0.18612034618854523, 0.035631872713565826, -0.04947563633322716, 0.06266374886035919, 0.053733572363853455, 0.11885900795459747, 0.12879317998886108, 0.1965758353471756, 0.14122870564460754, 0.024902427569031715, 0.11012924462556839, 0.0329442135989666, -0.03412274643778801, -0.1490483283996582, 0.031151408329606056, -0.05289508029818535, 0.1155051738023758, -0.019113697111606598, 0.24716302752494812, 0.06406799703836441, -0.1579589694738388, 0.0632362961769104, -0.06138550862669945, -0.08040082454681396, -0.10332414507865906, -0.06329834461212158, -0.08037661761045456, -0.14411550760269165, 0.0025336265098303556, -0.13426759839057922, 0.005474053788930178, 0.09475982934236526, 0.00956919975578785, -0.04215778410434723, 0.13856680691242218, 0.014715434052050114, 0.019868774339556694, 0.09109906852245331, 0.008550958707928658, -0.06558398902416229, -0.13376803696155548, -0.05702768638730049, -0.012237808667123318, -0.016983939334750175, 0.041399698704481125, -0.05078297480940819, -0.06258729100227356, 0.025758864358067513, -0.017688779160380363, -0.10116159915924072, 0.008657144382596016, 0.0059991255402565, 0.06315010040998459, 0.04520856589078903, 0.0013192961923778057, 0.022395290434360504, 0.0016427388181909919, 0.19095507264137268, -0.07483170181512833, -0.027863137423992157, -0.10452951490879059, 0.22176119685173035, 0.022867584601044655, -0.01507311686873436, 0.033104050904512405, -0.07178811728954315, -0.004575121216475964, 0.2504042088985443, 0.20701393485069275, -0.08614639192819595, -0.005991039797663689, 0.0022048470564186573, 0.0023594223894178867, -0.048026230186223984, 0.09589774906635284, 0.15208660066127777, 0.020511502400040627, -0.09726826101541519, -0.023576488718390465, -0.05884680524468422, -0.02264833264052868, -0.017206816002726555, 0.057174116373062134, 0.06049402058124542, 0.01096320990473032, -0.043680958449840546, 0.05078990012407303, -0.08657506108283997, -0.10151320695877075, 0.07617861777544022, -0.218962162733078, -0.1548580676317215, -0.017603712156414986, 0.09968241304159164, 0.028940385207533836, 0.07209303230047226, -0.016413703560829163, -0.002903409069404006, 0.1146613210439682, -0.020544353872537613, -0.11848846077919006, -0.07315097004175186, 0.09725384414196014, -0.1285727471113205, 0.20411114394664764, -0.0641099214553833, 0.040724605321884155, 0.12453281879425049, 0.0713607668876648, -0.05386510491371155, 0.07286467403173447, 0.04819166287779808, -0.05544179677963257, 0.006526752840727568, 0.10101494193077087, -0.030234023928642273, 0.0763757973909378, 0.0484880730509758, -0.14932294189929962, 0.025702571496367455, -0.04272696375846863, -0.06709374487400055, -0.04611178860068321, -0.007201050408184528, -0.06554511189460754, 0.12083408981561661, 0.2213602513074875, -0.02479848638176918, -0.0025440179742872715, -0.07133164256811142, 0.0059598591178655624, 0.04853558540344238, 0.008797635324299335, -0.055767010897397995, -0.20521077513694763, 0.012924603186547756, 0.06720889359712601, -0.014371490105986595, -0.25709110498428345, -0.10266827791929245, 0.0026639732532203197, -0.06915947794914246, -0.08955221623182297, 0.06035466864705086, 0.06729795038700104, 0.059065282344818115, -0.04664875566959381, -0.05752357468008995, -0.0617380253970623, 0.16832412779331207, -0.139162078499794, -0.08459077030420303 ]
null
null
transformers
# Model Trained Using AutoNLP - Problem type: Summarization - Model ID: 483413089 - CO2 Emissions (in grams): 210.6348731063569 ## Validation Metrics - Loss: 1.8478657007217407 - Rouge1: 50.5981 - Rouge2: 26.2167 - RougeL: 46.0513 - RougeLsum: 46.061 - Gen Len: 13.5987 ## Usage You can use cURL to access this model: ``` $ curl -X POST -H "Authorization: Bearer YOUR_HUGGINGFACE_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoNLP"}' https://api-inference.huggingface.co/gborn/autonlp-news-summarization-483413089 ```
{"language": "en", "tags": "autonlp", "datasets": ["gborn/autonlp-data-news-summarization"], "widget": [{"text": "I love AutoNLP \ud83e\udd17"}], "co2_eq_emissions": 210.6348731063569}
text2text-generation
gborn/autonlp-news-summarization-483413089
[ "transformers", "pytorch", "pegasus", "text2text-generation", "autonlp", "en", "dataset:gborn/autonlp-data-news-summarization", "co2_eq_emissions", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #pegasus #text2text-generation #autonlp #en #dataset-gborn/autonlp-data-news-summarization #co2_eq_emissions #autotrain_compatible #endpoints_compatible #region-us
# Model Trained Using AutoNLP - Problem type: Summarization - Model ID: 483413089 - CO2 Emissions (in grams): 210.6348731063569 ## Validation Metrics - Loss: 1.8478657007217407 - Rouge1: 50.5981 - Rouge2: 26.2167 - RougeL: 46.0513 - RougeLsum: 46.061 - Gen Len: 13.5987 ## Usage You can use cURL to access this model:
[ "# Model Trained Using AutoNLP\n\n- Problem type: Summarization\n- Model ID: 483413089\n- CO2 Emissions (in grams): 210.6348731063569", "## Validation Metrics\n\n- Loss: 1.8478657007217407\n- Rouge1: 50.5981\n- Rouge2: 26.2167\n- RougeL: 46.0513\n- RougeLsum: 46.061\n- Gen Len: 13.5987", "## Usage\n\nYou can use cURL to access this model:" ]
[ "TAGS\n#transformers #pytorch #pegasus #text2text-generation #autonlp #en #dataset-gborn/autonlp-data-news-summarization #co2_eq_emissions #autotrain_compatible #endpoints_compatible #region-us \n", "# Model Trained Using AutoNLP\n\n- Problem type: Summarization\n- Model ID: 483413089\n- CO2 Emissions (in grams): 210.6348731063569", "## Validation Metrics\n\n- Loss: 1.8478657007217407\n- Rouge1: 50.5981\n- Rouge2: 26.2167\n- RougeL: 46.0513\n- RougeLsum: 46.061\n- Gen Len: 13.5987", "## Usage\n\nYou can use cURL to access this model:" ]
[ 73, 41, 55, 13 ]
[ "passage: TAGS\n#transformers #pytorch #pegasus #text2text-generation #autonlp #en #dataset-gborn/autonlp-data-news-summarization #co2_eq_emissions #autotrain_compatible #endpoints_compatible #region-us \n# Model Trained Using AutoNLP\n\n- Problem type: Summarization\n- Model ID: 483413089\n- CO2 Emissions (in grams): 210.6348731063569## Validation Metrics\n\n- Loss: 1.8478657007217407\n- Rouge1: 50.5981\n- Rouge2: 26.2167\n- RougeL: 46.0513\n- RougeLsum: 46.061\n- Gen Len: 13.5987## Usage\n\nYou can use cURL to access this model:" ]
[ -0.18963724374771118, 0.14177680015563965, -0.0015962355537340045, 0.042236339300870895, 0.042988188564777374, 0.024453485384583473, 0.0912882387638092, 0.07097578793764114, -0.0007423714268952608, 0.0212081465870142, 0.1576870083808899, 0.11553992331027985, -0.0044862451031804085, 0.17767363786697388, -0.04440278187394142, -0.17041738331317902, 0.0809633880853653, 0.04028085991740227, 0.048067834228277206, 0.12367583066225052, 0.12416387349367142, -0.07030951976776123, 0.12783533334732056, 0.08227983862161636, -0.12921637296676636, -0.02245134674012661, 0.03989275544881821, -0.07706528902053833, 0.1380751132965088, 0.11447792500257492, 0.1251063048839569, 0.08238168060779572, 0.1287820041179657, -0.08657367527484894, 0.006678940262645483, -0.03228491544723511, -0.05315898731350899, 0.1220608577132225, 0.05505913496017456, -0.03771636262536049, 0.011912861838936806, -0.016101829707622528, 0.004956198390573263, 0.03111487813293934, -0.09155871719121933, -0.0436684675514698, -0.06917550414800644, -0.05803220346570015, 0.13022209703922272, 0.1190868616104126, -0.016871647909283638, 0.2565743923187256, -0.16753605008125305, 0.026824483647942543, 0.16100317239761353, -0.17713741958141327, -0.00676005519926548, 0.12254515290260315, 0.006764431018382311, -0.1349656879901886, -0.0427745096385479, 0.10833894461393356, 0.09783534705638885, -0.017667923122644424, 0.033554717898368835, -0.0681493803858757, -0.010884964838624, 0.01693691313266754, -0.12189904600381851, -0.01874421536922455, 0.22172805666923523, 0.06976601481437683, -0.07100307941436768, 0.011923284269869328, -0.05507894232869148, -0.08670512586832047, -0.056867148727178574, -0.08807395398616791, 0.004906500689685345, -0.05912420526146889, -0.026885686442255974, 0.03735485300421715, -0.15867792069911957, -0.04517007991671562, -0.14276166260242462, 0.06105521321296692, -0.04618347808718681, 0.029849164187908173, -0.03679730370640755, 0.13470210134983063, -0.17455032467842102, -0.0753309577703476, -0.04293690249323845, -0.06342560052871704, -0.07241877913475037, -0.018178557977080345, -0.036581434309482574, 0.10113362967967987, 0.0038556335493922234, 0.19014576077461243, -0.006311364006251097, -0.035326454788446426, 0.1257343739271164, 0.019981689751148224, 0.0321187786757946, 0.14615389704704285, -0.1308596134185791, -0.08480818569660187, 0.06321217864751816, -0.06317944079637527, 0.03131905943155289, -0.054341722279787064, -0.10470618307590485, -0.09341181814670563, 0.03342786431312561, 0.017166510224342346, 0.045155830681324005, -0.005965941119939089, -0.11513855308294296, -0.018208080902695656, 0.1658620536327362, -0.010566860437393188, 0.0265651848167181, -0.026854969561100006, -0.02620634436607361, 0.07975698262453079, 0.10456864535808563, 0.05827584117650986, -0.022791927680373192, 0.09656409919261932, -0.14541415870189667, -0.0308515764772892, -0.028842521831393242, -0.06659968942403793, 0.05261996388435364, -0.06361725926399231, 0.035001832991838455, -0.19243201613426208, -0.11082212626934052, 0.019700873643159866, -0.03380009904503822, -0.04571076110005379, -0.0784144401550293, -0.05868011340498924, -0.02966317906975746, 0.0474284403026104, 0.011150781996548176, 0.007189470808953047, -0.0551285557448864, -0.006074067205190659, -0.005509498063474894, 0.0427294597029686, -0.16832786798477173, 0.00937036145478487, -0.0877927914261818, 0.006161322817206383, -0.09710833430290222, 0.03868245333433151, 0.013707509264349937, -0.03220823034644127, -0.13027670979499817, -0.056602299213409424, 0.026923321187496185, -0.011484252288937569, 0.10276799649000168, 0.2258668690919876, -0.07362871617078781, -0.06726699322462082, 0.05048443004488945, -0.06117206811904907, -0.0734144002199173, 0.08123406022787094, -0.029648926109075546, 0.013675234280526638, 0.040325827896595, -0.05803290754556656, 0.10261386632919312, -0.1241023987531662, -0.030866915360093117, 0.07498861849308014, -0.04076885059475899, -0.12966828048229218, 0.09160885214805603, -0.00928440596908331, -0.19277122616767883, -0.003014716086909175, 0.02772432006895542, 0.04368509724736214, -0.11223310232162476, -0.1071925237774849, -0.022839834913611412, 0.008146259002387524, 0.0579524040222168, -0.0634547621011734, 0.0559687614440918, -0.00015845482994336635, -0.09527445584535599, -0.07506995648145676, 0.10979697108268738, 0.023560263216495514, 0.004924912471324205, -0.08670730143785477, 0.10290364176034927, -0.15335999429225922, -0.033204980194568634, -0.12587910890579224, -0.05929802358150482, -0.027107976377010345, 0.000002247538986921427, -0.04162058234214783, 0.040495965629816055, 0.011888467706739902, 0.05191716179251671, -0.021622398868203163, 0.013823337852954865, 0.00814215000718832, -0.012259946204721928, -0.11822880059480667, -0.12028084695339203, 0.0061048343777656555, -0.017958881333470345, 0.24616952240467072, -0.10672953724861145, -0.016891498118638992, -0.019684797152876854, 0.08387978374958038, -0.038720857352018356, 0.03844626247882843, -0.00572951789945364, -0.000806764408480376, -0.07913457602262497, 0.02140856347978115, 0.01159744430333376, -0.0061217667534947395, -0.16324783861637115, 0.10618426650762558, -0.11922166496515274, 0.11612561345100403, 0.15020808577537537, -0.06093400716781616, -0.07086090743541718, 0.0004203467397019267, -0.0028994453605264425, -0.0027419747784733772, -0.11068765074014664, -0.04530210793018341, 0.02305525355041027, -0.0017384373350068927, 0.08486873656511307, -0.06360702961683273, -0.036147184669971466, 0.09542117267847061, -0.05821855366230011, 0.006220792420208454, 0.12638415396213531, 0.193215474486351, -0.1302652209997177, 0.07761168479919434, 0.12891604006290436, -0.08962095528841019, 0.0012033721432089806, 0.049445994198322296, -0.07147835940122604, -0.03820377588272095, -0.12430260330438614, 0.01381267886608839, 0.11066671460866928, -0.013269813731312752, 0.10506916791200638, 0.09542170912027359, -0.027085738256573677, 0.004500306211411953, -0.14404772222042084, -0.04914841428399086, 0.011436975561082363, 0.028106248006224632, -0.10923504084348679, 0.06601661443710327, -0.011592480354011059, 0.13957230746746063, -0.002751081483438611, -0.13709338009357452, 0.029148908331990242, 0.036452505737543106, -0.1517365723848343, 0.27967822551727295, -0.06725998222827911, -0.28140124678611755, -0.12560153007507324, 0.009631345979869366, -0.001096467487514019, 0.03557294234633446, 0.06872639060020447, -0.07008076459169388, -0.1026712954044342, -0.045539818704128265, 0.002372155897319317, 0.02307223714888096, 0.09761565923690796, -0.03908976539969444, -0.06784866005182266, -0.027803149074316025, -0.104636549949646, -0.01943214051425457, -0.04493318125605583, 0.002311182441189885, 0.12999694049358368, -0.08735892921686172, 0.12202510237693787, 0.165326327085495, -0.021092213690280914, -0.030596770346164703, 0.041785188019275665, 0.2715023458003998, -0.07920283824205399, 0.04579203575849533, 0.11633323132991791, 0.05927536264061928, 0.04131226986646652, 0.09970643371343613, 0.03411347419023514, -0.06278309226036072, -0.00449585122987628, -0.003135982435196638, -0.060092899948358536, -0.2261085957288742, -0.14740800857543945, -0.007882166653871536, -0.004233391955494881, 0.034783534705638885, 0.007786792702972889, 0.13351120054721832, 0.15959054231643677, -0.029527388513088226, 0.042573802173137665, -0.0728415921330452, 0.0733153373003006, 0.12552012503147125, 0.003046804340556264, 0.14819534122943878, -0.06478734314441681, -0.11086739599704742, 0.12073525041341782, -0.07452349364757538, 0.1128670945763588, 0.1175106093287468, 0.0024294990580528975, -0.013542628847062588, 0.09144216030836105, 0.09442966431379318, 0.16800017654895782, 0.10585445165634155, -0.0660250261425972, -0.03491034731268883, -0.055270854383707047, 0.002089237328618765, 0.08093336224555969, 0.06845796853303909, -0.010733944363892078, -0.09695485234260559, 0.03752446547150612, 0.010934850201010704, 0.07373624294996262, 0.18558265268802643, -0.39892926812171936, -0.061958037316799164, -0.0008366230176761746, 0.0286983884871006, -0.07538896799087524, -0.045508865267038345, -0.036302849650382996, -0.15489445626735687, 0.0698719248175621, 0.013873419724404812, 0.08341603726148605, 0.012926386669278145, 0.020627984777092934, -0.09105553478002548, 0.008756687864661217, -0.03832653909921646, 0.08117016404867172, -0.1999719738960266, 0.2878740727901459, 0.052968792617321014, -0.0482906773686409, -0.06063799560070038, 0.00039542181184515357, 0.0024795872159302235, 0.18348519504070282, 0.1704932302236557, 0.03804326057434082, 0.035228289663791656, -0.10519816726446152, -0.22857272624969482, 0.08028123527765274, -0.026411956176161766, -0.09692848473787308, 0.03844969719648361, 0.03837030008435249, -0.07968796044588089, 0.020799431949853897, -0.03611889109015465, -0.15562103688716888, -0.06983970105648041, 0.09615296870470047, 0.09522131830453873, -0.04593909531831741, -0.0031558400951325893, -0.1322690099477768, 0.002734583104029298, 0.18432016670703888, -0.028476456180214882, -0.037576157599687576, -0.14453116059303284, 0.02830641344189644, 0.12960393726825714, -0.09441162645816803, 0.06652683764696121, -0.03099747747182846, 0.09725383669137955, -0.028931843116879463, -0.058789126574993134, 0.14436426758766174, -0.10465913265943527, -0.10726940631866455, -0.023554401472210884, 0.1452772319316864, 0.0556531697511673, 0.08251214772462845, 0.0669591873884201, 0.015902144834399223, -0.08204102516174316, -0.15727540850639343, 0.02026994340121746, 0.049632422626018524, 0.04221213608980179, 0.013252492062747478, 0.037075914442539215, -0.09831909090280533, -0.004143517930060625, 0.01634310744702816, 0.16906841099262238, 0.2201651930809021, -0.1077035665512085, 0.032737769186496735, 0.18232232332229614, -0.021961983293294907, -0.25856560468673706, -0.021223055198788643, -0.006295024883002043, 0.07414507120847702, -0.08528106659650803, -0.08602145314216614, 0.09912080317735672, 0.1272285282611847, -0.060988619923591614, 0.019622860476374626, -0.2302410900592804, -0.15096348524093628, 0.19035984575748444, -0.030329570174217224, 0.29659274220466614, -0.007753382436931133, -0.015735246241092682, -0.09318651258945465, -0.22275382280349731, 0.19269315898418427, -0.035937923938035965, 0.07252278923988342, -0.019161788746714592, 0.08010479807853699, 0.03910534828901291, -0.02967917174100876, 0.2100054919719696, 0.06284783035516739, -0.0039092316292226315, 0.017912620678544044, -0.07024896144866943, 0.05143212154507637, -0.030098501592874527, 0.10442985594272614, 0.03803879767656326, 0.05271764099597931, -0.1300104707479477, -0.03529186546802521, -0.01980910263955593, 0.14006568491458893, -0.042702820152044296, -0.06799019128084183, -0.025835681706666946, -0.008696015924215317, -0.031814221292734146, -0.040054429322481155, 0.09173287451267242, 0.0009446852491237223, -0.00503590889275074, 0.038938798010349274, 0.16397725045681, -0.06068624183535576, -0.0123527180403471, 0.016538802534341812, -0.07892992347478867, 0.10971319675445557, -0.16507191956043243, 0.05366985872387886, 0.12513068318367004, 0.0019170779269188643, 0.02579294703900814, 0.04046839475631714, -0.07273885607719421, -0.006096874363720417, 0.10870931297540665, -0.18621401488780975, 0.02763674408197403, -0.05049968883395195, 0.027649519965052605, -0.05224302038550377, 0.10396506637334824, 0.15342161059379578, -0.024162136018276215, -0.057120971381664276, 0.01097638439387083, -0.00958719290792942, -0.04651149734854698, 0.1955314576625824, 0.03821084275841713, 0.07414177805185318, -0.13202233612537384, 0.03518186882138252, 0.018689729273319244, -0.040699493139982224, -0.043763622641563416, 0.012726935558021069, -0.12114734202623367, -0.10497577488422394, -0.013102912344038486, 0.09889201074838638, -0.3400309681892395, -0.04634638503193855, -0.04035872966051102, -0.07882975786924362, 0.0711628869175911, 0.17250818014144897, 0.10030855983495712, 0.05930585414171219, 0.015376817435026169, -0.1381557583808899, -0.11005382239818573, -0.022209182381629944, 0.09666359424591064, 0.05122723430395126, -0.0385877750813961, -0.03652500733733177, -0.017687350511550903, 0.15182572603225708, -0.06256306916475296, 0.003086786950007081, -0.10629243403673172, -0.01730407029390335, -0.0832662507891655, 0.010770353488624096, -0.049226634204387665, -0.015210894867777824, -0.03978068381547928, -0.06791511923074722, -0.06700380891561508, 0.022677090018987656, -0.063674196600914, -0.007989511825144291, -0.016765154898166656, 0.03806265816092491, -0.04751582443714142, -0.022983497008681297, 0.05077306926250458, -0.019790111109614372, 0.11477116495370865, 0.10588186234235764, 0.0669347271323204, 0.06927160918712616, -0.16648472845554352, -0.009890181943774223, 0.09721744060516357, 0.019198328256607056, 0.10327368229627609, -0.13274534046649933, 0.0424983985722065, 0.027530137449502945, 0.038324274122714996, -0.004695971496403217, 0.0049495031125843525, -0.12842737138271332, 0.06023383140563965, -0.06064171344041824, -0.11620914936065674, -0.07831761240959167, -0.04341142624616623, 0.06726900488138199, 0.03501168638467789, 0.06148456037044525, 0.00002132118424924556, 0.0729856863617897, -0.08738414943218231, 0.056541819125413895, -0.08378113806247711, -0.05090751498937607, -0.10676585137844086, -0.04722943156957626, 0.06794972717761993, -0.006802418734878302, 0.14934605360031128, -0.06277845799922943, 0.16586089134216309, -0.012999498285353184, 0.0660402923822403, 0.07265118509531021, 0.0016466445522382855, 0.10752848535776138, 0.13107173144817352, 0.026911411434412003, 0.013687603175640106, 0.1227966845035553, 0.09334629029035568, 0.0009857438271865249, 0.12671422958374023, -0.07150392979383469, 0.058898940682411194, 0.18856091797351837, -0.014798665419220924, -0.11429848521947861, -0.02989109233021736, -0.05485297366976738, -0.08507625013589859, -0.005782022140920162, 0.019126122817397118, 0.020488154143095016, 0.09591595828533173, -0.0505213737487793, -0.038734544068574905, 0.013920923694968224, -0.039761364459991455, -0.23572158813476562, -0.06926290690898895, -0.13267770409584045, -0.07797348499298096, -0.02171965129673481, -0.1050262525677681, -0.04514998197555542, 0.07604385912418365, 0.07210554927587509, -0.031426455825567245, 0.04522458091378212, -0.03571275621652603, -0.03294434770941734, 0.02416578307747841, 0.03995085135102272, 0.059608153998851776, -0.06709082424640656, -0.008365041576325893, 0.025080598890781403, 0.07187748700380325, 0.001359744230285287, -0.03321896120905876, 0.03384155407547951, 0.12550611793994904, 0.022227948531508446, -0.09743505716323853, -0.06507470458745956, 0.015785569325089455, 0.05714606121182442, -0.014593956060707569, 0.014724099077284336, 0.0452946312725544, 0.01982404850423336, 0.14817744493484497, -0.060695622116327286, -0.008555373176932335, -0.12187909334897995, 0.22710368037223816, -0.037534136325120926, 0.08008977025747299, 0.018505670130252838, -0.04330557584762573, -0.010656538419425488, 0.18195532262325287, 0.185167595744133, -0.0032076204661279917, 0.023565970361232758, 0.0013793348334729671, 0.016376659274101257, 0.03912829980254173, -0.03991800174117088, 0.05315525457262993, 0.19655035436153412, -0.1395287811756134, -0.0461636558175087, -0.0168374665081501, 0.008436932228505611, 0.026859043166041374, 0.02892826497554779, -0.010040048509836197, -0.03382515534758568, -0.04483805224299431, 0.08791666477918625, -0.06559360772371292, 0.05367466062307358, 0.032844070345163345, -0.11476733535528183, -0.12975360453128815, 0.019742704927921295, -0.0686686560511589, 0.022440364584326744, 0.1066802516579628, -0.10090148448944092, -0.07289863377809525, 0.13438741862773895, 0.03179393336176872, -0.20420977473258972, -0.12299290299415588, 0.03988238796591759, 0.103300541639328, 0.17662183940410614, 0.02002067305147648, 0.1437128484249115, 0.10197199881076813, 0.047284409403800964, -0.11680013686418533, 0.07800398766994476, 0.0409194678068161, -0.06867989897727966, 0.11852189153432846, -0.012605306692421436, -0.006381961051374674, 0.06089339777827263, 0.026417480781674385, -0.15100909769535065, 0.05452575534582138, -0.08004456013441086, -0.005264665000140667, -0.04321504756808281, 0.013639538548886776, -0.08367163687944412, 0.11154770106077194, 0.0765620768070221, -0.08756458759307861, -0.0690431073307991, -0.020303945988416672, 0.11117115616798401, 0.05721338093280792, -0.13754865527153015, 0.004419086501002312, -0.12149303406476974, 0.09853630512952805, -0.011848857626318932, 0.04624149575829506, -0.09583728015422821, -0.01648629456758499, -0.0698632076382637, -0.04314102977514267, -0.02861538529396057, 0.032545167952775955, -0.018289217725396156, 0.016691157594323158, -0.02795870415866375, -0.11413033306598663, -0.004840186797082424, 0.05788051337003708, -0.08699283003807068, -0.191297709941864 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-cased-finetuned-cola This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the GLUE COLA dataset. It achieves the following results on the evaluation set: - Loss: 0.6747 - Matthews Correlation: 0.5957 The model was fine-tuned to compare [google/fnet-base](https://huggingface.co/google/fnet-base) as introduced in [this paper](https://arxiv.org/abs/2105.03824) against [bert-base-cased](https://huggingface.co/bert-base-cased). ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure This model is trained using the [run_glue](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) script. The following command was used: ```bash #!/usr/bin/bash python ../run_glue.py \\n --model_name_or_path bert-base-cased \\n --task_name cola \\n --do_train \\n --do_eval \\n --max_seq_length 512 \\n --per_device_train_batch_size 16 \\n --learning_rate 2e-5 \\n --num_train_epochs 3 \\n --output_dir bert-base-cased-finetuned-cola \\n --push_to_hub \\n --hub_strategy all_checkpoints \\n --logging_strategy epoch \\n --save_strategy epoch \\n --evaluation_strategy epoch \\n``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.4921 | 1.0 | 535 | 0.5283 | 0.5068 | | 0.2837 | 2.0 | 1070 | 0.5133 | 0.5521 | | 0.1775 | 3.0 | 1605 | 0.6747 | 0.5957 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer", "fnet-bert-base-comparison"], "datasets": ["glue"], "metrics": ["matthews_correlation"], "model-index": [{"name": "bert-base-cased-finetuned-cola", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE COLA", "type": "glue", "args": "cola"}, "metrics": [{"type": "matthews_correlation", "value": 0.5956649094312695, "name": "Matthews Correlation"}]}]}]}
text-classification
gchhablani/bert-base-cased-finetuned-cola
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "fnet-bert-base-comparison", "en", "dataset:glue", "arxiv:2105.03824", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2105.03824" ]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
bert-base-cased-finetuned-cola ============================== This model is a fine-tuned version of bert-base-cased on the GLUE COLA dataset. It achieves the following results on the evaluation set: * Loss: 0.6747 * Matthews Correlation: 0.5957 The model was fine-tuned to compare google/fnet-base as introduced in this paper against bert-base-cased. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ This model is trained using the run\_glue script. The following command was used: ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 87, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.12161670625209808, 0.1586534082889557, -0.0035777657758444548, 0.11875102669000626, 0.11533597856760025, -0.0031049344688653946, 0.13364577293395996, 0.1597096472978592, -0.08531280606985092, 0.0648089274764061, 0.15738050639629364, 0.1568128913640976, 0.02761896513402462, 0.18884018063545227, -0.052745409309864044, -0.23438897728919983, 0.03223801031708717, 0.07453203946352005, -0.03721422329545021, 0.13727734982967377, 0.10004161298274994, -0.12226498872041702, 0.09843768924474716, 0.03329526633024216, -0.20268045365810394, -0.0181635320186615, 0.0035198922269046307, -0.08516480028629303, 0.11188160628080368, 0.0131202582269907, 0.08862993866205215, 0.031522486358881, 0.03825753182172775, -0.1487365961074829, 0.005022411700338125, 0.046756207942962646, 0.0042826407589018345, 0.11839430034160614, 0.03855927288532257, -0.01641850546002388, 0.0520799495279789, -0.0885375365614891, 0.05395820736885071, 0.028443723917007446, -0.1181475892663002, -0.26681092381477356, -0.0919037014245987, 0.06938058882951736, 0.046251799911260605, 0.07366777211427689, -0.00005182146924198605, 0.16483885049819946, -0.005088786594569683, 0.10346852988004684, 0.27087369561195374, -0.3125593364238739, -0.05947171151638031, 0.02776799350976944, 0.011863073334097862, 0.05958991125226021, -0.09307480603456497, -0.02570064552128315, 0.05127633363008499, 0.04167178273200989, 0.18514986336231232, -0.016228215768933296, -0.00003337187081342563, -0.02189422771334648, -0.1382196545600891, -0.07627017796039581, 0.20115217566490173, 0.05067397281527519, -0.05151061341166496, -0.07212591916322708, -0.07642883062362671, -0.16918812692165375, -0.03298860415816307, -0.020403943955898285, 0.046687398105859756, -0.035265833139419556, -0.06775462627410889, -0.019117340445518494, -0.07634665071964264, -0.04710827395319939, -0.038415029644966125, 0.15217410027980804, 0.04655686020851135, 0.022712748497724533, -0.025998499244451523, 0.09033547341823578, -0.015903212130069733, -0.15715816617012024, -0.005602593533694744, 0.011772151105105877, 0.037549275904893875, -0.022561458870768547, -0.036397535353899, -0.08168396353721619, 0.012048215605318546, 0.14117512106895447, -0.08264761418104172, 0.06482455879449844, 0.011765061877667904, 0.0496395118534565, -0.08314628899097443, 0.1905018389225006, -0.024279993027448654, 0.0037186690606176853, 0.02406053990125656, 0.08622891455888748, 0.04516247287392616, -0.026812154799699783, -0.11238549649715424, 0.0233257245272398, 0.14243075251579285, 0.007431734818965197, -0.039754197001457214, 0.07417906820774078, -0.059125080704689026, -0.04582800716161728, 0.04539237171411514, -0.11030158400535583, 0.0141338761895895, -0.009373044595122337, -0.08056624233722687, -0.03472790867090225, 0.02697722613811493, -0.011787989176809788, -0.032092269510030746, 0.06212129071354866, -0.09913354367017746, 0.008189168758690357, -0.06144571676850319, -0.10696113854646683, 0.015179385431110859, -0.11808769404888153, -0.001081367488950491, -0.11012967675924301, -0.13974249362945557, -0.006976991426199675, 0.053723063319921494, -0.023807602003216743, -0.0772162452340126, -0.05605267360806465, -0.07926763594150543, 0.029673025012016296, -0.0168012622743845, 0.04827882722020149, -0.06232791766524315, 0.08783834427595139, 0.04711596667766571, 0.07783091068267822, -0.03511257842183113, 0.04865354299545288, -0.0857672467827797, 0.04486498609185219, -0.20575650036334991, 0.06056206300854683, -0.057986579835414886, 0.06628572940826416, -0.1133461743593216, -0.11155343800783157, 0.0248954389244318, -0.03267720714211464, 0.0798325464129448, 0.10354934632778168, -0.14306844770908356, -0.0819404348731041, 0.1941177397966385, -0.08345154672861099, -0.13623358309268951, 0.11876475811004639, -0.04778644070029259, 0.023907052353024483, 0.06198955699801445, 0.22297649085521698, 0.0731961652636528, -0.057169198989868164, -0.026051243767142296, -0.017552388831973076, 0.051131829619407654, -0.07172699272632599, 0.08445700258016586, -0.004041696432977915, 0.028126448392868042, 0.030312296003103256, -0.035424020141363144, 0.032627593725919724, -0.08209208399057388, -0.08335229009389877, -0.05411062017083168, -0.0801575630903244, 0.06567159295082092, 0.04338371008634567, 0.08023004233837128, -0.11180076748132706, -0.09540257602930069, 0.034219104796648026, 0.08448748290538788, -0.08299247920513153, 0.045332007110118866, -0.0944017842411995, 0.12313508242368698, -0.07442712783813477, -0.005775441415607929, -0.18186262249946594, -0.004674280993640423, 0.04966714233160019, -0.028262924402952194, 0.004111959598958492, -0.025413617491722107, 0.06383029371500015, 0.051281414926052094, -0.046544935554265976, -0.04452786594629288, -0.037859585136175156, -0.007022486068308353, -0.1164112463593483, -0.17888443171977997, -0.04770522564649582, -0.03426934406161308, 0.11073228716850281, -0.15757550299167633, 0.05754205211997032, 0.0591144785284996, 0.10811145603656769, 0.02597060613334179, -0.03133917599916458, -0.009403175674378872, 0.04480119049549103, -0.04116583615541458, -0.07460780441761017, 0.07281174510717392, 0.03331552818417549, -0.09998483955860138, -0.02359735779464245, -0.11360644549131393, 0.1753949373960495, 0.12714703381061554, -0.022101160138845444, -0.05310492590069771, -0.009392490610480309, -0.06040368974208832, -0.023967178538441658, -0.010677943006157875, 0.013884914107620716, 0.17248490452766418, 0.005367063917219639, 0.17223170399665833, -0.10302618145942688, -0.05416569113731384, 0.04557802155613899, -0.0306826401501894, -0.013940786942839622, 0.10292626917362213, 0.01127760112285614, -0.10022857785224915, 0.15135939419269562, 0.1410094052553177, -0.055681660771369934, 0.12143083661794662, -0.05604437366127968, -0.047931358218193054, -0.039293039590120316, -0.0005570283392444253, 0.018249090760946274, 0.09346067160367966, -0.11466828733682632, -0.018797609955072403, 0.039005741477012634, 0.028456371277570724, 0.011234947480261326, -0.1804705411195755, 0.001110660727135837, 0.04256647825241089, -0.0585462749004364, 0.007875440642237663, -0.005216784775257111, -0.005512913689017296, 0.10080158710479736, 0.02539101243019104, -0.0727960467338562, 0.052135586738586426, 0.010471811518073082, -0.07134804874658585, 0.19735264778137207, -0.09606501460075378, -0.19630330801010132, -0.12494390457868576, -0.06156739220023155, -0.09005486220121384, 0.006349241826683283, 0.06912596523761749, -0.07998423278331757, -0.024601668119430542, -0.09020604938268661, -0.027059001848101616, -0.014606335200369358, 0.03885342925786972, 0.07165474444627762, -0.02355276234447956, 0.10255801677703857, -0.120688796043396, -0.030849842354655266, -0.03027579002082348, 0.0014742627972736955, 0.05496053025126457, 0.006049624178558588, 0.10471925884485245, 0.11509934067726135, -0.02873915433883667, 0.05114046484231949, -0.03214579448103905, 0.2341459095478058, -0.04930265247821808, -0.02795281447470188, 0.12977902591228485, -0.005955296568572521, 0.08535684645175934, 0.0899367704987526, 0.045135561376810074, -0.09218605607748032, -0.00756901316344738, 0.005072117783129215, -0.040117885917425156, -0.21103355288505554, -0.035214588046073914, -0.04211072623729706, 0.020072607323527336, 0.12289147824048996, 0.04161820560693741, 0.05459532514214516, 0.0637349858880043, 0.029591871425509453, 0.0592803992331028, -0.028413070365786552, 0.1009601503610611, 0.12848083674907684, 0.05054585635662079, 0.13528481125831604, -0.03867393359541893, -0.03375968337059021, 0.040527984499931335, 0.0007192929624579847, 0.20465855300426483, -0.014075764454901218, 0.1890448033809662, 0.04658591002225876, 0.18518702685832977, 0.010942237451672554, 0.06792156398296356, -0.0207637008279562, -0.004309804644435644, -0.017503520473837852, -0.04292628914117813, -0.060658156871795654, 0.009485099464654922, -0.04635186865925789, 0.07441109418869019, -0.12235406786203384, 0.019649730995297432, 0.060750771313905716, 0.291361004114151, 0.02242939919233322, -0.3738984167575836, -0.1097446084022522, -0.011077821254730225, -0.025019990280270576, -0.04325328394770622, 0.011022142134606838, 0.09511391818523407, -0.09566101431846619, 0.06279890984296799, -0.08686861395835876, 0.08955012261867523, -0.073078453540802, 0.03659684211015701, 0.049923211336135864, 0.09388187527656555, 0.005718675907701254, 0.057970207184553146, -0.27198654413223267, 0.2506929934024811, 0.017948467284440994, 0.05225548893213272, -0.06244153901934624, 0.014453649520874023, 0.022270919755101204, 0.0580802857875824, 0.08689375221729279, -0.0023560896515846252, -0.03474327549338341, -0.17138437926769257, -0.10784955322742462, 0.012887988239526749, 0.07278495281934738, -0.04178544878959656, 0.0889785885810852, -0.010356182232499123, 0.003321166383102536, 0.04448032006621361, -0.003272048430517316, -0.033030617982149124, -0.09141189604997635, 0.021423611789941788, 0.06062246859073639, -0.022481713443994522, -0.0803140178322792, -0.11677858233451843, -0.08531743288040161, 0.1759343147277832, -0.006428628694266081, -0.07254543155431747, -0.12288008630275726, 0.061379604041576385, 0.06829093396663666, -0.09375333040952682, 0.04493061453104019, -0.02001975290477276, 0.12447743117809296, 0.009351304732263088, -0.0690632313489914, 0.09878260642290115, -0.047859106212854385, -0.16231606900691986, -0.0379788912832737, 0.13346640765666962, 0.02778930962085724, 0.055701445788145065, -0.007568121422082186, 0.02814381755888462, -0.0244494266808033, -0.07651051878929138, 0.03854655846953392, 0.007937511429190636, 0.09600245952606201, -0.019526541233062744, -0.019290292635560036, 0.03250190243124962, -0.07478196173906326, -0.004258669447153807, 0.1966393142938614, 0.2564740777015686, -0.10714545100927353, 0.043486859649419785, 0.03290016949176788, -0.05048590898513794, -0.1577579379081726, 0.013250388205051422, 0.07308094948530197, 0.00450450275093317, 0.0024772710166871548, -0.17275027930736542, 0.055431436747312546, 0.09000243246555328, -0.01831885054707527, 0.07066475600004196, -0.29239320755004883, -0.11998265236616135, 0.10719365626573563, 0.12865246832370758, 0.10155823826789856, -0.14214134216308594, -0.050149016082286835, -0.013447506353259087, -0.13887791335582733, 0.11616533249616623, -0.0724763497710228, 0.11383344233036041, -0.046218473464250565, 0.05384330451488495, 0.0073478645645082, -0.05154034122824669, 0.12863507866859436, 0.02274453639984131, 0.07456807792186737, -0.052454277873039246, -0.012860242277383804, 0.09835619479417801, -0.07845889031887054, 0.06105131283402443, -0.09937947243452072, 0.045091770589351654, -0.12510208785533905, -0.015082119032740593, -0.07675468176603317, 0.02839023247361183, -0.028737343847751617, -0.042841967195272446, -0.05252382904291153, 0.01206414494663477, 0.07457873225212097, -0.0033069076016545296, 0.17504702508449554, 0.04161931201815605, 0.1356934756040573, 0.18485517799854279, 0.07381550967693329, -0.11268056184053421, -0.10376991331577301, -0.00886673852801323, -0.02009434811770916, 0.05579818785190582, -0.1618359386920929, 0.042422011494636536, 0.13992659747600555, 0.006874611135572195, 0.12767653167247772, 0.0688827782869339, -0.04798436909914017, 0.003899581264704466, 0.0452854298055172, -0.1783701628446579, -0.09679107367992401, -0.009332024492323399, -0.012513089925050735, -0.13734038174152374, 0.07031508535146713, 0.11065638810396194, -0.06522715091705322, -0.020677953958511353, 0.0015312001341953874, 0.001479341764934361, -0.023513123393058777, 0.17326204478740692, 0.0655597522854805, 0.06481005996465683, -0.10098203271627426, 0.0951208770275116, 0.04609247297048569, -0.07288358360528946, 0.04003791883587837, 0.05164699628949165, -0.1138383150100708, -0.02253587916493416, 0.03539628908038139, 0.1590377688407898, -0.03593802452087402, -0.046655960381031036, -0.17098021507263184, -0.10509005934000015, 0.08731745183467865, 0.1274595707654953, 0.10466798394918442, 0.020139098167419434, -0.048447709530591965, -0.006609742529690266, -0.10675366967916489, 0.09348485618829727, 0.058594346046447754, 0.07109634578227997, -0.15644440054893494, 0.1105559915304184, 0.001110950019210577, 0.0607583224773407, -0.012919686734676361, 0.001354161067865789, -0.09134423732757568, -0.0003082616603933275, -0.08251616358757019, 0.0122391851618886, -0.04241093248128891, -0.003554311813786626, -0.02001729980111122, -0.05084189772605896, -0.05258030444383621, 0.031079620122909546, -0.10778090357780457, -0.03766832500696182, 0.029691802337765694, 0.04227648675441742, -0.1072600930929184, -0.030516671016812325, 0.011589708738029003, -0.08385424315929413, 0.09281181544065475, 0.04818247631192207, -0.005318017210811377, 0.013172389008104801, -0.024976933375000954, 0.00860611442476511, 0.06548852473497391, 0.0009176498278975487, 0.05490829423069954, -0.12593981623649597, -0.005958652589470148, 0.0027340054512023926, 0.0033404123969376087, 0.022175608202815056, 0.11886222660541534, -0.1292734295129776, -0.010102422907948494, -0.019927645102143288, -0.0311695858836174, -0.07569055259227753, 0.05729673057794571, 0.10719096660614014, 0.0378112718462944, 0.1875617355108261, -0.06822221726179123, 0.018444180488586426, -0.2068903148174286, -0.0002452793996781111, 0.005979603622108698, -0.1454673707485199, -0.07120571285486221, -0.036396682262420654, 0.05290095508098602, -0.06443101167678833, 0.11634373664855957, 0.0008753861766308546, -0.010869303718209267, 0.037613291293382645, -0.03043281100690365, -0.017883090302348137, 0.007872702553868294, 0.17467574775218964, 0.01212515588849783, -0.03913277015089989, 0.11452601850032806, 0.02858971431851387, 0.10287530720233917, 0.11307211220264435, 0.16108684241771698, 0.13502591848373413, 0.025834301486611366, 0.0998968780040741, 0.025341186672449112, -0.01848791167140007, -0.19047077000141144, 0.07276903837919235, -0.03531571105122566, 0.12698744237422943, 0.008672770112752914, 0.19214610755443573, 0.09776225686073303, -0.15829795598983765, 0.05093172565102577, -0.02248046174645424, -0.08756396174430847, -0.09562913328409195, -0.09577185660600662, -0.08325570076704025, -0.15142379701137543, -0.009702731855213642, -0.11629122495651245, 0.0066182236187160015, 0.07528655976057053, -0.0008431302267126739, -0.01898612640798092, 0.14263658225536346, 0.009304924868047237, 0.001253073220141232, 0.07858073711395264, -0.006525959819555283, -0.05684375762939453, -0.07991866022348404, -0.07751739770174026, 0.003022255375981331, 0.01993376761674881, 0.05594443157315254, -0.02513107657432556, 0.00735208997502923, 0.03755185008049011, -0.028129376471042633, -0.10968209058046341, 0.010150511749088764, 0.02177766151726246, 0.051356241106987, 0.02320592664182186, 0.00956298690289259, -0.01430518552660942, -0.012962219305336475, 0.17568282783031464, -0.06385239958763123, -0.01689169742166996, -0.10847058892250061, 0.1899847388267517, 0.03989950940012932, -0.03787495940923691, 0.03915709629654884, -0.06612126529216766, -0.011634225957095623, 0.1944432258605957, 0.18756258487701416, -0.021927885711193085, -0.0002661887847352773, -0.004998415242880583, -0.01534326747059822, 0.0016246484592556953, 0.08369015157222748, 0.11981640011072159, -0.013047742657363415, -0.06906230002641678, -0.01819325052201748, -0.06562758982181549, -0.0067472620867192745, -0.04779968410730362, 0.0645025223493576, 0.02037656493484974, 0.008185519836843014, -0.05136895924806595, 0.023522306233644485, -0.034733060747385025, -0.07609675824642181, 0.038163378834724426, -0.21621042490005493, -0.15695160627365112, -0.006919717416167259, 0.04431783780455589, -0.0026687439531087875, 0.06132279336452484, -0.0016218317905440927, 0.012163124978542328, 0.0941525548696518, -0.020845945924520493, -0.08206836134195328, -0.08182139694690704, 0.08768419921398163, -0.15437477827072144, 0.20841164886951447, -0.04061716049909592, 0.04299676790833473, 0.13137340545654297, 0.04467977210879326, -0.1047448068857193, 0.03841583803296089, 0.03423437103629112, -0.00853665266185999, 0.004708867520093918, 0.09485199302434921, -0.016044387593865395, 0.07863204181194305, 0.03675774112343788, -0.09839759767055511, -0.03282250836491585, -0.06005345657467842, -0.02700952999293804, -0.04377070441842079, -0.049499623477458954, -0.05170126631855965, 0.11996490508317947, 0.18235653638839722, -0.051337070763111115, -0.025411048904061317, -0.0625823587179184, 0.016907107084989548, 0.07732902467250824, -0.01597549393773079, -0.04855141416192055, -0.2462962567806244, 0.010467390529811382, 0.05714378505945206, -0.019154073670506477, -0.23835007846355438, -0.11273518949747086, -0.0007911180146038532, -0.06173158064484596, -0.08129233121871948, 0.08054714649915695, 0.043887872248888016, 0.05588071793317795, -0.0629199966788292, 0.008542655035853386, -0.0903623104095459, 0.16071398556232452, -0.15241222083568573, -0.07646507024765015 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-cased-finetuned-mnli This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the GLUE MNLI dataset. It achieves the following results on the evaluation set: - Loss: 0.5721 - Accuracy: 0.8410 The model was fine-tuned to compare [google/fnet-base](https://huggingface.co/google/fnet-base) as introduced in [this paper](https://arxiv.org/abs/2105.03824) against [bert-base-cased](https://huggingface.co/bert-base-cased). ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure This model is trained using the [run_glue](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) script. The following command was used: ```bash #!/usr/bin/bash python ../run_glue.py \\n --model_name_or_path bert-base-cased \\n --task_name mnli \\n --do_train \\n --do_eval \\n --max_seq_length 512 \\n --per_device_train_batch_size 16 \\n --learning_rate 2e-5 \\n --num_train_epochs 3 \\n --output_dir bert-base-cased-finetuned-mnli \\n --push_to_hub \\n --hub_strategy all_checkpoints \\n --logging_strategy epoch \\n --save_strategy epoch \\n --evaluation_strategy epoch \\n``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:--------:| | 0.5323 | 1.0 | 24544 | 0.4431 | 0.8302 | | 0.3447 | 2.0 | 49088 | 0.4725 | 0.8353 | | 0.2267 | 3.0 | 73632 | 0.5887 | 0.8368 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer", "fnet-bert-base-comparison"], "datasets": ["glue"], "metrics": ["accuracy"], "model-index": [{"name": "bert-base-cased-finetuned-mnli", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE MNLI", "type": "glue", "args": "mnli"}, "metrics": [{"type": "accuracy", "value": 0.8410292921074044, "name": "Accuracy"}]}]}]}
text-classification
gchhablani/bert-base-cased-finetuned-mnli
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "fnet-bert-base-comparison", "en", "dataset:glue", "arxiv:2105.03824", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2105.03824" ]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
bert-base-cased-finetuned-mnli ============================== This model is a fine-tuned version of bert-base-cased on the GLUE MNLI dataset. It achieves the following results on the evaluation set: * Loss: 0.5721 * Accuracy: 0.8410 The model was fine-tuned to compare google/fnet-base as introduced in this paper against bert-base-cased. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ This model is trained using the run\_glue script. The following command was used: ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 87, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.12161670625209808, 0.1586534082889557, -0.0035777657758444548, 0.11875102669000626, 0.11533597856760025, -0.0031049344688653946, 0.13364577293395996, 0.1597096472978592, -0.08531280606985092, 0.0648089274764061, 0.15738050639629364, 0.1568128913640976, 0.02761896513402462, 0.18884018063545227, -0.052745409309864044, -0.23438897728919983, 0.03223801031708717, 0.07453203946352005, -0.03721422329545021, 0.13727734982967377, 0.10004161298274994, -0.12226498872041702, 0.09843768924474716, 0.03329526633024216, -0.20268045365810394, -0.0181635320186615, 0.0035198922269046307, -0.08516480028629303, 0.11188160628080368, 0.0131202582269907, 0.08862993866205215, 0.031522486358881, 0.03825753182172775, -0.1487365961074829, 0.005022411700338125, 0.046756207942962646, 0.0042826407589018345, 0.11839430034160614, 0.03855927288532257, -0.01641850546002388, 0.0520799495279789, -0.0885375365614891, 0.05395820736885071, 0.028443723917007446, -0.1181475892663002, -0.26681092381477356, -0.0919037014245987, 0.06938058882951736, 0.046251799911260605, 0.07366777211427689, -0.00005182146924198605, 0.16483885049819946, -0.005088786594569683, 0.10346852988004684, 0.27087369561195374, -0.3125593364238739, -0.05947171151638031, 0.02776799350976944, 0.011863073334097862, 0.05958991125226021, -0.09307480603456497, -0.02570064552128315, 0.05127633363008499, 0.04167178273200989, 0.18514986336231232, -0.016228215768933296, -0.00003337187081342563, -0.02189422771334648, -0.1382196545600891, -0.07627017796039581, 0.20115217566490173, 0.05067397281527519, -0.05151061341166496, -0.07212591916322708, -0.07642883062362671, -0.16918812692165375, -0.03298860415816307, -0.020403943955898285, 0.046687398105859756, -0.035265833139419556, -0.06775462627410889, -0.019117340445518494, -0.07634665071964264, -0.04710827395319939, -0.038415029644966125, 0.15217410027980804, 0.04655686020851135, 0.022712748497724533, -0.025998499244451523, 0.09033547341823578, -0.015903212130069733, -0.15715816617012024, -0.005602593533694744, 0.011772151105105877, 0.037549275904893875, -0.022561458870768547, -0.036397535353899, -0.08168396353721619, 0.012048215605318546, 0.14117512106895447, -0.08264761418104172, 0.06482455879449844, 0.011765061877667904, 0.0496395118534565, -0.08314628899097443, 0.1905018389225006, -0.024279993027448654, 0.0037186690606176853, 0.02406053990125656, 0.08622891455888748, 0.04516247287392616, -0.026812154799699783, -0.11238549649715424, 0.0233257245272398, 0.14243075251579285, 0.007431734818965197, -0.039754197001457214, 0.07417906820774078, -0.059125080704689026, -0.04582800716161728, 0.04539237171411514, -0.11030158400535583, 0.0141338761895895, -0.009373044595122337, -0.08056624233722687, -0.03472790867090225, 0.02697722613811493, -0.011787989176809788, -0.032092269510030746, 0.06212129071354866, -0.09913354367017746, 0.008189168758690357, -0.06144571676850319, -0.10696113854646683, 0.015179385431110859, -0.11808769404888153, -0.001081367488950491, -0.11012967675924301, -0.13974249362945557, -0.006976991426199675, 0.053723063319921494, -0.023807602003216743, -0.0772162452340126, -0.05605267360806465, -0.07926763594150543, 0.029673025012016296, -0.0168012622743845, 0.04827882722020149, -0.06232791766524315, 0.08783834427595139, 0.04711596667766571, 0.07783091068267822, -0.03511257842183113, 0.04865354299545288, -0.0857672467827797, 0.04486498609185219, -0.20575650036334991, 0.06056206300854683, -0.057986579835414886, 0.06628572940826416, -0.1133461743593216, -0.11155343800783157, 0.0248954389244318, -0.03267720714211464, 0.0798325464129448, 0.10354934632778168, -0.14306844770908356, -0.0819404348731041, 0.1941177397966385, -0.08345154672861099, -0.13623358309268951, 0.11876475811004639, -0.04778644070029259, 0.023907052353024483, 0.06198955699801445, 0.22297649085521698, 0.0731961652636528, -0.057169198989868164, -0.026051243767142296, -0.017552388831973076, 0.051131829619407654, -0.07172699272632599, 0.08445700258016586, -0.004041696432977915, 0.028126448392868042, 0.030312296003103256, -0.035424020141363144, 0.032627593725919724, -0.08209208399057388, -0.08335229009389877, -0.05411062017083168, -0.0801575630903244, 0.06567159295082092, 0.04338371008634567, 0.08023004233837128, -0.11180076748132706, -0.09540257602930069, 0.034219104796648026, 0.08448748290538788, -0.08299247920513153, 0.045332007110118866, -0.0944017842411995, 0.12313508242368698, -0.07442712783813477, -0.005775441415607929, -0.18186262249946594, -0.004674280993640423, 0.04966714233160019, -0.028262924402952194, 0.004111959598958492, -0.025413617491722107, 0.06383029371500015, 0.051281414926052094, -0.046544935554265976, -0.04452786594629288, -0.037859585136175156, -0.007022486068308353, -0.1164112463593483, -0.17888443171977997, -0.04770522564649582, -0.03426934406161308, 0.11073228716850281, -0.15757550299167633, 0.05754205211997032, 0.0591144785284996, 0.10811145603656769, 0.02597060613334179, -0.03133917599916458, -0.009403175674378872, 0.04480119049549103, -0.04116583615541458, -0.07460780441761017, 0.07281174510717392, 0.03331552818417549, -0.09998483955860138, -0.02359735779464245, -0.11360644549131393, 0.1753949373960495, 0.12714703381061554, -0.022101160138845444, -0.05310492590069771, -0.009392490610480309, -0.06040368974208832, -0.023967178538441658, -0.010677943006157875, 0.013884914107620716, 0.17248490452766418, 0.005367063917219639, 0.17223170399665833, -0.10302618145942688, -0.05416569113731384, 0.04557802155613899, -0.0306826401501894, -0.013940786942839622, 0.10292626917362213, 0.01127760112285614, -0.10022857785224915, 0.15135939419269562, 0.1410094052553177, -0.055681660771369934, 0.12143083661794662, -0.05604437366127968, -0.047931358218193054, -0.039293039590120316, -0.0005570283392444253, 0.018249090760946274, 0.09346067160367966, -0.11466828733682632, -0.018797609955072403, 0.039005741477012634, 0.028456371277570724, 0.011234947480261326, -0.1804705411195755, 0.001110660727135837, 0.04256647825241089, -0.0585462749004364, 0.007875440642237663, -0.005216784775257111, -0.005512913689017296, 0.10080158710479736, 0.02539101243019104, -0.0727960467338562, 0.052135586738586426, 0.010471811518073082, -0.07134804874658585, 0.19735264778137207, -0.09606501460075378, -0.19630330801010132, -0.12494390457868576, -0.06156739220023155, -0.09005486220121384, 0.006349241826683283, 0.06912596523761749, -0.07998423278331757, -0.024601668119430542, -0.09020604938268661, -0.027059001848101616, -0.014606335200369358, 0.03885342925786972, 0.07165474444627762, -0.02355276234447956, 0.10255801677703857, -0.120688796043396, -0.030849842354655266, -0.03027579002082348, 0.0014742627972736955, 0.05496053025126457, 0.006049624178558588, 0.10471925884485245, 0.11509934067726135, -0.02873915433883667, 0.05114046484231949, -0.03214579448103905, 0.2341459095478058, -0.04930265247821808, -0.02795281447470188, 0.12977902591228485, -0.005955296568572521, 0.08535684645175934, 0.0899367704987526, 0.045135561376810074, -0.09218605607748032, -0.00756901316344738, 0.005072117783129215, -0.040117885917425156, -0.21103355288505554, -0.035214588046073914, -0.04211072623729706, 0.020072607323527336, 0.12289147824048996, 0.04161820560693741, 0.05459532514214516, 0.0637349858880043, 0.029591871425509453, 0.0592803992331028, -0.028413070365786552, 0.1009601503610611, 0.12848083674907684, 0.05054585635662079, 0.13528481125831604, -0.03867393359541893, -0.03375968337059021, 0.040527984499931335, 0.0007192929624579847, 0.20465855300426483, -0.014075764454901218, 0.1890448033809662, 0.04658591002225876, 0.18518702685832977, 0.010942237451672554, 0.06792156398296356, -0.0207637008279562, -0.004309804644435644, -0.017503520473837852, -0.04292628914117813, -0.060658156871795654, 0.009485099464654922, -0.04635186865925789, 0.07441109418869019, -0.12235406786203384, 0.019649730995297432, 0.060750771313905716, 0.291361004114151, 0.02242939919233322, -0.3738984167575836, -0.1097446084022522, -0.011077821254730225, -0.025019990280270576, -0.04325328394770622, 0.011022142134606838, 0.09511391818523407, -0.09566101431846619, 0.06279890984296799, -0.08686861395835876, 0.08955012261867523, -0.073078453540802, 0.03659684211015701, 0.049923211336135864, 0.09388187527656555, 0.005718675907701254, 0.057970207184553146, -0.27198654413223267, 0.2506929934024811, 0.017948467284440994, 0.05225548893213272, -0.06244153901934624, 0.014453649520874023, 0.022270919755101204, 0.0580802857875824, 0.08689375221729279, -0.0023560896515846252, -0.03474327549338341, -0.17138437926769257, -0.10784955322742462, 0.012887988239526749, 0.07278495281934738, -0.04178544878959656, 0.0889785885810852, -0.010356182232499123, 0.003321166383102536, 0.04448032006621361, -0.003272048430517316, -0.033030617982149124, -0.09141189604997635, 0.021423611789941788, 0.06062246859073639, -0.022481713443994522, -0.0803140178322792, -0.11677858233451843, -0.08531743288040161, 0.1759343147277832, -0.006428628694266081, -0.07254543155431747, -0.12288008630275726, 0.061379604041576385, 0.06829093396663666, -0.09375333040952682, 0.04493061453104019, -0.02001975290477276, 0.12447743117809296, 0.009351304732263088, -0.0690632313489914, 0.09878260642290115, -0.047859106212854385, -0.16231606900691986, -0.0379788912832737, 0.13346640765666962, 0.02778930962085724, 0.055701445788145065, -0.007568121422082186, 0.02814381755888462, -0.0244494266808033, -0.07651051878929138, 0.03854655846953392, 0.007937511429190636, 0.09600245952606201, -0.019526541233062744, -0.019290292635560036, 0.03250190243124962, -0.07478196173906326, -0.004258669447153807, 0.1966393142938614, 0.2564740777015686, -0.10714545100927353, 0.043486859649419785, 0.03290016949176788, -0.05048590898513794, -0.1577579379081726, 0.013250388205051422, 0.07308094948530197, 0.00450450275093317, 0.0024772710166871548, -0.17275027930736542, 0.055431436747312546, 0.09000243246555328, -0.01831885054707527, 0.07066475600004196, -0.29239320755004883, -0.11998265236616135, 0.10719365626573563, 0.12865246832370758, 0.10155823826789856, -0.14214134216308594, -0.050149016082286835, -0.013447506353259087, -0.13887791335582733, 0.11616533249616623, -0.0724763497710228, 0.11383344233036041, -0.046218473464250565, 0.05384330451488495, 0.0073478645645082, -0.05154034122824669, 0.12863507866859436, 0.02274453639984131, 0.07456807792186737, -0.052454277873039246, -0.012860242277383804, 0.09835619479417801, -0.07845889031887054, 0.06105131283402443, -0.09937947243452072, 0.045091770589351654, -0.12510208785533905, -0.015082119032740593, -0.07675468176603317, 0.02839023247361183, -0.028737343847751617, -0.042841967195272446, -0.05252382904291153, 0.01206414494663477, 0.07457873225212097, -0.0033069076016545296, 0.17504702508449554, 0.04161931201815605, 0.1356934756040573, 0.18485517799854279, 0.07381550967693329, -0.11268056184053421, -0.10376991331577301, -0.00886673852801323, -0.02009434811770916, 0.05579818785190582, -0.1618359386920929, 0.042422011494636536, 0.13992659747600555, 0.006874611135572195, 0.12767653167247772, 0.0688827782869339, -0.04798436909914017, 0.003899581264704466, 0.0452854298055172, -0.1783701628446579, -0.09679107367992401, -0.009332024492323399, -0.012513089925050735, -0.13734038174152374, 0.07031508535146713, 0.11065638810396194, -0.06522715091705322, -0.020677953958511353, 0.0015312001341953874, 0.001479341764934361, -0.023513123393058777, 0.17326204478740692, 0.0655597522854805, 0.06481005996465683, -0.10098203271627426, 0.0951208770275116, 0.04609247297048569, -0.07288358360528946, 0.04003791883587837, 0.05164699628949165, -0.1138383150100708, -0.02253587916493416, 0.03539628908038139, 0.1590377688407898, -0.03593802452087402, -0.046655960381031036, -0.17098021507263184, -0.10509005934000015, 0.08731745183467865, 0.1274595707654953, 0.10466798394918442, 0.020139098167419434, -0.048447709530591965, -0.006609742529690266, -0.10675366967916489, 0.09348485618829727, 0.058594346046447754, 0.07109634578227997, -0.15644440054893494, 0.1105559915304184, 0.001110950019210577, 0.0607583224773407, -0.012919686734676361, 0.001354161067865789, -0.09134423732757568, -0.0003082616603933275, -0.08251616358757019, 0.0122391851618886, -0.04241093248128891, -0.003554311813786626, -0.02001729980111122, -0.05084189772605896, -0.05258030444383621, 0.031079620122909546, -0.10778090357780457, -0.03766832500696182, 0.029691802337765694, 0.04227648675441742, -0.1072600930929184, -0.030516671016812325, 0.011589708738029003, -0.08385424315929413, 0.09281181544065475, 0.04818247631192207, -0.005318017210811377, 0.013172389008104801, -0.024976933375000954, 0.00860611442476511, 0.06548852473497391, 0.0009176498278975487, 0.05490829423069954, -0.12593981623649597, -0.005958652589470148, 0.0027340054512023926, 0.0033404123969376087, 0.022175608202815056, 0.11886222660541534, -0.1292734295129776, -0.010102422907948494, -0.019927645102143288, -0.0311695858836174, -0.07569055259227753, 0.05729673057794571, 0.10719096660614014, 0.0378112718462944, 0.1875617355108261, -0.06822221726179123, 0.018444180488586426, -0.2068903148174286, -0.0002452793996781111, 0.005979603622108698, -0.1454673707485199, -0.07120571285486221, -0.036396682262420654, 0.05290095508098602, -0.06443101167678833, 0.11634373664855957, 0.0008753861766308546, -0.010869303718209267, 0.037613291293382645, -0.03043281100690365, -0.017883090302348137, 0.007872702553868294, 0.17467574775218964, 0.01212515588849783, -0.03913277015089989, 0.11452601850032806, 0.02858971431851387, 0.10287530720233917, 0.11307211220264435, 0.16108684241771698, 0.13502591848373413, 0.025834301486611366, 0.0998968780040741, 0.025341186672449112, -0.01848791167140007, -0.19047077000141144, 0.07276903837919235, -0.03531571105122566, 0.12698744237422943, 0.008672770112752914, 0.19214610755443573, 0.09776225686073303, -0.15829795598983765, 0.05093172565102577, -0.02248046174645424, -0.08756396174430847, -0.09562913328409195, -0.09577185660600662, -0.08325570076704025, -0.15142379701137543, -0.009702731855213642, -0.11629122495651245, 0.0066182236187160015, 0.07528655976057053, -0.0008431302267126739, -0.01898612640798092, 0.14263658225536346, 0.009304924868047237, 0.001253073220141232, 0.07858073711395264, -0.006525959819555283, -0.05684375762939453, -0.07991866022348404, -0.07751739770174026, 0.003022255375981331, 0.01993376761674881, 0.05594443157315254, -0.02513107657432556, 0.00735208997502923, 0.03755185008049011, -0.028129376471042633, -0.10968209058046341, 0.010150511749088764, 0.02177766151726246, 0.051356241106987, 0.02320592664182186, 0.00956298690289259, -0.01430518552660942, -0.012962219305336475, 0.17568282783031464, -0.06385239958763123, -0.01689169742166996, -0.10847058892250061, 0.1899847388267517, 0.03989950940012932, -0.03787495940923691, 0.03915709629654884, -0.06612126529216766, -0.011634225957095623, 0.1944432258605957, 0.18756258487701416, -0.021927885711193085, -0.0002661887847352773, -0.004998415242880583, -0.01534326747059822, 0.0016246484592556953, 0.08369015157222748, 0.11981640011072159, -0.013047742657363415, -0.06906230002641678, -0.01819325052201748, -0.06562758982181549, -0.0067472620867192745, -0.04779968410730362, 0.0645025223493576, 0.02037656493484974, 0.008185519836843014, -0.05136895924806595, 0.023522306233644485, -0.034733060747385025, -0.07609675824642181, 0.038163378834724426, -0.21621042490005493, -0.15695160627365112, -0.006919717416167259, 0.04431783780455589, -0.0026687439531087875, 0.06132279336452484, -0.0016218317905440927, 0.012163124978542328, 0.0941525548696518, -0.020845945924520493, -0.08206836134195328, -0.08182139694690704, 0.08768419921398163, -0.15437477827072144, 0.20841164886951447, -0.04061716049909592, 0.04299676790833473, 0.13137340545654297, 0.04467977210879326, -0.1047448068857193, 0.03841583803296089, 0.03423437103629112, -0.00853665266185999, 0.004708867520093918, 0.09485199302434921, -0.016044387593865395, 0.07863204181194305, 0.03675774112343788, -0.09839759767055511, -0.03282250836491585, -0.06005345657467842, -0.02700952999293804, -0.04377070441842079, -0.049499623477458954, -0.05170126631855965, 0.11996490508317947, 0.18235653638839722, -0.051337070763111115, -0.025411048904061317, -0.0625823587179184, 0.016907107084989548, 0.07732902467250824, -0.01597549393773079, -0.04855141416192055, -0.2462962567806244, 0.010467390529811382, 0.05714378505945206, -0.019154073670506477, -0.23835007846355438, -0.11273518949747086, -0.0007911180146038532, -0.06173158064484596, -0.08129233121871948, 0.08054714649915695, 0.043887872248888016, 0.05588071793317795, -0.0629199966788292, 0.008542655035853386, -0.0903623104095459, 0.16071398556232452, -0.15241222083568573, -0.07646507024765015 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-cased-finetuned-mrpc This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the GLUE MRPC dataset. It achieves the following results on the evaluation set: - Loss: 0.7132 - Accuracy: 0.8603 - F1: 0.9026 - Combined Score: 0.8814 The model was fine-tuned to compare [google/fnet-base](https://huggingface.co/google/fnet-base) as introduced in [this paper](https://arxiv.org/abs/2105.03824) against [bert-base-cased](https://huggingface.co/bert-base-cased). ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure This model is trained using the [run_glue](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) script. The following command was used: ```bash #!/usr/bin/bash python ../run_glue.py \\n --model_name_or_path bert-base-cased \\n --task_name mrpc \\n --do_train \\n --do_eval \\n --max_seq_length 512 \\n --per_device_train_batch_size 16 \\n --learning_rate 2e-5 \\n --num_train_epochs 5 \\n --output_dir bert-base-cased-finetuned-mrpc \\n --push_to_hub \\n --hub_strategy all_checkpoints \\n --logging_strategy epoch \\n --save_strategy epoch \\n --evaluation_strategy epoch \\n``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Combined Score | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:--------------:| | 0.5981 | 1.0 | 230 | 0.4580 | 0.7892 | 0.8562 | 0.8227 | | 0.3739 | 2.0 | 460 | 0.3806 | 0.8480 | 0.8942 | 0.8711 | | 0.1991 | 3.0 | 690 | 0.4879 | 0.8529 | 0.8958 | 0.8744 | | 0.1286 | 4.0 | 920 | 0.6342 | 0.8529 | 0.8986 | 0.8758 | | 0.0812 | 5.0 | 1150 | 0.7132 | 0.8603 | 0.9026 | 0.8814 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer", "fnet-bert-base-comparison"], "datasets": ["glue"], "metrics": ["accuracy", "f1"], "model-index": [{"name": "bert-base-cased-finetuned-mrpc", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE MRPC", "type": "glue", "args": "mrpc"}, "metrics": [{"type": "accuracy", "value": 0.8602941176470589, "name": "Accuracy"}, {"type": "f1", "value": 0.9025641025641027, "name": "F1"}]}]}]}
text-classification
gchhablani/bert-base-cased-finetuned-mrpc
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "fnet-bert-base-comparison", "en", "dataset:glue", "arxiv:2105.03824", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2105.03824" ]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
bert-base-cased-finetuned-mrpc ============================== This model is a fine-tuned version of bert-base-cased on the GLUE MRPC dataset. It achieves the following results on the evaluation set: * Loss: 0.7132 * Accuracy: 0.8603 * F1: 0.9026 * Combined Score: 0.8814 The model was fine-tuned to compare google/fnet-base as introduced in this paper against bert-base-cased. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ This model is trained using the run\_glue script. The following command was used: ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 5.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 87, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.12273520231246948, 0.15941818058490753, -0.0036010530311614275, 0.1185033991932869, 0.1152205765247345, -0.003181234933435917, 0.1323397010564804, 0.1594047248363495, -0.08407840877771378, 0.06581338495016098, 0.15726500749588013, 0.15550516545772552, 0.02782674878835678, 0.1900021880865097, -0.052922964096069336, -0.23428578674793243, 0.03135466203093529, 0.07575137168169022, -0.037131719291210175, 0.13710589706897736, 0.09941554814577103, -0.12357978522777557, 0.09802382439374924, 0.0332266166806221, -0.20414406061172485, -0.01862412318587303, 0.0045390985906124115, -0.08521460741758347, 0.1119740903377533, 0.013080747798085213, 0.08856029063463211, 0.031364813446998596, 0.037303902208805084, -0.14865486323833466, 0.0052331616170704365, 0.04732818529009819, 0.004134282469749451, 0.11925225704908371, 0.037881553173065186, -0.01573019288480282, 0.05293995887041092, -0.08882270753383636, 0.05314220115542412, 0.02836529351770878, -0.11822879314422607, -0.270174115896225, -0.0905454233288765, 0.07019355893135071, 0.04569752886891365, 0.073348268866539, -0.00039739516796544194, 0.16413745284080505, -0.004999269265681505, 0.10369167476892471, 0.2695405185222626, -0.31407949328422546, -0.0589446984231472, 0.02677224390208721, 0.011416743509471416, 0.06014813482761383, -0.09477435052394867, -0.02597678266465664, 0.05175976827740669, 0.039981659501791, 0.1861463189125061, -0.015982169657945633, 0.0002565350441727787, -0.022086571902036667, -0.1378646194934845, -0.07677428424358368, 0.20211897790431976, 0.051418669521808624, -0.05063481628894806, -0.07162654399871826, -0.07740622013807297, -0.16897781193256378, -0.03351329267024994, -0.02044130302965641, 0.047230564057826996, -0.03537306562066078, -0.06571278721094131, -0.017890259623527527, -0.07657642662525177, -0.04722020775079727, -0.03871086984872818, 0.15120668709278107, 0.04658329114317894, 0.022782612591981888, -0.027851099148392677, 0.0897502526640892, -0.018068252131342888, -0.1580222249031067, -0.0053072599694132805, 0.011906379833817482, 0.03783579543232918, -0.02219948172569275, -0.03605634346604347, -0.08193022012710571, 0.012833919376134872, 0.14057014882564545, -0.08208564668893814, 0.06448864191770554, 0.010675122030079365, 0.04922620579600334, -0.0835108757019043, 0.1901181936264038, -0.023104993626475334, 0.0022616602946072817, 0.02452436089515686, 0.08614446967840195, 0.044317811727523804, -0.026830751448869705, -0.11209622025489807, 0.022753268480300903, 0.14248020946979523, 0.006461338605731726, -0.03920947015285492, 0.07400575280189514, -0.058990851044654846, -0.045090094208717346, 0.04464469477534294, -0.11016041040420532, 0.014915427193045616, -0.008828986436128616, -0.08084052056074142, -0.035866234451532364, 0.026778962463140488, -0.013091707602143288, -0.03300216794013977, 0.06295855343341827, -0.10010455548763275, 0.007508574984967709, -0.06212785467505455, -0.10758529603481293, 0.014876498840749264, -0.11965788900852203, -0.0002283805952174589, -0.10969529300928116, -0.1379294991493225, -0.0071915821172297, 0.05261197313666344, -0.023271607235074043, -0.07799466699361801, -0.05656461790204048, -0.08003277331590652, 0.029941730201244354, -0.01649371162056923, 0.04931098595261574, -0.06223886460065842, 0.08887401223182678, 0.0474187433719635, 0.0786278024315834, -0.03450461104512215, 0.04901968315243721, -0.08649168163537979, 0.043723590672016144, -0.2052707076072693, 0.05972237139940262, -0.058257415890693665, 0.066720150411129, -0.11373113840818405, -0.11091870814561844, 0.023934224620461464, -0.03338725119829178, 0.07939267158508301, 0.10428828746080399, -0.14210061728954315, -0.08317490667104721, 0.19722555577754974, -0.08388625085353851, -0.13603731989860535, 0.11822353303432465, -0.04774384945631027, 0.023914366960525513, 0.06245804950594902, 0.22421255707740784, 0.07384224981069565, -0.056651849299669266, -0.02626190148293972, -0.01773103140294552, 0.05105559527873993, -0.07211291790008545, 0.08372409641742706, -0.0034855830017477274, 0.027296477928757668, 0.03050689958035946, -0.03478715568780899, 0.03305815905332565, -0.08226604759693146, -0.08269429206848145, -0.0552789568901062, -0.08013436943292618, 0.06636259704828262, 0.04476308450102806, 0.08076454699039459, -0.11116137355566025, -0.09508826583623886, 0.03346242383122444, 0.08442533016204834, -0.08160518854856491, 0.04427807778120041, -0.09438317269086838, 0.12457689642906189, -0.07490513473749161, -0.004501476418226957, -0.18299716711044312, -0.002926086075603962, 0.050260331481695175, -0.030341053381562233, 0.004212009720504284, -0.02511434070765972, 0.06390994042158127, 0.05089034512639046, -0.045860640704631805, -0.04464670270681381, -0.03862137347459793, -0.0073234583251178265, -0.1176871806383133, -0.18063300848007202, -0.046826720237731934, -0.03420059010386467, 0.11011023819446564, -0.15884621441364288, 0.057446498423814774, 0.06037832051515579, 0.1089540496468544, 0.027162769809365273, -0.031170153990387917, -0.009540140628814697, 0.04513327404856682, -0.04029618948698044, -0.07336514443159103, 0.07257451117038727, 0.03279393911361694, -0.09915802627801895, -0.024446843191981316, -0.11312498152256012, 0.17786890268325806, 0.1279948204755783, -0.022059347480535507, -0.0518190823495388, -0.008839407935738564, -0.059505634009838104, -0.024196699261665344, -0.010522348806262016, 0.015498658642172813, 0.1704971343278885, 0.00443031033501029, 0.1728103756904602, -0.10287037491798401, -0.05360252782702446, 0.04714669659733772, -0.029268475249409676, -0.012326378375291824, 0.10431425273418427, 0.01162251178175211, -0.10040758550167084, 0.1517464816570282, 0.1416548192501068, -0.05495679751038551, 0.12223485112190247, -0.05711975693702698, -0.04896477609872818, -0.03903970867395401, -0.0010107593843713403, 0.017619730904698372, 0.0949004516005516, -0.11745531111955643, -0.019590117037296295, 0.039691485464572906, 0.02760290540754795, 0.010583777911961079, -0.18035614490509033, 0.002068496774882078, 0.04255614057183266, -0.0577472522854805, 0.005149621982127428, -0.004730795975774527, -0.0052475999109447, 0.10132960975170135, 0.024889452382922173, -0.07262793928384781, 0.050802480429410934, 0.010588769800961018, -0.07018314301967621, 0.19748838245868683, -0.09537739306688309, -0.19580714404582977, -0.12369107455015182, -0.05933047831058502, -0.08906487375497818, 0.005777008831501007, 0.07029293477535248, -0.08024737238883972, -0.024308767169713974, -0.08881144970655441, -0.02706608735024929, -0.013862849213182926, 0.03982379660010338, 0.07027624547481537, -0.022989053279161453, 0.10219364613294601, -0.12064753472805023, -0.030045047402381897, -0.030844224616885185, 0.0021685869432985783, 0.05443640425801277, 0.007260463200509548, 0.10514117032289505, 0.11515062302350998, -0.029252836480736732, 0.05099339410662651, -0.03219757601618767, 0.23458045721054077, -0.05094539746642113, -0.02754887565970421, 0.12972086668014526, -0.005566699430346489, 0.08485884964466095, 0.09015154838562012, 0.04523105174303055, -0.09123662859201431, -0.008441068232059479, 0.004572180565446615, -0.03901638463139534, -0.21114090085029602, -0.03490867465734482, -0.042593251913785934, 0.020444147288799286, 0.12356333434581757, 0.04155367612838745, 0.05562235042452812, 0.06334526091814041, 0.029814587906003, 0.06015196070075035, -0.02873038686811924, 0.10066144913434982, 0.129075288772583, 0.05018611624836922, 0.13519997894763947, -0.03983941674232483, -0.03396950662136078, 0.03950963914394379, -0.00009943571058101952, 0.20626559853553772, -0.01341833733022213, 0.190015509724617, 0.04551733657717705, 0.18698570132255554, 0.011998365633189678, 0.06770443171262741, -0.020937258377671242, -0.0041463072411715984, -0.016936300322413445, -0.04249386116862297, -0.05993194878101349, 0.010172491893172264, -0.046779412776231766, 0.07449685782194138, -0.1229928657412529, 0.018185555934906006, 0.06067291647195816, 0.29110443592071533, 0.02225201576948166, -0.37252581119537354, -0.11070704460144043, -0.011908737011253834, -0.025519277900457382, -0.04423600062727928, 0.010994509793817997, 0.09587650746107101, -0.09470661729574203, 0.061872486025094986, -0.08669182658195496, 0.0905299112200737, -0.07252071052789688, 0.03704323247075081, 0.04851524531841278, 0.09357307106256485, 0.005450603552162647, 0.05804775655269623, -0.27323558926582336, 0.25258272886276245, 0.018488362431526184, 0.052274398505687714, -0.06330030411481857, 0.013847973197698593, 0.021581653505563736, 0.05988399684429169, 0.08734140545129776, -0.00242057116702199, -0.03574739396572113, -0.170226588845253, -0.10684733092784882, 0.013181532733142376, 0.07300505042076111, -0.043209258466959, 0.08819505572319031, -0.009992691688239574, 0.0025730605702847242, 0.04417675733566284, -0.005149500444531441, -0.03280147910118103, -0.09067709743976593, 0.02149878814816475, 0.06143028661608696, -0.021822771057486534, -0.08020330965518951, -0.11618226766586304, -0.08372586220502853, 0.17271628975868225, -0.011341128498315811, -0.07161374390125275, -0.12234587222337723, 0.06291403621435165, 0.06748344749212265, -0.09436832368373871, 0.045931633561849594, -0.020597733557224274, 0.12422186136245728, 0.009217459708452225, -0.06901761889457703, 0.10053239017724991, -0.04681336134672165, -0.16271468997001648, -0.037926286458969116, 0.13365566730499268, 0.027268672361969948, 0.05532421916723251, -0.008006083779036999, 0.028334422037005424, -0.025608304888010025, -0.07614213973283768, 0.03917950391769409, 0.007323337718844414, 0.09472580999135971, -0.020243126899003983, -0.018931115046143532, 0.029467495158314705, -0.07463947683572769, -0.004165452439337969, 0.19488921761512756, 0.25613391399383545, -0.10813448578119278, 0.043561406433582306, 0.033728744834661484, -0.05141662433743477, -0.1578151136636734, 0.014114146120846272, 0.07223153859376907, 0.004969296511262655, 0.001109614851884544, -0.17397440969944, 0.05568954348564148, 0.0886269137263298, -0.0185370035469532, 0.07366597652435303, -0.2923673987388611, -0.11986135691404343, 0.10729062557220459, 0.12743747234344482, 0.1041778177022934, -0.14241072535514832, -0.050853945314884186, -0.012729215435683727, -0.13892170786857605, 0.11515681445598602, -0.07430049777030945, 0.113226979970932, -0.04666992276906967, 0.05292979255318642, 0.007788409013301134, -0.05154074728488922, 0.1275826096534729, 0.02271832711994648, 0.07553572952747345, -0.05197565257549286, -0.013126295059919357, 0.09809998422861099, -0.0778113305568695, 0.060851484537124634, -0.09939137101173401, 0.045361850410699844, -0.1248331144452095, -0.014556867070496082, -0.07716278731822968, 0.029143769294023514, -0.028787555173039436, -0.04176540672779083, -0.05256284028291702, 0.01147465780377388, 0.07403411716222763, -0.0028210324235260487, 0.17671364545822144, 0.04185609146952629, 0.13671880960464478, 0.18551644682884216, 0.07253473252058029, -0.11256355047225952, -0.10412265360355377, -0.008216392248868942, -0.019994845613837242, 0.05785110220313072, -0.16473303735256195, 0.04240985959768295, 0.140421062707901, 0.008228769525885582, 0.1269446760416031, 0.06872107833623886, -0.047829966992139816, 0.0033522218000143766, 0.045486725866794586, -0.1793903410434723, -0.10016090422868729, -0.009445465169847012, -0.012818184681236744, -0.13622497022151947, 0.07125774770975113, 0.110176682472229, -0.06657290458679199, -0.02074158936738968, 0.0009991518454626203, 0.0010803861077874899, -0.022899242118000984, 0.17322438955307007, 0.06538136303424835, 0.06537455320358276, -0.10131470859050751, 0.09524796903133392, 0.047929782420396805, -0.07653660327196121, 0.03941129520535469, 0.05052856728434563, -0.11436935514211655, -0.022952113300561905, 0.035996194928884506, 0.16025100648403168, -0.034298527985811234, -0.046940721571445465, -0.17082402110099792, -0.10452842712402344, 0.08786427974700928, 0.13057833909988403, 0.10420487821102142, 0.01995844952762127, -0.04781729355454445, -0.005857453215867281, -0.10732704401016235, 0.09478437900543213, 0.05822419375181198, 0.07136266678571701, -0.15628288686275482, 0.11073862761259079, 0.001596992602571845, 0.05998965725302696, -0.013029854744672775, 0.0014995939563959837, -0.09175609797239304, -0.0005578352720476687, -0.08042430132627487, 0.012492680922150612, -0.04159858450293541, -0.0036286425311118364, -0.020429277792572975, -0.05159774795174599, -0.05193471163511276, 0.03137846663594246, -0.10849754512310028, -0.037209369242191315, 0.030224846675992012, 0.04291409254074097, -0.10771522670984268, -0.030150866135954857, 0.011944940313696861, -0.08316744118928909, 0.0927014946937561, 0.04819602519273758, -0.005198898259550333, 0.01361685898154974, -0.024452082812786102, 0.00996569823473692, 0.06474380940198898, 0.0011315169977024198, 0.054949041455984116, -0.1251956820487976, -0.004955892451107502, 0.0020057805813848972, 0.003167948219925165, 0.021521935239434242, 0.11819131672382355, -0.13000352680683136, -0.01095547154545784, -0.01931416429579258, -0.030984848737716675, -0.07567927986383438, 0.05678528547286987, 0.10669084638357162, 0.036463551223278046, 0.18688152730464935, -0.06743103265762329, 0.017906490713357925, -0.20741963386535645, 0.00013413703709375113, 0.005770268850028515, -0.14702105522155762, -0.07122698426246643, -0.036772824823856354, 0.05338381603360176, -0.06538230180740356, 0.11566061526536942, 0.001493913820013404, -0.009878304786980152, 0.0378260537981987, -0.031193168833851814, -0.020588120445609093, 0.008561814203858376, 0.17474886775016785, 0.0126771479845047, -0.03905124589800835, 0.11461129784584045, 0.02905210293829441, 0.10191743075847626, 0.11596836149692535, 0.16053743660449982, 0.13562360405921936, 0.02690793201327324, 0.09993529319763184, 0.025183329358696938, -0.01807461678981781, -0.19147835671901703, 0.07456520944833755, -0.03571565821766853, 0.128890722990036, 0.007677143905311823, 0.1920078843832016, 0.09843330085277557, -0.15698692202568054, 0.051954105496406555, -0.022671973332762718, -0.08706733584403992, -0.09513752907514572, -0.09581948071718216, -0.0837426707148552, -0.1526874601840973, -0.010047889314591885, -0.1159338429570198, 0.006987566594034433, 0.07303724437952042, -0.0017146181780844927, -0.019828321412205696, 0.14302708208560944, 0.010929194279015064, 0.0013329503126442432, 0.08076636493206024, -0.006675939541310072, -0.05830436199903488, -0.08042674511671066, -0.07730001956224442, 0.003484914544969797, 0.018351571634411812, 0.05460543558001518, -0.025369582697749138, 0.00580788915976882, 0.03771538287401199, -0.027528436854481697, -0.10909802466630936, 0.010541521944105625, 0.021785886958241463, 0.050911594182252884, 0.023273441940546036, 0.009273657575249672, -0.014442682266235352, -0.013307257555425167, 0.17752113938331604, -0.06385718286037445, -0.015753794461488724, -0.10788676887750626, 0.18873083591461182, 0.04108770191669464, -0.03788184002041817, 0.03930654376745224, -0.06607408076524734, -0.010708626359701157, 0.1941891610622406, 0.1851033717393875, -0.020034559071063995, 0.00002634760494402144, -0.0045935907401144505, -0.015512681566178799, 0.001854269066825509, 0.08318820595741272, 0.12001276016235352, -0.012533774599432945, -0.06875097006559372, -0.017599457874894142, -0.0655461773276329, -0.006694902200251818, -0.04739747196435928, 0.06480608880519867, 0.020269542932510376, 0.006782840937376022, -0.051228221505880356, 0.023403694853186607, -0.03428596630692482, -0.0766749382019043, 0.03951941058039665, -0.21706217527389526, -0.15715551376342773, -0.007559482473880053, 0.04624676704406738, -0.001891341176815331, 0.059971537441015244, -0.0012983903288841248, 0.011913720518350601, 0.09471691399812698, -0.02078905515372753, -0.08147251605987549, -0.08469844609498978, 0.08651077002286911, -0.15671049058437347, 0.20816466212272644, -0.04045833647251129, 0.04214433580636978, 0.13087542355060577, 0.0450337752699852, -0.10505655407905579, 0.03753887861967087, 0.03393945470452309, -0.00948834978044033, 0.004219671245664358, 0.09632031619548798, -0.015130841173231602, 0.07932081073522568, 0.036358702927827835, -0.0976419672369957, -0.03288884833455086, -0.06169722229242325, -0.02750725857913494, -0.043688997626304626, -0.048681341111660004, -0.050978995859622955, 0.11931037157773972, 0.18234147131443024, -0.051180802285671234, -0.025530055165290833, -0.0633423700928688, 0.01658000238239765, 0.07789728045463562, -0.014490681700408459, -0.048653412610292435, -0.24489381909370422, 0.0111764557659626, 0.058221664279699326, -0.018459200859069824, -0.2366642951965332, -0.11184651404619217, -0.0018209918634966016, -0.06137614697217941, -0.08097735792398453, 0.0797455832362175, 0.045839566737413406, 0.05556558072566986, -0.06334666907787323, 0.004335817415267229, -0.08912677317857742, 0.16122785210609436, -0.15241484344005585, -0.07673730701208115 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-cased-finetuned-qnli This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the GLUE QNLI dataset. It achieves the following results on the evaluation set: - Loss: 0.3986 - Accuracy: 0.9099 The model was fine-tuned to compare [google/fnet-base](https://huggingface.co/google/fnet-base) as introduced in [this paper](https://arxiv.org/abs/2105.03824) against [bert-base-cased](https://huggingface.co/bert-base-cased). ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure This model is trained using the [run_glue](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) script. The following command was used: ```bash #!/usr/bin/bash python ../run_glue.py \\n --model_name_or_path bert-base-cased \\n --task_name qnli \\n --do_train \\n --do_eval \\n --max_seq_length 512 \\n --per_device_train_batch_size 16 \\n --learning_rate 2e-5 \\n --num_train_epochs 3 \\n --output_dir bert-base-cased-finetuned-qnli \\n --push_to_hub \\n --hub_strategy all_checkpoints \\n --logging_strategy epoch \\n --save_strategy epoch \\n --evaluation_strategy epoch \\n``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Accuracy | Validation Loss | |:-------------:|:-----:|:-----:|:--------:|:---------------:| | 0.337 | 1.0 | 6547 | 0.9013 | 0.2448 | | 0.1971 | 2.0 | 13094 | 0.9143 | 0.2839 | | 0.1175 | 3.0 | 19641 | 0.9099 | 0.3986 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer", "fnet-bert-base-comparison"], "datasets": ["glue"], "metrics": ["accuracy"], "model-index": [{"name": "bert-base-cased-finetuned-qnli", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE QNLI", "type": "glue", "args": "qnli"}, "metrics": [{"type": "accuracy", "value": 0.9099395936298736, "name": "Accuracy"}]}]}]}
text-classification
gchhablani/bert-base-cased-finetuned-qnli
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "fnet-bert-base-comparison", "en", "dataset:glue", "arxiv:2105.03824", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2105.03824" ]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
bert-base-cased-finetuned-qnli ============================== This model is a fine-tuned version of bert-base-cased on the GLUE QNLI dataset. It achieves the following results on the evaluation set: * Loss: 0.3986 * Accuracy: 0.9099 The model was fine-tuned to compare google/fnet-base as introduced in this paper against bert-base-cased. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ This model is trained using the run\_glue script. The following command was used: ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 87, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.12161670625209808, 0.1586534082889557, -0.0035777657758444548, 0.11875102669000626, 0.11533597856760025, -0.0031049344688653946, 0.13364577293395996, 0.1597096472978592, -0.08531280606985092, 0.0648089274764061, 0.15738050639629364, 0.1568128913640976, 0.02761896513402462, 0.18884018063545227, -0.052745409309864044, -0.23438897728919983, 0.03223801031708717, 0.07453203946352005, -0.03721422329545021, 0.13727734982967377, 0.10004161298274994, -0.12226498872041702, 0.09843768924474716, 0.03329526633024216, -0.20268045365810394, -0.0181635320186615, 0.0035198922269046307, -0.08516480028629303, 0.11188160628080368, 0.0131202582269907, 0.08862993866205215, 0.031522486358881, 0.03825753182172775, -0.1487365961074829, 0.005022411700338125, 0.046756207942962646, 0.0042826407589018345, 0.11839430034160614, 0.03855927288532257, -0.01641850546002388, 0.0520799495279789, -0.0885375365614891, 0.05395820736885071, 0.028443723917007446, -0.1181475892663002, -0.26681092381477356, -0.0919037014245987, 0.06938058882951736, 0.046251799911260605, 0.07366777211427689, -0.00005182146924198605, 0.16483885049819946, -0.005088786594569683, 0.10346852988004684, 0.27087369561195374, -0.3125593364238739, -0.05947171151638031, 0.02776799350976944, 0.011863073334097862, 0.05958991125226021, -0.09307480603456497, -0.02570064552128315, 0.05127633363008499, 0.04167178273200989, 0.18514986336231232, -0.016228215768933296, -0.00003337187081342563, -0.02189422771334648, -0.1382196545600891, -0.07627017796039581, 0.20115217566490173, 0.05067397281527519, -0.05151061341166496, -0.07212591916322708, -0.07642883062362671, -0.16918812692165375, -0.03298860415816307, -0.020403943955898285, 0.046687398105859756, -0.035265833139419556, -0.06775462627410889, -0.019117340445518494, -0.07634665071964264, -0.04710827395319939, -0.038415029644966125, 0.15217410027980804, 0.04655686020851135, 0.022712748497724533, -0.025998499244451523, 0.09033547341823578, -0.015903212130069733, -0.15715816617012024, -0.005602593533694744, 0.011772151105105877, 0.037549275904893875, -0.022561458870768547, -0.036397535353899, -0.08168396353721619, 0.012048215605318546, 0.14117512106895447, -0.08264761418104172, 0.06482455879449844, 0.011765061877667904, 0.0496395118534565, -0.08314628899097443, 0.1905018389225006, -0.024279993027448654, 0.0037186690606176853, 0.02406053990125656, 0.08622891455888748, 0.04516247287392616, -0.026812154799699783, -0.11238549649715424, 0.0233257245272398, 0.14243075251579285, 0.007431734818965197, -0.039754197001457214, 0.07417906820774078, -0.059125080704689026, -0.04582800716161728, 0.04539237171411514, -0.11030158400535583, 0.0141338761895895, -0.009373044595122337, -0.08056624233722687, -0.03472790867090225, 0.02697722613811493, -0.011787989176809788, -0.032092269510030746, 0.06212129071354866, -0.09913354367017746, 0.008189168758690357, -0.06144571676850319, -0.10696113854646683, 0.015179385431110859, -0.11808769404888153, -0.001081367488950491, -0.11012967675924301, -0.13974249362945557, -0.006976991426199675, 0.053723063319921494, -0.023807602003216743, -0.0772162452340126, -0.05605267360806465, -0.07926763594150543, 0.029673025012016296, -0.0168012622743845, 0.04827882722020149, -0.06232791766524315, 0.08783834427595139, 0.04711596667766571, 0.07783091068267822, -0.03511257842183113, 0.04865354299545288, -0.0857672467827797, 0.04486498609185219, -0.20575650036334991, 0.06056206300854683, -0.057986579835414886, 0.06628572940826416, -0.1133461743593216, -0.11155343800783157, 0.0248954389244318, -0.03267720714211464, 0.0798325464129448, 0.10354934632778168, -0.14306844770908356, -0.0819404348731041, 0.1941177397966385, -0.08345154672861099, -0.13623358309268951, 0.11876475811004639, -0.04778644070029259, 0.023907052353024483, 0.06198955699801445, 0.22297649085521698, 0.0731961652636528, -0.057169198989868164, -0.026051243767142296, -0.017552388831973076, 0.051131829619407654, -0.07172699272632599, 0.08445700258016586, -0.004041696432977915, 0.028126448392868042, 0.030312296003103256, -0.035424020141363144, 0.032627593725919724, -0.08209208399057388, -0.08335229009389877, -0.05411062017083168, -0.0801575630903244, 0.06567159295082092, 0.04338371008634567, 0.08023004233837128, -0.11180076748132706, -0.09540257602930069, 0.034219104796648026, 0.08448748290538788, -0.08299247920513153, 0.045332007110118866, -0.0944017842411995, 0.12313508242368698, -0.07442712783813477, -0.005775441415607929, -0.18186262249946594, -0.004674280993640423, 0.04966714233160019, -0.028262924402952194, 0.004111959598958492, -0.025413617491722107, 0.06383029371500015, 0.051281414926052094, -0.046544935554265976, -0.04452786594629288, -0.037859585136175156, -0.007022486068308353, -0.1164112463593483, -0.17888443171977997, -0.04770522564649582, -0.03426934406161308, 0.11073228716850281, -0.15757550299167633, 0.05754205211997032, 0.0591144785284996, 0.10811145603656769, 0.02597060613334179, -0.03133917599916458, -0.009403175674378872, 0.04480119049549103, -0.04116583615541458, -0.07460780441761017, 0.07281174510717392, 0.03331552818417549, -0.09998483955860138, -0.02359735779464245, -0.11360644549131393, 0.1753949373960495, 0.12714703381061554, -0.022101160138845444, -0.05310492590069771, -0.009392490610480309, -0.06040368974208832, -0.023967178538441658, -0.010677943006157875, 0.013884914107620716, 0.17248490452766418, 0.005367063917219639, 0.17223170399665833, -0.10302618145942688, -0.05416569113731384, 0.04557802155613899, -0.0306826401501894, -0.013940786942839622, 0.10292626917362213, 0.01127760112285614, -0.10022857785224915, 0.15135939419269562, 0.1410094052553177, -0.055681660771369934, 0.12143083661794662, -0.05604437366127968, -0.047931358218193054, -0.039293039590120316, -0.0005570283392444253, 0.018249090760946274, 0.09346067160367966, -0.11466828733682632, -0.018797609955072403, 0.039005741477012634, 0.028456371277570724, 0.011234947480261326, -0.1804705411195755, 0.001110660727135837, 0.04256647825241089, -0.0585462749004364, 0.007875440642237663, -0.005216784775257111, -0.005512913689017296, 0.10080158710479736, 0.02539101243019104, -0.0727960467338562, 0.052135586738586426, 0.010471811518073082, -0.07134804874658585, 0.19735264778137207, -0.09606501460075378, -0.19630330801010132, -0.12494390457868576, -0.06156739220023155, -0.09005486220121384, 0.006349241826683283, 0.06912596523761749, -0.07998423278331757, -0.024601668119430542, -0.09020604938268661, -0.027059001848101616, -0.014606335200369358, 0.03885342925786972, 0.07165474444627762, -0.02355276234447956, 0.10255801677703857, -0.120688796043396, -0.030849842354655266, -0.03027579002082348, 0.0014742627972736955, 0.05496053025126457, 0.006049624178558588, 0.10471925884485245, 0.11509934067726135, -0.02873915433883667, 0.05114046484231949, -0.03214579448103905, 0.2341459095478058, -0.04930265247821808, -0.02795281447470188, 0.12977902591228485, -0.005955296568572521, 0.08535684645175934, 0.0899367704987526, 0.045135561376810074, -0.09218605607748032, -0.00756901316344738, 0.005072117783129215, -0.040117885917425156, -0.21103355288505554, -0.035214588046073914, -0.04211072623729706, 0.020072607323527336, 0.12289147824048996, 0.04161820560693741, 0.05459532514214516, 0.0637349858880043, 0.029591871425509453, 0.0592803992331028, -0.028413070365786552, 0.1009601503610611, 0.12848083674907684, 0.05054585635662079, 0.13528481125831604, -0.03867393359541893, -0.03375968337059021, 0.040527984499931335, 0.0007192929624579847, 0.20465855300426483, -0.014075764454901218, 0.1890448033809662, 0.04658591002225876, 0.18518702685832977, 0.010942237451672554, 0.06792156398296356, -0.0207637008279562, -0.004309804644435644, -0.017503520473837852, -0.04292628914117813, -0.060658156871795654, 0.009485099464654922, -0.04635186865925789, 0.07441109418869019, -0.12235406786203384, 0.019649730995297432, 0.060750771313905716, 0.291361004114151, 0.02242939919233322, -0.3738984167575836, -0.1097446084022522, -0.011077821254730225, -0.025019990280270576, -0.04325328394770622, 0.011022142134606838, 0.09511391818523407, -0.09566101431846619, 0.06279890984296799, -0.08686861395835876, 0.08955012261867523, -0.073078453540802, 0.03659684211015701, 0.049923211336135864, 0.09388187527656555, 0.005718675907701254, 0.057970207184553146, -0.27198654413223267, 0.2506929934024811, 0.017948467284440994, 0.05225548893213272, -0.06244153901934624, 0.014453649520874023, 0.022270919755101204, 0.0580802857875824, 0.08689375221729279, -0.0023560896515846252, -0.03474327549338341, -0.17138437926769257, -0.10784955322742462, 0.012887988239526749, 0.07278495281934738, -0.04178544878959656, 0.0889785885810852, -0.010356182232499123, 0.003321166383102536, 0.04448032006621361, -0.003272048430517316, -0.033030617982149124, -0.09141189604997635, 0.021423611789941788, 0.06062246859073639, -0.022481713443994522, -0.0803140178322792, -0.11677858233451843, -0.08531743288040161, 0.1759343147277832, -0.006428628694266081, -0.07254543155431747, -0.12288008630275726, 0.061379604041576385, 0.06829093396663666, -0.09375333040952682, 0.04493061453104019, -0.02001975290477276, 0.12447743117809296, 0.009351304732263088, -0.0690632313489914, 0.09878260642290115, -0.047859106212854385, -0.16231606900691986, -0.0379788912832737, 0.13346640765666962, 0.02778930962085724, 0.055701445788145065, -0.007568121422082186, 0.02814381755888462, -0.0244494266808033, -0.07651051878929138, 0.03854655846953392, 0.007937511429190636, 0.09600245952606201, -0.019526541233062744, -0.019290292635560036, 0.03250190243124962, -0.07478196173906326, -0.004258669447153807, 0.1966393142938614, 0.2564740777015686, -0.10714545100927353, 0.043486859649419785, 0.03290016949176788, -0.05048590898513794, -0.1577579379081726, 0.013250388205051422, 0.07308094948530197, 0.00450450275093317, 0.0024772710166871548, -0.17275027930736542, 0.055431436747312546, 0.09000243246555328, -0.01831885054707527, 0.07066475600004196, -0.29239320755004883, -0.11998265236616135, 0.10719365626573563, 0.12865246832370758, 0.10155823826789856, -0.14214134216308594, -0.050149016082286835, -0.013447506353259087, -0.13887791335582733, 0.11616533249616623, -0.0724763497710228, 0.11383344233036041, -0.046218473464250565, 0.05384330451488495, 0.0073478645645082, -0.05154034122824669, 0.12863507866859436, 0.02274453639984131, 0.07456807792186737, -0.052454277873039246, -0.012860242277383804, 0.09835619479417801, -0.07845889031887054, 0.06105131283402443, -0.09937947243452072, 0.045091770589351654, -0.12510208785533905, -0.015082119032740593, -0.07675468176603317, 0.02839023247361183, -0.028737343847751617, -0.042841967195272446, -0.05252382904291153, 0.01206414494663477, 0.07457873225212097, -0.0033069076016545296, 0.17504702508449554, 0.04161931201815605, 0.1356934756040573, 0.18485517799854279, 0.07381550967693329, -0.11268056184053421, -0.10376991331577301, -0.00886673852801323, -0.02009434811770916, 0.05579818785190582, -0.1618359386920929, 0.042422011494636536, 0.13992659747600555, 0.006874611135572195, 0.12767653167247772, 0.0688827782869339, -0.04798436909914017, 0.003899581264704466, 0.0452854298055172, -0.1783701628446579, -0.09679107367992401, -0.009332024492323399, -0.012513089925050735, -0.13734038174152374, 0.07031508535146713, 0.11065638810396194, -0.06522715091705322, -0.020677953958511353, 0.0015312001341953874, 0.001479341764934361, -0.023513123393058777, 0.17326204478740692, 0.0655597522854805, 0.06481005996465683, -0.10098203271627426, 0.0951208770275116, 0.04609247297048569, -0.07288358360528946, 0.04003791883587837, 0.05164699628949165, -0.1138383150100708, -0.02253587916493416, 0.03539628908038139, 0.1590377688407898, -0.03593802452087402, -0.046655960381031036, -0.17098021507263184, -0.10509005934000015, 0.08731745183467865, 0.1274595707654953, 0.10466798394918442, 0.020139098167419434, -0.048447709530591965, -0.006609742529690266, -0.10675366967916489, 0.09348485618829727, 0.058594346046447754, 0.07109634578227997, -0.15644440054893494, 0.1105559915304184, 0.001110950019210577, 0.0607583224773407, -0.012919686734676361, 0.001354161067865789, -0.09134423732757568, -0.0003082616603933275, -0.08251616358757019, 0.0122391851618886, -0.04241093248128891, -0.003554311813786626, -0.02001729980111122, -0.05084189772605896, -0.05258030444383621, 0.031079620122909546, -0.10778090357780457, -0.03766832500696182, 0.029691802337765694, 0.04227648675441742, -0.1072600930929184, -0.030516671016812325, 0.011589708738029003, -0.08385424315929413, 0.09281181544065475, 0.04818247631192207, -0.005318017210811377, 0.013172389008104801, -0.024976933375000954, 0.00860611442476511, 0.06548852473497391, 0.0009176498278975487, 0.05490829423069954, -0.12593981623649597, -0.005958652589470148, 0.0027340054512023926, 0.0033404123969376087, 0.022175608202815056, 0.11886222660541534, -0.1292734295129776, -0.010102422907948494, -0.019927645102143288, -0.0311695858836174, -0.07569055259227753, 0.05729673057794571, 0.10719096660614014, 0.0378112718462944, 0.1875617355108261, -0.06822221726179123, 0.018444180488586426, -0.2068903148174286, -0.0002452793996781111, 0.005979603622108698, -0.1454673707485199, -0.07120571285486221, -0.036396682262420654, 0.05290095508098602, -0.06443101167678833, 0.11634373664855957, 0.0008753861766308546, -0.010869303718209267, 0.037613291293382645, -0.03043281100690365, -0.017883090302348137, 0.007872702553868294, 0.17467574775218964, 0.01212515588849783, -0.03913277015089989, 0.11452601850032806, 0.02858971431851387, 0.10287530720233917, 0.11307211220264435, 0.16108684241771698, 0.13502591848373413, 0.025834301486611366, 0.0998968780040741, 0.025341186672449112, -0.01848791167140007, -0.19047077000141144, 0.07276903837919235, -0.03531571105122566, 0.12698744237422943, 0.008672770112752914, 0.19214610755443573, 0.09776225686073303, -0.15829795598983765, 0.05093172565102577, -0.02248046174645424, -0.08756396174430847, -0.09562913328409195, -0.09577185660600662, -0.08325570076704025, -0.15142379701137543, -0.009702731855213642, -0.11629122495651245, 0.0066182236187160015, 0.07528655976057053, -0.0008431302267126739, -0.01898612640798092, 0.14263658225536346, 0.009304924868047237, 0.001253073220141232, 0.07858073711395264, -0.006525959819555283, -0.05684375762939453, -0.07991866022348404, -0.07751739770174026, 0.003022255375981331, 0.01993376761674881, 0.05594443157315254, -0.02513107657432556, 0.00735208997502923, 0.03755185008049011, -0.028129376471042633, -0.10968209058046341, 0.010150511749088764, 0.02177766151726246, 0.051356241106987, 0.02320592664182186, 0.00956298690289259, -0.01430518552660942, -0.012962219305336475, 0.17568282783031464, -0.06385239958763123, -0.01689169742166996, -0.10847058892250061, 0.1899847388267517, 0.03989950940012932, -0.03787495940923691, 0.03915709629654884, -0.06612126529216766, -0.011634225957095623, 0.1944432258605957, 0.18756258487701416, -0.021927885711193085, -0.0002661887847352773, -0.004998415242880583, -0.01534326747059822, 0.0016246484592556953, 0.08369015157222748, 0.11981640011072159, -0.013047742657363415, -0.06906230002641678, -0.01819325052201748, -0.06562758982181549, -0.0067472620867192745, -0.04779968410730362, 0.0645025223493576, 0.02037656493484974, 0.008185519836843014, -0.05136895924806595, 0.023522306233644485, -0.034733060747385025, -0.07609675824642181, 0.038163378834724426, -0.21621042490005493, -0.15695160627365112, -0.006919717416167259, 0.04431783780455589, -0.0026687439531087875, 0.06132279336452484, -0.0016218317905440927, 0.012163124978542328, 0.0941525548696518, -0.020845945924520493, -0.08206836134195328, -0.08182139694690704, 0.08768419921398163, -0.15437477827072144, 0.20841164886951447, -0.04061716049909592, 0.04299676790833473, 0.13137340545654297, 0.04467977210879326, -0.1047448068857193, 0.03841583803296089, 0.03423437103629112, -0.00853665266185999, 0.004708867520093918, 0.09485199302434921, -0.016044387593865395, 0.07863204181194305, 0.03675774112343788, -0.09839759767055511, -0.03282250836491585, -0.06005345657467842, -0.02700952999293804, -0.04377070441842079, -0.049499623477458954, -0.05170126631855965, 0.11996490508317947, 0.18235653638839722, -0.051337070763111115, -0.025411048904061317, -0.0625823587179184, 0.016907107084989548, 0.07732902467250824, -0.01597549393773079, -0.04855141416192055, -0.2462962567806244, 0.010467390529811382, 0.05714378505945206, -0.019154073670506477, -0.23835007846355438, -0.11273518949747086, -0.0007911180146038532, -0.06173158064484596, -0.08129233121871948, 0.08054714649915695, 0.043887872248888016, 0.05588071793317795, -0.0629199966788292, 0.008542655035853386, -0.0903623104095459, 0.16071398556232452, -0.15241222083568573, -0.07646507024765015 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-cased-finetuned-qqp This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the GLUE QQP dataset. It achieves the following results on the evaluation set: - Loss: 0.3752 - Accuracy: 0.9084 - F1: 0.8768 - Combined Score: 0.8926 The model was fine-tuned to compare [google/fnet-base](https://huggingface.co/google/fnet-base) as introduced in [this paper](https://arxiv.org/abs/2105.03824) against [bert-base-cased](https://huggingface.co/bert-base-cased). ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure This model is trained using the [run_glue](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) script. The following command was used: ```bash #!/usr/bin/bash python ../run_glue.py \\n --model_name_or_path bert-base-cased \\n --task_name qqp \\n --do_train \\n --do_eval \\n --max_seq_length 512 \\n --per_device_train_batch_size 16 \\n --learning_rate 2e-5 \\n --num_train_epochs 3 \\n --output_dir bert-base-cased-finetuned-qqp \\n --push_to_hub \\n --hub_strategy all_checkpoints \\n --logging_strategy epoch \\n --save_strategy epoch \\n --evaluation_strategy epoch \\n``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Combined Score | |:-------------:|:-----:|:-----:|:---------------:|:--------:|:------:|:--------------:| | 0.308 | 1.0 | 22741 | 0.2548 | 0.8925 | 0.8556 | 0.8740 | | 0.201 | 2.0 | 45482 | 0.2881 | 0.9032 | 0.8698 | 0.8865 | | 0.1416 | 3.0 | 68223 | 0.3752 | 0.9084 | 0.8768 | 0.8926 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer", "fnet-bert-base-comparison"], "datasets": ["glue"], "metrics": ["accuracy", "f1"], "model-index": [{"name": "bert-base-cased-finetuned-qqp", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE QQP", "type": "glue", "args": "qqp"}, "metrics": [{"type": "accuracy", "value": 0.9083848627256987, "name": "Accuracy"}, {"type": "f1", "value": 0.8767633750332712, "name": "F1"}]}]}]}
text-classification
gchhablani/bert-base-cased-finetuned-qqp
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "fnet-bert-base-comparison", "en", "dataset:glue", "arxiv:2105.03824", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2105.03824" ]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
bert-base-cased-finetuned-qqp ============================= This model is a fine-tuned version of bert-base-cased on the GLUE QQP dataset. It achieves the following results on the evaluation set: * Loss: 0.3752 * Accuracy: 0.9084 * F1: 0.8768 * Combined Score: 0.8926 The model was fine-tuned to compare google/fnet-base as introduced in this paper against bert-base-cased. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ This model is trained using the run\_glue script. The following command was used: ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 87, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.12161670625209808, 0.1586534082889557, -0.0035777657758444548, 0.11875102669000626, 0.11533597856760025, -0.0031049344688653946, 0.13364577293395996, 0.1597096472978592, -0.08531280606985092, 0.0648089274764061, 0.15738050639629364, 0.1568128913640976, 0.02761896513402462, 0.18884018063545227, -0.052745409309864044, -0.23438897728919983, 0.03223801031708717, 0.07453203946352005, -0.03721422329545021, 0.13727734982967377, 0.10004161298274994, -0.12226498872041702, 0.09843768924474716, 0.03329526633024216, -0.20268045365810394, -0.0181635320186615, 0.0035198922269046307, -0.08516480028629303, 0.11188160628080368, 0.0131202582269907, 0.08862993866205215, 0.031522486358881, 0.03825753182172775, -0.1487365961074829, 0.005022411700338125, 0.046756207942962646, 0.0042826407589018345, 0.11839430034160614, 0.03855927288532257, -0.01641850546002388, 0.0520799495279789, -0.0885375365614891, 0.05395820736885071, 0.028443723917007446, -0.1181475892663002, -0.26681092381477356, -0.0919037014245987, 0.06938058882951736, 0.046251799911260605, 0.07366777211427689, -0.00005182146924198605, 0.16483885049819946, -0.005088786594569683, 0.10346852988004684, 0.27087369561195374, -0.3125593364238739, -0.05947171151638031, 0.02776799350976944, 0.011863073334097862, 0.05958991125226021, -0.09307480603456497, -0.02570064552128315, 0.05127633363008499, 0.04167178273200989, 0.18514986336231232, -0.016228215768933296, -0.00003337187081342563, -0.02189422771334648, -0.1382196545600891, -0.07627017796039581, 0.20115217566490173, 0.05067397281527519, -0.05151061341166496, -0.07212591916322708, -0.07642883062362671, -0.16918812692165375, -0.03298860415816307, -0.020403943955898285, 0.046687398105859756, -0.035265833139419556, -0.06775462627410889, -0.019117340445518494, -0.07634665071964264, -0.04710827395319939, -0.038415029644966125, 0.15217410027980804, 0.04655686020851135, 0.022712748497724533, -0.025998499244451523, 0.09033547341823578, -0.015903212130069733, -0.15715816617012024, -0.005602593533694744, 0.011772151105105877, 0.037549275904893875, -0.022561458870768547, -0.036397535353899, -0.08168396353721619, 0.012048215605318546, 0.14117512106895447, -0.08264761418104172, 0.06482455879449844, 0.011765061877667904, 0.0496395118534565, -0.08314628899097443, 0.1905018389225006, -0.024279993027448654, 0.0037186690606176853, 0.02406053990125656, 0.08622891455888748, 0.04516247287392616, -0.026812154799699783, -0.11238549649715424, 0.0233257245272398, 0.14243075251579285, 0.007431734818965197, -0.039754197001457214, 0.07417906820774078, -0.059125080704689026, -0.04582800716161728, 0.04539237171411514, -0.11030158400535583, 0.0141338761895895, -0.009373044595122337, -0.08056624233722687, -0.03472790867090225, 0.02697722613811493, -0.011787989176809788, -0.032092269510030746, 0.06212129071354866, -0.09913354367017746, 0.008189168758690357, -0.06144571676850319, -0.10696113854646683, 0.015179385431110859, -0.11808769404888153, -0.001081367488950491, -0.11012967675924301, -0.13974249362945557, -0.006976991426199675, 0.053723063319921494, -0.023807602003216743, -0.0772162452340126, -0.05605267360806465, -0.07926763594150543, 0.029673025012016296, -0.0168012622743845, 0.04827882722020149, -0.06232791766524315, 0.08783834427595139, 0.04711596667766571, 0.07783091068267822, -0.03511257842183113, 0.04865354299545288, -0.0857672467827797, 0.04486498609185219, -0.20575650036334991, 0.06056206300854683, -0.057986579835414886, 0.06628572940826416, -0.1133461743593216, -0.11155343800783157, 0.0248954389244318, -0.03267720714211464, 0.0798325464129448, 0.10354934632778168, -0.14306844770908356, -0.0819404348731041, 0.1941177397966385, -0.08345154672861099, -0.13623358309268951, 0.11876475811004639, -0.04778644070029259, 0.023907052353024483, 0.06198955699801445, 0.22297649085521698, 0.0731961652636528, -0.057169198989868164, -0.026051243767142296, -0.017552388831973076, 0.051131829619407654, -0.07172699272632599, 0.08445700258016586, -0.004041696432977915, 0.028126448392868042, 0.030312296003103256, -0.035424020141363144, 0.032627593725919724, -0.08209208399057388, -0.08335229009389877, -0.05411062017083168, -0.0801575630903244, 0.06567159295082092, 0.04338371008634567, 0.08023004233837128, -0.11180076748132706, -0.09540257602930069, 0.034219104796648026, 0.08448748290538788, -0.08299247920513153, 0.045332007110118866, -0.0944017842411995, 0.12313508242368698, -0.07442712783813477, -0.005775441415607929, -0.18186262249946594, -0.004674280993640423, 0.04966714233160019, -0.028262924402952194, 0.004111959598958492, -0.025413617491722107, 0.06383029371500015, 0.051281414926052094, -0.046544935554265976, -0.04452786594629288, -0.037859585136175156, -0.007022486068308353, -0.1164112463593483, -0.17888443171977997, -0.04770522564649582, -0.03426934406161308, 0.11073228716850281, -0.15757550299167633, 0.05754205211997032, 0.0591144785284996, 0.10811145603656769, 0.02597060613334179, -0.03133917599916458, -0.009403175674378872, 0.04480119049549103, -0.04116583615541458, -0.07460780441761017, 0.07281174510717392, 0.03331552818417549, -0.09998483955860138, -0.02359735779464245, -0.11360644549131393, 0.1753949373960495, 0.12714703381061554, -0.022101160138845444, -0.05310492590069771, -0.009392490610480309, -0.06040368974208832, -0.023967178538441658, -0.010677943006157875, 0.013884914107620716, 0.17248490452766418, 0.005367063917219639, 0.17223170399665833, -0.10302618145942688, -0.05416569113731384, 0.04557802155613899, -0.0306826401501894, -0.013940786942839622, 0.10292626917362213, 0.01127760112285614, -0.10022857785224915, 0.15135939419269562, 0.1410094052553177, -0.055681660771369934, 0.12143083661794662, -0.05604437366127968, -0.047931358218193054, -0.039293039590120316, -0.0005570283392444253, 0.018249090760946274, 0.09346067160367966, -0.11466828733682632, -0.018797609955072403, 0.039005741477012634, 0.028456371277570724, 0.011234947480261326, -0.1804705411195755, 0.001110660727135837, 0.04256647825241089, -0.0585462749004364, 0.007875440642237663, -0.005216784775257111, -0.005512913689017296, 0.10080158710479736, 0.02539101243019104, -0.0727960467338562, 0.052135586738586426, 0.010471811518073082, -0.07134804874658585, 0.19735264778137207, -0.09606501460075378, -0.19630330801010132, -0.12494390457868576, -0.06156739220023155, -0.09005486220121384, 0.006349241826683283, 0.06912596523761749, -0.07998423278331757, -0.024601668119430542, -0.09020604938268661, -0.027059001848101616, -0.014606335200369358, 0.03885342925786972, 0.07165474444627762, -0.02355276234447956, 0.10255801677703857, -0.120688796043396, -0.030849842354655266, -0.03027579002082348, 0.0014742627972736955, 0.05496053025126457, 0.006049624178558588, 0.10471925884485245, 0.11509934067726135, -0.02873915433883667, 0.05114046484231949, -0.03214579448103905, 0.2341459095478058, -0.04930265247821808, -0.02795281447470188, 0.12977902591228485, -0.005955296568572521, 0.08535684645175934, 0.0899367704987526, 0.045135561376810074, -0.09218605607748032, -0.00756901316344738, 0.005072117783129215, -0.040117885917425156, -0.21103355288505554, -0.035214588046073914, -0.04211072623729706, 0.020072607323527336, 0.12289147824048996, 0.04161820560693741, 0.05459532514214516, 0.0637349858880043, 0.029591871425509453, 0.0592803992331028, -0.028413070365786552, 0.1009601503610611, 0.12848083674907684, 0.05054585635662079, 0.13528481125831604, -0.03867393359541893, -0.03375968337059021, 0.040527984499931335, 0.0007192929624579847, 0.20465855300426483, -0.014075764454901218, 0.1890448033809662, 0.04658591002225876, 0.18518702685832977, 0.010942237451672554, 0.06792156398296356, -0.0207637008279562, -0.004309804644435644, -0.017503520473837852, -0.04292628914117813, -0.060658156871795654, 0.009485099464654922, -0.04635186865925789, 0.07441109418869019, -0.12235406786203384, 0.019649730995297432, 0.060750771313905716, 0.291361004114151, 0.02242939919233322, -0.3738984167575836, -0.1097446084022522, -0.011077821254730225, -0.025019990280270576, -0.04325328394770622, 0.011022142134606838, 0.09511391818523407, -0.09566101431846619, 0.06279890984296799, -0.08686861395835876, 0.08955012261867523, -0.073078453540802, 0.03659684211015701, 0.049923211336135864, 0.09388187527656555, 0.005718675907701254, 0.057970207184553146, -0.27198654413223267, 0.2506929934024811, 0.017948467284440994, 0.05225548893213272, -0.06244153901934624, 0.014453649520874023, 0.022270919755101204, 0.0580802857875824, 0.08689375221729279, -0.0023560896515846252, -0.03474327549338341, -0.17138437926769257, -0.10784955322742462, 0.012887988239526749, 0.07278495281934738, -0.04178544878959656, 0.0889785885810852, -0.010356182232499123, 0.003321166383102536, 0.04448032006621361, -0.003272048430517316, -0.033030617982149124, -0.09141189604997635, 0.021423611789941788, 0.06062246859073639, -0.022481713443994522, -0.0803140178322792, -0.11677858233451843, -0.08531743288040161, 0.1759343147277832, -0.006428628694266081, -0.07254543155431747, -0.12288008630275726, 0.061379604041576385, 0.06829093396663666, -0.09375333040952682, 0.04493061453104019, -0.02001975290477276, 0.12447743117809296, 0.009351304732263088, -0.0690632313489914, 0.09878260642290115, -0.047859106212854385, -0.16231606900691986, -0.0379788912832737, 0.13346640765666962, 0.02778930962085724, 0.055701445788145065, -0.007568121422082186, 0.02814381755888462, -0.0244494266808033, -0.07651051878929138, 0.03854655846953392, 0.007937511429190636, 0.09600245952606201, -0.019526541233062744, -0.019290292635560036, 0.03250190243124962, -0.07478196173906326, -0.004258669447153807, 0.1966393142938614, 0.2564740777015686, -0.10714545100927353, 0.043486859649419785, 0.03290016949176788, -0.05048590898513794, -0.1577579379081726, 0.013250388205051422, 0.07308094948530197, 0.00450450275093317, 0.0024772710166871548, -0.17275027930736542, 0.055431436747312546, 0.09000243246555328, -0.01831885054707527, 0.07066475600004196, -0.29239320755004883, -0.11998265236616135, 0.10719365626573563, 0.12865246832370758, 0.10155823826789856, -0.14214134216308594, -0.050149016082286835, -0.013447506353259087, -0.13887791335582733, 0.11616533249616623, -0.0724763497710228, 0.11383344233036041, -0.046218473464250565, 0.05384330451488495, 0.0073478645645082, -0.05154034122824669, 0.12863507866859436, 0.02274453639984131, 0.07456807792186737, -0.052454277873039246, -0.012860242277383804, 0.09835619479417801, -0.07845889031887054, 0.06105131283402443, -0.09937947243452072, 0.045091770589351654, -0.12510208785533905, -0.015082119032740593, -0.07675468176603317, 0.02839023247361183, -0.028737343847751617, -0.042841967195272446, -0.05252382904291153, 0.01206414494663477, 0.07457873225212097, -0.0033069076016545296, 0.17504702508449554, 0.04161931201815605, 0.1356934756040573, 0.18485517799854279, 0.07381550967693329, -0.11268056184053421, -0.10376991331577301, -0.00886673852801323, -0.02009434811770916, 0.05579818785190582, -0.1618359386920929, 0.042422011494636536, 0.13992659747600555, 0.006874611135572195, 0.12767653167247772, 0.0688827782869339, -0.04798436909914017, 0.003899581264704466, 0.0452854298055172, -0.1783701628446579, -0.09679107367992401, -0.009332024492323399, -0.012513089925050735, -0.13734038174152374, 0.07031508535146713, 0.11065638810396194, -0.06522715091705322, -0.020677953958511353, 0.0015312001341953874, 0.001479341764934361, -0.023513123393058777, 0.17326204478740692, 0.0655597522854805, 0.06481005996465683, -0.10098203271627426, 0.0951208770275116, 0.04609247297048569, -0.07288358360528946, 0.04003791883587837, 0.05164699628949165, -0.1138383150100708, -0.02253587916493416, 0.03539628908038139, 0.1590377688407898, -0.03593802452087402, -0.046655960381031036, -0.17098021507263184, -0.10509005934000015, 0.08731745183467865, 0.1274595707654953, 0.10466798394918442, 0.020139098167419434, -0.048447709530591965, -0.006609742529690266, -0.10675366967916489, 0.09348485618829727, 0.058594346046447754, 0.07109634578227997, -0.15644440054893494, 0.1105559915304184, 0.001110950019210577, 0.0607583224773407, -0.012919686734676361, 0.001354161067865789, -0.09134423732757568, -0.0003082616603933275, -0.08251616358757019, 0.0122391851618886, -0.04241093248128891, -0.003554311813786626, -0.02001729980111122, -0.05084189772605896, -0.05258030444383621, 0.031079620122909546, -0.10778090357780457, -0.03766832500696182, 0.029691802337765694, 0.04227648675441742, -0.1072600930929184, -0.030516671016812325, 0.011589708738029003, -0.08385424315929413, 0.09281181544065475, 0.04818247631192207, -0.005318017210811377, 0.013172389008104801, -0.024976933375000954, 0.00860611442476511, 0.06548852473497391, 0.0009176498278975487, 0.05490829423069954, -0.12593981623649597, -0.005958652589470148, 0.0027340054512023926, 0.0033404123969376087, 0.022175608202815056, 0.11886222660541534, -0.1292734295129776, -0.010102422907948494, -0.019927645102143288, -0.0311695858836174, -0.07569055259227753, 0.05729673057794571, 0.10719096660614014, 0.0378112718462944, 0.1875617355108261, -0.06822221726179123, 0.018444180488586426, -0.2068903148174286, -0.0002452793996781111, 0.005979603622108698, -0.1454673707485199, -0.07120571285486221, -0.036396682262420654, 0.05290095508098602, -0.06443101167678833, 0.11634373664855957, 0.0008753861766308546, -0.010869303718209267, 0.037613291293382645, -0.03043281100690365, -0.017883090302348137, 0.007872702553868294, 0.17467574775218964, 0.01212515588849783, -0.03913277015089989, 0.11452601850032806, 0.02858971431851387, 0.10287530720233917, 0.11307211220264435, 0.16108684241771698, 0.13502591848373413, 0.025834301486611366, 0.0998968780040741, 0.025341186672449112, -0.01848791167140007, -0.19047077000141144, 0.07276903837919235, -0.03531571105122566, 0.12698744237422943, 0.008672770112752914, 0.19214610755443573, 0.09776225686073303, -0.15829795598983765, 0.05093172565102577, -0.02248046174645424, -0.08756396174430847, -0.09562913328409195, -0.09577185660600662, -0.08325570076704025, -0.15142379701137543, -0.009702731855213642, -0.11629122495651245, 0.0066182236187160015, 0.07528655976057053, -0.0008431302267126739, -0.01898612640798092, 0.14263658225536346, 0.009304924868047237, 0.001253073220141232, 0.07858073711395264, -0.006525959819555283, -0.05684375762939453, -0.07991866022348404, -0.07751739770174026, 0.003022255375981331, 0.01993376761674881, 0.05594443157315254, -0.02513107657432556, 0.00735208997502923, 0.03755185008049011, -0.028129376471042633, -0.10968209058046341, 0.010150511749088764, 0.02177766151726246, 0.051356241106987, 0.02320592664182186, 0.00956298690289259, -0.01430518552660942, -0.012962219305336475, 0.17568282783031464, -0.06385239958763123, -0.01689169742166996, -0.10847058892250061, 0.1899847388267517, 0.03989950940012932, -0.03787495940923691, 0.03915709629654884, -0.06612126529216766, -0.011634225957095623, 0.1944432258605957, 0.18756258487701416, -0.021927885711193085, -0.0002661887847352773, -0.004998415242880583, -0.01534326747059822, 0.0016246484592556953, 0.08369015157222748, 0.11981640011072159, -0.013047742657363415, -0.06906230002641678, -0.01819325052201748, -0.06562758982181549, -0.0067472620867192745, -0.04779968410730362, 0.0645025223493576, 0.02037656493484974, 0.008185519836843014, -0.05136895924806595, 0.023522306233644485, -0.034733060747385025, -0.07609675824642181, 0.038163378834724426, -0.21621042490005493, -0.15695160627365112, -0.006919717416167259, 0.04431783780455589, -0.0026687439531087875, 0.06132279336452484, -0.0016218317905440927, 0.012163124978542328, 0.0941525548696518, -0.020845945924520493, -0.08206836134195328, -0.08182139694690704, 0.08768419921398163, -0.15437477827072144, 0.20841164886951447, -0.04061716049909592, 0.04299676790833473, 0.13137340545654297, 0.04467977210879326, -0.1047448068857193, 0.03841583803296089, 0.03423437103629112, -0.00853665266185999, 0.004708867520093918, 0.09485199302434921, -0.016044387593865395, 0.07863204181194305, 0.03675774112343788, -0.09839759767055511, -0.03282250836491585, -0.06005345657467842, -0.02700952999293804, -0.04377070441842079, -0.049499623477458954, -0.05170126631855965, 0.11996490508317947, 0.18235653638839722, -0.051337070763111115, -0.025411048904061317, -0.0625823587179184, 0.016907107084989548, 0.07732902467250824, -0.01597549393773079, -0.04855141416192055, -0.2462962567806244, 0.010467390529811382, 0.05714378505945206, -0.019154073670506477, -0.23835007846355438, -0.11273518949747086, -0.0007911180146038532, -0.06173158064484596, -0.08129233121871948, 0.08054714649915695, 0.043887872248888016, 0.05588071793317795, -0.0629199966788292, 0.008542655035853386, -0.0903623104095459, 0.16071398556232452, -0.15241222083568573, -0.07646507024765015 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-cased-finetuned-rte This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the GLUE RTE dataset. It achieves the following results on the evaluation set: - Loss: 0.7260 - Accuracy: 0.6715 The model was fine-tuned to compare [google/fnet-base](https://huggingface.co/google/fnet-base) as introduced in [this paper](https://arxiv.org/abs/2105.03824) against [bert-base-cased](https://huggingface.co/bert-base-cased). ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure This model is trained using the [run_glue](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) script. The following command was used: ```bash #!/usr/bin/bash python ../run_glue.py \\n --model_name_or_path bert-base-cased \\n --task_name rte \\n --do_train \\n --do_eval \\n --max_seq_length 512 \\n --per_device_train_batch_size 16 \\n --learning_rate 2e-5 \\n --num_train_epochs 3 \\n --output_dir bert-base-cased-finetuned-rte \\n --push_to_hub \\n --hub_strategy all_checkpoints \\n --logging_strategy epoch \\n --save_strategy epoch \\n --evaluation_strategy epoch \\n``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6915 | 1.0 | 156 | 0.6491 | 0.6606 | | 0.55 | 2.0 | 312 | 0.6737 | 0.6570 | | 0.3955 | 3.0 | 468 | 0.7260 | 0.6715 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer", "fnet-bert-base-comparison"], "datasets": ["glue"], "metrics": ["accuracy"], "model-index": [{"name": "bert-base-cased-finetuned-rte", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE RTE", "type": "glue", "args": "rte"}, "metrics": [{"type": "accuracy", "value": 0.6714801444043321, "name": "Accuracy"}]}]}]}
text-classification
gchhablani/bert-base-cased-finetuned-rte
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "fnet-bert-base-comparison", "en", "dataset:glue", "arxiv:2105.03824", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2105.03824" ]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
bert-base-cased-finetuned-rte ============================= This model is a fine-tuned version of bert-base-cased on the GLUE RTE dataset. It achieves the following results on the evaluation set: * Loss: 0.7260 * Accuracy: 0.6715 The model was fine-tuned to compare google/fnet-base as introduced in this paper against bert-base-cased. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ This model is trained using the run\_glue script. The following command was used: ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 87, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.12161670625209808, 0.1586534082889557, -0.0035777657758444548, 0.11875102669000626, 0.11533597856760025, -0.0031049344688653946, 0.13364577293395996, 0.1597096472978592, -0.08531280606985092, 0.0648089274764061, 0.15738050639629364, 0.1568128913640976, 0.02761896513402462, 0.18884018063545227, -0.052745409309864044, -0.23438897728919983, 0.03223801031708717, 0.07453203946352005, -0.03721422329545021, 0.13727734982967377, 0.10004161298274994, -0.12226498872041702, 0.09843768924474716, 0.03329526633024216, -0.20268045365810394, -0.0181635320186615, 0.0035198922269046307, -0.08516480028629303, 0.11188160628080368, 0.0131202582269907, 0.08862993866205215, 0.031522486358881, 0.03825753182172775, -0.1487365961074829, 0.005022411700338125, 0.046756207942962646, 0.0042826407589018345, 0.11839430034160614, 0.03855927288532257, -0.01641850546002388, 0.0520799495279789, -0.0885375365614891, 0.05395820736885071, 0.028443723917007446, -0.1181475892663002, -0.26681092381477356, -0.0919037014245987, 0.06938058882951736, 0.046251799911260605, 0.07366777211427689, -0.00005182146924198605, 0.16483885049819946, -0.005088786594569683, 0.10346852988004684, 0.27087369561195374, -0.3125593364238739, -0.05947171151638031, 0.02776799350976944, 0.011863073334097862, 0.05958991125226021, -0.09307480603456497, -0.02570064552128315, 0.05127633363008499, 0.04167178273200989, 0.18514986336231232, -0.016228215768933296, -0.00003337187081342563, -0.02189422771334648, -0.1382196545600891, -0.07627017796039581, 0.20115217566490173, 0.05067397281527519, -0.05151061341166496, -0.07212591916322708, -0.07642883062362671, -0.16918812692165375, -0.03298860415816307, -0.020403943955898285, 0.046687398105859756, -0.035265833139419556, -0.06775462627410889, -0.019117340445518494, -0.07634665071964264, -0.04710827395319939, -0.038415029644966125, 0.15217410027980804, 0.04655686020851135, 0.022712748497724533, -0.025998499244451523, 0.09033547341823578, -0.015903212130069733, -0.15715816617012024, -0.005602593533694744, 0.011772151105105877, 0.037549275904893875, -0.022561458870768547, -0.036397535353899, -0.08168396353721619, 0.012048215605318546, 0.14117512106895447, -0.08264761418104172, 0.06482455879449844, 0.011765061877667904, 0.0496395118534565, -0.08314628899097443, 0.1905018389225006, -0.024279993027448654, 0.0037186690606176853, 0.02406053990125656, 0.08622891455888748, 0.04516247287392616, -0.026812154799699783, -0.11238549649715424, 0.0233257245272398, 0.14243075251579285, 0.007431734818965197, -0.039754197001457214, 0.07417906820774078, -0.059125080704689026, -0.04582800716161728, 0.04539237171411514, -0.11030158400535583, 0.0141338761895895, -0.009373044595122337, -0.08056624233722687, -0.03472790867090225, 0.02697722613811493, -0.011787989176809788, -0.032092269510030746, 0.06212129071354866, -0.09913354367017746, 0.008189168758690357, -0.06144571676850319, -0.10696113854646683, 0.015179385431110859, -0.11808769404888153, -0.001081367488950491, -0.11012967675924301, -0.13974249362945557, -0.006976991426199675, 0.053723063319921494, -0.023807602003216743, -0.0772162452340126, -0.05605267360806465, -0.07926763594150543, 0.029673025012016296, -0.0168012622743845, 0.04827882722020149, -0.06232791766524315, 0.08783834427595139, 0.04711596667766571, 0.07783091068267822, -0.03511257842183113, 0.04865354299545288, -0.0857672467827797, 0.04486498609185219, -0.20575650036334991, 0.06056206300854683, -0.057986579835414886, 0.06628572940826416, -0.1133461743593216, -0.11155343800783157, 0.0248954389244318, -0.03267720714211464, 0.0798325464129448, 0.10354934632778168, -0.14306844770908356, -0.0819404348731041, 0.1941177397966385, -0.08345154672861099, -0.13623358309268951, 0.11876475811004639, -0.04778644070029259, 0.023907052353024483, 0.06198955699801445, 0.22297649085521698, 0.0731961652636528, -0.057169198989868164, -0.026051243767142296, -0.017552388831973076, 0.051131829619407654, -0.07172699272632599, 0.08445700258016586, -0.004041696432977915, 0.028126448392868042, 0.030312296003103256, -0.035424020141363144, 0.032627593725919724, -0.08209208399057388, -0.08335229009389877, -0.05411062017083168, -0.0801575630903244, 0.06567159295082092, 0.04338371008634567, 0.08023004233837128, -0.11180076748132706, -0.09540257602930069, 0.034219104796648026, 0.08448748290538788, -0.08299247920513153, 0.045332007110118866, -0.0944017842411995, 0.12313508242368698, -0.07442712783813477, -0.005775441415607929, -0.18186262249946594, -0.004674280993640423, 0.04966714233160019, -0.028262924402952194, 0.004111959598958492, -0.025413617491722107, 0.06383029371500015, 0.051281414926052094, -0.046544935554265976, -0.04452786594629288, -0.037859585136175156, -0.007022486068308353, -0.1164112463593483, -0.17888443171977997, -0.04770522564649582, -0.03426934406161308, 0.11073228716850281, -0.15757550299167633, 0.05754205211997032, 0.0591144785284996, 0.10811145603656769, 0.02597060613334179, -0.03133917599916458, -0.009403175674378872, 0.04480119049549103, -0.04116583615541458, -0.07460780441761017, 0.07281174510717392, 0.03331552818417549, -0.09998483955860138, -0.02359735779464245, -0.11360644549131393, 0.1753949373960495, 0.12714703381061554, -0.022101160138845444, -0.05310492590069771, -0.009392490610480309, -0.06040368974208832, -0.023967178538441658, -0.010677943006157875, 0.013884914107620716, 0.17248490452766418, 0.005367063917219639, 0.17223170399665833, -0.10302618145942688, -0.05416569113731384, 0.04557802155613899, -0.0306826401501894, -0.013940786942839622, 0.10292626917362213, 0.01127760112285614, -0.10022857785224915, 0.15135939419269562, 0.1410094052553177, -0.055681660771369934, 0.12143083661794662, -0.05604437366127968, -0.047931358218193054, -0.039293039590120316, -0.0005570283392444253, 0.018249090760946274, 0.09346067160367966, -0.11466828733682632, -0.018797609955072403, 0.039005741477012634, 0.028456371277570724, 0.011234947480261326, -0.1804705411195755, 0.001110660727135837, 0.04256647825241089, -0.0585462749004364, 0.007875440642237663, -0.005216784775257111, -0.005512913689017296, 0.10080158710479736, 0.02539101243019104, -0.0727960467338562, 0.052135586738586426, 0.010471811518073082, -0.07134804874658585, 0.19735264778137207, -0.09606501460075378, -0.19630330801010132, -0.12494390457868576, -0.06156739220023155, -0.09005486220121384, 0.006349241826683283, 0.06912596523761749, -0.07998423278331757, -0.024601668119430542, -0.09020604938268661, -0.027059001848101616, -0.014606335200369358, 0.03885342925786972, 0.07165474444627762, -0.02355276234447956, 0.10255801677703857, -0.120688796043396, -0.030849842354655266, -0.03027579002082348, 0.0014742627972736955, 0.05496053025126457, 0.006049624178558588, 0.10471925884485245, 0.11509934067726135, -0.02873915433883667, 0.05114046484231949, -0.03214579448103905, 0.2341459095478058, -0.04930265247821808, -0.02795281447470188, 0.12977902591228485, -0.005955296568572521, 0.08535684645175934, 0.0899367704987526, 0.045135561376810074, -0.09218605607748032, -0.00756901316344738, 0.005072117783129215, -0.040117885917425156, -0.21103355288505554, -0.035214588046073914, -0.04211072623729706, 0.020072607323527336, 0.12289147824048996, 0.04161820560693741, 0.05459532514214516, 0.0637349858880043, 0.029591871425509453, 0.0592803992331028, -0.028413070365786552, 0.1009601503610611, 0.12848083674907684, 0.05054585635662079, 0.13528481125831604, -0.03867393359541893, -0.03375968337059021, 0.040527984499931335, 0.0007192929624579847, 0.20465855300426483, -0.014075764454901218, 0.1890448033809662, 0.04658591002225876, 0.18518702685832977, 0.010942237451672554, 0.06792156398296356, -0.0207637008279562, -0.004309804644435644, -0.017503520473837852, -0.04292628914117813, -0.060658156871795654, 0.009485099464654922, -0.04635186865925789, 0.07441109418869019, -0.12235406786203384, 0.019649730995297432, 0.060750771313905716, 0.291361004114151, 0.02242939919233322, -0.3738984167575836, -0.1097446084022522, -0.011077821254730225, -0.025019990280270576, -0.04325328394770622, 0.011022142134606838, 0.09511391818523407, -0.09566101431846619, 0.06279890984296799, -0.08686861395835876, 0.08955012261867523, -0.073078453540802, 0.03659684211015701, 0.049923211336135864, 0.09388187527656555, 0.005718675907701254, 0.057970207184553146, -0.27198654413223267, 0.2506929934024811, 0.017948467284440994, 0.05225548893213272, -0.06244153901934624, 0.014453649520874023, 0.022270919755101204, 0.0580802857875824, 0.08689375221729279, -0.0023560896515846252, -0.03474327549338341, -0.17138437926769257, -0.10784955322742462, 0.012887988239526749, 0.07278495281934738, -0.04178544878959656, 0.0889785885810852, -0.010356182232499123, 0.003321166383102536, 0.04448032006621361, -0.003272048430517316, -0.033030617982149124, -0.09141189604997635, 0.021423611789941788, 0.06062246859073639, -0.022481713443994522, -0.0803140178322792, -0.11677858233451843, -0.08531743288040161, 0.1759343147277832, -0.006428628694266081, -0.07254543155431747, -0.12288008630275726, 0.061379604041576385, 0.06829093396663666, -0.09375333040952682, 0.04493061453104019, -0.02001975290477276, 0.12447743117809296, 0.009351304732263088, -0.0690632313489914, 0.09878260642290115, -0.047859106212854385, -0.16231606900691986, -0.0379788912832737, 0.13346640765666962, 0.02778930962085724, 0.055701445788145065, -0.007568121422082186, 0.02814381755888462, -0.0244494266808033, -0.07651051878929138, 0.03854655846953392, 0.007937511429190636, 0.09600245952606201, -0.019526541233062744, -0.019290292635560036, 0.03250190243124962, -0.07478196173906326, -0.004258669447153807, 0.1966393142938614, 0.2564740777015686, -0.10714545100927353, 0.043486859649419785, 0.03290016949176788, -0.05048590898513794, -0.1577579379081726, 0.013250388205051422, 0.07308094948530197, 0.00450450275093317, 0.0024772710166871548, -0.17275027930736542, 0.055431436747312546, 0.09000243246555328, -0.01831885054707527, 0.07066475600004196, -0.29239320755004883, -0.11998265236616135, 0.10719365626573563, 0.12865246832370758, 0.10155823826789856, -0.14214134216308594, -0.050149016082286835, -0.013447506353259087, -0.13887791335582733, 0.11616533249616623, -0.0724763497710228, 0.11383344233036041, -0.046218473464250565, 0.05384330451488495, 0.0073478645645082, -0.05154034122824669, 0.12863507866859436, 0.02274453639984131, 0.07456807792186737, -0.052454277873039246, -0.012860242277383804, 0.09835619479417801, -0.07845889031887054, 0.06105131283402443, -0.09937947243452072, 0.045091770589351654, -0.12510208785533905, -0.015082119032740593, -0.07675468176603317, 0.02839023247361183, -0.028737343847751617, -0.042841967195272446, -0.05252382904291153, 0.01206414494663477, 0.07457873225212097, -0.0033069076016545296, 0.17504702508449554, 0.04161931201815605, 0.1356934756040573, 0.18485517799854279, 0.07381550967693329, -0.11268056184053421, -0.10376991331577301, -0.00886673852801323, -0.02009434811770916, 0.05579818785190582, -0.1618359386920929, 0.042422011494636536, 0.13992659747600555, 0.006874611135572195, 0.12767653167247772, 0.0688827782869339, -0.04798436909914017, 0.003899581264704466, 0.0452854298055172, -0.1783701628446579, -0.09679107367992401, -0.009332024492323399, -0.012513089925050735, -0.13734038174152374, 0.07031508535146713, 0.11065638810396194, -0.06522715091705322, -0.020677953958511353, 0.0015312001341953874, 0.001479341764934361, -0.023513123393058777, 0.17326204478740692, 0.0655597522854805, 0.06481005996465683, -0.10098203271627426, 0.0951208770275116, 0.04609247297048569, -0.07288358360528946, 0.04003791883587837, 0.05164699628949165, -0.1138383150100708, -0.02253587916493416, 0.03539628908038139, 0.1590377688407898, -0.03593802452087402, -0.046655960381031036, -0.17098021507263184, -0.10509005934000015, 0.08731745183467865, 0.1274595707654953, 0.10466798394918442, 0.020139098167419434, -0.048447709530591965, -0.006609742529690266, -0.10675366967916489, 0.09348485618829727, 0.058594346046447754, 0.07109634578227997, -0.15644440054893494, 0.1105559915304184, 0.001110950019210577, 0.0607583224773407, -0.012919686734676361, 0.001354161067865789, -0.09134423732757568, -0.0003082616603933275, -0.08251616358757019, 0.0122391851618886, -0.04241093248128891, -0.003554311813786626, -0.02001729980111122, -0.05084189772605896, -0.05258030444383621, 0.031079620122909546, -0.10778090357780457, -0.03766832500696182, 0.029691802337765694, 0.04227648675441742, -0.1072600930929184, -0.030516671016812325, 0.011589708738029003, -0.08385424315929413, 0.09281181544065475, 0.04818247631192207, -0.005318017210811377, 0.013172389008104801, -0.024976933375000954, 0.00860611442476511, 0.06548852473497391, 0.0009176498278975487, 0.05490829423069954, -0.12593981623649597, -0.005958652589470148, 0.0027340054512023926, 0.0033404123969376087, 0.022175608202815056, 0.11886222660541534, -0.1292734295129776, -0.010102422907948494, -0.019927645102143288, -0.0311695858836174, -0.07569055259227753, 0.05729673057794571, 0.10719096660614014, 0.0378112718462944, 0.1875617355108261, -0.06822221726179123, 0.018444180488586426, -0.2068903148174286, -0.0002452793996781111, 0.005979603622108698, -0.1454673707485199, -0.07120571285486221, -0.036396682262420654, 0.05290095508098602, -0.06443101167678833, 0.11634373664855957, 0.0008753861766308546, -0.010869303718209267, 0.037613291293382645, -0.03043281100690365, -0.017883090302348137, 0.007872702553868294, 0.17467574775218964, 0.01212515588849783, -0.03913277015089989, 0.11452601850032806, 0.02858971431851387, 0.10287530720233917, 0.11307211220264435, 0.16108684241771698, 0.13502591848373413, 0.025834301486611366, 0.0998968780040741, 0.025341186672449112, -0.01848791167140007, -0.19047077000141144, 0.07276903837919235, -0.03531571105122566, 0.12698744237422943, 0.008672770112752914, 0.19214610755443573, 0.09776225686073303, -0.15829795598983765, 0.05093172565102577, -0.02248046174645424, -0.08756396174430847, -0.09562913328409195, -0.09577185660600662, -0.08325570076704025, -0.15142379701137543, -0.009702731855213642, -0.11629122495651245, 0.0066182236187160015, 0.07528655976057053, -0.0008431302267126739, -0.01898612640798092, 0.14263658225536346, 0.009304924868047237, 0.001253073220141232, 0.07858073711395264, -0.006525959819555283, -0.05684375762939453, -0.07991866022348404, -0.07751739770174026, 0.003022255375981331, 0.01993376761674881, 0.05594443157315254, -0.02513107657432556, 0.00735208997502923, 0.03755185008049011, -0.028129376471042633, -0.10968209058046341, 0.010150511749088764, 0.02177766151726246, 0.051356241106987, 0.02320592664182186, 0.00956298690289259, -0.01430518552660942, -0.012962219305336475, 0.17568282783031464, -0.06385239958763123, -0.01689169742166996, -0.10847058892250061, 0.1899847388267517, 0.03989950940012932, -0.03787495940923691, 0.03915709629654884, -0.06612126529216766, -0.011634225957095623, 0.1944432258605957, 0.18756258487701416, -0.021927885711193085, -0.0002661887847352773, -0.004998415242880583, -0.01534326747059822, 0.0016246484592556953, 0.08369015157222748, 0.11981640011072159, -0.013047742657363415, -0.06906230002641678, -0.01819325052201748, -0.06562758982181549, -0.0067472620867192745, -0.04779968410730362, 0.0645025223493576, 0.02037656493484974, 0.008185519836843014, -0.05136895924806595, 0.023522306233644485, -0.034733060747385025, -0.07609675824642181, 0.038163378834724426, -0.21621042490005493, -0.15695160627365112, -0.006919717416167259, 0.04431783780455589, -0.0026687439531087875, 0.06132279336452484, -0.0016218317905440927, 0.012163124978542328, 0.0941525548696518, -0.020845945924520493, -0.08206836134195328, -0.08182139694690704, 0.08768419921398163, -0.15437477827072144, 0.20841164886951447, -0.04061716049909592, 0.04299676790833473, 0.13137340545654297, 0.04467977210879326, -0.1047448068857193, 0.03841583803296089, 0.03423437103629112, -0.00853665266185999, 0.004708867520093918, 0.09485199302434921, -0.016044387593865395, 0.07863204181194305, 0.03675774112343788, -0.09839759767055511, -0.03282250836491585, -0.06005345657467842, -0.02700952999293804, -0.04377070441842079, -0.049499623477458954, -0.05170126631855965, 0.11996490508317947, 0.18235653638839722, -0.051337070763111115, -0.025411048904061317, -0.0625823587179184, 0.016907107084989548, 0.07732902467250824, -0.01597549393773079, -0.04855141416192055, -0.2462962567806244, 0.010467390529811382, 0.05714378505945206, -0.019154073670506477, -0.23835007846355438, -0.11273518949747086, -0.0007911180146038532, -0.06173158064484596, -0.08129233121871948, 0.08054714649915695, 0.043887872248888016, 0.05588071793317795, -0.0629199966788292, 0.008542655035853386, -0.0903623104095459, 0.16071398556232452, -0.15241222083568573, -0.07646507024765015 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-cased-finetuned-sst2 This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the GLUE SST2 dataset. It achieves the following results on the evaluation set: - Loss: 0.3649 - Accuracy: 0.9232 The model was fine-tuned to compare [google/fnet-base](https://huggingface.co/google/fnet-base) as introduced in [this paper](https://arxiv.org/abs/2105.03824) against [bert-base-cased](https://huggingface.co/bert-base-cased). ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure This model is trained using the [run_glue](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) script. The following command was used: ```bash #!/usr/bin/bash python ../run_glue.py \\n --model_name_or_path bert-base-cased \\n --task_name sst2 \\n --do_train \\n --do_eval \\n --max_seq_length 512 \\n --per_device_train_batch_size 16 \\n --learning_rate 2e-5 \\n --num_train_epochs 3 \\n --output_dir bert-base-cased-finetuned-sst2 \\n --push_to_hub \\n --hub_strategy all_checkpoints \\n --logging_strategy epoch \\n --save_strategy epoch \\n --evaluation_strategy epoch \\n``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Accuracy | Validation Loss | |:-------------:|:-----:|:-----:|:--------:|:---------------:| | 0.233 | 1.0 | 4210 | 0.9174 | 0.2841 | | 0.1261 | 2.0 | 8420 | 0.9278 | 0.3310 | | 0.0768 | 3.0 | 12630 | 0.9232 | 0.3649 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer", "fnet-bert-base-comparison"], "datasets": ["glue"], "metrics": ["accuracy"], "model-index": [{"name": "bert-base-cased-finetuned-sst2", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE SST2", "type": "glue", "args": "sst2"}, "metrics": [{"type": "accuracy", "value": 0.9231651376146789, "name": "Accuracy"}]}]}]}
text-classification
gchhablani/bert-base-cased-finetuned-sst2
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "fnet-bert-base-comparison", "en", "dataset:glue", "arxiv:2105.03824", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2105.03824" ]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
bert-base-cased-finetuned-sst2 ============================== This model is a fine-tuned version of bert-base-cased on the GLUE SST2 dataset. It achieves the following results on the evaluation set: * Loss: 0.3649 * Accuracy: 0.9232 The model was fine-tuned to compare google/fnet-base as introduced in this paper against bert-base-cased. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ This model is trained using the run\_glue script. The following command was used: ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 87, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.12161670625209808, 0.1586534082889557, -0.0035777657758444548, 0.11875102669000626, 0.11533597856760025, -0.0031049344688653946, 0.13364577293395996, 0.1597096472978592, -0.08531280606985092, 0.0648089274764061, 0.15738050639629364, 0.1568128913640976, 0.02761896513402462, 0.18884018063545227, -0.052745409309864044, -0.23438897728919983, 0.03223801031708717, 0.07453203946352005, -0.03721422329545021, 0.13727734982967377, 0.10004161298274994, -0.12226498872041702, 0.09843768924474716, 0.03329526633024216, -0.20268045365810394, -0.0181635320186615, 0.0035198922269046307, -0.08516480028629303, 0.11188160628080368, 0.0131202582269907, 0.08862993866205215, 0.031522486358881, 0.03825753182172775, -0.1487365961074829, 0.005022411700338125, 0.046756207942962646, 0.0042826407589018345, 0.11839430034160614, 0.03855927288532257, -0.01641850546002388, 0.0520799495279789, -0.0885375365614891, 0.05395820736885071, 0.028443723917007446, -0.1181475892663002, -0.26681092381477356, -0.0919037014245987, 0.06938058882951736, 0.046251799911260605, 0.07366777211427689, -0.00005182146924198605, 0.16483885049819946, -0.005088786594569683, 0.10346852988004684, 0.27087369561195374, -0.3125593364238739, -0.05947171151638031, 0.02776799350976944, 0.011863073334097862, 0.05958991125226021, -0.09307480603456497, -0.02570064552128315, 0.05127633363008499, 0.04167178273200989, 0.18514986336231232, -0.016228215768933296, -0.00003337187081342563, -0.02189422771334648, -0.1382196545600891, -0.07627017796039581, 0.20115217566490173, 0.05067397281527519, -0.05151061341166496, -0.07212591916322708, -0.07642883062362671, -0.16918812692165375, -0.03298860415816307, -0.020403943955898285, 0.046687398105859756, -0.035265833139419556, -0.06775462627410889, -0.019117340445518494, -0.07634665071964264, -0.04710827395319939, -0.038415029644966125, 0.15217410027980804, 0.04655686020851135, 0.022712748497724533, -0.025998499244451523, 0.09033547341823578, -0.015903212130069733, -0.15715816617012024, -0.005602593533694744, 0.011772151105105877, 0.037549275904893875, -0.022561458870768547, -0.036397535353899, -0.08168396353721619, 0.012048215605318546, 0.14117512106895447, -0.08264761418104172, 0.06482455879449844, 0.011765061877667904, 0.0496395118534565, -0.08314628899097443, 0.1905018389225006, -0.024279993027448654, 0.0037186690606176853, 0.02406053990125656, 0.08622891455888748, 0.04516247287392616, -0.026812154799699783, -0.11238549649715424, 0.0233257245272398, 0.14243075251579285, 0.007431734818965197, -0.039754197001457214, 0.07417906820774078, -0.059125080704689026, -0.04582800716161728, 0.04539237171411514, -0.11030158400535583, 0.0141338761895895, -0.009373044595122337, -0.08056624233722687, -0.03472790867090225, 0.02697722613811493, -0.011787989176809788, -0.032092269510030746, 0.06212129071354866, -0.09913354367017746, 0.008189168758690357, -0.06144571676850319, -0.10696113854646683, 0.015179385431110859, -0.11808769404888153, -0.001081367488950491, -0.11012967675924301, -0.13974249362945557, -0.006976991426199675, 0.053723063319921494, -0.023807602003216743, -0.0772162452340126, -0.05605267360806465, -0.07926763594150543, 0.029673025012016296, -0.0168012622743845, 0.04827882722020149, -0.06232791766524315, 0.08783834427595139, 0.04711596667766571, 0.07783091068267822, -0.03511257842183113, 0.04865354299545288, -0.0857672467827797, 0.04486498609185219, -0.20575650036334991, 0.06056206300854683, -0.057986579835414886, 0.06628572940826416, -0.1133461743593216, -0.11155343800783157, 0.0248954389244318, -0.03267720714211464, 0.0798325464129448, 0.10354934632778168, -0.14306844770908356, -0.0819404348731041, 0.1941177397966385, -0.08345154672861099, -0.13623358309268951, 0.11876475811004639, -0.04778644070029259, 0.023907052353024483, 0.06198955699801445, 0.22297649085521698, 0.0731961652636528, -0.057169198989868164, -0.026051243767142296, -0.017552388831973076, 0.051131829619407654, -0.07172699272632599, 0.08445700258016586, -0.004041696432977915, 0.028126448392868042, 0.030312296003103256, -0.035424020141363144, 0.032627593725919724, -0.08209208399057388, -0.08335229009389877, -0.05411062017083168, -0.0801575630903244, 0.06567159295082092, 0.04338371008634567, 0.08023004233837128, -0.11180076748132706, -0.09540257602930069, 0.034219104796648026, 0.08448748290538788, -0.08299247920513153, 0.045332007110118866, -0.0944017842411995, 0.12313508242368698, -0.07442712783813477, -0.005775441415607929, -0.18186262249946594, -0.004674280993640423, 0.04966714233160019, -0.028262924402952194, 0.004111959598958492, -0.025413617491722107, 0.06383029371500015, 0.051281414926052094, -0.046544935554265976, -0.04452786594629288, -0.037859585136175156, -0.007022486068308353, -0.1164112463593483, -0.17888443171977997, -0.04770522564649582, -0.03426934406161308, 0.11073228716850281, -0.15757550299167633, 0.05754205211997032, 0.0591144785284996, 0.10811145603656769, 0.02597060613334179, -0.03133917599916458, -0.009403175674378872, 0.04480119049549103, -0.04116583615541458, -0.07460780441761017, 0.07281174510717392, 0.03331552818417549, -0.09998483955860138, -0.02359735779464245, -0.11360644549131393, 0.1753949373960495, 0.12714703381061554, -0.022101160138845444, -0.05310492590069771, -0.009392490610480309, -0.06040368974208832, -0.023967178538441658, -0.010677943006157875, 0.013884914107620716, 0.17248490452766418, 0.005367063917219639, 0.17223170399665833, -0.10302618145942688, -0.05416569113731384, 0.04557802155613899, -0.0306826401501894, -0.013940786942839622, 0.10292626917362213, 0.01127760112285614, -0.10022857785224915, 0.15135939419269562, 0.1410094052553177, -0.055681660771369934, 0.12143083661794662, -0.05604437366127968, -0.047931358218193054, -0.039293039590120316, -0.0005570283392444253, 0.018249090760946274, 0.09346067160367966, -0.11466828733682632, -0.018797609955072403, 0.039005741477012634, 0.028456371277570724, 0.011234947480261326, -0.1804705411195755, 0.001110660727135837, 0.04256647825241089, -0.0585462749004364, 0.007875440642237663, -0.005216784775257111, -0.005512913689017296, 0.10080158710479736, 0.02539101243019104, -0.0727960467338562, 0.052135586738586426, 0.010471811518073082, -0.07134804874658585, 0.19735264778137207, -0.09606501460075378, -0.19630330801010132, -0.12494390457868576, -0.06156739220023155, -0.09005486220121384, 0.006349241826683283, 0.06912596523761749, -0.07998423278331757, -0.024601668119430542, -0.09020604938268661, -0.027059001848101616, -0.014606335200369358, 0.03885342925786972, 0.07165474444627762, -0.02355276234447956, 0.10255801677703857, -0.120688796043396, -0.030849842354655266, -0.03027579002082348, 0.0014742627972736955, 0.05496053025126457, 0.006049624178558588, 0.10471925884485245, 0.11509934067726135, -0.02873915433883667, 0.05114046484231949, -0.03214579448103905, 0.2341459095478058, -0.04930265247821808, -0.02795281447470188, 0.12977902591228485, -0.005955296568572521, 0.08535684645175934, 0.0899367704987526, 0.045135561376810074, -0.09218605607748032, -0.00756901316344738, 0.005072117783129215, -0.040117885917425156, -0.21103355288505554, -0.035214588046073914, -0.04211072623729706, 0.020072607323527336, 0.12289147824048996, 0.04161820560693741, 0.05459532514214516, 0.0637349858880043, 0.029591871425509453, 0.0592803992331028, -0.028413070365786552, 0.1009601503610611, 0.12848083674907684, 0.05054585635662079, 0.13528481125831604, -0.03867393359541893, -0.03375968337059021, 0.040527984499931335, 0.0007192929624579847, 0.20465855300426483, -0.014075764454901218, 0.1890448033809662, 0.04658591002225876, 0.18518702685832977, 0.010942237451672554, 0.06792156398296356, -0.0207637008279562, -0.004309804644435644, -0.017503520473837852, -0.04292628914117813, -0.060658156871795654, 0.009485099464654922, -0.04635186865925789, 0.07441109418869019, -0.12235406786203384, 0.019649730995297432, 0.060750771313905716, 0.291361004114151, 0.02242939919233322, -0.3738984167575836, -0.1097446084022522, -0.011077821254730225, -0.025019990280270576, -0.04325328394770622, 0.011022142134606838, 0.09511391818523407, -0.09566101431846619, 0.06279890984296799, -0.08686861395835876, 0.08955012261867523, -0.073078453540802, 0.03659684211015701, 0.049923211336135864, 0.09388187527656555, 0.005718675907701254, 0.057970207184553146, -0.27198654413223267, 0.2506929934024811, 0.017948467284440994, 0.05225548893213272, -0.06244153901934624, 0.014453649520874023, 0.022270919755101204, 0.0580802857875824, 0.08689375221729279, -0.0023560896515846252, -0.03474327549338341, -0.17138437926769257, -0.10784955322742462, 0.012887988239526749, 0.07278495281934738, -0.04178544878959656, 0.0889785885810852, -0.010356182232499123, 0.003321166383102536, 0.04448032006621361, -0.003272048430517316, -0.033030617982149124, -0.09141189604997635, 0.021423611789941788, 0.06062246859073639, -0.022481713443994522, -0.0803140178322792, -0.11677858233451843, -0.08531743288040161, 0.1759343147277832, -0.006428628694266081, -0.07254543155431747, -0.12288008630275726, 0.061379604041576385, 0.06829093396663666, -0.09375333040952682, 0.04493061453104019, -0.02001975290477276, 0.12447743117809296, 0.009351304732263088, -0.0690632313489914, 0.09878260642290115, -0.047859106212854385, -0.16231606900691986, -0.0379788912832737, 0.13346640765666962, 0.02778930962085724, 0.055701445788145065, -0.007568121422082186, 0.02814381755888462, -0.0244494266808033, -0.07651051878929138, 0.03854655846953392, 0.007937511429190636, 0.09600245952606201, -0.019526541233062744, -0.019290292635560036, 0.03250190243124962, -0.07478196173906326, -0.004258669447153807, 0.1966393142938614, 0.2564740777015686, -0.10714545100927353, 0.043486859649419785, 0.03290016949176788, -0.05048590898513794, -0.1577579379081726, 0.013250388205051422, 0.07308094948530197, 0.00450450275093317, 0.0024772710166871548, -0.17275027930736542, 0.055431436747312546, 0.09000243246555328, -0.01831885054707527, 0.07066475600004196, -0.29239320755004883, -0.11998265236616135, 0.10719365626573563, 0.12865246832370758, 0.10155823826789856, -0.14214134216308594, -0.050149016082286835, -0.013447506353259087, -0.13887791335582733, 0.11616533249616623, -0.0724763497710228, 0.11383344233036041, -0.046218473464250565, 0.05384330451488495, 0.0073478645645082, -0.05154034122824669, 0.12863507866859436, 0.02274453639984131, 0.07456807792186737, -0.052454277873039246, -0.012860242277383804, 0.09835619479417801, -0.07845889031887054, 0.06105131283402443, -0.09937947243452072, 0.045091770589351654, -0.12510208785533905, -0.015082119032740593, -0.07675468176603317, 0.02839023247361183, -0.028737343847751617, -0.042841967195272446, -0.05252382904291153, 0.01206414494663477, 0.07457873225212097, -0.0033069076016545296, 0.17504702508449554, 0.04161931201815605, 0.1356934756040573, 0.18485517799854279, 0.07381550967693329, -0.11268056184053421, -0.10376991331577301, -0.00886673852801323, -0.02009434811770916, 0.05579818785190582, -0.1618359386920929, 0.042422011494636536, 0.13992659747600555, 0.006874611135572195, 0.12767653167247772, 0.0688827782869339, -0.04798436909914017, 0.003899581264704466, 0.0452854298055172, -0.1783701628446579, -0.09679107367992401, -0.009332024492323399, -0.012513089925050735, -0.13734038174152374, 0.07031508535146713, 0.11065638810396194, -0.06522715091705322, -0.020677953958511353, 0.0015312001341953874, 0.001479341764934361, -0.023513123393058777, 0.17326204478740692, 0.0655597522854805, 0.06481005996465683, -0.10098203271627426, 0.0951208770275116, 0.04609247297048569, -0.07288358360528946, 0.04003791883587837, 0.05164699628949165, -0.1138383150100708, -0.02253587916493416, 0.03539628908038139, 0.1590377688407898, -0.03593802452087402, -0.046655960381031036, -0.17098021507263184, -0.10509005934000015, 0.08731745183467865, 0.1274595707654953, 0.10466798394918442, 0.020139098167419434, -0.048447709530591965, -0.006609742529690266, -0.10675366967916489, 0.09348485618829727, 0.058594346046447754, 0.07109634578227997, -0.15644440054893494, 0.1105559915304184, 0.001110950019210577, 0.0607583224773407, -0.012919686734676361, 0.001354161067865789, -0.09134423732757568, -0.0003082616603933275, -0.08251616358757019, 0.0122391851618886, -0.04241093248128891, -0.003554311813786626, -0.02001729980111122, -0.05084189772605896, -0.05258030444383621, 0.031079620122909546, -0.10778090357780457, -0.03766832500696182, 0.029691802337765694, 0.04227648675441742, -0.1072600930929184, -0.030516671016812325, 0.011589708738029003, -0.08385424315929413, 0.09281181544065475, 0.04818247631192207, -0.005318017210811377, 0.013172389008104801, -0.024976933375000954, 0.00860611442476511, 0.06548852473497391, 0.0009176498278975487, 0.05490829423069954, -0.12593981623649597, -0.005958652589470148, 0.0027340054512023926, 0.0033404123969376087, 0.022175608202815056, 0.11886222660541534, -0.1292734295129776, -0.010102422907948494, -0.019927645102143288, -0.0311695858836174, -0.07569055259227753, 0.05729673057794571, 0.10719096660614014, 0.0378112718462944, 0.1875617355108261, -0.06822221726179123, 0.018444180488586426, -0.2068903148174286, -0.0002452793996781111, 0.005979603622108698, -0.1454673707485199, -0.07120571285486221, -0.036396682262420654, 0.05290095508098602, -0.06443101167678833, 0.11634373664855957, 0.0008753861766308546, -0.010869303718209267, 0.037613291293382645, -0.03043281100690365, -0.017883090302348137, 0.007872702553868294, 0.17467574775218964, 0.01212515588849783, -0.03913277015089989, 0.11452601850032806, 0.02858971431851387, 0.10287530720233917, 0.11307211220264435, 0.16108684241771698, 0.13502591848373413, 0.025834301486611366, 0.0998968780040741, 0.025341186672449112, -0.01848791167140007, -0.19047077000141144, 0.07276903837919235, -0.03531571105122566, 0.12698744237422943, 0.008672770112752914, 0.19214610755443573, 0.09776225686073303, -0.15829795598983765, 0.05093172565102577, -0.02248046174645424, -0.08756396174430847, -0.09562913328409195, -0.09577185660600662, -0.08325570076704025, -0.15142379701137543, -0.009702731855213642, -0.11629122495651245, 0.0066182236187160015, 0.07528655976057053, -0.0008431302267126739, -0.01898612640798092, 0.14263658225536346, 0.009304924868047237, 0.001253073220141232, 0.07858073711395264, -0.006525959819555283, -0.05684375762939453, -0.07991866022348404, -0.07751739770174026, 0.003022255375981331, 0.01993376761674881, 0.05594443157315254, -0.02513107657432556, 0.00735208997502923, 0.03755185008049011, -0.028129376471042633, -0.10968209058046341, 0.010150511749088764, 0.02177766151726246, 0.051356241106987, 0.02320592664182186, 0.00956298690289259, -0.01430518552660942, -0.012962219305336475, 0.17568282783031464, -0.06385239958763123, -0.01689169742166996, -0.10847058892250061, 0.1899847388267517, 0.03989950940012932, -0.03787495940923691, 0.03915709629654884, -0.06612126529216766, -0.011634225957095623, 0.1944432258605957, 0.18756258487701416, -0.021927885711193085, -0.0002661887847352773, -0.004998415242880583, -0.01534326747059822, 0.0016246484592556953, 0.08369015157222748, 0.11981640011072159, -0.013047742657363415, -0.06906230002641678, -0.01819325052201748, -0.06562758982181549, -0.0067472620867192745, -0.04779968410730362, 0.0645025223493576, 0.02037656493484974, 0.008185519836843014, -0.05136895924806595, 0.023522306233644485, -0.034733060747385025, -0.07609675824642181, 0.038163378834724426, -0.21621042490005493, -0.15695160627365112, -0.006919717416167259, 0.04431783780455589, -0.0026687439531087875, 0.06132279336452484, -0.0016218317905440927, 0.012163124978542328, 0.0941525548696518, -0.020845945924520493, -0.08206836134195328, -0.08182139694690704, 0.08768419921398163, -0.15437477827072144, 0.20841164886951447, -0.04061716049909592, 0.04299676790833473, 0.13137340545654297, 0.04467977210879326, -0.1047448068857193, 0.03841583803296089, 0.03423437103629112, -0.00853665266185999, 0.004708867520093918, 0.09485199302434921, -0.016044387593865395, 0.07863204181194305, 0.03675774112343788, -0.09839759767055511, -0.03282250836491585, -0.06005345657467842, -0.02700952999293804, -0.04377070441842079, -0.049499623477458954, -0.05170126631855965, 0.11996490508317947, 0.18235653638839722, -0.051337070763111115, -0.025411048904061317, -0.0625823587179184, 0.016907107084989548, 0.07732902467250824, -0.01597549393773079, -0.04855141416192055, -0.2462962567806244, 0.010467390529811382, 0.05714378505945206, -0.019154073670506477, -0.23835007846355438, -0.11273518949747086, -0.0007911180146038532, -0.06173158064484596, -0.08129233121871948, 0.08054714649915695, 0.043887872248888016, 0.05588071793317795, -0.0629199966788292, 0.008542655035853386, -0.0903623104095459, 0.16071398556232452, -0.15241222083568573, -0.07646507024765015 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-cased-finetuned-stsb This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the GLUE STSB dataset. It achieves the following results on the evaluation set: - Loss: 0.4861 - Pearson: 0.8926 - Spearmanr: 0.8898 - Combined Score: 0.8912 The model was fine-tuned to compare [google/fnet-base](https://huggingface.co/google/fnet-base) as introduced in [this paper](https://arxiv.org/abs/2105.03824) against [bert-base-cased](https://huggingface.co/bert-base-cased). ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure This model is trained using the [run_glue](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) script. The following command was used: ```bash #!/usr/bin/bash python ../run_glue.py \\n --model_name_or_path bert-base-cased \\n --task_name stsb \\n --do_train \\n --do_eval \\n --max_seq_length 512 \\n --per_device_train_batch_size 16 \\n --learning_rate 2e-5 \\n --num_train_epochs 3 \\n --output_dir bert-base-cased-finetuned-stsb \\n --push_to_hub \\n --hub_strategy all_checkpoints \\n --logging_strategy epoch \\n --save_strategy epoch \\n --evaluation_strategy epoch \\n``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Combined Score | Validation Loss | Pearson | Spearmanr | |:-------------:|:-----:|:----:|:--------------:|:---------------:|:-------:|:---------:| | 1.1174 | 1.0 | 360 | 0.8816 | 0.5000 | 0.8832 | 0.8800 | | 0.3835 | 2.0 | 720 | 0.8901 | 0.4672 | 0.8915 | 0.8888 | | 0.2388 | 3.0 | 1080 | 0.8912 | 0.4861 | 0.8926 | 0.8898 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer", "fnet-bert-base-comparison"], "datasets": ["glue"], "metrics": ["spearmanr"], "model-index": [{"name": "bert-base-cased-finetuned-stsb", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE STSB", "type": "glue", "args": "stsb"}, "metrics": [{"type": "spearmanr", "value": 0.8897907271421561, "name": "Spearmanr"}]}]}]}
text-classification
gchhablani/bert-base-cased-finetuned-stsb
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "fnet-bert-base-comparison", "en", "dataset:glue", "arxiv:2105.03824", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2105.03824" ]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
bert-base-cased-finetuned-stsb ============================== This model is a fine-tuned version of bert-base-cased on the GLUE STSB dataset. It achieves the following results on the evaluation set: * Loss: 0.4861 * Pearson: 0.8926 * Spearmanr: 0.8898 * Combined Score: 0.8912 The model was fine-tuned to compare google/fnet-base as introduced in this paper against bert-base-cased. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ This model is trained using the run\_glue script. The following command was used: ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 87, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.12161670625209808, 0.1586534082889557, -0.0035777657758444548, 0.11875102669000626, 0.11533597856760025, -0.0031049344688653946, 0.13364577293395996, 0.1597096472978592, -0.08531280606985092, 0.0648089274764061, 0.15738050639629364, 0.1568128913640976, 0.02761896513402462, 0.18884018063545227, -0.052745409309864044, -0.23438897728919983, 0.03223801031708717, 0.07453203946352005, -0.03721422329545021, 0.13727734982967377, 0.10004161298274994, -0.12226498872041702, 0.09843768924474716, 0.03329526633024216, -0.20268045365810394, -0.0181635320186615, 0.0035198922269046307, -0.08516480028629303, 0.11188160628080368, 0.0131202582269907, 0.08862993866205215, 0.031522486358881, 0.03825753182172775, -0.1487365961074829, 0.005022411700338125, 0.046756207942962646, 0.0042826407589018345, 0.11839430034160614, 0.03855927288532257, -0.01641850546002388, 0.0520799495279789, -0.0885375365614891, 0.05395820736885071, 0.028443723917007446, -0.1181475892663002, -0.26681092381477356, -0.0919037014245987, 0.06938058882951736, 0.046251799911260605, 0.07366777211427689, -0.00005182146924198605, 0.16483885049819946, -0.005088786594569683, 0.10346852988004684, 0.27087369561195374, -0.3125593364238739, -0.05947171151638031, 0.02776799350976944, 0.011863073334097862, 0.05958991125226021, -0.09307480603456497, -0.02570064552128315, 0.05127633363008499, 0.04167178273200989, 0.18514986336231232, -0.016228215768933296, -0.00003337187081342563, -0.02189422771334648, -0.1382196545600891, -0.07627017796039581, 0.20115217566490173, 0.05067397281527519, -0.05151061341166496, -0.07212591916322708, -0.07642883062362671, -0.16918812692165375, -0.03298860415816307, -0.020403943955898285, 0.046687398105859756, -0.035265833139419556, -0.06775462627410889, -0.019117340445518494, -0.07634665071964264, -0.04710827395319939, -0.038415029644966125, 0.15217410027980804, 0.04655686020851135, 0.022712748497724533, -0.025998499244451523, 0.09033547341823578, -0.015903212130069733, -0.15715816617012024, -0.005602593533694744, 0.011772151105105877, 0.037549275904893875, -0.022561458870768547, -0.036397535353899, -0.08168396353721619, 0.012048215605318546, 0.14117512106895447, -0.08264761418104172, 0.06482455879449844, 0.011765061877667904, 0.0496395118534565, -0.08314628899097443, 0.1905018389225006, -0.024279993027448654, 0.0037186690606176853, 0.02406053990125656, 0.08622891455888748, 0.04516247287392616, -0.026812154799699783, -0.11238549649715424, 0.0233257245272398, 0.14243075251579285, 0.007431734818965197, -0.039754197001457214, 0.07417906820774078, -0.059125080704689026, -0.04582800716161728, 0.04539237171411514, -0.11030158400535583, 0.0141338761895895, -0.009373044595122337, -0.08056624233722687, -0.03472790867090225, 0.02697722613811493, -0.011787989176809788, -0.032092269510030746, 0.06212129071354866, -0.09913354367017746, 0.008189168758690357, -0.06144571676850319, -0.10696113854646683, 0.015179385431110859, -0.11808769404888153, -0.001081367488950491, -0.11012967675924301, -0.13974249362945557, -0.006976991426199675, 0.053723063319921494, -0.023807602003216743, -0.0772162452340126, -0.05605267360806465, -0.07926763594150543, 0.029673025012016296, -0.0168012622743845, 0.04827882722020149, -0.06232791766524315, 0.08783834427595139, 0.04711596667766571, 0.07783091068267822, -0.03511257842183113, 0.04865354299545288, -0.0857672467827797, 0.04486498609185219, -0.20575650036334991, 0.06056206300854683, -0.057986579835414886, 0.06628572940826416, -0.1133461743593216, -0.11155343800783157, 0.0248954389244318, -0.03267720714211464, 0.0798325464129448, 0.10354934632778168, -0.14306844770908356, -0.0819404348731041, 0.1941177397966385, -0.08345154672861099, -0.13623358309268951, 0.11876475811004639, -0.04778644070029259, 0.023907052353024483, 0.06198955699801445, 0.22297649085521698, 0.0731961652636528, -0.057169198989868164, -0.026051243767142296, -0.017552388831973076, 0.051131829619407654, -0.07172699272632599, 0.08445700258016586, -0.004041696432977915, 0.028126448392868042, 0.030312296003103256, -0.035424020141363144, 0.032627593725919724, -0.08209208399057388, -0.08335229009389877, -0.05411062017083168, -0.0801575630903244, 0.06567159295082092, 0.04338371008634567, 0.08023004233837128, -0.11180076748132706, -0.09540257602930069, 0.034219104796648026, 0.08448748290538788, -0.08299247920513153, 0.045332007110118866, -0.0944017842411995, 0.12313508242368698, -0.07442712783813477, -0.005775441415607929, -0.18186262249946594, -0.004674280993640423, 0.04966714233160019, -0.028262924402952194, 0.004111959598958492, -0.025413617491722107, 0.06383029371500015, 0.051281414926052094, -0.046544935554265976, -0.04452786594629288, -0.037859585136175156, -0.007022486068308353, -0.1164112463593483, -0.17888443171977997, -0.04770522564649582, -0.03426934406161308, 0.11073228716850281, -0.15757550299167633, 0.05754205211997032, 0.0591144785284996, 0.10811145603656769, 0.02597060613334179, -0.03133917599916458, -0.009403175674378872, 0.04480119049549103, -0.04116583615541458, -0.07460780441761017, 0.07281174510717392, 0.03331552818417549, -0.09998483955860138, -0.02359735779464245, -0.11360644549131393, 0.1753949373960495, 0.12714703381061554, -0.022101160138845444, -0.05310492590069771, -0.009392490610480309, -0.06040368974208832, -0.023967178538441658, -0.010677943006157875, 0.013884914107620716, 0.17248490452766418, 0.005367063917219639, 0.17223170399665833, -0.10302618145942688, -0.05416569113731384, 0.04557802155613899, -0.0306826401501894, -0.013940786942839622, 0.10292626917362213, 0.01127760112285614, -0.10022857785224915, 0.15135939419269562, 0.1410094052553177, -0.055681660771369934, 0.12143083661794662, -0.05604437366127968, -0.047931358218193054, -0.039293039590120316, -0.0005570283392444253, 0.018249090760946274, 0.09346067160367966, -0.11466828733682632, -0.018797609955072403, 0.039005741477012634, 0.028456371277570724, 0.011234947480261326, -0.1804705411195755, 0.001110660727135837, 0.04256647825241089, -0.0585462749004364, 0.007875440642237663, -0.005216784775257111, -0.005512913689017296, 0.10080158710479736, 0.02539101243019104, -0.0727960467338562, 0.052135586738586426, 0.010471811518073082, -0.07134804874658585, 0.19735264778137207, -0.09606501460075378, -0.19630330801010132, -0.12494390457868576, -0.06156739220023155, -0.09005486220121384, 0.006349241826683283, 0.06912596523761749, -0.07998423278331757, -0.024601668119430542, -0.09020604938268661, -0.027059001848101616, -0.014606335200369358, 0.03885342925786972, 0.07165474444627762, -0.02355276234447956, 0.10255801677703857, -0.120688796043396, -0.030849842354655266, -0.03027579002082348, 0.0014742627972736955, 0.05496053025126457, 0.006049624178558588, 0.10471925884485245, 0.11509934067726135, -0.02873915433883667, 0.05114046484231949, -0.03214579448103905, 0.2341459095478058, -0.04930265247821808, -0.02795281447470188, 0.12977902591228485, -0.005955296568572521, 0.08535684645175934, 0.0899367704987526, 0.045135561376810074, -0.09218605607748032, -0.00756901316344738, 0.005072117783129215, -0.040117885917425156, -0.21103355288505554, -0.035214588046073914, -0.04211072623729706, 0.020072607323527336, 0.12289147824048996, 0.04161820560693741, 0.05459532514214516, 0.0637349858880043, 0.029591871425509453, 0.0592803992331028, -0.028413070365786552, 0.1009601503610611, 0.12848083674907684, 0.05054585635662079, 0.13528481125831604, -0.03867393359541893, -0.03375968337059021, 0.040527984499931335, 0.0007192929624579847, 0.20465855300426483, -0.014075764454901218, 0.1890448033809662, 0.04658591002225876, 0.18518702685832977, 0.010942237451672554, 0.06792156398296356, -0.0207637008279562, -0.004309804644435644, -0.017503520473837852, -0.04292628914117813, -0.060658156871795654, 0.009485099464654922, -0.04635186865925789, 0.07441109418869019, -0.12235406786203384, 0.019649730995297432, 0.060750771313905716, 0.291361004114151, 0.02242939919233322, -0.3738984167575836, -0.1097446084022522, -0.011077821254730225, -0.025019990280270576, -0.04325328394770622, 0.011022142134606838, 0.09511391818523407, -0.09566101431846619, 0.06279890984296799, -0.08686861395835876, 0.08955012261867523, -0.073078453540802, 0.03659684211015701, 0.049923211336135864, 0.09388187527656555, 0.005718675907701254, 0.057970207184553146, -0.27198654413223267, 0.2506929934024811, 0.017948467284440994, 0.05225548893213272, -0.06244153901934624, 0.014453649520874023, 0.022270919755101204, 0.0580802857875824, 0.08689375221729279, -0.0023560896515846252, -0.03474327549338341, -0.17138437926769257, -0.10784955322742462, 0.012887988239526749, 0.07278495281934738, -0.04178544878959656, 0.0889785885810852, -0.010356182232499123, 0.003321166383102536, 0.04448032006621361, -0.003272048430517316, -0.033030617982149124, -0.09141189604997635, 0.021423611789941788, 0.06062246859073639, -0.022481713443994522, -0.0803140178322792, -0.11677858233451843, -0.08531743288040161, 0.1759343147277832, -0.006428628694266081, -0.07254543155431747, -0.12288008630275726, 0.061379604041576385, 0.06829093396663666, -0.09375333040952682, 0.04493061453104019, -0.02001975290477276, 0.12447743117809296, 0.009351304732263088, -0.0690632313489914, 0.09878260642290115, -0.047859106212854385, -0.16231606900691986, -0.0379788912832737, 0.13346640765666962, 0.02778930962085724, 0.055701445788145065, -0.007568121422082186, 0.02814381755888462, -0.0244494266808033, -0.07651051878929138, 0.03854655846953392, 0.007937511429190636, 0.09600245952606201, -0.019526541233062744, -0.019290292635560036, 0.03250190243124962, -0.07478196173906326, -0.004258669447153807, 0.1966393142938614, 0.2564740777015686, -0.10714545100927353, 0.043486859649419785, 0.03290016949176788, -0.05048590898513794, -0.1577579379081726, 0.013250388205051422, 0.07308094948530197, 0.00450450275093317, 0.0024772710166871548, -0.17275027930736542, 0.055431436747312546, 0.09000243246555328, -0.01831885054707527, 0.07066475600004196, -0.29239320755004883, -0.11998265236616135, 0.10719365626573563, 0.12865246832370758, 0.10155823826789856, -0.14214134216308594, -0.050149016082286835, -0.013447506353259087, -0.13887791335582733, 0.11616533249616623, -0.0724763497710228, 0.11383344233036041, -0.046218473464250565, 0.05384330451488495, 0.0073478645645082, -0.05154034122824669, 0.12863507866859436, 0.02274453639984131, 0.07456807792186737, -0.052454277873039246, -0.012860242277383804, 0.09835619479417801, -0.07845889031887054, 0.06105131283402443, -0.09937947243452072, 0.045091770589351654, -0.12510208785533905, -0.015082119032740593, -0.07675468176603317, 0.02839023247361183, -0.028737343847751617, -0.042841967195272446, -0.05252382904291153, 0.01206414494663477, 0.07457873225212097, -0.0033069076016545296, 0.17504702508449554, 0.04161931201815605, 0.1356934756040573, 0.18485517799854279, 0.07381550967693329, -0.11268056184053421, -0.10376991331577301, -0.00886673852801323, -0.02009434811770916, 0.05579818785190582, -0.1618359386920929, 0.042422011494636536, 0.13992659747600555, 0.006874611135572195, 0.12767653167247772, 0.0688827782869339, -0.04798436909914017, 0.003899581264704466, 0.0452854298055172, -0.1783701628446579, -0.09679107367992401, -0.009332024492323399, -0.012513089925050735, -0.13734038174152374, 0.07031508535146713, 0.11065638810396194, -0.06522715091705322, -0.020677953958511353, 0.0015312001341953874, 0.001479341764934361, -0.023513123393058777, 0.17326204478740692, 0.0655597522854805, 0.06481005996465683, -0.10098203271627426, 0.0951208770275116, 0.04609247297048569, -0.07288358360528946, 0.04003791883587837, 0.05164699628949165, -0.1138383150100708, -0.02253587916493416, 0.03539628908038139, 0.1590377688407898, -0.03593802452087402, -0.046655960381031036, -0.17098021507263184, -0.10509005934000015, 0.08731745183467865, 0.1274595707654953, 0.10466798394918442, 0.020139098167419434, -0.048447709530591965, -0.006609742529690266, -0.10675366967916489, 0.09348485618829727, 0.058594346046447754, 0.07109634578227997, -0.15644440054893494, 0.1105559915304184, 0.001110950019210577, 0.0607583224773407, -0.012919686734676361, 0.001354161067865789, -0.09134423732757568, -0.0003082616603933275, -0.08251616358757019, 0.0122391851618886, -0.04241093248128891, -0.003554311813786626, -0.02001729980111122, -0.05084189772605896, -0.05258030444383621, 0.031079620122909546, -0.10778090357780457, -0.03766832500696182, 0.029691802337765694, 0.04227648675441742, -0.1072600930929184, -0.030516671016812325, 0.011589708738029003, -0.08385424315929413, 0.09281181544065475, 0.04818247631192207, -0.005318017210811377, 0.013172389008104801, -0.024976933375000954, 0.00860611442476511, 0.06548852473497391, 0.0009176498278975487, 0.05490829423069954, -0.12593981623649597, -0.005958652589470148, 0.0027340054512023926, 0.0033404123969376087, 0.022175608202815056, 0.11886222660541534, -0.1292734295129776, -0.010102422907948494, -0.019927645102143288, -0.0311695858836174, -0.07569055259227753, 0.05729673057794571, 0.10719096660614014, 0.0378112718462944, 0.1875617355108261, -0.06822221726179123, 0.018444180488586426, -0.2068903148174286, -0.0002452793996781111, 0.005979603622108698, -0.1454673707485199, -0.07120571285486221, -0.036396682262420654, 0.05290095508098602, -0.06443101167678833, 0.11634373664855957, 0.0008753861766308546, -0.010869303718209267, 0.037613291293382645, -0.03043281100690365, -0.017883090302348137, 0.007872702553868294, 0.17467574775218964, 0.01212515588849783, -0.03913277015089989, 0.11452601850032806, 0.02858971431851387, 0.10287530720233917, 0.11307211220264435, 0.16108684241771698, 0.13502591848373413, 0.025834301486611366, 0.0998968780040741, 0.025341186672449112, -0.01848791167140007, -0.19047077000141144, 0.07276903837919235, -0.03531571105122566, 0.12698744237422943, 0.008672770112752914, 0.19214610755443573, 0.09776225686073303, -0.15829795598983765, 0.05093172565102577, -0.02248046174645424, -0.08756396174430847, -0.09562913328409195, -0.09577185660600662, -0.08325570076704025, -0.15142379701137543, -0.009702731855213642, -0.11629122495651245, 0.0066182236187160015, 0.07528655976057053, -0.0008431302267126739, -0.01898612640798092, 0.14263658225536346, 0.009304924868047237, 0.001253073220141232, 0.07858073711395264, -0.006525959819555283, -0.05684375762939453, -0.07991866022348404, -0.07751739770174026, 0.003022255375981331, 0.01993376761674881, 0.05594443157315254, -0.02513107657432556, 0.00735208997502923, 0.03755185008049011, -0.028129376471042633, -0.10968209058046341, 0.010150511749088764, 0.02177766151726246, 0.051356241106987, 0.02320592664182186, 0.00956298690289259, -0.01430518552660942, -0.012962219305336475, 0.17568282783031464, -0.06385239958763123, -0.01689169742166996, -0.10847058892250061, 0.1899847388267517, 0.03989950940012932, -0.03787495940923691, 0.03915709629654884, -0.06612126529216766, -0.011634225957095623, 0.1944432258605957, 0.18756258487701416, -0.021927885711193085, -0.0002661887847352773, -0.004998415242880583, -0.01534326747059822, 0.0016246484592556953, 0.08369015157222748, 0.11981640011072159, -0.013047742657363415, -0.06906230002641678, -0.01819325052201748, -0.06562758982181549, -0.0067472620867192745, -0.04779968410730362, 0.0645025223493576, 0.02037656493484974, 0.008185519836843014, -0.05136895924806595, 0.023522306233644485, -0.034733060747385025, -0.07609675824642181, 0.038163378834724426, -0.21621042490005493, -0.15695160627365112, -0.006919717416167259, 0.04431783780455589, -0.0026687439531087875, 0.06132279336452484, -0.0016218317905440927, 0.012163124978542328, 0.0941525548696518, -0.020845945924520493, -0.08206836134195328, -0.08182139694690704, 0.08768419921398163, -0.15437477827072144, 0.20841164886951447, -0.04061716049909592, 0.04299676790833473, 0.13137340545654297, 0.04467977210879326, -0.1047448068857193, 0.03841583803296089, 0.03423437103629112, -0.00853665266185999, 0.004708867520093918, 0.09485199302434921, -0.016044387593865395, 0.07863204181194305, 0.03675774112343788, -0.09839759767055511, -0.03282250836491585, -0.06005345657467842, -0.02700952999293804, -0.04377070441842079, -0.049499623477458954, -0.05170126631855965, 0.11996490508317947, 0.18235653638839722, -0.051337070763111115, -0.025411048904061317, -0.0625823587179184, 0.016907107084989548, 0.07732902467250824, -0.01597549393773079, -0.04855141416192055, -0.2462962567806244, 0.010467390529811382, 0.05714378505945206, -0.019154073670506477, -0.23835007846355438, -0.11273518949747086, -0.0007911180146038532, -0.06173158064484596, -0.08129233121871948, 0.08054714649915695, 0.043887872248888016, 0.05588071793317795, -0.0629199966788292, 0.008542655035853386, -0.0903623104095459, 0.16071398556232452, -0.15241222083568573, -0.07646507024765015 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-cased-finetuned-wnli This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the GLUE WNLI dataset. It achieves the following results on the evaluation set: - Loss: 0.6996 - Accuracy: 0.4648 The model was fine-tuned to compare [google/fnet-base](https://huggingface.co/google/fnet-base) as introduced in [this paper](https://arxiv.org/abs/2105.03824) against [bert-base-cased](https://huggingface.co/bert-base-cased). ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure This model is trained using the [run_glue](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) script. The following command was used: ```bash #!/usr/bin/bash python ../run_glue.py \\n --model_name_or_path bert-base-cased \\n --task_name wnli \\n --do_train \\n --do_eval \\n --max_seq_length 512 \\n --per_device_train_batch_size 16 \\n --learning_rate 2e-5 \\n --num_train_epochs 5 \\n --output_dir bert-base-cased-finetuned-wnli \\n --push_to_hub \\n --hub_strategy all_checkpoints \\n --logging_strategy epoch \\n --save_strategy epoch \\n --evaluation_strategy epoch \\n``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.7299 | 1.0 | 40 | 0.6923 | 0.5634 | | 0.6982 | 2.0 | 80 | 0.7027 | 0.3803 | | 0.6972 | 3.0 | 120 | 0.7005 | 0.4507 | | 0.6992 | 4.0 | 160 | 0.6977 | 0.5352 | | 0.699 | 5.0 | 200 | 0.6996 | 0.4648 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer", "fnet-bert-base-comparison"], "datasets": ["glue"], "metrics": ["accuracy"], "model-index": [{"name": "bert-base-cased-finetuned-wnli", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE WNLI", "type": "glue", "args": "wnli"}, "metrics": [{"type": "accuracy", "value": 0.4647887323943662, "name": "Accuracy"}]}]}]}
text-classification
gchhablani/bert-base-cased-finetuned-wnli
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "fnet-bert-base-comparison", "en", "dataset:glue", "arxiv:2105.03824", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2105.03824" ]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
bert-base-cased-finetuned-wnli ============================== This model is a fine-tuned version of bert-base-cased on the GLUE WNLI dataset. It achieves the following results on the evaluation set: * Loss: 0.6996 * Accuracy: 0.4648 The model was fine-tuned to compare google/fnet-base as introduced in this paper against bert-base-cased. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ This model is trained using the run\_glue script. The following command was used: ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 5.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 87, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.12273520231246948, 0.15941818058490753, -0.0036010530311614275, 0.1185033991932869, 0.1152205765247345, -0.003181234933435917, 0.1323397010564804, 0.1594047248363495, -0.08407840877771378, 0.06581338495016098, 0.15726500749588013, 0.15550516545772552, 0.02782674878835678, 0.1900021880865097, -0.052922964096069336, -0.23428578674793243, 0.03135466203093529, 0.07575137168169022, -0.037131719291210175, 0.13710589706897736, 0.09941554814577103, -0.12357978522777557, 0.09802382439374924, 0.0332266166806221, -0.20414406061172485, -0.01862412318587303, 0.0045390985906124115, -0.08521460741758347, 0.1119740903377533, 0.013080747798085213, 0.08856029063463211, 0.031364813446998596, 0.037303902208805084, -0.14865486323833466, 0.0052331616170704365, 0.04732818529009819, 0.004134282469749451, 0.11925225704908371, 0.037881553173065186, -0.01573019288480282, 0.05293995887041092, -0.08882270753383636, 0.05314220115542412, 0.02836529351770878, -0.11822879314422607, -0.270174115896225, -0.0905454233288765, 0.07019355893135071, 0.04569752886891365, 0.073348268866539, -0.00039739516796544194, 0.16413745284080505, -0.004999269265681505, 0.10369167476892471, 0.2695405185222626, -0.31407949328422546, -0.0589446984231472, 0.02677224390208721, 0.011416743509471416, 0.06014813482761383, -0.09477435052394867, -0.02597678266465664, 0.05175976827740669, 0.039981659501791, 0.1861463189125061, -0.015982169657945633, 0.0002565350441727787, -0.022086571902036667, -0.1378646194934845, -0.07677428424358368, 0.20211897790431976, 0.051418669521808624, -0.05063481628894806, -0.07162654399871826, -0.07740622013807297, -0.16897781193256378, -0.03351329267024994, -0.02044130302965641, 0.047230564057826996, -0.03537306562066078, -0.06571278721094131, -0.017890259623527527, -0.07657642662525177, -0.04722020775079727, -0.03871086984872818, 0.15120668709278107, 0.04658329114317894, 0.022782612591981888, -0.027851099148392677, 0.0897502526640892, -0.018068252131342888, -0.1580222249031067, -0.0053072599694132805, 0.011906379833817482, 0.03783579543232918, -0.02219948172569275, -0.03605634346604347, -0.08193022012710571, 0.012833919376134872, 0.14057014882564545, -0.08208564668893814, 0.06448864191770554, 0.010675122030079365, 0.04922620579600334, -0.0835108757019043, 0.1901181936264038, -0.023104993626475334, 0.0022616602946072817, 0.02452436089515686, 0.08614446967840195, 0.044317811727523804, -0.026830751448869705, -0.11209622025489807, 0.022753268480300903, 0.14248020946979523, 0.006461338605731726, -0.03920947015285492, 0.07400575280189514, -0.058990851044654846, -0.045090094208717346, 0.04464469477534294, -0.11016041040420532, 0.014915427193045616, -0.008828986436128616, -0.08084052056074142, -0.035866234451532364, 0.026778962463140488, -0.013091707602143288, -0.03300216794013977, 0.06295855343341827, -0.10010455548763275, 0.007508574984967709, -0.06212785467505455, -0.10758529603481293, 0.014876498840749264, -0.11965788900852203, -0.0002283805952174589, -0.10969529300928116, -0.1379294991493225, -0.0071915821172297, 0.05261197313666344, -0.023271607235074043, -0.07799466699361801, -0.05656461790204048, -0.08003277331590652, 0.029941730201244354, -0.01649371162056923, 0.04931098595261574, -0.06223886460065842, 0.08887401223182678, 0.0474187433719635, 0.0786278024315834, -0.03450461104512215, 0.04901968315243721, -0.08649168163537979, 0.043723590672016144, -0.2052707076072693, 0.05972237139940262, -0.058257415890693665, 0.066720150411129, -0.11373113840818405, -0.11091870814561844, 0.023934224620461464, -0.03338725119829178, 0.07939267158508301, 0.10428828746080399, -0.14210061728954315, -0.08317490667104721, 0.19722555577754974, -0.08388625085353851, -0.13603731989860535, 0.11822353303432465, -0.04774384945631027, 0.023914366960525513, 0.06245804950594902, 0.22421255707740784, 0.07384224981069565, -0.056651849299669266, -0.02626190148293972, -0.01773103140294552, 0.05105559527873993, -0.07211291790008545, 0.08372409641742706, -0.0034855830017477274, 0.027296477928757668, 0.03050689958035946, -0.03478715568780899, 0.03305815905332565, -0.08226604759693146, -0.08269429206848145, -0.0552789568901062, -0.08013436943292618, 0.06636259704828262, 0.04476308450102806, 0.08076454699039459, -0.11116137355566025, -0.09508826583623886, 0.03346242383122444, 0.08442533016204834, -0.08160518854856491, 0.04427807778120041, -0.09438317269086838, 0.12457689642906189, -0.07490513473749161, -0.004501476418226957, -0.18299716711044312, -0.002926086075603962, 0.050260331481695175, -0.030341053381562233, 0.004212009720504284, -0.02511434070765972, 0.06390994042158127, 0.05089034512639046, -0.045860640704631805, -0.04464670270681381, -0.03862137347459793, -0.0073234583251178265, -0.1176871806383133, -0.18063300848007202, -0.046826720237731934, -0.03420059010386467, 0.11011023819446564, -0.15884621441364288, 0.057446498423814774, 0.06037832051515579, 0.1089540496468544, 0.027162769809365273, -0.031170153990387917, -0.009540140628814697, 0.04513327404856682, -0.04029618948698044, -0.07336514443159103, 0.07257451117038727, 0.03279393911361694, -0.09915802627801895, -0.024446843191981316, -0.11312498152256012, 0.17786890268325806, 0.1279948204755783, -0.022059347480535507, -0.0518190823495388, -0.008839407935738564, -0.059505634009838104, -0.024196699261665344, -0.010522348806262016, 0.015498658642172813, 0.1704971343278885, 0.00443031033501029, 0.1728103756904602, -0.10287037491798401, -0.05360252782702446, 0.04714669659733772, -0.029268475249409676, -0.012326378375291824, 0.10431425273418427, 0.01162251178175211, -0.10040758550167084, 0.1517464816570282, 0.1416548192501068, -0.05495679751038551, 0.12223485112190247, -0.05711975693702698, -0.04896477609872818, -0.03903970867395401, -0.0010107593843713403, 0.017619730904698372, 0.0949004516005516, -0.11745531111955643, -0.019590117037296295, 0.039691485464572906, 0.02760290540754795, 0.010583777911961079, -0.18035614490509033, 0.002068496774882078, 0.04255614057183266, -0.0577472522854805, 0.005149621982127428, -0.004730795975774527, -0.0052475999109447, 0.10132960975170135, 0.024889452382922173, -0.07262793928384781, 0.050802480429410934, 0.010588769800961018, -0.07018314301967621, 0.19748838245868683, -0.09537739306688309, -0.19580714404582977, -0.12369107455015182, -0.05933047831058502, -0.08906487375497818, 0.005777008831501007, 0.07029293477535248, -0.08024737238883972, -0.024308767169713974, -0.08881144970655441, -0.02706608735024929, -0.013862849213182926, 0.03982379660010338, 0.07027624547481537, -0.022989053279161453, 0.10219364613294601, -0.12064753472805023, -0.030045047402381897, -0.030844224616885185, 0.0021685869432985783, 0.05443640425801277, 0.007260463200509548, 0.10514117032289505, 0.11515062302350998, -0.029252836480736732, 0.05099339410662651, -0.03219757601618767, 0.23458045721054077, -0.05094539746642113, -0.02754887565970421, 0.12972086668014526, -0.005566699430346489, 0.08485884964466095, 0.09015154838562012, 0.04523105174303055, -0.09123662859201431, -0.008441068232059479, 0.004572180565446615, -0.03901638463139534, -0.21114090085029602, -0.03490867465734482, -0.042593251913785934, 0.020444147288799286, 0.12356333434581757, 0.04155367612838745, 0.05562235042452812, 0.06334526091814041, 0.029814587906003, 0.06015196070075035, -0.02873038686811924, 0.10066144913434982, 0.129075288772583, 0.05018611624836922, 0.13519997894763947, -0.03983941674232483, -0.03396950662136078, 0.03950963914394379, -0.00009943571058101952, 0.20626559853553772, -0.01341833733022213, 0.190015509724617, 0.04551733657717705, 0.18698570132255554, 0.011998365633189678, 0.06770443171262741, -0.020937258377671242, -0.0041463072411715984, -0.016936300322413445, -0.04249386116862297, -0.05993194878101349, 0.010172491893172264, -0.046779412776231766, 0.07449685782194138, -0.1229928657412529, 0.018185555934906006, 0.06067291647195816, 0.29110443592071533, 0.02225201576948166, -0.37252581119537354, -0.11070704460144043, -0.011908737011253834, -0.025519277900457382, -0.04423600062727928, 0.010994509793817997, 0.09587650746107101, -0.09470661729574203, 0.061872486025094986, -0.08669182658195496, 0.0905299112200737, -0.07252071052789688, 0.03704323247075081, 0.04851524531841278, 0.09357307106256485, 0.005450603552162647, 0.05804775655269623, -0.27323558926582336, 0.25258272886276245, 0.018488362431526184, 0.052274398505687714, -0.06330030411481857, 0.013847973197698593, 0.021581653505563736, 0.05988399684429169, 0.08734140545129776, -0.00242057116702199, -0.03574739396572113, -0.170226588845253, -0.10684733092784882, 0.013181532733142376, 0.07300505042076111, -0.043209258466959, 0.08819505572319031, -0.009992691688239574, 0.0025730605702847242, 0.04417675733566284, -0.005149500444531441, -0.03280147910118103, -0.09067709743976593, 0.02149878814816475, 0.06143028661608696, -0.021822771057486534, -0.08020330965518951, -0.11618226766586304, -0.08372586220502853, 0.17271628975868225, -0.011341128498315811, -0.07161374390125275, -0.12234587222337723, 0.06291403621435165, 0.06748344749212265, -0.09436832368373871, 0.045931633561849594, -0.020597733557224274, 0.12422186136245728, 0.009217459708452225, -0.06901761889457703, 0.10053239017724991, -0.04681336134672165, -0.16271468997001648, -0.037926286458969116, 0.13365566730499268, 0.027268672361969948, 0.05532421916723251, -0.008006083779036999, 0.028334422037005424, -0.025608304888010025, -0.07614213973283768, 0.03917950391769409, 0.007323337718844414, 0.09472580999135971, -0.020243126899003983, -0.018931115046143532, 0.029467495158314705, -0.07463947683572769, -0.004165452439337969, 0.19488921761512756, 0.25613391399383545, -0.10813448578119278, 0.043561406433582306, 0.033728744834661484, -0.05141662433743477, -0.1578151136636734, 0.014114146120846272, 0.07223153859376907, 0.004969296511262655, 0.001109614851884544, -0.17397440969944, 0.05568954348564148, 0.0886269137263298, -0.0185370035469532, 0.07366597652435303, -0.2923673987388611, -0.11986135691404343, 0.10729062557220459, 0.12743747234344482, 0.1041778177022934, -0.14241072535514832, -0.050853945314884186, -0.012729215435683727, -0.13892170786857605, 0.11515681445598602, -0.07430049777030945, 0.113226979970932, -0.04666992276906967, 0.05292979255318642, 0.007788409013301134, -0.05154074728488922, 0.1275826096534729, 0.02271832711994648, 0.07553572952747345, -0.05197565257549286, -0.013126295059919357, 0.09809998422861099, -0.0778113305568695, 0.060851484537124634, -0.09939137101173401, 0.045361850410699844, -0.1248331144452095, -0.014556867070496082, -0.07716278731822968, 0.029143769294023514, -0.028787555173039436, -0.04176540672779083, -0.05256284028291702, 0.01147465780377388, 0.07403411716222763, -0.0028210324235260487, 0.17671364545822144, 0.04185609146952629, 0.13671880960464478, 0.18551644682884216, 0.07253473252058029, -0.11256355047225952, -0.10412265360355377, -0.008216392248868942, -0.019994845613837242, 0.05785110220313072, -0.16473303735256195, 0.04240985959768295, 0.140421062707901, 0.008228769525885582, 0.1269446760416031, 0.06872107833623886, -0.047829966992139816, 0.0033522218000143766, 0.045486725866794586, -0.1793903410434723, -0.10016090422868729, -0.009445465169847012, -0.012818184681236744, -0.13622497022151947, 0.07125774770975113, 0.110176682472229, -0.06657290458679199, -0.02074158936738968, 0.0009991518454626203, 0.0010803861077874899, -0.022899242118000984, 0.17322438955307007, 0.06538136303424835, 0.06537455320358276, -0.10131470859050751, 0.09524796903133392, 0.047929782420396805, -0.07653660327196121, 0.03941129520535469, 0.05052856728434563, -0.11436935514211655, -0.022952113300561905, 0.035996194928884506, 0.16025100648403168, -0.034298527985811234, -0.046940721571445465, -0.17082402110099792, -0.10452842712402344, 0.08786427974700928, 0.13057833909988403, 0.10420487821102142, 0.01995844952762127, -0.04781729355454445, -0.005857453215867281, -0.10732704401016235, 0.09478437900543213, 0.05822419375181198, 0.07136266678571701, -0.15628288686275482, 0.11073862761259079, 0.001596992602571845, 0.05998965725302696, -0.013029854744672775, 0.0014995939563959837, -0.09175609797239304, -0.0005578352720476687, -0.08042430132627487, 0.012492680922150612, -0.04159858450293541, -0.0036286425311118364, -0.020429277792572975, -0.05159774795174599, -0.05193471163511276, 0.03137846663594246, -0.10849754512310028, -0.037209369242191315, 0.030224846675992012, 0.04291409254074097, -0.10771522670984268, -0.030150866135954857, 0.011944940313696861, -0.08316744118928909, 0.0927014946937561, 0.04819602519273758, -0.005198898259550333, 0.01361685898154974, -0.024452082812786102, 0.00996569823473692, 0.06474380940198898, 0.0011315169977024198, 0.054949041455984116, -0.1251956820487976, -0.004955892451107502, 0.0020057805813848972, 0.003167948219925165, 0.021521935239434242, 0.11819131672382355, -0.13000352680683136, -0.01095547154545784, -0.01931416429579258, -0.030984848737716675, -0.07567927986383438, 0.05678528547286987, 0.10669084638357162, 0.036463551223278046, 0.18688152730464935, -0.06743103265762329, 0.017906490713357925, -0.20741963386535645, 0.00013413703709375113, 0.005770268850028515, -0.14702105522155762, -0.07122698426246643, -0.036772824823856354, 0.05338381603360176, -0.06538230180740356, 0.11566061526536942, 0.001493913820013404, -0.009878304786980152, 0.0378260537981987, -0.031193168833851814, -0.020588120445609093, 0.008561814203858376, 0.17474886775016785, 0.0126771479845047, -0.03905124589800835, 0.11461129784584045, 0.02905210293829441, 0.10191743075847626, 0.11596836149692535, 0.16053743660449982, 0.13562360405921936, 0.02690793201327324, 0.09993529319763184, 0.025183329358696938, -0.01807461678981781, -0.19147835671901703, 0.07456520944833755, -0.03571565821766853, 0.128890722990036, 0.007677143905311823, 0.1920078843832016, 0.09843330085277557, -0.15698692202568054, 0.051954105496406555, -0.022671973332762718, -0.08706733584403992, -0.09513752907514572, -0.09581948071718216, -0.0837426707148552, -0.1526874601840973, -0.010047889314591885, -0.1159338429570198, 0.006987566594034433, 0.07303724437952042, -0.0017146181780844927, -0.019828321412205696, 0.14302708208560944, 0.010929194279015064, 0.0013329503126442432, 0.08076636493206024, -0.006675939541310072, -0.05830436199903488, -0.08042674511671066, -0.07730001956224442, 0.003484914544969797, 0.018351571634411812, 0.05460543558001518, -0.025369582697749138, 0.00580788915976882, 0.03771538287401199, -0.027528436854481697, -0.10909802466630936, 0.010541521944105625, 0.021785886958241463, 0.050911594182252884, 0.023273441940546036, 0.009273657575249672, -0.014442682266235352, -0.013307257555425167, 0.17752113938331604, -0.06385718286037445, -0.015753794461488724, -0.10788676887750626, 0.18873083591461182, 0.04108770191669464, -0.03788184002041817, 0.03930654376745224, -0.06607408076524734, -0.010708626359701157, 0.1941891610622406, 0.1851033717393875, -0.020034559071063995, 0.00002634760494402144, -0.0045935907401144505, -0.015512681566178799, 0.001854269066825509, 0.08318820595741272, 0.12001276016235352, -0.012533774599432945, -0.06875097006559372, -0.017599457874894142, -0.0655461773276329, -0.006694902200251818, -0.04739747196435928, 0.06480608880519867, 0.020269542932510376, 0.006782840937376022, -0.051228221505880356, 0.023403694853186607, -0.03428596630692482, -0.0766749382019043, 0.03951941058039665, -0.21706217527389526, -0.15715551376342773, -0.007559482473880053, 0.04624676704406738, -0.001891341176815331, 0.059971537441015244, -0.0012983903288841248, 0.011913720518350601, 0.09471691399812698, -0.02078905515372753, -0.08147251605987549, -0.08469844609498978, 0.08651077002286911, -0.15671049058437347, 0.20816466212272644, -0.04045833647251129, 0.04214433580636978, 0.13087542355060577, 0.0450337752699852, -0.10505655407905579, 0.03753887861967087, 0.03393945470452309, -0.00948834978044033, 0.004219671245664358, 0.09632031619548798, -0.015130841173231602, 0.07932081073522568, 0.036358702927827835, -0.0976419672369957, -0.03288884833455086, -0.06169722229242325, -0.02750725857913494, -0.043688997626304626, -0.048681341111660004, -0.050978995859622955, 0.11931037157773972, 0.18234147131443024, -0.051180802285671234, -0.025530055165290833, -0.0633423700928688, 0.01658000238239765, 0.07789728045463562, -0.014490681700408459, -0.048653412610292435, -0.24489381909370422, 0.0111764557659626, 0.058221664279699326, -0.018459200859069824, -0.2366642951965332, -0.11184651404619217, -0.0018209918634966016, -0.06137614697217941, -0.08097735792398453, 0.0797455832362175, 0.045839566737413406, 0.05556558072566986, -0.06334666907787323, 0.004335817415267229, -0.08912677317857742, 0.16122785210609436, -0.15241484344005585, -0.07673730701208115 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-large-cased-finetuned-cola This model is a fine-tuned version of [bert-large-cased](https://huggingface.co/bert-large-cased) on the GLUE COLA dataset. It achieves the following results on the evaluation set: - Loss: 0.8385 - Matthews Correlation: 0.5957 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.5533 | 1.0 | 2138 | 0.7943 | 0.4439 | | 0.5004 | 2.0 | 4276 | 0.7272 | 0.5678 | | 0.2865 | 3.0 | 6414 | 0.8385 | 0.5957 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["glue"], "metrics": ["matthews_correlation"], "model-index": [{"name": "bert-large-cased-finetuned-cola", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE COLA", "type": "glue", "args": "cola"}, "metrics": [{"type": "matthews_correlation", "value": 0.5957317644481708, "name": "Matthews Correlation"}]}]}]}
text-classification
gchhablani/bert-large-cased-finetuned-cola
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "en", "dataset:glue", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #en #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
bert-large-cased-finetuned-cola =============================== This model is a fine-tuned version of bert-large-cased on the GLUE COLA dataset. It achieves the following results on the evaluation set: * Loss: 0.8385 * Matthews Correlation: 0.5957 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 4 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #en #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 67, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #en #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.09624065458774567, 0.09955063462257385, -0.0026663674507290125, 0.12233678996562958, 0.1606411635875702, 0.03215743228793144, 0.12267734110355377, 0.13195538520812988, -0.08486742526292801, 0.01870565116405487, 0.12272996455430984, 0.16391664743423462, 0.018688445910811424, 0.11550883203744888, -0.046508997678756714, -0.26966848969459534, -0.009110793471336365, 0.05400576815009117, -0.058758340775966644, 0.1358145773410797, 0.0948256254196167, -0.1238134354352951, 0.0893140360713005, 0.01236787810921669, -0.1854150891304016, 0.004355923272669315, 0.004461904987692833, -0.05912149325013161, 0.14575855433940887, 0.018412422388792038, 0.1167803704738617, 0.00040736416121944785, 0.08347548544406891, -0.19406671822071075, 0.009881878271698952, 0.05340686812996864, 0.004321607761085033, 0.09506513923406601, 0.04936794936656952, 0.008129656314849854, 0.11189798265695572, -0.07901789993047714, 0.055528104305267334, 0.02463972568511963, -0.11238809674978256, -0.2399875670671463, -0.07282808423042297, 0.037027355283498764, 0.0740681141614914, 0.1023164615035057, -0.005008150357753038, 0.12684179842472076, -0.06927302479743958, 0.09163445979356766, 0.2270132303237915, -0.29200083017349243, -0.062483884394168854, 0.05119242146611214, 0.010518276132643223, 0.054868414998054504, -0.09969478100538254, -0.028113335371017456, 0.04499640688300133, 0.0561663918197155, 0.1373552829027176, -0.031408995389938354, -0.09769929945468903, 0.008257018402218819, -0.14285437762737274, -0.03350107744336128, 0.1733582466840744, 0.03919161856174469, -0.02977021597325802, -0.054241716861724854, -0.06082460656762123, -0.1339312344789505, -0.03440126031637192, -0.016044603660702705, 0.04987446963787079, -0.0194453876465559, -0.04659074917435646, -0.012714596465229988, -0.11053911596536636, -0.06590517610311508, -0.07186423242092133, 0.12182352691888809, 0.03858835622668266, 0.01698452979326248, -0.030455898493528366, 0.11377494782209396, -0.007787707727402449, -0.12330934405326843, 0.021297359839081764, 0.01989939995110035, 0.025820735841989517, -0.0319552905857563, -0.05589330196380615, -0.061863210052251816, 0.010856718756258488, 0.12357568740844727, -0.05277571827173233, 0.0446060448884964, 0.0453709252178669, 0.04966704174876213, -0.0893254429101944, 0.18350131809711456, -0.032479770481586456, -0.023526914417743683, 0.009950077161192894, 0.044431187212467194, 0.024190394207835197, -0.011971864849328995, -0.124683678150177, 0.008309499360620975, 0.09545876830816269, 0.0057476288639009, -0.06085534766316414, 0.07939501851797104, -0.050725314766168594, -0.03213759511709213, -0.006347034126520157, -0.09149795025587082, 0.02164720930159092, 0.00242617167532444, -0.0726463720202446, -0.024878911674022675, 0.03534867987036705, 0.01344688143581152, -0.024611959233880043, 0.10715840011835098, -0.0847240537405014, 0.024549664929509163, -0.09281458705663681, -0.1103331670165062, 0.017382444813847542, -0.09248446673154831, 0.02165031060576439, -0.09901522099971771, -0.1825384795665741, -0.00994229968637228, 0.06636174023151398, -0.023078858852386475, -0.0583648756146431, -0.043868452310562134, -0.06445350497961044, 0.010027475655078888, -0.017435073852539062, 0.12104400992393494, -0.06494352966547012, 0.09195806086063385, 0.0227196104824543, 0.05753939226269722, -0.04597872123122215, 0.059839703142642975, -0.1019076481461525, 0.012001135386526585, -0.1567085236310959, 0.033876512199640274, -0.05160209909081459, 0.060489166527986526, -0.08846108615398407, -0.10703590512275696, 0.013825437054038048, -0.002660207450389862, 0.06004123389720917, 0.09042271971702576, -0.17369325459003448, -0.07627386599779129, 0.15298448503017426, -0.0758451521396637, -0.12591737508773804, 0.11832267045974731, -0.059722695499658585, 0.055540405213832855, 0.06174943968653679, 0.18307331204414368, 0.07782846689224243, -0.07455380260944366, 0.010968664661049843, 0.02781253680586815, 0.05276698246598244, -0.07263585925102234, 0.07463609427213669, 0.0035101727116853, 0.013453012332320213, 0.03918846324086189, -0.03828362748026848, 0.061971426010131836, -0.0842839851975441, -0.09340687096118927, -0.03942016512155533, -0.08057273179292679, 0.03859388828277588, 0.07305388897657394, 0.07229185104370117, -0.09514500945806503, -0.08145271986722946, 0.05218813195824623, 0.08160724490880966, -0.06189076602458954, 0.029457777738571167, -0.052497003227472305, 0.07722547650337219, -0.03631065785884857, -0.018118295818567276, -0.17934367060661316, -0.031864095479249954, 0.011508386582136154, -0.011600063182413578, 0.013546913862228394, 0.032294511795043945, 0.06229916214942932, 0.057009853422641754, -0.05466185137629509, -0.017526565119624138, -0.032779060304164886, 0.0032311424147337675, -0.1271989494562149, -0.19206227362155914, -0.034151606261730194, -0.021822931244969368, 0.13331127166748047, -0.19654668867588043, 0.04860207065939903, -0.007861302234232426, 0.07463932037353516, 0.010202854871749878, -0.005920075811445713, -0.04214789345860481, 0.06752505153417587, -0.044097017496824265, -0.05136480927467346, 0.08785995095968246, 0.01922564208507538, -0.08842770755290985, -0.04372691363096237, -0.099749356508255, 0.16481786966323853, 0.12885503470897675, -0.11912618577480316, -0.06534887850284576, -0.01675848662853241, -0.06664930284023285, -0.03183472156524658, -0.04601408913731575, 0.021434122696518898, 0.18623022735118866, -0.006579231470823288, 0.15221582353115082, -0.06983739137649536, -0.04125898331403732, 0.018391873687505722, -0.03462676703929901, 0.014082837849855423, 0.12281899154186249, 0.12765833735466003, -0.06600239872932434, 0.15375342965126038, 0.15929262340068817, -0.08576879650354385, 0.1411527842283249, -0.04225218668580055, -0.06259749084711075, -0.018176736310124397, -0.03686186298727989, -0.015194122679531574, 0.10063527524471283, -0.15178850293159485, 0.0021366802975535393, 0.03645699843764305, 0.025968225672841072, 0.026575788855552673, -0.221955344080925, -0.03538588434457779, 0.035139359533786774, -0.0405682697892189, 0.0001947561395354569, -0.006039177533239126, 0.0024743475951254368, 0.10182642191648483, 0.0037002163007855415, -0.08593712747097015, 0.045064978301525116, -0.0015891270013526082, -0.08569430559873581, 0.21371793746948242, -0.08121493458747864, -0.1776936799287796, -0.1258380264043808, -0.07995165139436722, -0.05537958815693855, 0.0003615658497437835, 0.0710969865322113, -0.08796952664852142, -0.032766811549663544, -0.07330936193466187, 0.02786816842854023, -0.0009106051875278354, 0.026721270754933357, 0.005710284691303968, 0.0019002188928425312, 0.07142426073551178, -0.11346574872732162, -0.01991988904774189, -0.06108444929122925, -0.04330765828490257, 0.04123419150710106, 0.03264953941106796, 0.10631082952022552, 0.14909066259860992, -0.013310462236404419, 0.019161483272910118, -0.028462272137403488, 0.2352350503206253, -0.05493570491671562, -0.020346906036138535, 0.15474484860897064, -0.012476389296352863, 0.05023787543177605, 0.11233063787221909, 0.07413924485445023, -0.07835007458925247, 0.004125483799725771, 0.036005981266498566, -0.040030352771282196, -0.22443436086177826, -0.05713476613163948, -0.05518857762217522, 0.014405518770217896, 0.09477297961711884, 0.02860194630920887, 0.0261562280356884, 0.0678858608007431, 0.04673674330115318, 0.06752058118581772, -0.036432940512895584, 0.0606808066368103, 0.14244720339775085, 0.029392071068286896, 0.12940694391727448, -0.04088454693555832, -0.05961482226848602, 0.04527316614985466, -0.008333292789757252, 0.21456016600131989, 0.00348348799161613, 0.12263227254152298, 0.06304474174976349, 0.16508528590202332, -0.006165117956697941, 0.07757799327373505, -0.014170979149639606, -0.032569751143455505, -0.02133876085281372, -0.04030373692512512, -0.03840538486838341, 0.02547278068959713, -0.06923425197601318, 0.06040016561746597, -0.11979921162128448, 0.0005183702451176941, 0.05842360109090805, 0.2486162930727005, 0.03629889711737633, -0.32320284843444824, -0.09831482172012329, 0.005307691637426615, -0.028546947985887527, -0.018132254481315613, 0.03376784175634384, 0.09025244414806366, -0.09573614597320557, 0.03267209231853485, -0.07967761158943176, 0.09565100073814392, -0.051767997443675995, 0.05253976210951805, 0.09077772498130798, 0.09408006072044373, 0.013921128585934639, 0.09444761276245117, -0.28585490584373474, 0.26606398820877075, -0.0011629497166723013, 0.05611091107130051, -0.07803214341402054, 0.008231346495449543, 0.039912108331918716, 0.06727396696805954, 0.0790090337395668, -0.013563624583184719, -0.02555840276181698, -0.17742381989955902, -0.06818605959415436, 0.03299332782626152, 0.06263915449380875, -0.04049598053097725, 0.08151790499687195, -0.03625113144516945, 0.011121245101094246, 0.07626894861459732, 0.00978172942996025, -0.05481652542948723, -0.10135668516159058, -0.0010370832169428468, 0.031113948673009872, -0.06206654757261276, -0.0615290105342865, -0.12261777371168137, -0.1238018348813057, 0.163645938038826, -0.01921737752854824, -0.04382816702127457, -0.109429270029068, 0.09022320061922073, 0.057270318269729614, -0.0875500738620758, 0.037784308195114136, -0.00005559223427553661, 0.08095261454582214, 0.028389230370521545, -0.08056071400642395, 0.10426200926303864, -0.07694460451602936, -0.15371933579444885, -0.06213109567761421, 0.1063443124294281, 0.03172723203897476, 0.06139996647834778, -0.012001262046396732, 0.00816765334457159, -0.04605482518672943, -0.09145131707191467, 0.023197924718260765, 0.0026915683411061764, 0.08504897356033325, 0.009198497980833054, -0.07460201531648636, 0.02006295882165432, -0.058423224836587906, -0.03237378969788551, 0.2043369561433792, 0.21232619881629944, -0.10105175524950027, 0.02852863445878029, 0.03403891995549202, -0.07227198779582977, -0.20590628683567047, 0.027527185156941414, 0.049684204161167145, -0.002057756530120969, 0.045846737921237946, -0.17838771641254425, 0.1388525664806366, 0.09775126725435257, -0.01811867393553257, 0.09962065517902374, -0.3128257095813751, -0.1232745572924614, 0.14224252104759216, 0.14292606711387634, 0.11552654206752777, -0.13620565831661224, -0.020663224160671234, -0.017764810472726822, -0.13955825567245483, 0.12315498292446136, -0.09684952348470688, 0.11743222922086716, -0.03810708969831467, 0.07901652157306671, 0.002770873950794339, -0.05992662534117699, 0.11887063831090927, 0.01920950599014759, 0.08838226646184921, -0.06334490329027176, -0.02773796208202839, 0.033923566341400146, -0.043468065559864044, 0.03498063236474991, -0.09636924415826797, 0.031842198222875595, -0.10165029019117355, -0.028140675276517868, -0.07138922810554504, 0.046002086251974106, -0.03914405405521393, -0.0767509937286377, -0.037220798432826996, 0.027214476838707924, 0.051858607679605484, -0.009808214381337166, 0.13950254023075104, 0.028160542249679565, 0.1445336639881134, 0.11847491562366486, 0.06573983281850815, -0.07295768707990646, -0.08793066442012787, -0.028920577839016914, -0.014843004755675793, 0.05528461933135986, -0.139312744140625, 0.024727309122681618, 0.15001529455184937, 0.014143741689622402, 0.1481979936361313, 0.08705729246139526, -0.022433988749980927, -0.003058880101889372, 0.05504049360752106, -0.17051774263381958, -0.09581425040960312, -0.013994061388075352, -0.06577996164560318, -0.11612004786729813, 0.053103070706129074, 0.10053814947605133, -0.07065427303314209, -0.005371682811528444, -0.0030203366186469793, 0.01780072972178459, -0.050984229892492294, 0.1861531287431717, 0.06146577000617981, 0.04397404193878174, -0.09816218912601471, 0.0785122737288475, 0.04132809489965439, -0.07178964465856552, 0.004344574641436338, 0.06782659143209457, -0.09094034880399704, -0.054981011897325516, 0.06647961586713791, 0.18478482961654663, -0.04806734248995781, -0.050944164395332336, -0.14548426866531372, -0.1260456144809723, 0.08039271086454391, 0.11902277171611786, 0.12157487869262695, 0.006039421074092388, -0.07286527752876282, 0.0045364717952907085, -0.10375693440437317, 0.10418010503053665, 0.044098515063524246, 0.059477634727954865, -0.14696674048900604, 0.1357119083404541, 0.020099440589547157, 0.05167369544506073, -0.02312709018588066, 0.02084973081946373, -0.0973726212978363, 0.00845542922616005, -0.10871388763189316, -0.018524587154388428, -0.027348605915904045, 0.009420540183782578, -0.00527747068554163, -0.05180655047297478, -0.05978180468082428, 0.017835335806012154, -0.1113293468952179, -0.024005228653550148, 0.02869814820587635, 0.06593389809131622, -0.10840646922588348, -0.04022184759378433, 0.022955162450671196, -0.061636220663785934, 0.07924428582191467, 0.04559391736984253, 0.010550598613917828, 0.05296533927321434, -0.13034309446811676, 0.016353514045476913, 0.0734846219420433, 0.029182346537709236, 0.05939546972513199, -0.09865947812795639, -0.009885996580123901, -0.0020042674150317907, 0.03302076458930969, 0.018970055505633354, 0.08167483657598495, -0.14078399538993835, 0.004283816087990999, -0.022713083773851395, -0.07938187569379807, -0.06684045493602753, 0.024325530976057053, 0.0822998434305191, 0.03159202262759209, 0.20085522532463074, -0.07683441042900085, 0.047416165471076965, -0.21537481248378754, 0.007678445894271135, -0.004576575011014938, -0.11387678980827332, -0.10230255126953125, -0.06447644531726837, 0.05446239933371544, -0.05925348401069641, 0.15299662947654724, 0.0506068617105484, 0.017128881067037582, 0.02823813073337078, -0.012935548089444637, 0.019529689103364944, 0.009635900147259235, 0.19693420827388763, 0.023763258010149002, -0.033243611454963684, 0.06285648047924042, 0.043830569833517075, 0.10904955863952637, 0.11531230807304382, 0.2001006007194519, 0.14637964963912964, 0.0006141560152173042, 0.0904940739274025, 0.039791226387023926, -0.05479295179247856, -0.17086173593997955, 0.047213006764650345, -0.02637372352182865, 0.11851751804351807, -0.01962713897228241, 0.21435046195983887, 0.06194686517119408, -0.16590678691864014, 0.04570665583014488, -0.05666576325893402, -0.0821349248290062, -0.11906159669160843, -0.05707874149084091, -0.07657578587532043, -0.13889333605766296, -0.005688000936061144, -0.11587034910917282, -0.00017965854203794152, 0.12667395174503326, 0.006447251420468092, -0.027460677549242973, 0.14922739565372467, 0.004153023939579725, 0.017027847468852997, 0.055918045341968536, 0.007438011933118105, -0.03976532444357872, -0.12369070947170258, -0.05787182226777077, -0.02172340266406536, -0.005461963824927807, 0.02974032424390316, -0.05976889654994011, -0.04034041613340378, 0.029146261513233185, -0.02425864152610302, -0.09368163347244263, 0.003910803701728582, 0.013014798983931541, 0.05448725447058678, 0.05280468985438347, 0.008613139390945435, 0.014966918155550957, -0.003064875490963459, 0.20565827190876007, -0.07258933782577515, -0.05836240574717522, -0.09818195551633835, 0.2406144142150879, 0.034694332629442215, -0.01956876739859581, 0.03710349649190903, -0.06559594720602036, -0.005295134615153074, 0.2500089406967163, 0.2241545021533966, -0.07982916384935379, -0.006788645870983601, 0.019206037744879723, -0.00931378174573183, -0.015787554904818535, 0.10749366879463196, 0.1416216641664505, 0.05335712432861328, -0.09138569980859756, -0.03638821467757225, -0.056785713881254196, -0.010250318795442581, -0.039071641862392426, 0.07402456551790237, 0.050159476697444916, 0.007921763695776463, -0.039261817932128906, 0.04983757436275482, -0.06186498701572418, -0.09983278810977936, 0.04922524839639664, -0.2108970433473587, -0.16345985233783722, -0.012340680696070194, 0.10004943609237671, -0.002300487132743001, 0.0640321671962738, -0.03366585075855255, 0.0027021048590540886, 0.08857354521751404, -0.018983889371156693, -0.09548420459032059, -0.07277628779411316, 0.09841836988925934, -0.11422606557607651, 0.22972609102725983, -0.04592173546552658, 0.055711861699819565, 0.12398207932710648, 0.06129786744713783, -0.07080429047346115, 0.061077386140823364, 0.04071512073278427, -0.04364478960633278, 0.02041429653763771, 0.06825253367424011, -0.036071088165044785, 0.06332977861166, 0.045397888869047165, -0.13932836055755615, 0.010228789411485195, -0.04427666217088699, -0.07257308810949326, -0.04142169654369354, -0.030986180528998375, -0.06465091556310654, 0.13108615577220917, 0.20706406235694885, -0.028249824419617653, -0.014149054884910583, -0.07465700060129166, 0.01549980603158474, 0.05391182750463486, 0.01404651440680027, -0.05135972052812576, -0.20725055038928986, 0.017412958666682243, 0.04823734983801842, -0.023168247193098068, -0.2485376000404358, -0.10177735984325409, 0.010998763144016266, -0.07118414342403412, -0.0975857675075531, 0.07177483290433884, 0.08087921142578125, 0.04843691363930702, -0.05996540188789368, -0.04442998021841049, -0.0790470689535141, 0.14195428788661957, -0.14670926332473755, -0.09249257296323776 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-large-cased-finetuned-mrpc This model is a fine-tuned version of [bert-large-cased](https://huggingface.co/bert-large-cased) on the GLUE MRPC dataset. It achieves the following results on the evaluation set: - Loss: 0.6274 - Accuracy: 0.6838 - F1: 0.8122 - Combined Score: 0.7480 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Combined Score | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:--------------:| | 0.6441 | 1.0 | 917 | 0.6370 | 0.6838 | 0.8122 | 0.7480 | | 0.6451 | 2.0 | 1834 | 0.6553 | 0.6838 | 0.8122 | 0.7480 | | 0.6428 | 3.0 | 2751 | 0.6332 | 0.6838 | 0.8122 | 0.7480 | | 0.6476 | 4.0 | 3668 | 0.6248 | 0.6838 | 0.8122 | 0.7480 | | 0.6499 | 5.0 | 4585 | 0.6274 | 0.6838 | 0.8122 | 0.7480 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["glue"], "metrics": ["accuracy", "f1"], "model-index": [{"name": "bert-large-cased-finetuned-mrpc", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE MRPC", "type": "glue", "args": "mrpc"}, "metrics": [{"type": "accuracy", "value": 0.6838235294117647, "name": "Accuracy"}, {"type": "f1", "value": 0.8122270742358079, "name": "F1"}]}]}]}
text-classification
gchhablani/bert-large-cased-finetuned-mrpc
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "en", "dataset:glue", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #en #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
bert-large-cased-finetuned-mrpc =============================== This model is a fine-tuned version of bert-large-cased on the GLUE MRPC dataset. It achieves the following results on the evaluation set: * Loss: 0.6274 * Accuracy: 0.6838 * F1: 0.8122 * Combined Score: 0.7480 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 4 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 5.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #en #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 67, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #en #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.0976102352142334, 0.09936647117137909, -0.002672189613804221, 0.12195486575365067, 0.16110917925834656, 0.03247614577412605, 0.1213773712515831, 0.13151462376117706, -0.08341013640165329, 0.019382942467927933, 0.12232600897550583, 0.16258969902992249, 0.01876535452902317, 0.11641276627779007, -0.04651031643152237, -0.26918312907218933, -0.010064840316772461, 0.05488419905304909, -0.05938786640763283, 0.13579422235488892, 0.09399029612541199, -0.12511876225471497, 0.0891716256737709, 0.012169326655566692, -0.1872284710407257, 0.004259427078068256, 0.005672164727002382, -0.05877799168229103, 0.14609383046627045, 0.018703630194067955, 0.11675883084535599, 0.00003582282806746662, 0.08291850984096527, -0.19420795142650604, 0.010098351165652275, 0.054083336144685745, 0.0043545495718717575, 0.09575458616018295, 0.048873189836740494, 0.009347987361252308, 0.11236174404621124, -0.0793292373418808, 0.05474558472633362, 0.024683725088834763, -0.11247405409812927, -0.24312597513198853, -0.07160093635320663, 0.03693250194191933, 0.07330705225467682, 0.10204507410526276, -0.005552090238779783, 0.12655803561210632, -0.06983516365289688, 0.0918550044298172, 0.22564134001731873, -0.29315733909606934, -0.061992332339286804, 0.05087697133421898, 0.009979608468711376, 0.055669113993644714, -0.10138673335313797, -0.028127353638410568, 0.04564734548330307, 0.05465557798743248, 0.1381852775812149, -0.030967101454734802, -0.09843979775905609, 0.00821664184331894, -0.1420469582080841, -0.033919233828783035, 0.1738097220659256, 0.03982063755393028, -0.028622567653656006, -0.054070934653282166, -0.061537183821201324, -0.13361498713493347, -0.03464250639081001, -0.01621822640299797, 0.05061670020222664, -0.019486509263515472, -0.045077063143253326, -0.011115099303424358, -0.1107238233089447, -0.06624976545572281, -0.07208099216222763, 0.12095384299755096, 0.03857453539967537, 0.016972262412309647, -0.03215523809194565, 0.11358600854873657, -0.010029386729001999, -0.12412283569574356, 0.021471858024597168, 0.019995957612991333, 0.025281477719545364, -0.03170512989163399, -0.05552344024181366, -0.06175937131047249, 0.011591022834181786, 0.12309639155864716, -0.05241638422012329, 0.04452582448720932, 0.04484933614730835, 0.048885200172662735, -0.08958490192890167, 0.18331816792488098, -0.03193177282810211, -0.0249992273747921, 0.009942438453435898, 0.043930359184741974, 0.02330765314400196, -0.012014612555503845, -0.12405231595039368, 0.00792987272143364, 0.09529060870409012, 0.004790466744452715, -0.06052672117948532, 0.07933007925748825, -0.050575483590364456, -0.031458478420972824, -0.00708750868216157, -0.09122146666049957, 0.022523390129208565, 0.0027903050649911165, -0.07296080142259598, -0.026073317974805832, 0.03473372757434845, 0.012356848455965519, -0.025106975808739662, 0.10841948539018631, -0.0857846587896347, 0.024185610935091972, -0.09363146871328354, -0.1108751893043518, 0.01706712134182453, -0.09391313046216965, 0.022464720532298088, -0.09843502193689346, -0.18116706609725952, -0.010625375434756279, 0.06536457687616348, -0.0228746198117733, -0.05853870138525963, -0.0439993217587471, -0.0652107298374176, 0.010579889640212059, -0.0174049511551857, 0.12262888997793198, -0.06501127034425735, 0.0928327888250351, 0.023376809433102608, 0.05793778598308563, -0.04520407319068909, 0.06017840653657913, -0.10293940454721451, 0.010917382314801216, -0.15611115097999573, 0.033262211829423904, -0.05175517499446869, 0.06136655434966087, -0.08839258551597595, -0.10605067759752274, 0.012814945541322231, -0.0031625761184841394, 0.05991736799478531, 0.09122243523597717, -0.17328718304634094, -0.07755474001169205, 0.15574303269386292, -0.0762687548995018, -0.1258247345685959, 0.11795607209205627, -0.05977392569184303, 0.055517252534627914, 0.06242389976978302, 0.1843721717596054, 0.07839464396238327, -0.07442736625671387, 0.01027555949985981, 0.02752048894762993, 0.05271969735622406, -0.07282748073339462, 0.07366485893726349, 0.004275749437510967, 0.012977390550076962, 0.039640795439481735, -0.03786756843328476, 0.06277041882276535, -0.0843513086438179, -0.09271667897701263, -0.04069691523909569, -0.08022435754537582, 0.03900573030114174, 0.07502502202987671, 0.07274094223976135, -0.09476830065250397, -0.08111522346735, 0.051697470247745514, 0.08157351613044739, -0.06079532206058502, 0.028261110186576843, -0.05223787948489189, 0.0780346617102623, -0.03667766600847244, -0.01701318845152855, -0.18064351379871368, -0.031141415238380432, 0.011841713450849056, -0.013485615141689777, 0.01334770955145359, 0.0326857827603817, 0.06271495670080185, 0.056571342051029205, -0.05454390496015549, -0.01736464537680149, -0.03354068100452423, 0.0028900050092488527, -0.1282014697790146, -0.193827822804451, -0.033352844417095184, -0.021876277402043343, 0.13311365246772766, -0.19806548953056335, 0.04835042729973793, -0.006943442393094301, 0.07547688484191895, 0.011464327573776245, -0.005951065104454756, -0.04243546351790428, 0.0677863359451294, -0.04325307533144951, -0.05036669969558716, 0.0874180793762207, 0.018765095621347427, -0.08722604811191559, -0.04445832967758179, -0.09914910793304443, 0.16732032597064972, 0.12972886860370636, -0.11873441189527512, -0.06437160819768906, -0.016277091577649117, -0.06576891988515854, -0.032300036400556564, -0.046140048652887344, 0.0227329321205616, 0.18421471118927002, -0.00741422176361084, 0.15248200297355652, -0.06974963843822479, -0.040814030915498734, 0.019723327830433846, -0.0331619456410408, 0.015886330977082253, 0.12413706630468369, 0.1282740831375122, -0.06559982895851135, 0.15418152511119843, 0.15941673517227173, -0.08487163484096527, 0.1422242969274521, -0.04333624616265297, -0.06347992271184921, -0.017862480133771896, -0.03773048520088196, -0.016069943085312843, 0.10228079557418823, -0.15387645363807678, 0.001617925358004868, 0.03709304332733154, 0.025371834635734558, 0.026067743077874184, -0.22200049459934235, -0.03455454111099243, 0.03493591025471687, -0.039809174835681915, -0.0024919509887695312, -0.005420095287263393, 0.0028411531820893288, 0.1023520976305008, 0.002888386370614171, -0.08515071123838425, 0.0433807298541069, -0.0013903152430430055, -0.08476655185222626, 0.2139444500207901, -0.0802430659532547, -0.1774560809135437, -0.12425795942544937, -0.07774058729410172, -0.05414064973592758, -0.0005127398180775344, 0.0720231905579567, -0.08906834572553635, -0.03252917900681496, -0.07175689935684204, 0.02817642316222191, 0.0004518343193922192, 0.027263330295681953, 0.004086782224476337, 0.002676589647307992, 0.07090707868337631, -0.1134081557393074, -0.019071435555815697, -0.06199377402663231, -0.04224175587296486, 0.04067744314670563, 0.033918965607881546, 0.10669896006584167, 0.14927758276462555, -0.013438706286251545, 0.0188836008310318, -0.028484776616096497, 0.2354217767715454, -0.056646913290023804, -0.02005748078227043, 0.15443594753742218, -0.011800444684922695, 0.04932769760489464, 0.11312438547611237, 0.07431694865226746, -0.07768645137548447, 0.0031734551303088665, 0.03593902289867401, -0.03924098610877991, -0.2247641384601593, -0.05702979862689972, -0.05578586831688881, 0.01433180458843708, 0.09506276994943619, 0.02831125073134899, 0.026894286274909973, 0.06739514321088791, 0.04697586968541145, 0.06784062087535858, -0.036935217678546906, 0.06005891412496567, 0.14293049275875092, 0.029087936505675316, 0.12923696637153625, -0.04220735654234886, -0.05971696600317955, 0.0444951131939888, -0.009248945862054825, 0.21686850488185883, 0.003934380132704973, 0.12343882769346237, 0.06182524934411049, 0.1663391888141632, -0.005540202371776104, 0.07694202661514282, -0.013981996104121208, -0.033022165298461914, -0.020576324313879013, -0.039612263441085815, -0.03762431442737579, 0.02585711143910885, -0.06962882727384567, 0.06022243946790695, -0.12031690776348114, -0.0010294135427102447, 0.05853549763560295, 0.2475212663412094, 0.036378465592861176, -0.3220648169517517, -0.09916582703590393, 0.004513485357165337, -0.0287352092564106, -0.019020449370145798, 0.033321842551231384, 0.09136277437210083, -0.09498634934425354, 0.03184245526790619, -0.07975354790687561, 0.0965612456202507, -0.050986163318157196, 0.05284735560417175, 0.0900365337729454, 0.094030000269413, 0.01364613976329565, 0.09443002194166183, -0.28663894534111023, 0.2680751085281372, -0.0006494453991763294, 0.05628465488553047, -0.07870705425739288, 0.007773632649332285, 0.03950111195445061, 0.068779356777668, 0.07901504635810852, -0.013699501752853394, -0.026843415573239326, -0.1764722615480423, -0.0673225000500679, 0.033501844853162766, 0.06271138787269592, -0.04167846962809563, 0.08069954812526703, -0.03599863499403, 0.010489752516150475, 0.0761018842458725, 0.007347069680690765, -0.054988957941532135, -0.10074210911989212, -0.0012152357958257198, 0.03176466003060341, -0.061484742909669876, -0.06114106625318527, -0.12227476388216019, -0.12258432805538177, 0.16115999221801758, -0.02296372316777706, -0.04294164106249809, -0.10914649069309235, 0.09212889522314072, 0.056863829493522644, -0.08795413374900818, 0.038500405848026276, -0.000327902875142172, 0.08047305047512054, 0.028799112886190414, -0.0805036798119545, 0.1060391217470169, -0.07636438310146332, -0.15390388667583466, -0.06225806474685669, 0.1062450110912323, 0.031391508877277374, 0.061169736087322235, -0.012482204474508762, 0.008419888094067574, -0.04702845960855484, -0.09128129482269287, 0.02387714385986328, 0.001966517185792327, 0.08349449932575226, 0.00866552535444498, -0.0738549456000328, 0.01695571094751358, -0.05823530629277229, -0.03262261301279068, 0.2026766836643219, 0.21211427450180054, -0.10201698541641235, 0.028370339423418045, 0.034632645547389984, -0.07315336912870407, -0.20590299367904663, 0.028807779774069786, 0.04927895590662956, -0.0014008330181241035, 0.044853344559669495, -0.1794818788766861, 0.13935346901416779, 0.0967443436384201, -0.018301600590348244, 0.10205823183059692, -0.3130754828453064, -0.12346244603395462, 0.14208893477916718, 0.1417371779680252, 0.11829537898302078, -0.13624699413776398, -0.02140231430530548, -0.017252175137400627, -0.1401212513446808, 0.12243872880935669, -0.09825469553470612, 0.11674070358276367, -0.03857799619436264, 0.07763806730508804, 0.0030884670559316874, -0.06019184738397598, 0.11799732595682144, 0.01921112649142742, 0.08952765166759491, -0.0628134161233902, -0.0275727566331625, 0.03402796760201454, -0.04281206801533699, 0.0345970094203949, -0.09624014049768448, 0.03211625665426254, -0.10117616504430771, -0.027533680200576782, -0.0721392035484314, 0.04651001840829849, -0.03904503583908081, -0.07578983902931213, -0.03744593262672424, 0.026871010661125183, 0.050932008773088455, -0.009576674550771713, 0.14048266410827637, 0.028415191918611526, 0.1454911231994629, 0.11825840920209885, 0.06443970650434494, -0.07301773130893707, -0.08845897018909454, -0.028305696323513985, -0.014447078108787537, 0.05691270902752876, -0.14205476641654968, 0.024711037054657936, 0.15047235786914825, 0.01565638557076454, 0.14734461903572083, 0.08675093203783035, -0.022020047530531883, -0.0034190728329122066, 0.05546081066131592, -0.17116686701774597, -0.0990695133805275, -0.014003586024045944, -0.06685681641101837, -0.11525572836399078, 0.053843580186367035, 0.10016743093729019, -0.07175523787736893, -0.005365441087633371, -0.0033988894429057837, 0.017403457313776016, -0.05040215700864792, 0.18709012866020203, 0.061101291328668594, 0.044216785579919815, -0.0989004448056221, 0.07840321958065033, 0.04248986020684242, -0.07499589771032333, 0.0035009197890758514, 0.06673578172922134, -0.09152981638908386, -0.05507204309105873, 0.06716036051511765, 0.18546466529369354, -0.04658083617687225, -0.05140521004796028, -0.14549706876277924, -0.12558671832084656, 0.08094824850559235, 0.12121469527482986, 0.1214657574892044, 0.005996201653033495, -0.0724189281463623, 0.005100664217025042, -0.10431451350450516, 0.10551860928535461, 0.04417169466614723, 0.060016483068466187, -0.14719988405704498, 0.13589566946029663, 0.02082657441496849, 0.05062829330563545, -0.023369889706373215, 0.020944442600011826, -0.09796511381864548, 0.008230148814618587, -0.10670613497495651, -0.018555371090769768, -0.026394350454211235, 0.009305293671786785, -0.005771094933152199, -0.052633434534072876, -0.059571847319602966, 0.0179905965924263, -0.11197777092456818, -0.02365105412900448, 0.029261741787195206, 0.0664503425359726, -0.10900621861219406, -0.03973565995693207, 0.02363383211195469, -0.061003491282463074, 0.07896764576435089, 0.04568549245595932, 0.011074677109718323, 0.05367050692439079, -0.13051925599575043, 0.017069417983293533, 0.07291415333747864, 0.029571840539574623, 0.059489961713552475, -0.09817635267972946, -0.008948633447289467, -0.0025318285916000605, 0.03328393027186394, 0.018514983355998993, 0.08086178451776505, -0.14140820503234863, 0.0036213670391589403, -0.022418564185500145, -0.0794878825545311, -0.0665006935596466, 0.023724470287561417, 0.08154180645942688, 0.03051106259226799, 0.20031949877738953, -0.07612800598144531, 0.04680617153644562, -0.21655263006687164, 0.007846522144973278, -0.004819144029170275, -0.11533224582672119, -0.1021881103515625, -0.0648970901966095, 0.05523134395480156, -0.059748657047748566, 0.15259039402008057, 0.05145854502916336, 0.018327929079532623, 0.028435297310352325, -0.013000483624637127, 0.017145715653896332, 0.010023065842688084, 0.19696204364299774, 0.024565931409597397, -0.03308612480759621, 0.0632064938545227, 0.04469598829746246, 0.10829045623540878, 0.11776725947856903, 0.20024071633815765, 0.1472475677728653, 0.0013614862691611052, 0.09062184393405914, 0.03958705812692642, -0.055089958012104034, -0.17143400013446808, 0.04928014799952507, -0.027800429612398148, 0.12028032541275024, -0.02070426754653454, 0.21418233215808868, 0.06252310425043106, -0.16507434844970703, 0.04673760011792183, -0.05689859017729759, -0.08169996738433838, -0.11859188228845596, -0.05599171668291092, -0.07707660645246506, -0.14051198959350586, -0.005772426258772612, -0.11554284393787384, 0.0002381563390372321, 0.12560053169727325, 0.005618050694465637, -0.028257902711629868, 0.1500902622938156, 0.00589335709810257, 0.01685897633433342, 0.05763181671500206, 0.007298589684069157, -0.0412118062376976, -0.12455623596906662, -0.05731077864766121, -0.021210240200161934, -0.0067823175340890884, 0.028453994542360306, -0.060132455080747604, -0.04206157848238945, 0.029383787885308266, -0.02374044805765152, -0.09295868128538132, 0.004053476732224226, 0.013348747044801712, 0.05407030135393143, 0.052432116121053696, 0.008429237641394138, 0.01524826418608427, -0.0033252707216888666, 0.2079176902770996, -0.07260357588529587, -0.05715000629425049, -0.09797889739274979, 0.23888063430786133, 0.03585147485136986, -0.019686749204993248, 0.03734346106648445, -0.06568225473165512, -0.004361408296972513, 0.24954579770565033, 0.22186478972434998, -0.07830173522233963, -0.006712665781378746, 0.019575312733650208, -0.009557166136801243, -0.016031969338655472, 0.10772424191236496, 0.14170996844768524, 0.054387882351875305, -0.09126259386539459, -0.03596600145101547, -0.05642533674836159, -0.010790226049721241, -0.03824600949883461, 0.07427041232585907, 0.049858734011650085, 0.006499612703919411, -0.038984883576631546, 0.05010038614273071, -0.06054513528943062, -0.0999545156955719, 0.05072363093495369, -0.21182700991630554, -0.16378560662269592, -0.012944746762514114, 0.10200632363557816, -0.0016564095858484507, 0.06280625611543655, -0.03329086676239967, 0.002569399308413267, 0.08935030549764633, -0.01901974156498909, -0.09497696161270142, -0.07604139298200607, 0.09724408388137817, -0.11604832112789154, 0.22954528033733368, -0.045974187552928925, 0.055315155535936356, 0.12348725646734238, 0.06152692064642906, -0.07137792557477951, 0.06000271812081337, 0.04052073881030083, -0.044816117733716965, 0.020156823098659515, 0.06924524903297424, -0.035139527171850204, 0.06370943784713745, 0.045128483325242996, -0.1386181116104126, 0.010564154013991356, -0.04600505903363228, -0.0727553740143776, -0.041528042405843735, -0.030076157301664352, -0.0639835074543953, 0.1305534392595291, 0.20732425153255463, -0.02794814482331276, -0.013891981914639473, -0.07531335204839706, 0.01536940410733223, 0.05437286198139191, 0.015924062579870224, -0.05145624652504921, -0.20643475651741028, 0.018344108015298843, 0.04973233491182327, -0.02281355671584606, -0.24700039625167847, -0.10108326375484467, 0.010105917230248451, -0.07113083451986313, -0.0970691591501236, 0.07135510444641113, 0.08287014812231064, 0.0481468103826046, -0.06011649966239929, -0.048794958740472794, -0.07777239382266998, 0.1425076127052307, -0.14664505422115326, -0.0928139016032219 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-large-cased-finetuned-rte This model is a fine-tuned version of [bert-large-cased](https://huggingface.co/bert-large-cased) on the GLUE RTE dataset. It achieves the following results on the evaluation set: - Loss: 1.5187 - Accuracy: 0.6643 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6969 | 1.0 | 623 | 0.7039 | 0.5343 | | 0.5903 | 2.0 | 1246 | 0.6461 | 0.7184 | | 0.4557 | 3.0 | 1869 | 1.5187 | 0.6643 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["glue"], "metrics": ["accuracy"], "model-index": [{"name": "bert-large-cased-finetuned-rte", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE RTE", "type": "glue", "args": "rte"}, "metrics": [{"type": "accuracy", "value": 0.6642599277978339, "name": "Accuracy"}]}]}]}
text-classification
gchhablani/bert-large-cased-finetuned-rte
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "en", "dataset:glue", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #en #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
bert-large-cased-finetuned-rte ============================== This model is a fine-tuned version of bert-large-cased on the GLUE RTE dataset. It achieves the following results on the evaluation set: * Loss: 1.5187 * Accuracy: 0.6643 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 4 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #en #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 67, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #en #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.09624065458774567, 0.09955063462257385, -0.0026663674507290125, 0.12233678996562958, 0.1606411635875702, 0.03215743228793144, 0.12267734110355377, 0.13195538520812988, -0.08486742526292801, 0.01870565116405487, 0.12272996455430984, 0.16391664743423462, 0.018688445910811424, 0.11550883203744888, -0.046508997678756714, -0.26966848969459534, -0.009110793471336365, 0.05400576815009117, -0.058758340775966644, 0.1358145773410797, 0.0948256254196167, -0.1238134354352951, 0.0893140360713005, 0.01236787810921669, -0.1854150891304016, 0.004355923272669315, 0.004461904987692833, -0.05912149325013161, 0.14575855433940887, 0.018412422388792038, 0.1167803704738617, 0.00040736416121944785, 0.08347548544406891, -0.19406671822071075, 0.009881878271698952, 0.05340686812996864, 0.004321607761085033, 0.09506513923406601, 0.04936794936656952, 0.008129656314849854, 0.11189798265695572, -0.07901789993047714, 0.055528104305267334, 0.02463972568511963, -0.11238809674978256, -0.2399875670671463, -0.07282808423042297, 0.037027355283498764, 0.0740681141614914, 0.1023164615035057, -0.005008150357753038, 0.12684179842472076, -0.06927302479743958, 0.09163445979356766, 0.2270132303237915, -0.29200083017349243, -0.062483884394168854, 0.05119242146611214, 0.010518276132643223, 0.054868414998054504, -0.09969478100538254, -0.028113335371017456, 0.04499640688300133, 0.0561663918197155, 0.1373552829027176, -0.031408995389938354, -0.09769929945468903, 0.008257018402218819, -0.14285437762737274, -0.03350107744336128, 0.1733582466840744, 0.03919161856174469, -0.02977021597325802, -0.054241716861724854, -0.06082460656762123, -0.1339312344789505, -0.03440126031637192, -0.016044603660702705, 0.04987446963787079, -0.0194453876465559, -0.04659074917435646, -0.012714596465229988, -0.11053911596536636, -0.06590517610311508, -0.07186423242092133, 0.12182352691888809, 0.03858835622668266, 0.01698452979326248, -0.030455898493528366, 0.11377494782209396, -0.007787707727402449, -0.12330934405326843, 0.021297359839081764, 0.01989939995110035, 0.025820735841989517, -0.0319552905857563, -0.05589330196380615, -0.061863210052251816, 0.010856718756258488, 0.12357568740844727, -0.05277571827173233, 0.0446060448884964, 0.0453709252178669, 0.04966704174876213, -0.0893254429101944, 0.18350131809711456, -0.032479770481586456, -0.023526914417743683, 0.009950077161192894, 0.044431187212467194, 0.024190394207835197, -0.011971864849328995, -0.124683678150177, 0.008309499360620975, 0.09545876830816269, 0.0057476288639009, -0.06085534766316414, 0.07939501851797104, -0.050725314766168594, -0.03213759511709213, -0.006347034126520157, -0.09149795025587082, 0.02164720930159092, 0.00242617167532444, -0.0726463720202446, -0.024878911674022675, 0.03534867987036705, 0.01344688143581152, -0.024611959233880043, 0.10715840011835098, -0.0847240537405014, 0.024549664929509163, -0.09281458705663681, -0.1103331670165062, 0.017382444813847542, -0.09248446673154831, 0.02165031060576439, -0.09901522099971771, -0.1825384795665741, -0.00994229968637228, 0.06636174023151398, -0.023078858852386475, -0.0583648756146431, -0.043868452310562134, -0.06445350497961044, 0.010027475655078888, -0.017435073852539062, 0.12104400992393494, -0.06494352966547012, 0.09195806086063385, 0.0227196104824543, 0.05753939226269722, -0.04597872123122215, 0.059839703142642975, -0.1019076481461525, 0.012001135386526585, -0.1567085236310959, 0.033876512199640274, -0.05160209909081459, 0.060489166527986526, -0.08846108615398407, -0.10703590512275696, 0.013825437054038048, -0.002660207450389862, 0.06004123389720917, 0.09042271971702576, -0.17369325459003448, -0.07627386599779129, 0.15298448503017426, -0.0758451521396637, -0.12591737508773804, 0.11832267045974731, -0.059722695499658585, 0.055540405213832855, 0.06174943968653679, 0.18307331204414368, 0.07782846689224243, -0.07455380260944366, 0.010968664661049843, 0.02781253680586815, 0.05276698246598244, -0.07263585925102234, 0.07463609427213669, 0.0035101727116853, 0.013453012332320213, 0.03918846324086189, -0.03828362748026848, 0.061971426010131836, -0.0842839851975441, -0.09340687096118927, -0.03942016512155533, -0.08057273179292679, 0.03859388828277588, 0.07305388897657394, 0.07229185104370117, -0.09514500945806503, -0.08145271986722946, 0.05218813195824623, 0.08160724490880966, -0.06189076602458954, 0.029457777738571167, -0.052497003227472305, 0.07722547650337219, -0.03631065785884857, -0.018118295818567276, -0.17934367060661316, -0.031864095479249954, 0.011508386582136154, -0.011600063182413578, 0.013546913862228394, 0.032294511795043945, 0.06229916214942932, 0.057009853422641754, -0.05466185137629509, -0.017526565119624138, -0.032779060304164886, 0.0032311424147337675, -0.1271989494562149, -0.19206227362155914, -0.034151606261730194, -0.021822931244969368, 0.13331127166748047, -0.19654668867588043, 0.04860207065939903, -0.007861302234232426, 0.07463932037353516, 0.010202854871749878, -0.005920075811445713, -0.04214789345860481, 0.06752505153417587, -0.044097017496824265, -0.05136480927467346, 0.08785995095968246, 0.01922564208507538, -0.08842770755290985, -0.04372691363096237, -0.099749356508255, 0.16481786966323853, 0.12885503470897675, -0.11912618577480316, -0.06534887850284576, -0.01675848662853241, -0.06664930284023285, -0.03183472156524658, -0.04601408913731575, 0.021434122696518898, 0.18623022735118866, -0.006579231470823288, 0.15221582353115082, -0.06983739137649536, -0.04125898331403732, 0.018391873687505722, -0.03462676703929901, 0.014082837849855423, 0.12281899154186249, 0.12765833735466003, -0.06600239872932434, 0.15375342965126038, 0.15929262340068817, -0.08576879650354385, 0.1411527842283249, -0.04225218668580055, -0.06259749084711075, -0.018176736310124397, -0.03686186298727989, -0.015194122679531574, 0.10063527524471283, -0.15178850293159485, 0.0021366802975535393, 0.03645699843764305, 0.025968225672841072, 0.026575788855552673, -0.221955344080925, -0.03538588434457779, 0.035139359533786774, -0.0405682697892189, 0.0001947561395354569, -0.006039177533239126, 0.0024743475951254368, 0.10182642191648483, 0.0037002163007855415, -0.08593712747097015, 0.045064978301525116, -0.0015891270013526082, -0.08569430559873581, 0.21371793746948242, -0.08121493458747864, -0.1776936799287796, -0.1258380264043808, -0.07995165139436722, -0.05537958815693855, 0.0003615658497437835, 0.0710969865322113, -0.08796952664852142, -0.032766811549663544, -0.07330936193466187, 0.02786816842854023, -0.0009106051875278354, 0.026721270754933357, 0.005710284691303968, 0.0019002188928425312, 0.07142426073551178, -0.11346574872732162, -0.01991988904774189, -0.06108444929122925, -0.04330765828490257, 0.04123419150710106, 0.03264953941106796, 0.10631082952022552, 0.14909066259860992, -0.013310462236404419, 0.019161483272910118, -0.028462272137403488, 0.2352350503206253, -0.05493570491671562, -0.020346906036138535, 0.15474484860897064, -0.012476389296352863, 0.05023787543177605, 0.11233063787221909, 0.07413924485445023, -0.07835007458925247, 0.004125483799725771, 0.036005981266498566, -0.040030352771282196, -0.22443436086177826, -0.05713476613163948, -0.05518857762217522, 0.014405518770217896, 0.09477297961711884, 0.02860194630920887, 0.0261562280356884, 0.0678858608007431, 0.04673674330115318, 0.06752058118581772, -0.036432940512895584, 0.0606808066368103, 0.14244720339775085, 0.029392071068286896, 0.12940694391727448, -0.04088454693555832, -0.05961482226848602, 0.04527316614985466, -0.008333292789757252, 0.21456016600131989, 0.00348348799161613, 0.12263227254152298, 0.06304474174976349, 0.16508528590202332, -0.006165117956697941, 0.07757799327373505, -0.014170979149639606, -0.032569751143455505, -0.02133876085281372, -0.04030373692512512, -0.03840538486838341, 0.02547278068959713, -0.06923425197601318, 0.06040016561746597, -0.11979921162128448, 0.0005183702451176941, 0.05842360109090805, 0.2486162930727005, 0.03629889711737633, -0.32320284843444824, -0.09831482172012329, 0.005307691637426615, -0.028546947985887527, -0.018132254481315613, 0.03376784175634384, 0.09025244414806366, -0.09573614597320557, 0.03267209231853485, -0.07967761158943176, 0.09565100073814392, -0.051767997443675995, 0.05253976210951805, 0.09077772498130798, 0.09408006072044373, 0.013921128585934639, 0.09444761276245117, -0.28585490584373474, 0.26606398820877075, -0.0011629497166723013, 0.05611091107130051, -0.07803214341402054, 0.008231346495449543, 0.039912108331918716, 0.06727396696805954, 0.0790090337395668, -0.013563624583184719, -0.02555840276181698, -0.17742381989955902, -0.06818605959415436, 0.03299332782626152, 0.06263915449380875, -0.04049598053097725, 0.08151790499687195, -0.03625113144516945, 0.011121245101094246, 0.07626894861459732, 0.00978172942996025, -0.05481652542948723, -0.10135668516159058, -0.0010370832169428468, 0.031113948673009872, -0.06206654757261276, -0.0615290105342865, -0.12261777371168137, -0.1238018348813057, 0.163645938038826, -0.01921737752854824, -0.04382816702127457, -0.109429270029068, 0.09022320061922073, 0.057270318269729614, -0.0875500738620758, 0.037784308195114136, -0.00005559223427553661, 0.08095261454582214, 0.028389230370521545, -0.08056071400642395, 0.10426200926303864, -0.07694460451602936, -0.15371933579444885, -0.06213109567761421, 0.1063443124294281, 0.03172723203897476, 0.06139996647834778, -0.012001262046396732, 0.00816765334457159, -0.04605482518672943, -0.09145131707191467, 0.023197924718260765, 0.0026915683411061764, 0.08504897356033325, 0.009198497980833054, -0.07460201531648636, 0.02006295882165432, -0.058423224836587906, -0.03237378969788551, 0.2043369561433792, 0.21232619881629944, -0.10105175524950027, 0.02852863445878029, 0.03403891995549202, -0.07227198779582977, -0.20590628683567047, 0.027527185156941414, 0.049684204161167145, -0.002057756530120969, 0.045846737921237946, -0.17838771641254425, 0.1388525664806366, 0.09775126725435257, -0.01811867393553257, 0.09962065517902374, -0.3128257095813751, -0.1232745572924614, 0.14224252104759216, 0.14292606711387634, 0.11552654206752777, -0.13620565831661224, -0.020663224160671234, -0.017764810472726822, -0.13955825567245483, 0.12315498292446136, -0.09684952348470688, 0.11743222922086716, -0.03810708969831467, 0.07901652157306671, 0.002770873950794339, -0.05992662534117699, 0.11887063831090927, 0.01920950599014759, 0.08838226646184921, -0.06334490329027176, -0.02773796208202839, 0.033923566341400146, -0.043468065559864044, 0.03498063236474991, -0.09636924415826797, 0.031842198222875595, -0.10165029019117355, -0.028140675276517868, -0.07138922810554504, 0.046002086251974106, -0.03914405405521393, -0.0767509937286377, -0.037220798432826996, 0.027214476838707924, 0.051858607679605484, -0.009808214381337166, 0.13950254023075104, 0.028160542249679565, 0.1445336639881134, 0.11847491562366486, 0.06573983281850815, -0.07295768707990646, -0.08793066442012787, -0.028920577839016914, -0.014843004755675793, 0.05528461933135986, -0.139312744140625, 0.024727309122681618, 0.15001529455184937, 0.014143741689622402, 0.1481979936361313, 0.08705729246139526, -0.022433988749980927, -0.003058880101889372, 0.05504049360752106, -0.17051774263381958, -0.09581425040960312, -0.013994061388075352, -0.06577996164560318, -0.11612004786729813, 0.053103070706129074, 0.10053814947605133, -0.07065427303314209, -0.005371682811528444, -0.0030203366186469793, 0.01780072972178459, -0.050984229892492294, 0.1861531287431717, 0.06146577000617981, 0.04397404193878174, -0.09816218912601471, 0.0785122737288475, 0.04132809489965439, -0.07178964465856552, 0.004344574641436338, 0.06782659143209457, -0.09094034880399704, -0.054981011897325516, 0.06647961586713791, 0.18478482961654663, -0.04806734248995781, -0.050944164395332336, -0.14548426866531372, -0.1260456144809723, 0.08039271086454391, 0.11902277171611786, 0.12157487869262695, 0.006039421074092388, -0.07286527752876282, 0.0045364717952907085, -0.10375693440437317, 0.10418010503053665, 0.044098515063524246, 0.059477634727954865, -0.14696674048900604, 0.1357119083404541, 0.020099440589547157, 0.05167369544506073, -0.02312709018588066, 0.02084973081946373, -0.0973726212978363, 0.00845542922616005, -0.10871388763189316, -0.018524587154388428, -0.027348605915904045, 0.009420540183782578, -0.00527747068554163, -0.05180655047297478, -0.05978180468082428, 0.017835335806012154, -0.1113293468952179, -0.024005228653550148, 0.02869814820587635, 0.06593389809131622, -0.10840646922588348, -0.04022184759378433, 0.022955162450671196, -0.061636220663785934, 0.07924428582191467, 0.04559391736984253, 0.010550598613917828, 0.05296533927321434, -0.13034309446811676, 0.016353514045476913, 0.0734846219420433, 0.029182346537709236, 0.05939546972513199, -0.09865947812795639, -0.009885996580123901, -0.0020042674150317907, 0.03302076458930969, 0.018970055505633354, 0.08167483657598495, -0.14078399538993835, 0.004283816087990999, -0.022713083773851395, -0.07938187569379807, -0.06684045493602753, 0.024325530976057053, 0.0822998434305191, 0.03159202262759209, 0.20085522532463074, -0.07683441042900085, 0.047416165471076965, -0.21537481248378754, 0.007678445894271135, -0.004576575011014938, -0.11387678980827332, -0.10230255126953125, -0.06447644531726837, 0.05446239933371544, -0.05925348401069641, 0.15299662947654724, 0.0506068617105484, 0.017128881067037582, 0.02823813073337078, -0.012935548089444637, 0.019529689103364944, 0.009635900147259235, 0.19693420827388763, 0.023763258010149002, -0.033243611454963684, 0.06285648047924042, 0.043830569833517075, 0.10904955863952637, 0.11531230807304382, 0.2001006007194519, 0.14637964963912964, 0.0006141560152173042, 0.0904940739274025, 0.039791226387023926, -0.05479295179247856, -0.17086173593997955, 0.047213006764650345, -0.02637372352182865, 0.11851751804351807, -0.01962713897228241, 0.21435046195983887, 0.06194686517119408, -0.16590678691864014, 0.04570665583014488, -0.05666576325893402, -0.0821349248290062, -0.11906159669160843, -0.05707874149084091, -0.07657578587532043, -0.13889333605766296, -0.005688000936061144, -0.11587034910917282, -0.00017965854203794152, 0.12667395174503326, 0.006447251420468092, -0.027460677549242973, 0.14922739565372467, 0.004153023939579725, 0.017027847468852997, 0.055918045341968536, 0.007438011933118105, -0.03976532444357872, -0.12369070947170258, -0.05787182226777077, -0.02172340266406536, -0.005461963824927807, 0.02974032424390316, -0.05976889654994011, -0.04034041613340378, 0.029146261513233185, -0.02425864152610302, -0.09368163347244263, 0.003910803701728582, 0.013014798983931541, 0.05448725447058678, 0.05280468985438347, 0.008613139390945435, 0.014966918155550957, -0.003064875490963459, 0.20565827190876007, -0.07258933782577515, -0.05836240574717522, -0.09818195551633835, 0.2406144142150879, 0.034694332629442215, -0.01956876739859581, 0.03710349649190903, -0.06559594720602036, -0.005295134615153074, 0.2500089406967163, 0.2241545021533966, -0.07982916384935379, -0.006788645870983601, 0.019206037744879723, -0.00931378174573183, -0.015787554904818535, 0.10749366879463196, 0.1416216641664505, 0.05335712432861328, -0.09138569980859756, -0.03638821467757225, -0.056785713881254196, -0.010250318795442581, -0.039071641862392426, 0.07402456551790237, 0.050159476697444916, 0.007921763695776463, -0.039261817932128906, 0.04983757436275482, -0.06186498701572418, -0.09983278810977936, 0.04922524839639664, -0.2108970433473587, -0.16345985233783722, -0.012340680696070194, 0.10004943609237671, -0.002300487132743001, 0.0640321671962738, -0.03366585075855255, 0.0027021048590540886, 0.08857354521751404, -0.018983889371156693, -0.09548420459032059, -0.07277628779411316, 0.09841836988925934, -0.11422606557607651, 0.22972609102725983, -0.04592173546552658, 0.055711861699819565, 0.12398207932710648, 0.06129786744713783, -0.07080429047346115, 0.061077386140823364, 0.04071512073278427, -0.04364478960633278, 0.02041429653763771, 0.06825253367424011, -0.036071088165044785, 0.06332977861166, 0.045397888869047165, -0.13932836055755615, 0.010228789411485195, -0.04427666217088699, -0.07257308810949326, -0.04142169654369354, -0.030986180528998375, -0.06465091556310654, 0.13108615577220917, 0.20706406235694885, -0.028249824419617653, -0.014149054884910583, -0.07465700060129166, 0.01549980603158474, 0.05391182750463486, 0.01404651440680027, -0.05135972052812576, -0.20725055038928986, 0.017412958666682243, 0.04823734983801842, -0.023168247193098068, -0.2485376000404358, -0.10177735984325409, 0.010998763144016266, -0.07118414342403412, -0.0975857675075531, 0.07177483290433884, 0.08087921142578125, 0.04843691363930702, -0.05996540188789368, -0.04442998021841049, -0.0790470689535141, 0.14195428788661957, -0.14670926332473755, -0.09249257296323776 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-large-cased-finetuned-wnli This model is a fine-tuned version of [bert-large-cased](https://huggingface.co/bert-large-cased) on the GLUE WNLI dataset. It achieves the following results on the evaluation set: - Loss: 0.7087 - Accuracy: 0.3521 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5.0 ### Training results | Training Loss | Epoch | Step | Accuracy | Validation Loss | |:-------------:|:-----:|:----:|:--------:|:---------------:| | 0.7114 | 1.0 | 159 | 0.5634 | 0.6923 | | 0.7141 | 2.0 | 318 | 0.5634 | 0.6895 | | 0.7063 | 3.0 | 477 | 0.5634 | 0.6930 | | 0.712 | 4.0 | 636 | 0.4507 | 0.7077 | | 0.7037 | 5.0 | 795 | 0.3521 | 0.7087 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["glue"], "metrics": ["accuracy"], "model-index": [{"name": "bert-large-cased-finetuned-wnli", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE WNLI", "type": "glue", "args": "wnli"}, "metrics": [{"type": "accuracy", "value": 0.352112676056338, "name": "Accuracy"}]}]}]}
text-classification
gchhablani/bert-large-cased-finetuned-wnli
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "en", "dataset:glue", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #en #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
bert-large-cased-finetuned-wnli =============================== This model is a fine-tuned version of bert-large-cased on the GLUE WNLI dataset. It achieves the following results on the evaluation set: * Loss: 0.7087 * Accuracy: 0.3521 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 4 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 5.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #en #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 67, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #en #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.0976102352142334, 0.09936647117137909, -0.002672189613804221, 0.12195486575365067, 0.16110917925834656, 0.03247614577412605, 0.1213773712515831, 0.13151462376117706, -0.08341013640165329, 0.019382942467927933, 0.12232600897550583, 0.16258969902992249, 0.01876535452902317, 0.11641276627779007, -0.04651031643152237, -0.26918312907218933, -0.010064840316772461, 0.05488419905304909, -0.05938786640763283, 0.13579422235488892, 0.09399029612541199, -0.12511876225471497, 0.0891716256737709, 0.012169326655566692, -0.1872284710407257, 0.004259427078068256, 0.005672164727002382, -0.05877799168229103, 0.14609383046627045, 0.018703630194067955, 0.11675883084535599, 0.00003582282806746662, 0.08291850984096527, -0.19420795142650604, 0.010098351165652275, 0.054083336144685745, 0.0043545495718717575, 0.09575458616018295, 0.048873189836740494, 0.009347987361252308, 0.11236174404621124, -0.0793292373418808, 0.05474558472633362, 0.024683725088834763, -0.11247405409812927, -0.24312597513198853, -0.07160093635320663, 0.03693250194191933, 0.07330705225467682, 0.10204507410526276, -0.005552090238779783, 0.12655803561210632, -0.06983516365289688, 0.0918550044298172, 0.22564134001731873, -0.29315733909606934, -0.061992332339286804, 0.05087697133421898, 0.009979608468711376, 0.055669113993644714, -0.10138673335313797, -0.028127353638410568, 0.04564734548330307, 0.05465557798743248, 0.1381852775812149, -0.030967101454734802, -0.09843979775905609, 0.00821664184331894, -0.1420469582080841, -0.033919233828783035, 0.1738097220659256, 0.03982063755393028, -0.028622567653656006, -0.054070934653282166, -0.061537183821201324, -0.13361498713493347, -0.03464250639081001, -0.01621822640299797, 0.05061670020222664, -0.019486509263515472, -0.045077063143253326, -0.011115099303424358, -0.1107238233089447, -0.06624976545572281, -0.07208099216222763, 0.12095384299755096, 0.03857453539967537, 0.016972262412309647, -0.03215523809194565, 0.11358600854873657, -0.010029386729001999, -0.12412283569574356, 0.021471858024597168, 0.019995957612991333, 0.025281477719545364, -0.03170512989163399, -0.05552344024181366, -0.06175937131047249, 0.011591022834181786, 0.12309639155864716, -0.05241638422012329, 0.04452582448720932, 0.04484933614730835, 0.048885200172662735, -0.08958490192890167, 0.18331816792488098, -0.03193177282810211, -0.0249992273747921, 0.009942438453435898, 0.043930359184741974, 0.02330765314400196, -0.012014612555503845, -0.12405231595039368, 0.00792987272143364, 0.09529060870409012, 0.004790466744452715, -0.06052672117948532, 0.07933007925748825, -0.050575483590364456, -0.031458478420972824, -0.00708750868216157, -0.09122146666049957, 0.022523390129208565, 0.0027903050649911165, -0.07296080142259598, -0.026073317974805832, 0.03473372757434845, 0.012356848455965519, -0.025106975808739662, 0.10841948539018631, -0.0857846587896347, 0.024185610935091972, -0.09363146871328354, -0.1108751893043518, 0.01706712134182453, -0.09391313046216965, 0.022464720532298088, -0.09843502193689346, -0.18116706609725952, -0.010625375434756279, 0.06536457687616348, -0.0228746198117733, -0.05853870138525963, -0.0439993217587471, -0.0652107298374176, 0.010579889640212059, -0.0174049511551857, 0.12262888997793198, -0.06501127034425735, 0.0928327888250351, 0.023376809433102608, 0.05793778598308563, -0.04520407319068909, 0.06017840653657913, -0.10293940454721451, 0.010917382314801216, -0.15611115097999573, 0.033262211829423904, -0.05175517499446869, 0.06136655434966087, -0.08839258551597595, -0.10605067759752274, 0.012814945541322231, -0.0031625761184841394, 0.05991736799478531, 0.09122243523597717, -0.17328718304634094, -0.07755474001169205, 0.15574303269386292, -0.0762687548995018, -0.1258247345685959, 0.11795607209205627, -0.05977392569184303, 0.055517252534627914, 0.06242389976978302, 0.1843721717596054, 0.07839464396238327, -0.07442736625671387, 0.01027555949985981, 0.02752048894762993, 0.05271969735622406, -0.07282748073339462, 0.07366485893726349, 0.004275749437510967, 0.012977390550076962, 0.039640795439481735, -0.03786756843328476, 0.06277041882276535, -0.0843513086438179, -0.09271667897701263, -0.04069691523909569, -0.08022435754537582, 0.03900573030114174, 0.07502502202987671, 0.07274094223976135, -0.09476830065250397, -0.08111522346735, 0.051697470247745514, 0.08157351613044739, -0.06079532206058502, 0.028261110186576843, -0.05223787948489189, 0.0780346617102623, -0.03667766600847244, -0.01701318845152855, -0.18064351379871368, -0.031141415238380432, 0.011841713450849056, -0.013485615141689777, 0.01334770955145359, 0.0326857827603817, 0.06271495670080185, 0.056571342051029205, -0.05454390496015549, -0.01736464537680149, -0.03354068100452423, 0.0028900050092488527, -0.1282014697790146, -0.193827822804451, -0.033352844417095184, -0.021876277402043343, 0.13311365246772766, -0.19806548953056335, 0.04835042729973793, -0.006943442393094301, 0.07547688484191895, 0.011464327573776245, -0.005951065104454756, -0.04243546351790428, 0.0677863359451294, -0.04325307533144951, -0.05036669969558716, 0.0874180793762207, 0.018765095621347427, -0.08722604811191559, -0.04445832967758179, -0.09914910793304443, 0.16732032597064972, 0.12972886860370636, -0.11873441189527512, -0.06437160819768906, -0.016277091577649117, -0.06576891988515854, -0.032300036400556564, -0.046140048652887344, 0.0227329321205616, 0.18421471118927002, -0.00741422176361084, 0.15248200297355652, -0.06974963843822479, -0.040814030915498734, 0.019723327830433846, -0.0331619456410408, 0.015886330977082253, 0.12413706630468369, 0.1282740831375122, -0.06559982895851135, 0.15418152511119843, 0.15941673517227173, -0.08487163484096527, 0.1422242969274521, -0.04333624616265297, -0.06347992271184921, -0.017862480133771896, -0.03773048520088196, -0.016069943085312843, 0.10228079557418823, -0.15387645363807678, 0.001617925358004868, 0.03709304332733154, 0.025371834635734558, 0.026067743077874184, -0.22200049459934235, -0.03455454111099243, 0.03493591025471687, -0.039809174835681915, -0.0024919509887695312, -0.005420095287263393, 0.0028411531820893288, 0.1023520976305008, 0.002888386370614171, -0.08515071123838425, 0.0433807298541069, -0.0013903152430430055, -0.08476655185222626, 0.2139444500207901, -0.0802430659532547, -0.1774560809135437, -0.12425795942544937, -0.07774058729410172, -0.05414064973592758, -0.0005127398180775344, 0.0720231905579567, -0.08906834572553635, -0.03252917900681496, -0.07175689935684204, 0.02817642316222191, 0.0004518343193922192, 0.027263330295681953, 0.004086782224476337, 0.002676589647307992, 0.07090707868337631, -0.1134081557393074, -0.019071435555815697, -0.06199377402663231, -0.04224175587296486, 0.04067744314670563, 0.033918965607881546, 0.10669896006584167, 0.14927758276462555, -0.013438706286251545, 0.0188836008310318, -0.028484776616096497, 0.2354217767715454, -0.056646913290023804, -0.02005748078227043, 0.15443594753742218, -0.011800444684922695, 0.04932769760489464, 0.11312438547611237, 0.07431694865226746, -0.07768645137548447, 0.0031734551303088665, 0.03593902289867401, -0.03924098610877991, -0.2247641384601593, -0.05702979862689972, -0.05578586831688881, 0.01433180458843708, 0.09506276994943619, 0.02831125073134899, 0.026894286274909973, 0.06739514321088791, 0.04697586968541145, 0.06784062087535858, -0.036935217678546906, 0.06005891412496567, 0.14293049275875092, 0.029087936505675316, 0.12923696637153625, -0.04220735654234886, -0.05971696600317955, 0.0444951131939888, -0.009248945862054825, 0.21686850488185883, 0.003934380132704973, 0.12343882769346237, 0.06182524934411049, 0.1663391888141632, -0.005540202371776104, 0.07694202661514282, -0.013981996104121208, -0.033022165298461914, -0.020576324313879013, -0.039612263441085815, -0.03762431442737579, 0.02585711143910885, -0.06962882727384567, 0.06022243946790695, -0.12031690776348114, -0.0010294135427102447, 0.05853549763560295, 0.2475212663412094, 0.036378465592861176, -0.3220648169517517, -0.09916582703590393, 0.004513485357165337, -0.0287352092564106, -0.019020449370145798, 0.033321842551231384, 0.09136277437210083, -0.09498634934425354, 0.03184245526790619, -0.07975354790687561, 0.0965612456202507, -0.050986163318157196, 0.05284735560417175, 0.0900365337729454, 0.094030000269413, 0.01364613976329565, 0.09443002194166183, -0.28663894534111023, 0.2680751085281372, -0.0006494453991763294, 0.05628465488553047, -0.07870705425739288, 0.007773632649332285, 0.03950111195445061, 0.068779356777668, 0.07901504635810852, -0.013699501752853394, -0.026843415573239326, -0.1764722615480423, -0.0673225000500679, 0.033501844853162766, 0.06271138787269592, -0.04167846962809563, 0.08069954812526703, -0.03599863499403, 0.010489752516150475, 0.0761018842458725, 0.007347069680690765, -0.054988957941532135, -0.10074210911989212, -0.0012152357958257198, 0.03176466003060341, -0.061484742909669876, -0.06114106625318527, -0.12227476388216019, -0.12258432805538177, 0.16115999221801758, -0.02296372316777706, -0.04294164106249809, -0.10914649069309235, 0.09212889522314072, 0.056863829493522644, -0.08795413374900818, 0.038500405848026276, -0.000327902875142172, 0.08047305047512054, 0.028799112886190414, -0.0805036798119545, 0.1060391217470169, -0.07636438310146332, -0.15390388667583466, -0.06225806474685669, 0.1062450110912323, 0.031391508877277374, 0.061169736087322235, -0.012482204474508762, 0.008419888094067574, -0.04702845960855484, -0.09128129482269287, 0.02387714385986328, 0.001966517185792327, 0.08349449932575226, 0.00866552535444498, -0.0738549456000328, 0.01695571094751358, -0.05823530629277229, -0.03262261301279068, 0.2026766836643219, 0.21211427450180054, -0.10201698541641235, 0.028370339423418045, 0.034632645547389984, -0.07315336912870407, -0.20590299367904663, 0.028807779774069786, 0.04927895590662956, -0.0014008330181241035, 0.044853344559669495, -0.1794818788766861, 0.13935346901416779, 0.0967443436384201, -0.018301600590348244, 0.10205823183059692, -0.3130754828453064, -0.12346244603395462, 0.14208893477916718, 0.1417371779680252, 0.11829537898302078, -0.13624699413776398, -0.02140231430530548, -0.017252175137400627, -0.1401212513446808, 0.12243872880935669, -0.09825469553470612, 0.11674070358276367, -0.03857799619436264, 0.07763806730508804, 0.0030884670559316874, -0.06019184738397598, 0.11799732595682144, 0.01921112649142742, 0.08952765166759491, -0.0628134161233902, -0.0275727566331625, 0.03402796760201454, -0.04281206801533699, 0.0345970094203949, -0.09624014049768448, 0.03211625665426254, -0.10117616504430771, -0.027533680200576782, -0.0721392035484314, 0.04651001840829849, -0.03904503583908081, -0.07578983902931213, -0.03744593262672424, 0.026871010661125183, 0.050932008773088455, -0.009576674550771713, 0.14048266410827637, 0.028415191918611526, 0.1454911231994629, 0.11825840920209885, 0.06443970650434494, -0.07301773130893707, -0.08845897018909454, -0.028305696323513985, -0.014447078108787537, 0.05691270902752876, -0.14205476641654968, 0.024711037054657936, 0.15047235786914825, 0.01565638557076454, 0.14734461903572083, 0.08675093203783035, -0.022020047530531883, -0.0034190728329122066, 0.05546081066131592, -0.17116686701774597, -0.0990695133805275, -0.014003586024045944, -0.06685681641101837, -0.11525572836399078, 0.053843580186367035, 0.10016743093729019, -0.07175523787736893, -0.005365441087633371, -0.0033988894429057837, 0.017403457313776016, -0.05040215700864792, 0.18709012866020203, 0.061101291328668594, 0.044216785579919815, -0.0989004448056221, 0.07840321958065033, 0.04248986020684242, -0.07499589771032333, 0.0035009197890758514, 0.06673578172922134, -0.09152981638908386, -0.05507204309105873, 0.06716036051511765, 0.18546466529369354, -0.04658083617687225, -0.05140521004796028, -0.14549706876277924, -0.12558671832084656, 0.08094824850559235, 0.12121469527482986, 0.1214657574892044, 0.005996201653033495, -0.0724189281463623, 0.005100664217025042, -0.10431451350450516, 0.10551860928535461, 0.04417169466614723, 0.060016483068466187, -0.14719988405704498, 0.13589566946029663, 0.02082657441496849, 0.05062829330563545, -0.023369889706373215, 0.020944442600011826, -0.09796511381864548, 0.008230148814618587, -0.10670613497495651, -0.018555371090769768, -0.026394350454211235, 0.009305293671786785, -0.005771094933152199, -0.052633434534072876, -0.059571847319602966, 0.0179905965924263, -0.11197777092456818, -0.02365105412900448, 0.029261741787195206, 0.0664503425359726, -0.10900621861219406, -0.03973565995693207, 0.02363383211195469, -0.061003491282463074, 0.07896764576435089, 0.04568549245595932, 0.011074677109718323, 0.05367050692439079, -0.13051925599575043, 0.017069417983293533, 0.07291415333747864, 0.029571840539574623, 0.059489961713552475, -0.09817635267972946, -0.008948633447289467, -0.0025318285916000605, 0.03328393027186394, 0.018514983355998993, 0.08086178451776505, -0.14140820503234863, 0.0036213670391589403, -0.022418564185500145, -0.0794878825545311, -0.0665006935596466, 0.023724470287561417, 0.08154180645942688, 0.03051106259226799, 0.20031949877738953, -0.07612800598144531, 0.04680617153644562, -0.21655263006687164, 0.007846522144973278, -0.004819144029170275, -0.11533224582672119, -0.1021881103515625, -0.0648970901966095, 0.05523134395480156, -0.059748657047748566, 0.15259039402008057, 0.05145854502916336, 0.018327929079532623, 0.028435297310352325, -0.013000483624637127, 0.017145715653896332, 0.010023065842688084, 0.19696204364299774, 0.024565931409597397, -0.03308612480759621, 0.0632064938545227, 0.04469598829746246, 0.10829045623540878, 0.11776725947856903, 0.20024071633815765, 0.1472475677728653, 0.0013614862691611052, 0.09062184393405914, 0.03958705812692642, -0.055089958012104034, -0.17143400013446808, 0.04928014799952507, -0.027800429612398148, 0.12028032541275024, -0.02070426754653454, 0.21418233215808868, 0.06252310425043106, -0.16507434844970703, 0.04673760011792183, -0.05689859017729759, -0.08169996738433838, -0.11859188228845596, -0.05599171668291092, -0.07707660645246506, -0.14051198959350586, -0.005772426258772612, -0.11554284393787384, 0.0002381563390372321, 0.12560053169727325, 0.005618050694465637, -0.028257902711629868, 0.1500902622938156, 0.00589335709810257, 0.01685897633433342, 0.05763181671500206, 0.007298589684069157, -0.0412118062376976, -0.12455623596906662, -0.05731077864766121, -0.021210240200161934, -0.0067823175340890884, 0.028453994542360306, -0.060132455080747604, -0.04206157848238945, 0.029383787885308266, -0.02374044805765152, -0.09295868128538132, 0.004053476732224226, 0.013348747044801712, 0.05407030135393143, 0.052432116121053696, 0.008429237641394138, 0.01524826418608427, -0.0033252707216888666, 0.2079176902770996, -0.07260357588529587, -0.05715000629425049, -0.09797889739274979, 0.23888063430786133, 0.03585147485136986, -0.019686749204993248, 0.03734346106648445, -0.06568225473165512, -0.004361408296972513, 0.24954579770565033, 0.22186478972434998, -0.07830173522233963, -0.006712665781378746, 0.019575312733650208, -0.009557166136801243, -0.016031969338655472, 0.10772424191236496, 0.14170996844768524, 0.054387882351875305, -0.09126259386539459, -0.03596600145101547, -0.05642533674836159, -0.010790226049721241, -0.03824600949883461, 0.07427041232585907, 0.049858734011650085, 0.006499612703919411, -0.038984883576631546, 0.05010038614273071, -0.06054513528943062, -0.0999545156955719, 0.05072363093495369, -0.21182700991630554, -0.16378560662269592, -0.012944746762514114, 0.10200632363557816, -0.0016564095858484507, 0.06280625611543655, -0.03329086676239967, 0.002569399308413267, 0.08935030549764633, -0.01901974156498909, -0.09497696161270142, -0.07604139298200607, 0.09724408388137817, -0.11604832112789154, 0.22954528033733368, -0.045974187552928925, 0.055315155535936356, 0.12348725646734238, 0.06152692064642906, -0.07137792557477951, 0.06000271812081337, 0.04052073881030083, -0.044816117733716965, 0.020156823098659515, 0.06924524903297424, -0.035139527171850204, 0.06370943784713745, 0.045128483325242996, -0.1386181116104126, 0.010564154013991356, -0.04600505903363228, -0.0727553740143776, -0.041528042405843735, -0.030076157301664352, -0.0639835074543953, 0.1305534392595291, 0.20732425153255463, -0.02794814482331276, -0.013891981914639473, -0.07531335204839706, 0.01536940410733223, 0.05437286198139191, 0.015924062579870224, -0.05145624652504921, -0.20643475651741028, 0.018344108015298843, 0.04973233491182327, -0.02281355671584606, -0.24700039625167847, -0.10108326375484467, 0.010105917230248451, -0.07113083451986313, -0.0970691591501236, 0.07135510444641113, 0.08287014812231064, 0.0481468103826046, -0.06011649966239929, -0.048794958740472794, -0.07777239382266998, 0.1425076127052307, -0.14664505422115326, -0.0928139016032219 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # fnet-base-finetuned-cola This model is a fine-tuned version of [google/fnet-base](https://huggingface.co/google/fnet-base) on the GLUE COLA dataset. It achieves the following results on the evaluation set: - Loss: 0.5929 - Matthews Correlation: 0.3594 The model was fine-tuned to compare [google/fnet-base](https://huggingface.co/google/fnet-base) as introduced in [this paper](https://arxiv.org/abs/2105.03824) against [bert-base-cased](https://huggingface.co/bert-base-cased). ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure This model is trained using the [run_glue](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) script. The following command was used: ```bash #!/usr/bin/bash python ../run_glue.py \\n --model_name_or_path google/fnet-base \\n --task_name cola \\n --do_train \\n --do_eval \\n --max_seq_length 512 \\n --per_device_train_batch_size 16 \\n --learning_rate 2e-5 \\n --num_train_epochs 3 \\n --output_dir fnet-base-finetuned-cola \\n --push_to_hub \\n --hub_strategy all_checkpoints \\n --logging_strategy epoch \\n --save_strategy epoch \\n --evaluation_strategy epoch \\n``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.5895 | 1.0 | 535 | 0.6146 | 0.1699 | | 0.4656 | 2.0 | 1070 | 0.5667 | 0.3047 | | 0.3329 | 3.0 | 1605 | 0.5929 | 0.3594 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer", "fnet-bert-base-comparison"], "datasets": ["glue"], "metrics": ["matthews_correlation"], "model-index": [{"name": "fnet-base-finetuned-cola", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE COLA", "type": "glue", "args": "cola"}, "metrics": [{"type": "matthews_correlation", "value": 0.35940659235571387, "name": "Matthews Correlation"}]}]}]}
text-classification
gchhablani/fnet-base-finetuned-cola
[ "transformers", "pytorch", "tensorboard", "fnet", "text-classification", "generated_from_trainer", "fnet-bert-base-comparison", "en", "dataset:glue", "arxiv:2105.03824", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2105.03824" ]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #fnet #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
fnet-base-finetuned-cola ======================== This model is a fine-tuned version of google/fnet-base on the GLUE COLA dataset. It achieves the following results on the evaluation set: * Loss: 0.5929 * Matthews Correlation: 0.3594 The model was fine-tuned to compare google/fnet-base as introduced in this paper against bert-base-cased. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ This model is trained using the run\_glue script. The following command was used: ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #fnet #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 88, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #fnet #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.1281575709581375, 0.15705668926239014, -0.0038979940582066774, 0.11255989223718643, 0.11293601244688034, -0.008617952466011047, 0.13138744235038757, 0.1563730537891388, -0.1094919964671135, 0.05590956658124924, 0.1562153697013855, 0.15063650906085968, 0.03187095746397972, 0.18714870512485504, -0.05448957160115242, -0.23206175863742828, 0.030974131077528, 0.07522018253803253, -0.05673399940133095, 0.13729923963546753, 0.09973176568746567, -0.11717205494642258, 0.09386496245861053, 0.041752949357032776, -0.20222944021224976, -0.015307755209505558, -0.0008101215353235602, -0.08269140124320984, 0.11289821565151215, 0.02265351265668869, 0.08919014781713486, 0.03242024779319763, 0.03147246316075325, -0.1466279774904251, 0.007479942869395018, 0.055549394339323044, 0.0026599117554724216, 0.11788346618413925, 0.03937623277306557, -0.014160647056996822, 0.07863719761371613, -0.09087078273296356, 0.046092916280031204, 0.026117689907550812, -0.10794810950756073, -0.2786095142364502, -0.08820846676826477, 0.07192650437355042, 0.042675234377384186, 0.0806850865483284, 0.001147752278484404, 0.17639178037643433, -0.007389090955257416, 0.11422055959701538, 0.2610849142074585, -0.3091604709625244, -0.061695314943790436, 0.024227019399404526, 0.009995708242058754, 0.06274503469467163, -0.08376740664243698, -0.031031707301735878, 0.04775812104344368, 0.04504694044589996, 0.17843511700630188, -0.011732942424714565, -0.007466572802513838, -0.025801893323659897, -0.14371579885482788, -0.07526436448097229, 0.19688530266284943, 0.04901329055428505, -0.048377830535173416, -0.06737475097179413, -0.08691056817770004, -0.16888527572155, -0.03213018923997879, -0.018070582300424576, 0.04411018267273903, -0.0378425307571888, -0.060805536806583405, -0.029689611867070198, -0.07623965293169022, -0.042504265904426575, -0.042007382959127426, 0.15460588037967682, 0.05135969817638397, 0.03275564685463905, -0.03648478910326958, 0.08528219908475876, -0.016176627948880196, -0.15604232251644135, -0.0033359206281602383, 0.008210553787648678, 0.032338954508304596, -0.01879490725696087, -0.03313892334699631, -0.09054049104452133, 0.012796161696314812, 0.12658193707466125, -0.09896866977214813, 0.0699896514415741, 0.014115668833255768, 0.05199545621871948, -0.0832819789648056, 0.18305599689483643, -0.03047999180853367, 0.014890378341078758, 0.025784319266676903, 0.0863259807229042, 0.045972537249326706, -0.02756713703274727, -0.10970760136842728, 0.026695188134908676, 0.13855181634426117, 0.006172229070216417, -0.03736800700426102, 0.07175908982753754, -0.04684898629784584, -0.04453456029295921, 0.056356143206357956, -0.11108183860778809, 0.012465346604585648, 0.0007881103083491325, -0.08858437836170197, -0.03560861945152283, 0.026957271620631218, -0.012466381303966045, -0.04220158979296684, 0.05419426038861275, -0.0943819209933281, 0.009732727892696857, -0.05831371992826462, -0.1146949976682663, 0.013262145221233368, -0.11782565712928772, 0.0005620947922579944, -0.11043663322925568, -0.13709686696529388, -0.008001538924872875, 0.05364624038338661, -0.021408166736364365, -0.0810130313038826, -0.055306028574705124, -0.09359939396381378, 0.02663268707692623, -0.015616829507052898, 0.05365283414721489, -0.06507189571857452, 0.08449108898639679, 0.049848102033138275, 0.07236256450414658, -0.03752796724438667, 0.04536004737019539, -0.07848503440618515, 0.0464148111641407, -0.20472228527069092, 0.07016371190547943, -0.064207062125206, 0.06198544800281525, -0.1068328395485878, -0.11787286400794983, 0.030897412449121475, -0.03552781045436859, 0.08240031450986862, 0.09914099425077438, -0.14732807874679565, -0.08291295915842056, 0.19005797803401947, -0.08365947753190994, -0.12725262343883514, 0.12337704747915268, -0.04553937166929245, 0.0029195500537753105, 0.05737419053912163, 0.2257438600063324, 0.0889558345079422, -0.0515265017747879, -0.02331758849322796, -0.0051168533973395824, 0.04387769103050232, -0.07846325635910034, 0.08052174746990204, 0.006332707591354847, 0.04367716237902641, 0.02254851721227169, -0.029947886243462563, 0.04004385322332382, -0.08367564529180527, -0.08474674820899963, -0.05332755297422409, -0.08056257665157318, 0.056404486298561096, 0.04767722263932228, 0.08231600373983383, -0.11374399065971375, -0.094481460750103, 0.02864118292927742, 0.09102724492549896, -0.08461901545524597, 0.04763771593570709, -0.09820594638586044, 0.12363352626562119, -0.06146226450800896, -0.004213334526866674, -0.18391123414039612, -0.009381342679262161, 0.0438222773373127, -0.023048149421811104, -0.0010736447293311357, -0.017531832680106163, 0.0634535625576973, 0.05774131044745445, -0.039399147033691406, -0.0408325120806694, -0.033703941851854324, -0.009450753219425678, -0.11244195699691772, -0.18618963658809662, -0.04849305376410484, -0.034880541265010834, 0.11852548271417618, -0.17111770808696747, 0.06201837584376335, 0.06472091376781464, 0.11162162572145462, 0.02296465076506138, -0.02818126603960991, -0.0076725417748093605, 0.041496098041534424, -0.039293959736824036, -0.0763016790151596, 0.06855007261037827, 0.03613912686705589, -0.0994340106844902, -0.021636441349983215, -0.11005594581365585, 0.18589626252651215, 0.12711931765079498, -0.015304340980947018, -0.046145275235176086, -0.0037261333782225847, -0.06693658232688904, -0.0252322256565094, 0.00010479353659320623, 0.02630312740802765, 0.18688923120498657, 0.007668614853173494, 0.17470361292362213, -0.10448402166366577, -0.054649028927087784, 0.04829766973853111, -0.027804838493466377, -0.012250205501914024, 0.11106427013874054, 0.000900426646694541, -0.11244355887174606, 0.14791768789291382, 0.12212676554918289, -0.06123341992497444, 0.1253422647714615, -0.06269760429859161, -0.042313165962696075, -0.03154568001627922, 0.0070209321565926075, 0.016009759157896042, 0.09681913256645203, -0.12338456511497498, -0.019060473889112473, 0.03688754513859749, 0.026065392419695854, 0.022046322003006935, -0.18368276953697205, 0.0016608438454568386, 0.046017080545425415, -0.06232096627354622, -0.0001330666127614677, -0.01134573481976986, -0.0014391003642231226, 0.09911268204450607, 0.018035314977169037, -0.08723576366901398, 0.04871227592229843, 0.008767826482653618, -0.07373929023742676, 0.20202793180942535, -0.10205624997615814, -0.19618737697601318, -0.11792904138565063, -0.0520332008600235, -0.09464605152606964, -0.00033909876947291195, 0.06213480234146118, -0.07916445285081863, -0.022073807194828987, -0.09520765393972397, -0.03123565763235092, -0.01737889274954796, 0.03542753681540489, 0.07313533127307892, -0.02203269489109516, 0.10872673243284225, -0.1269455999135971, -0.02306712418794632, -0.0358131118118763, 0.008673591539263725, 0.055851276963949203, 0.009536282159388065, 0.09724908322095871, 0.1170153021812439, -0.033818624913692474, 0.05197916179895401, -0.028282886371016502, 0.24284608662128448, -0.053264424204826355, -0.02976847253739834, 0.12733469903469086, -0.0040460084564983845, 0.08623170107603073, 0.08496475964784622, 0.04557865113019943, -0.08525460958480835, -0.01157840620726347, 0.00595470005646348, -0.03817666694521904, -0.21714363992214203, -0.036886055022478104, -0.04140350595116615, 0.02079310268163681, 0.10649195313453674, 0.04592718183994293, 0.043558333069086075, 0.057499635964632034, 0.023834481835365295, 0.04960625246167183, -0.02408022992312908, 0.09037080407142639, 0.1317465603351593, 0.04990197345614433, 0.13514135777950287, -0.04064502939581871, -0.03152943775057793, 0.039630502462387085, -0.002475862158462405, 0.19950714707374573, -0.02927609719336033, 0.1829756200313568, 0.04599745571613312, 0.1883370578289032, 0.011177031323313713, 0.06271716207265854, -0.024891022592782974, -0.006736312061548233, -0.009674341417849064, -0.0412435419857502, -0.05800659582018852, 0.0024435357190668583, -0.04474163055419922, 0.07293359190225601, -0.117747962474823, 0.040190309286117554, 0.06010587513446808, 0.29027169942855835, 0.022407572716474533, -0.37493225932121277, -0.11181744188070297, -0.016597270965576172, -0.031712714582681656, -0.04552937671542168, 0.010863134637475014, 0.11205030232667923, -0.09440278261899948, 0.06324154883623123, -0.08447496592998505, 0.09077057987451553, -0.07360748946666718, 0.03601958230137825, 0.05221237242221832, 0.09796657413244247, 0.006393694784492254, 0.06149907410144806, -0.26434561610221863, 0.25069257616996765, 0.018371766433119774, 0.05161698907613754, -0.06146470457315445, 0.015569853596389294, 0.02125987596809864, 0.06923633813858032, 0.07857105880975723, 0.0007500528590753675, -0.02767954021692276, -0.16410666704177856, -0.11515739560127258, 0.016607077792286873, 0.07020550966262817, -0.02121029421687126, 0.08431950211524963, -0.0010764591861516237, 0.004067974630743265, 0.04312531650066376, -0.008349082432687283, -0.027957666665315628, -0.09104249626398087, 0.016294151544570923, 0.06238500773906708, -0.04210955649614334, -0.08233670145273209, -0.1198863536119461, -0.08412427455186844, 0.1865774542093277, -0.008332050405442715, -0.0823029950261116, -0.12076611816883087, 0.06184208020567894, 0.06703545898199081, -0.09185026586055756, 0.042081426829099655, -0.017016833648085594, 0.12382151186466217, 0.003344601020216942, -0.07244402170181274, 0.09798984974622726, -0.04916444793343544, -0.16366566717624664, -0.03181470185518265, 0.13521257042884827, 0.034096188843250275, 0.05754973739385605, -0.012549868784844875, 0.019313739612698555, -0.026896705850958824, -0.0765208750963211, 0.03267578408122063, 0.002446891972795129, 0.09423114359378815, -0.015256565995514393, 0.0032036362681537867, 0.028266185894608498, -0.07761038839817047, 0.005427155178040266, 0.19771218299865723, 0.2606421709060669, -0.1043548434972763, 0.03651893138885498, 0.027820715680718422, -0.04207947105169296, -0.1520952433347702, 0.017168138176202774, 0.08070502430200577, 0.007420012727379799, -0.010656541213393211, -0.1764611452817917, 0.05012203007936478, 0.08809427171945572, -0.01788947917521, 0.08023256808519363, -0.29865652322769165, -0.11554376780986786, 0.11132355779409409, 0.12564530968666077, 0.09193845093250275, -0.13787555694580078, -0.04598912596702576, -0.0075193652883172035, -0.12548284232616425, 0.1158299669623375, -0.0956348180770874, 0.10974547266960144, -0.04320225492119789, 0.05354843661189079, 0.007337958551943302, -0.052445266395807266, 0.12752960622310638, 0.01908416859805584, 0.0783768892288208, -0.0478614941239357, 0.003130370983853936, 0.10798399150371552, -0.0758182629942894, 0.06092115119099617, -0.10174112766981125, 0.04936999827623367, -0.12913434207439423, -0.01087163481861353, -0.08215732127428055, 0.02880539558827877, -0.029542503878474236, -0.03560418263077736, -0.058059368282556534, 0.009000681340694427, 0.07482077181339264, -0.0021140656899660826, 0.17803633213043213, 0.047338198870420456, 0.12804824113845825, 0.2045678347349167, 0.07827051728963852, -0.11816922575235367, -0.10215629637241364, -0.0021673408336937428, -0.009365938603878021, 0.0566975399851799, -0.15991105139255524, 0.03947640210390091, 0.1441853940486908, 0.004068439826369286, 0.11697175353765488, 0.06757407635450363, -0.058885324746370316, -0.00002974093695229385, 0.04177243635058403, -0.17993171513080597, -0.10290152579545975, -0.014580353163182735, -0.012066712602972984, -0.13221070170402527, 0.07946693897247314, 0.11126816272735596, -0.07109289616346359, -0.018744582310318947, 0.0011641265591606498, -0.0008233633125200868, -0.025426305830478668, 0.18059362471103668, 0.0713382288813591, 0.07033462822437286, -0.10616692155599594, 0.09460733085870743, 0.0462157279253006, -0.0725565180182457, 0.04061013460159302, 0.06810374557971954, -0.11764568090438843, -0.022825488820672035, 0.037216611206531525, 0.14856918156147003, -0.027179433032870293, -0.05499974638223648, -0.17361967265605927, -0.1086450144648552, 0.08841122686862946, 0.13100364804267883, 0.10317601263523102, 0.017121080309152603, -0.04885329306125641, -0.009807179681956768, -0.11012768745422363, 0.09655274450778961, 0.05846559628844261, 0.0660039484500885, -0.16399246454238892, 0.11977937817573547, -0.0003578494652174413, 0.07169348001480103, -0.012192845344543457, 0.001448655384592712, -0.0843949168920517, 0.004430484492331743, -0.09232339262962341, 0.012558856047689915, -0.04001414775848389, 0.00048618309665471315, -0.022317882627248764, -0.04360825940966606, -0.0546182319521904, 0.03691510483622551, -0.10589302331209183, -0.03589541092514992, 0.030760591849684715, 0.02635170891880989, -0.11072052270174026, -0.031555552035570145, 0.0018842692952603102, -0.08543401211500168, 0.09491559118032455, 0.053661517798900604, -0.007116627413779497, 0.009966611862182617, -0.003835842479020357, 0.002103820675984025, 0.0713348463177681, 0.007812848314642906, 0.06204405426979065, -0.11863148212432861, -0.004516935441643, 0.004596107639372349, 0.001546198152936995, 0.019018597900867462, 0.12120135128498077, -0.12007837742567062, -0.014022485353052616, -0.022460926324129105, -0.022654950618743896, -0.0753025934100151, 0.05987093597650528, 0.10378355532884598, 0.041243866086006165, 0.19044536352157593, -0.06453124433755875, 0.015417633578181267, -0.2059982270002365, -0.0010957218473777175, 0.009144444018602371, -0.1515446901321411, -0.05387648567557335, -0.025656811892986298, 0.05819327011704445, -0.07224544882774353, 0.11503966152667999, 0.004568712320178747, -0.01583913527429104, 0.03687077388167381, -0.044159963726997375, -0.016699714586138725, 0.008686945773661137, 0.16097450256347656, 0.013965492136776447, -0.037191856652498245, 0.12147503346204758, 0.023390216752886772, 0.1060599610209465, 0.12157408893108368, 0.1685730367898941, 0.12814003229141235, 0.026453150436282158, 0.10737672448158264, 0.032515060156583786, -0.030444564297795296, -0.18784967064857483, 0.06720752269029617, -0.04578623175621033, 0.13226191699504852, 0.006823073141276836, 0.17648069560527802, 0.11036133766174316, -0.15301413834095, 0.049124132841825485, -0.021638398990035057, -0.09448863565921783, -0.10050928592681885, -0.09679119288921356, -0.07743744552135468, -0.15149177610874176, -0.01236912701278925, -0.11908751726150513, -0.001917796558700502, 0.06672867387533188, 0.0058901347219944, -0.021161947399377823, 0.13932976126670837, 0.023673847317695618, -0.0022931741550564766, 0.0803397074341774, -0.0035189997870475054, -0.05596965178847313, -0.07125981897115707, -0.07409729808568954, 0.013776594772934914, 0.013752772472798824, 0.05871286615729332, -0.02784240059554577, 0.009659827686846256, 0.051329080015420914, -0.02629757858812809, -0.10809190571308136, 0.009442382492125034, 0.021296720951795578, 0.054612502455711365, 0.020937973633408546, 0.013347321189939976, -0.011912370100617409, -0.017500825226306915, 0.17539910972118378, -0.06457335501909256, -0.022280963137745857, -0.10723160207271576, 0.20094625651836395, 0.036107197403907776, -0.03869583457708359, 0.040720485150814056, -0.06905907392501831, -0.020572103559970856, 0.18964605033397675, 0.20552918314933777, -0.026153136044740677, 0.0006357349921017885, -0.004643670748919249, -0.01797187142074108, -0.007700672838836908, 0.08245248347520828, 0.13317352533340454, -0.0074555822648108006, -0.06725732237100601, -0.026400936767458916, -0.06664355099201202, -0.004993689712136984, -0.048690151423215866, 0.06622577458620071, 0.027980800718069077, 0.012009805999696255, -0.046264056116342545, 0.022495536133646965, -0.042431484907865524, -0.07840681821107864, 0.030966809019446373, -0.20489253103733063, -0.15987305343151093, -0.008346200920641422, 0.03857562318444252, 0.0004496309265960008, 0.062375955283641815, -0.0009975176071748137, 0.012196631170809269, 0.0908537283539772, -0.02237413451075554, -0.08646247535943985, -0.07932375371456146, 0.09022057801485062, -0.17671364545822144, 0.2005920559167862, -0.04314102232456207, 0.049184780567884445, 0.1326240599155426, 0.04266604781150818, -0.10272050648927689, 0.0379687063395977, 0.03511279076337814, -0.008898447267711163, -0.004219701513648033, 0.09855833649635315, -0.012647976167500019, 0.07372915744781494, 0.03857356309890747, -0.09686969965696335, -0.039119016379117966, -0.07064490765333176, -0.0207537654787302, -0.040169861167669296, -0.052368659526109695, -0.05137869715690613, 0.11046379059553146, 0.19000554084777832, -0.049329470843076706, -0.017489463090896606, -0.06654652208089828, 0.005614506546407938, 0.07303186506032944, -0.011176830157637596, -0.04266827553510666, -0.253805011510849, 0.010820526629686356, 0.046582285314798355, -0.017032980918884277, -0.2275182157754898, -0.1029951274394989, -0.004221665672957897, -0.06776802986860275, -0.07437815517187119, 0.07514068484306335, 0.04294868931174278, 0.05498082563281059, -0.06696301698684692, 0.028071772307157516, -0.09679107367992401, 0.15699414908885956, -0.15248030424118042, -0.07556789368391037 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # fnet-base-finetuned-mnli This model is a fine-tuned version of [google/fnet-base](https://huggingface.co/google/fnet-base) on the GLUE MNLI dataset. It achieves the following results on the evaluation set: - Loss: 0.6443 - Accuracy: 0.7675 The model was fine-tuned to compare [google/fnet-base](https://huggingface.co/google/fnet-base) as introduced in [this paper](https://arxiv.org/abs/2105.03824) against [bert-base-cased](https://huggingface.co/bert-base-cased). ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure This model is trained using the [run_glue](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) script. The following command was used: ```bash #!/usr/bin/bash python ../run_glue.py \\n --model_name_or_path google/fnet-base \\n --task_name mnli \\n --do_train \\n --do_eval \\n --max_seq_length 512 \\n --per_device_train_batch_size 16 \\n --learning_rate 2e-5 \\n --num_train_epochs 3 \\n --output_dir fnet-base-finetuned-mnli \\n --push_to_hub \\n --hub_strategy all_checkpoints \\n --logging_strategy epoch \\n --save_strategy epoch \\n --evaluation_strategy epoch \\n``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:--------:| | 0.7143 | 1.0 | 24544 | 0.6169 | 0.7504 | | 0.5407 | 2.0 | 49088 | 0.6218 | 0.7627 | | 0.4178 | 3.0 | 73632 | 0.6564 | 0.7658 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer", "fnet-bert-base-comparison"], "datasets": ["glue"], "metrics": ["accuracy"], "model-index": [{"name": "fnet-base-finetuned-mnli", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE MNLI", "type": "glue", "args": "mnli"}, "metrics": [{"type": "accuracy", "value": 0.7674938974776241, "name": "Accuracy"}]}]}]}
text-classification
gchhablani/fnet-base-finetuned-mnli
[ "transformers", "pytorch", "tensorboard", "fnet", "text-classification", "generated_from_trainer", "fnet-bert-base-comparison", "en", "dataset:glue", "arxiv:2105.03824", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2105.03824" ]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #fnet #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
fnet-base-finetuned-mnli ======================== This model is a fine-tuned version of google/fnet-base on the GLUE MNLI dataset. It achieves the following results on the evaluation set: * Loss: 0.6443 * Accuracy: 0.7675 The model was fine-tuned to compare google/fnet-base as introduced in this paper against bert-base-cased. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ This model is trained using the run\_glue script. The following command was used: ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #fnet #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 88, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #fnet #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.1281575709581375, 0.15705668926239014, -0.0038979940582066774, 0.11255989223718643, 0.11293601244688034, -0.008617952466011047, 0.13138744235038757, 0.1563730537891388, -0.1094919964671135, 0.05590956658124924, 0.1562153697013855, 0.15063650906085968, 0.03187095746397972, 0.18714870512485504, -0.05448957160115242, -0.23206175863742828, 0.030974131077528, 0.07522018253803253, -0.05673399940133095, 0.13729923963546753, 0.09973176568746567, -0.11717205494642258, 0.09386496245861053, 0.041752949357032776, -0.20222944021224976, -0.015307755209505558, -0.0008101215353235602, -0.08269140124320984, 0.11289821565151215, 0.02265351265668869, 0.08919014781713486, 0.03242024779319763, 0.03147246316075325, -0.1466279774904251, 0.007479942869395018, 0.055549394339323044, 0.0026599117554724216, 0.11788346618413925, 0.03937623277306557, -0.014160647056996822, 0.07863719761371613, -0.09087078273296356, 0.046092916280031204, 0.026117689907550812, -0.10794810950756073, -0.2786095142364502, -0.08820846676826477, 0.07192650437355042, 0.042675234377384186, 0.0806850865483284, 0.001147752278484404, 0.17639178037643433, -0.007389090955257416, 0.11422055959701538, 0.2610849142074585, -0.3091604709625244, -0.061695314943790436, 0.024227019399404526, 0.009995708242058754, 0.06274503469467163, -0.08376740664243698, -0.031031707301735878, 0.04775812104344368, 0.04504694044589996, 0.17843511700630188, -0.011732942424714565, -0.007466572802513838, -0.025801893323659897, -0.14371579885482788, -0.07526436448097229, 0.19688530266284943, 0.04901329055428505, -0.048377830535173416, -0.06737475097179413, -0.08691056817770004, -0.16888527572155, -0.03213018923997879, -0.018070582300424576, 0.04411018267273903, -0.0378425307571888, -0.060805536806583405, -0.029689611867070198, -0.07623965293169022, -0.042504265904426575, -0.042007382959127426, 0.15460588037967682, 0.05135969817638397, 0.03275564685463905, -0.03648478910326958, 0.08528219908475876, -0.016176627948880196, -0.15604232251644135, -0.0033359206281602383, 0.008210553787648678, 0.032338954508304596, -0.01879490725696087, -0.03313892334699631, -0.09054049104452133, 0.012796161696314812, 0.12658193707466125, -0.09896866977214813, 0.0699896514415741, 0.014115668833255768, 0.05199545621871948, -0.0832819789648056, 0.18305599689483643, -0.03047999180853367, 0.014890378341078758, 0.025784319266676903, 0.0863259807229042, 0.045972537249326706, -0.02756713703274727, -0.10970760136842728, 0.026695188134908676, 0.13855181634426117, 0.006172229070216417, -0.03736800700426102, 0.07175908982753754, -0.04684898629784584, -0.04453456029295921, 0.056356143206357956, -0.11108183860778809, 0.012465346604585648, 0.0007881103083491325, -0.08858437836170197, -0.03560861945152283, 0.026957271620631218, -0.012466381303966045, -0.04220158979296684, 0.05419426038861275, -0.0943819209933281, 0.009732727892696857, -0.05831371992826462, -0.1146949976682663, 0.013262145221233368, -0.11782565712928772, 0.0005620947922579944, -0.11043663322925568, -0.13709686696529388, -0.008001538924872875, 0.05364624038338661, -0.021408166736364365, -0.0810130313038826, -0.055306028574705124, -0.09359939396381378, 0.02663268707692623, -0.015616829507052898, 0.05365283414721489, -0.06507189571857452, 0.08449108898639679, 0.049848102033138275, 0.07236256450414658, -0.03752796724438667, 0.04536004737019539, -0.07848503440618515, 0.0464148111641407, -0.20472228527069092, 0.07016371190547943, -0.064207062125206, 0.06198544800281525, -0.1068328395485878, -0.11787286400794983, 0.030897412449121475, -0.03552781045436859, 0.08240031450986862, 0.09914099425077438, -0.14732807874679565, -0.08291295915842056, 0.19005797803401947, -0.08365947753190994, -0.12725262343883514, 0.12337704747915268, -0.04553937166929245, 0.0029195500537753105, 0.05737419053912163, 0.2257438600063324, 0.0889558345079422, -0.0515265017747879, -0.02331758849322796, -0.0051168533973395824, 0.04387769103050232, -0.07846325635910034, 0.08052174746990204, 0.006332707591354847, 0.04367716237902641, 0.02254851721227169, -0.029947886243462563, 0.04004385322332382, -0.08367564529180527, -0.08474674820899963, -0.05332755297422409, -0.08056257665157318, 0.056404486298561096, 0.04767722263932228, 0.08231600373983383, -0.11374399065971375, -0.094481460750103, 0.02864118292927742, 0.09102724492549896, -0.08461901545524597, 0.04763771593570709, -0.09820594638586044, 0.12363352626562119, -0.06146226450800896, -0.004213334526866674, -0.18391123414039612, -0.009381342679262161, 0.0438222773373127, -0.023048149421811104, -0.0010736447293311357, -0.017531832680106163, 0.0634535625576973, 0.05774131044745445, -0.039399147033691406, -0.0408325120806694, -0.033703941851854324, -0.009450753219425678, -0.11244195699691772, -0.18618963658809662, -0.04849305376410484, -0.034880541265010834, 0.11852548271417618, -0.17111770808696747, 0.06201837584376335, 0.06472091376781464, 0.11162162572145462, 0.02296465076506138, -0.02818126603960991, -0.0076725417748093605, 0.041496098041534424, -0.039293959736824036, -0.0763016790151596, 0.06855007261037827, 0.03613912686705589, -0.0994340106844902, -0.021636441349983215, -0.11005594581365585, 0.18589626252651215, 0.12711931765079498, -0.015304340980947018, -0.046145275235176086, -0.0037261333782225847, -0.06693658232688904, -0.0252322256565094, 0.00010479353659320623, 0.02630312740802765, 0.18688923120498657, 0.007668614853173494, 0.17470361292362213, -0.10448402166366577, -0.054649028927087784, 0.04829766973853111, -0.027804838493466377, -0.012250205501914024, 0.11106427013874054, 0.000900426646694541, -0.11244355887174606, 0.14791768789291382, 0.12212676554918289, -0.06123341992497444, 0.1253422647714615, -0.06269760429859161, -0.042313165962696075, -0.03154568001627922, 0.0070209321565926075, 0.016009759157896042, 0.09681913256645203, -0.12338456511497498, -0.019060473889112473, 0.03688754513859749, 0.026065392419695854, 0.022046322003006935, -0.18368276953697205, 0.0016608438454568386, 0.046017080545425415, -0.06232096627354622, -0.0001330666127614677, -0.01134573481976986, -0.0014391003642231226, 0.09911268204450607, 0.018035314977169037, -0.08723576366901398, 0.04871227592229843, 0.008767826482653618, -0.07373929023742676, 0.20202793180942535, -0.10205624997615814, -0.19618737697601318, -0.11792904138565063, -0.0520332008600235, -0.09464605152606964, -0.00033909876947291195, 0.06213480234146118, -0.07916445285081863, -0.022073807194828987, -0.09520765393972397, -0.03123565763235092, -0.01737889274954796, 0.03542753681540489, 0.07313533127307892, -0.02203269489109516, 0.10872673243284225, -0.1269455999135971, -0.02306712418794632, -0.0358131118118763, 0.008673591539263725, 0.055851276963949203, 0.009536282159388065, 0.09724908322095871, 0.1170153021812439, -0.033818624913692474, 0.05197916179895401, -0.028282886371016502, 0.24284608662128448, -0.053264424204826355, -0.02976847253739834, 0.12733469903469086, -0.0040460084564983845, 0.08623170107603073, 0.08496475964784622, 0.04557865113019943, -0.08525460958480835, -0.01157840620726347, 0.00595470005646348, -0.03817666694521904, -0.21714363992214203, -0.036886055022478104, -0.04140350595116615, 0.02079310268163681, 0.10649195313453674, 0.04592718183994293, 0.043558333069086075, 0.057499635964632034, 0.023834481835365295, 0.04960625246167183, -0.02408022992312908, 0.09037080407142639, 0.1317465603351593, 0.04990197345614433, 0.13514135777950287, -0.04064502939581871, -0.03152943775057793, 0.039630502462387085, -0.002475862158462405, 0.19950714707374573, -0.02927609719336033, 0.1829756200313568, 0.04599745571613312, 0.1883370578289032, 0.011177031323313713, 0.06271716207265854, -0.024891022592782974, -0.006736312061548233, -0.009674341417849064, -0.0412435419857502, -0.05800659582018852, 0.0024435357190668583, -0.04474163055419922, 0.07293359190225601, -0.117747962474823, 0.040190309286117554, 0.06010587513446808, 0.29027169942855835, 0.022407572716474533, -0.37493225932121277, -0.11181744188070297, -0.016597270965576172, -0.031712714582681656, -0.04552937671542168, 0.010863134637475014, 0.11205030232667923, -0.09440278261899948, 0.06324154883623123, -0.08447496592998505, 0.09077057987451553, -0.07360748946666718, 0.03601958230137825, 0.05221237242221832, 0.09796657413244247, 0.006393694784492254, 0.06149907410144806, -0.26434561610221863, 0.25069257616996765, 0.018371766433119774, 0.05161698907613754, -0.06146470457315445, 0.015569853596389294, 0.02125987596809864, 0.06923633813858032, 0.07857105880975723, 0.0007500528590753675, -0.02767954021692276, -0.16410666704177856, -0.11515739560127258, 0.016607077792286873, 0.07020550966262817, -0.02121029421687126, 0.08431950211524963, -0.0010764591861516237, 0.004067974630743265, 0.04312531650066376, -0.008349082432687283, -0.027957666665315628, -0.09104249626398087, 0.016294151544570923, 0.06238500773906708, -0.04210955649614334, -0.08233670145273209, -0.1198863536119461, -0.08412427455186844, 0.1865774542093277, -0.008332050405442715, -0.0823029950261116, -0.12076611816883087, 0.06184208020567894, 0.06703545898199081, -0.09185026586055756, 0.042081426829099655, -0.017016833648085594, 0.12382151186466217, 0.003344601020216942, -0.07244402170181274, 0.09798984974622726, -0.04916444793343544, -0.16366566717624664, -0.03181470185518265, 0.13521257042884827, 0.034096188843250275, 0.05754973739385605, -0.012549868784844875, 0.019313739612698555, -0.026896705850958824, -0.0765208750963211, 0.03267578408122063, 0.002446891972795129, 0.09423114359378815, -0.015256565995514393, 0.0032036362681537867, 0.028266185894608498, -0.07761038839817047, 0.005427155178040266, 0.19771218299865723, 0.2606421709060669, -0.1043548434972763, 0.03651893138885498, 0.027820715680718422, -0.04207947105169296, -0.1520952433347702, 0.017168138176202774, 0.08070502430200577, 0.007420012727379799, -0.010656541213393211, -0.1764611452817917, 0.05012203007936478, 0.08809427171945572, -0.01788947917521, 0.08023256808519363, -0.29865652322769165, -0.11554376780986786, 0.11132355779409409, 0.12564530968666077, 0.09193845093250275, -0.13787555694580078, -0.04598912596702576, -0.0075193652883172035, -0.12548284232616425, 0.1158299669623375, -0.0956348180770874, 0.10974547266960144, -0.04320225492119789, 0.05354843661189079, 0.007337958551943302, -0.052445266395807266, 0.12752960622310638, 0.01908416859805584, 0.0783768892288208, -0.0478614941239357, 0.003130370983853936, 0.10798399150371552, -0.0758182629942894, 0.06092115119099617, -0.10174112766981125, 0.04936999827623367, -0.12913434207439423, -0.01087163481861353, -0.08215732127428055, 0.02880539558827877, -0.029542503878474236, -0.03560418263077736, -0.058059368282556534, 0.009000681340694427, 0.07482077181339264, -0.0021140656899660826, 0.17803633213043213, 0.047338198870420456, 0.12804824113845825, 0.2045678347349167, 0.07827051728963852, -0.11816922575235367, -0.10215629637241364, -0.0021673408336937428, -0.009365938603878021, 0.0566975399851799, -0.15991105139255524, 0.03947640210390091, 0.1441853940486908, 0.004068439826369286, 0.11697175353765488, 0.06757407635450363, -0.058885324746370316, -0.00002974093695229385, 0.04177243635058403, -0.17993171513080597, -0.10290152579545975, -0.014580353163182735, -0.012066712602972984, -0.13221070170402527, 0.07946693897247314, 0.11126816272735596, -0.07109289616346359, -0.018744582310318947, 0.0011641265591606498, -0.0008233633125200868, -0.025426305830478668, 0.18059362471103668, 0.0713382288813591, 0.07033462822437286, -0.10616692155599594, 0.09460733085870743, 0.0462157279253006, -0.0725565180182457, 0.04061013460159302, 0.06810374557971954, -0.11764568090438843, -0.022825488820672035, 0.037216611206531525, 0.14856918156147003, -0.027179433032870293, -0.05499974638223648, -0.17361967265605927, -0.1086450144648552, 0.08841122686862946, 0.13100364804267883, 0.10317601263523102, 0.017121080309152603, -0.04885329306125641, -0.009807179681956768, -0.11012768745422363, 0.09655274450778961, 0.05846559628844261, 0.0660039484500885, -0.16399246454238892, 0.11977937817573547, -0.0003578494652174413, 0.07169348001480103, -0.012192845344543457, 0.001448655384592712, -0.0843949168920517, 0.004430484492331743, -0.09232339262962341, 0.012558856047689915, -0.04001414775848389, 0.00048618309665471315, -0.022317882627248764, -0.04360825940966606, -0.0546182319521904, 0.03691510483622551, -0.10589302331209183, -0.03589541092514992, 0.030760591849684715, 0.02635170891880989, -0.11072052270174026, -0.031555552035570145, 0.0018842692952603102, -0.08543401211500168, 0.09491559118032455, 0.053661517798900604, -0.007116627413779497, 0.009966611862182617, -0.003835842479020357, 0.002103820675984025, 0.0713348463177681, 0.007812848314642906, 0.06204405426979065, -0.11863148212432861, -0.004516935441643, 0.004596107639372349, 0.001546198152936995, 0.019018597900867462, 0.12120135128498077, -0.12007837742567062, -0.014022485353052616, -0.022460926324129105, -0.022654950618743896, -0.0753025934100151, 0.05987093597650528, 0.10378355532884598, 0.041243866086006165, 0.19044536352157593, -0.06453124433755875, 0.015417633578181267, -0.2059982270002365, -0.0010957218473777175, 0.009144444018602371, -0.1515446901321411, -0.05387648567557335, -0.025656811892986298, 0.05819327011704445, -0.07224544882774353, 0.11503966152667999, 0.004568712320178747, -0.01583913527429104, 0.03687077388167381, -0.044159963726997375, -0.016699714586138725, 0.008686945773661137, 0.16097450256347656, 0.013965492136776447, -0.037191856652498245, 0.12147503346204758, 0.023390216752886772, 0.1060599610209465, 0.12157408893108368, 0.1685730367898941, 0.12814003229141235, 0.026453150436282158, 0.10737672448158264, 0.032515060156583786, -0.030444564297795296, -0.18784967064857483, 0.06720752269029617, -0.04578623175621033, 0.13226191699504852, 0.006823073141276836, 0.17648069560527802, 0.11036133766174316, -0.15301413834095, 0.049124132841825485, -0.021638398990035057, -0.09448863565921783, -0.10050928592681885, -0.09679119288921356, -0.07743744552135468, -0.15149177610874176, -0.01236912701278925, -0.11908751726150513, -0.001917796558700502, 0.06672867387533188, 0.0058901347219944, -0.021161947399377823, 0.13932976126670837, 0.023673847317695618, -0.0022931741550564766, 0.0803397074341774, -0.0035189997870475054, -0.05596965178847313, -0.07125981897115707, -0.07409729808568954, 0.013776594772934914, 0.013752772472798824, 0.05871286615729332, -0.02784240059554577, 0.009659827686846256, 0.051329080015420914, -0.02629757858812809, -0.10809190571308136, 0.009442382492125034, 0.021296720951795578, 0.054612502455711365, 0.020937973633408546, 0.013347321189939976, -0.011912370100617409, -0.017500825226306915, 0.17539910972118378, -0.06457335501909256, -0.022280963137745857, -0.10723160207271576, 0.20094625651836395, 0.036107197403907776, -0.03869583457708359, 0.040720485150814056, -0.06905907392501831, -0.020572103559970856, 0.18964605033397675, 0.20552918314933777, -0.026153136044740677, 0.0006357349921017885, -0.004643670748919249, -0.01797187142074108, -0.007700672838836908, 0.08245248347520828, 0.13317352533340454, -0.0074555822648108006, -0.06725732237100601, -0.026400936767458916, -0.06664355099201202, -0.004993689712136984, -0.048690151423215866, 0.06622577458620071, 0.027980800718069077, 0.012009805999696255, -0.046264056116342545, 0.022495536133646965, -0.042431484907865524, -0.07840681821107864, 0.030966809019446373, -0.20489253103733063, -0.15987305343151093, -0.008346200920641422, 0.03857562318444252, 0.0004496309265960008, 0.062375955283641815, -0.0009975176071748137, 0.012196631170809269, 0.0908537283539772, -0.02237413451075554, -0.08646247535943985, -0.07932375371456146, 0.09022057801485062, -0.17671364545822144, 0.2005920559167862, -0.04314102232456207, 0.049184780567884445, 0.1326240599155426, 0.04266604781150818, -0.10272050648927689, 0.0379687063395977, 0.03511279076337814, -0.008898447267711163, -0.004219701513648033, 0.09855833649635315, -0.012647976167500019, 0.07372915744781494, 0.03857356309890747, -0.09686969965696335, -0.039119016379117966, -0.07064490765333176, -0.0207537654787302, -0.040169861167669296, -0.052368659526109695, -0.05137869715690613, 0.11046379059553146, 0.19000554084777832, -0.049329470843076706, -0.017489463090896606, -0.06654652208089828, 0.005614506546407938, 0.07303186506032944, -0.011176830157637596, -0.04266827553510666, -0.253805011510849, 0.010820526629686356, 0.046582285314798355, -0.017032980918884277, -0.2275182157754898, -0.1029951274394989, -0.004221665672957897, -0.06776802986860275, -0.07437815517187119, 0.07514068484306335, 0.04294868931174278, 0.05498082563281059, -0.06696301698684692, 0.028071772307157516, -0.09679107367992401, 0.15699414908885956, -0.15248030424118042, -0.07556789368391037 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # fnet-base-finetuned-mrpc This model is a fine-tuned version of [google/fnet-base](https://huggingface.co/google/fnet-base) on the GLUE MRPC dataset. It achieves the following results on the evaluation set: - Loss: 0.9653 - Accuracy: 0.7721 - F1: 0.8502 - Combined Score: 0.8112 The model was fine-tuned to compare [google/fnet-base](https://huggingface.co/google/fnet-base) as introduced in [this paper](https://arxiv.org/abs/2105.03824) against [bert-base-cased](https://huggingface.co/bert-base-cased). ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure This model is trained using the [run_glue](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) script. The following command was used: ```bash #!/usr/bin/bash python ../run_glue.py \\n --model_name_or_path google/fnet-base \\n --task_name mrpc \\n --do_train \\n --do_eval \\n --max_seq_length 512 \\n --per_device_train_batch_size 16 \\n --learning_rate 2e-5 \\n --num_train_epochs 5 \\n --output_dir fnet-base-finetuned-mrpc \\n --push_to_hub \\n --hub_strategy all_checkpoints \\n --logging_strategy epoch \\n --save_strategy epoch \\n --evaluation_strategy epoch \\n``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Combined Score | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:--------------:| | 0.544 | 1.0 | 230 | 0.5272 | 0.7328 | 0.8300 | 0.7814 | | 0.4034 | 2.0 | 460 | 0.6211 | 0.7255 | 0.8298 | 0.7776 | | 0.2602 | 3.0 | 690 | 0.9110 | 0.7230 | 0.8306 | 0.7768 | | 0.1688 | 4.0 | 920 | 0.8640 | 0.7696 | 0.8489 | 0.8092 | | 0.0913 | 5.0 | 1150 | 0.9653 | 0.7721 | 0.8502 | 0.8112 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer", "fnet-bert-base-comparison"], "datasets": ["glue"], "metrics": ["accuracy", "f1"], "model-index": [{"name": "fnet-base-finetuned-mrpc", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE MRPC", "type": "glue", "args": "mrpc"}, "metrics": [{"type": "accuracy", "value": 0.7720588235294118, "name": "Accuracy"}, {"type": "f1", "value": 0.8502415458937198, "name": "F1"}]}]}]}
text-classification
gchhablani/fnet-base-finetuned-mrpc
[ "transformers", "pytorch", "tensorboard", "fnet", "text-classification", "generated_from_trainer", "fnet-bert-base-comparison", "en", "dataset:glue", "arxiv:2105.03824", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2105.03824" ]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #fnet #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
fnet-base-finetuned-mrpc ======================== This model is a fine-tuned version of google/fnet-base on the GLUE MRPC dataset. It achieves the following results on the evaluation set: * Loss: 0.9653 * Accuracy: 0.7721 * F1: 0.8502 * Combined Score: 0.8112 The model was fine-tuned to compare google/fnet-base as introduced in this paper against bert-base-cased. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ This model is trained using the run\_glue script. The following command was used: ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 5.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #fnet #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 88, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #fnet #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.1294347047805786, 0.15776823461055756, -0.003917243331670761, 0.11227939277887344, 0.11287269741296768, -0.00867973268032074, 0.13006214797496796, 0.15599386394023895, -0.10833613574504852, 0.05695851519703865, 0.15625059604644775, 0.14928914606571198, 0.031990837305784225, 0.18835246562957764, -0.054748013615608215, -0.2318466156721115, 0.030150001868605614, 0.0764521062374115, -0.05639369785785675, 0.13710667192935944, 0.0989472046494484, -0.11860295385122299, 0.09347143769264221, 0.04175164923071861, -0.2037433683872223, -0.01586032286286354, 0.00032077240757644176, -0.08281534165143967, 0.11295787245035172, 0.022520827129483223, 0.08912119269371033, 0.03228167071938515, 0.03050021082162857, -0.1462295800447464, 0.007699727546423674, 0.05618569254875183, 0.0025992360897362232, 0.11889952421188354, 0.03863006457686424, -0.013525396585464478, 0.07942727953195572, -0.09125499427318573, 0.04518340900540352, 0.02592192031443119, -0.1080249696969986, -0.2819972038269043, -0.08688072115182877, 0.07268423587083817, 0.04202193394303322, 0.0803443118929863, 0.0008178367861546576, 0.17575936019420624, -0.007282824721187353, 0.1144152283668518, 0.2596985697746277, -0.3110000193119049, -0.06116798520088196, 0.02316102758049965, 0.009494342841207981, 0.06345420330762863, -0.08541600406169891, -0.031247681006789207, 0.048284538090229034, 0.04325330629944801, 0.1794067770242691, -0.011458450928330421, -0.007149836979806423, -0.025910677388310432, -0.14341889321804047, -0.0757712796330452, 0.19806008040905, 0.04967270791530609, -0.047597505152225494, -0.06660094112157822, -0.08804943412542343, -0.1688326746225357, -0.03264717012643814, -0.01801031455397606, 0.04462956637144089, -0.03791727125644684, -0.0589226633310318, -0.028534654527902603, -0.0764043927192688, -0.042577486485242844, -0.042390674352645874, 0.15392567217350006, 0.05139201879501343, 0.032932400703430176, -0.03843959793448448, 0.08476367592811584, -0.018588144332170486, -0.15685492753982544, -0.003073766129091382, 0.00831099133938551, 0.032669827342033386, -0.018471626564860344, -0.03288761153817177, -0.09086716920137405, 0.013486159965395927, 0.12598945200443268, -0.09839624166488647, 0.06968382745981216, 0.012926696799695492, 0.051633335649967194, -0.08377796411514282, 0.18277320265769958, -0.029339900240302086, 0.013655705377459526, 0.02643161453306675, 0.08627751469612122, 0.045253686606884, -0.027602221816778183, -0.10944601148366928, 0.026082325726747513, 0.138589009642601, 0.005100409034639597, -0.03699839860200882, 0.07159482687711716, -0.046671394258737564, -0.04377603530883789, 0.055555734783411026, -0.11096695065498352, 0.013195594772696495, 0.0012843089643865824, -0.08894145488739014, -0.036723051220178604, 0.026743998751044273, -0.01389769185334444, -0.043164223432540894, 0.05528248846530914, -0.09538555145263672, 0.009111303836107254, -0.059108372777700424, -0.11533742398023605, 0.01289414893835783, -0.11949043720960617, 0.0013519255444407463, -0.109888456761837, -0.1352292001247406, -0.008098151534795761, 0.05253935232758522, -0.02088603377342224, -0.08179537206888199, -0.05587518587708473, -0.09437811374664307, 0.027021823450922966, -0.01534327119588852, 0.0546845868229866, -0.0649983137845993, 0.08562931418418884, 0.05025998502969742, 0.0732530802488327, -0.03711319342255592, 0.045747533440589905, -0.07925648242235184, 0.04529445618391037, -0.20424382388591766, 0.06939747184515, -0.06434474140405655, 0.06229047104716301, -0.10715609043836594, -0.117185078561306, 0.029869243502616882, -0.03632883355021477, 0.08210492134094238, 0.09986089169979095, -0.1464376151561737, -0.08419061452150345, 0.19336776435375214, -0.08402501791715622, -0.1270311027765274, 0.12291108816862106, -0.04557105898857117, 0.002800210379064083, 0.05785496532917023, 0.22697393596172333, 0.08953249454498291, -0.05094270780682564, -0.023491637781262398, -0.005242811515927315, 0.043609559535980225, -0.07905299216508865, 0.07979247719049454, 0.006977790035307407, 0.04260342940688133, 0.022792303934693336, -0.029363535344600677, 0.04053501412272453, -0.08387166261672974, -0.08405832201242447, -0.054650142788887024, -0.08046764135360718, 0.05724374204874039, 0.04899771884083748, 0.08289357274770737, -0.11314118653535843, -0.09415262937545776, 0.027980055660009384, 0.09097029268741608, -0.08321060240268707, 0.04656745120882988, -0.09828027337789536, 0.1251775026321411, -0.0622311495244503, -0.0028876904398202896, -0.18514208495616913, -0.0072322809137403965, 0.04435204342007637, -0.02490418776869774, -0.000853859179187566, -0.01722896471619606, 0.06357277929782867, 0.05737774819135666, -0.038729820400476456, -0.04096408560872078, -0.034338995814323425, -0.009837169200181961, -0.11381691694259644, -0.1878732591867447, -0.04747845232486725, -0.03478720784187317, 0.11789482831954956, -0.17237557470798492, 0.06186696141958237, 0.06596499681472778, 0.11241570860147476, 0.024192502722144127, -0.02798614464700222, -0.0078679658472538, 0.04189065843820572, -0.03845973685383797, -0.07504423707723618, 0.06836918741464615, 0.03562292829155922, -0.09856722503900528, -0.022597767412662506, -0.10956339538097382, 0.1883760541677475, 0.12811586260795593, -0.01528948824852705, -0.04474876448512077, -0.0030113121028989553, -0.06595233082771301, -0.025376109406352043, 0.0001992616307688877, 0.028176143765449524, 0.18498769402503967, 0.006765210535377264, 0.17534932494163513, -0.10435620695352554, -0.05398137867450714, 0.04991708695888519, -0.026386620476841927, -0.010547580197453499, 0.1123865395784378, 0.0013204298447817564, -0.11275755614042282, 0.14809755980968475, 0.12280955165624619, -0.06073061749339104, 0.12605755031108856, -0.06380902230739594, -0.043400008231401443, -0.03124178946018219, 0.006571766920387745, 0.015356606803834438, 0.09826050698757172, -0.12633495032787323, -0.019942620769143105, 0.03755846247076988, 0.025127369910478592, 0.0213486235588789, -0.1836351752281189, 0.0026766532100737095, 0.0459781289100647, -0.06142926961183548, -0.002965846098959446, -0.010835528373718262, -0.0011737890308722854, 0.09970453381538391, 0.017555667087435722, -0.08726724237203598, 0.04739336669445038, 0.009003933519124985, -0.07242876291275024, 0.20206153392791748, -0.10128787904977798, -0.19561681151390076, -0.1166614294052124, -0.04983629286289215, -0.0937039852142334, -0.0009387843892909586, 0.06328993290662766, -0.07944092154502869, -0.021747061982750893, -0.09374503046274185, -0.031456977128982544, -0.016678253188729286, 0.03649067506194115, 0.07167337089776993, -0.02152007259428501, 0.10843265056610107, -0.1268416792154312, -0.022187968716025352, -0.036466311663389206, 0.009407520294189453, 0.055414751172065735, 0.010888321325182915, 0.09761083126068115, 0.11708344519138336, -0.03449703007936478, 0.051886748522520065, -0.02834194153547287, 0.2431972175836563, -0.0549328438937664, -0.029489416629076004, 0.12720809876918793, -0.0036414244677871466, 0.08575236052274704, 0.0850614607334137, 0.04563970863819122, -0.08425978571176529, -0.012521781027317047, 0.00530668580904603, -0.037102844566106796, -0.21723859012126923, -0.0365767739713192, -0.04194740578532219, 0.021104680374264717, 0.10723866522312164, 0.04582393541932106, 0.04446491226553917, 0.05713071674108505, 0.02403874322772026, 0.05055225268006325, -0.02454719878733158, 0.09011561423540115, 0.1325213462114334, 0.04956340417265892, 0.13514669239521027, -0.041864071041345596, -0.031858645379543304, 0.03866573050618172, -0.0034419160801917315, 0.2011357694864273, -0.028623657301068306, 0.18386606872081757, 0.04475536197423935, 0.19028805196285248, 0.012321713380515575, 0.06256132572889328, -0.0251434538513422, -0.006521121598780155, -0.009183596819639206, -0.04077095165848732, -0.057392667979002, 0.0030708396807312965, -0.04522627592086792, 0.07294609397649765, -0.11834508180618286, 0.03892838582396507, 0.06002010405063629, 0.29006633162498474, 0.022332480177283287, -0.37384721636772156, -0.1128094345331192, -0.01751832291483879, -0.03198004513978958, -0.04660427197813988, 0.010848252102732658, 0.11287738382816315, -0.09319595247507095, 0.06235712766647339, -0.08436550945043564, 0.09173392504453659, -0.07310128211975098, 0.036443956196308136, 0.0506972000002861, 0.09780693054199219, 0.006120195612311363, 0.06156545877456665, -0.2656126320362091, 0.25246939063072205, 0.01898382045328617, 0.05165523290634155, -0.06228072568774223, 0.014799673110246658, 0.020547624677419662, 0.07127467542886734, 0.07885634899139404, 0.0006794908549636602, -0.028727702796459198, -0.16280202567577362, -0.11423774808645248, 0.01696728728711605, 0.07045066356658936, -0.02243482507765293, 0.08354400843381882, -0.0007569814915768802, 0.0032902355305850506, 0.04280740022659302, -0.010342855006456375, -0.027652321383357048, -0.09026366472244263, 0.016356363892555237, 0.06345270574092865, -0.041483666747808456, -0.08211879432201385, -0.11932772397994995, -0.08241575956344604, 0.1833699345588684, -0.01318553276360035, -0.08128651231527328, -0.12019886076450348, 0.06338709592819214, 0.06619332730770111, -0.09245581924915314, 0.042988114058971405, -0.017607081681489944, 0.1234922781586647, 0.0032546466682106256, -0.07252337038516998, 0.09986564517021179, -0.04792711138725281, -0.16395272314548492, -0.03170733153820038, 0.13534678518772125, 0.03350100293755531, 0.057126160711050034, -0.013107032515108585, 0.019576072692871094, -0.02810630574822426, -0.07615172117948532, 0.033417459577322006, 0.0019211877370253205, 0.09283176064491272, -0.016100896522402763, 0.0036490904167294502, 0.02513909339904785, -0.07741467654705048, 0.005653811618685722, 0.1960393786430359, 0.2601768672466278, -0.1054224744439125, 0.03657129034399986, 0.028584623709321022, -0.04292646795511246, -0.15217222273349762, 0.018080569803714752, 0.07971378415822983, 0.007898523472249508, -0.012302767485380173, -0.17773962020874023, 0.050347067415714264, 0.08671995997428894, -0.018084589391946793, 0.08345312625169754, -0.29868802428245544, -0.11544010043144226, 0.11149483919143677, 0.12423747777938843, 0.09513396769762039, -0.1379803866147995, -0.04662996158003807, -0.006842230912297964, -0.12511427700519562, 0.11477506905794144, -0.09774457663297653, 0.10926564782857895, -0.04369644820690155, 0.05271214619278908, 0.0077659860253334045, -0.0524640791118145, 0.12633778154850006, 0.01897483877837658, 0.07942014187574387, -0.04729471355676651, 0.0028797932900488377, 0.10791897773742676, -0.07520560175180435, 0.06073889136314392, -0.10178165137767792, 0.04971356689929962, -0.12877590954303741, -0.010309635661542416, -0.082453154027462, 0.0295798871666193, -0.029397226870059967, -0.03437960892915726, -0.05814070999622345, 0.008260298520326614, 0.07413773238658905, -0.0016154507175087929, 0.17978344857692719, 0.04773912951350212, 0.12875743210315704, 0.20528703927993774, 0.07694612443447113, -0.11820431053638458, -0.10233582556247711, -0.0013466635718941689, -0.00923001766204834, 0.05888792872428894, -0.16294585168361664, 0.03940162807703018, 0.1446576565504074, 0.005404313560575247, 0.11615433543920517, 0.06740670651197433, -0.05869709327816963, -0.00048802205128595233, 0.0418846420943737, -0.18120436370372772, -0.1063351184129715, -0.014679894782602787, -0.012210211716592312, -0.13097167015075684, 0.08067499846220016, 0.1108740046620369, -0.07254352420568466, -0.018809350207448006, 0.000591395131777972, -0.0013111312873661518, -0.02480907179415226, 0.1805391013622284, 0.07104653865098953, 0.07086271047592163, -0.10655967146158218, 0.09469111263751984, 0.04809468612074852, -0.07629316300153732, 0.040003977715969086, 0.06678993254899979, -0.11828560382127762, -0.02318212017416954, 0.03783298283815384, 0.14980897307395935, -0.025542154908180237, -0.05531641095876694, -0.1733895242214203, -0.1079295352101326, 0.08900009095668793, 0.1341186910867691, 0.10272374749183655, 0.016925804316997528, -0.04834591969847679, -0.009077857248485088, -0.11072597652673721, 0.09780983626842499, 0.05804075300693512, 0.06632594019174576, -0.16388888657093048, 0.12005561590194702, 0.00006893343379488215, 0.07082127034664154, -0.01233591791242361, 0.0014726483495905995, -0.08466386049985886, 0.004197877366095781, -0.0903686061501503, 0.012840780429542065, -0.03902212157845497, 0.00046823700540699065, -0.02279692143201828, -0.04429886117577553, -0.05395001173019409, 0.03722048178315163, -0.10660385340452194, -0.035357195883989334, 0.03133080527186394, 0.02680857852101326, -0.11128820478916168, -0.031047474592924118, 0.002136401366442442, -0.08466892689466476, 0.0947682186961174, 0.05381849408149719, -0.007051385007798672, 0.010299705900251865, -0.0031958429608494043, 0.003333411179482937, 0.07069303095340729, 0.00807509757578373, 0.062125466763973236, -0.11772271990776062, -0.0033852620981633663, 0.003918701782822609, 0.0015556140569970012, 0.018365269526839256, 0.12068989127874374, -0.1207766979932785, -0.014938155189156532, -0.021776234731078148, -0.022403638809919357, -0.07529860734939575, 0.05935674160718918, 0.10323891043663025, 0.04002806171774864, 0.1898861676454544, -0.06376539915800095, 0.014657828956842422, -0.20644508302211761, -0.0007039525080472231, 0.008912970311939716, -0.15322327613830566, -0.05396716296672821, -0.026020385324954987, 0.05872125178575516, -0.07323415577411652, 0.11428996920585632, 0.005394491832703352, -0.014936160296201706, 0.03706309571862221, -0.04498806968331337, -0.019579971209168434, 0.009365580976009369, 0.16106203198432922, 0.014660797081887722, -0.037143524736166, 0.12152601033449173, 0.0238568764179945, 0.10519958287477493, 0.12467579543590546, 0.1681404858827591, 0.12871408462524414, 0.027525760233402252, 0.10744131356477737, 0.03219867870211601, -0.03000696748495102, -0.1889362335205078, 0.0689808651804924, -0.046327803283929825, 0.1342916637659073, 0.005838233977556229, 0.1764954775571823, 0.11104877293109894, -0.15142668783664703, 0.05000590160489082, -0.02190127596259117, -0.09401895105838776, -0.10003934055566788, -0.09683806449174881, -0.07801129668951035, -0.1528715342283249, -0.012698844075202942, -0.11861390620470047, -0.00156453310046345, 0.0643860250711441, 0.004968239460140467, -0.022064432501792908, 0.13960690796375275, 0.025364646688103676, -0.0021634565200656652, 0.08261991292238235, -0.003702986752614379, -0.05754280835390091, -0.0716606080532074, -0.07397709786891937, 0.014353367500007153, 0.01204715110361576, 0.05748964473605156, -0.02806028723716736, 0.008202285505831242, 0.05155220627784729, -0.025701100006699562, -0.10751278698444366, 0.009841586463153362, 0.021482164040207863, 0.05424986407160759, 0.020782899111509323, 0.013040878809988499, -0.012086511589586735, -0.017886120826005936, 0.1773153394460678, -0.06466637551784515, -0.021009275689721107, -0.10658837109804153, 0.19975638389587402, 0.03737352043390274, -0.03866131603717804, 0.04084712266921997, -0.06903629750013351, -0.019598117098212242, 0.1893388032913208, 0.20313337445259094, -0.02429080381989479, 0.0008256763103418052, -0.0044187759049236774, -0.01814758963882923, -0.007503405679017305, 0.0820104330778122, 0.1334698349237442, -0.0071779401041567326, -0.06694548577070236, -0.025810882449150085, -0.06658927351236343, -0.004836720880120993, -0.0482044480741024, 0.066391222178936, 0.027743671089410782, 0.010540149174630642, -0.046184465289115906, 0.022228645160794258, -0.04199829697608948, -0.07882942259311676, 0.03226859122514725, -0.20561571419239044, -0.16009460389614105, -0.009006503038108349, 0.04061219468712807, 0.001191785093396902, 0.06098309904336929, -0.0008008016739040613, 0.011895515024662018, 0.09127578884363174, -0.022390378639101982, -0.08587329834699631, -0.08225788921117783, 0.08901824802160263, -0.1791907101869583, 0.20023761689662933, -0.0428975448012352, 0.048247937113046646, 0.13208943605422974, 0.04311031848192215, -0.10308555513620377, 0.03702815622091293, 0.034668512642383575, -0.009859585203230381, -0.0046586100943386555, 0.10009314119815826, -0.011757995933294296, 0.07440435141324997, 0.03818235918879509, -0.09599507600069046, -0.039281949400901794, -0.0724983662366867, -0.02112523838877678, -0.04013347998261452, -0.051686834543943405, -0.05065472051501274, 0.10975289344787598, 0.189988374710083, -0.04916232079267502, -0.017519403249025345, -0.06736118346452713, 0.00519917719066143, 0.07361429184675217, -0.009850573725998402, -0.04280546307563782, -0.25230857729911804, 0.011408863589167595, 0.047809310257434845, -0.016395779326558113, -0.22591710090637207, -0.10197950899600983, -0.005344944540411234, -0.06733774393796921, -0.07399795949459076, 0.07431349903345108, 0.04495703801512718, 0.05470620095729828, -0.0674663856625557, 0.023926597088575363, -0.09551819413900375, 0.15745952725410461, -0.15249231457710266, -0.07582691311836243 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # fnet-base-finetuned-qnli This model is a fine-tuned version of [google/fnet-base](https://huggingface.co/google/fnet-base) on the GLUE QNLI dataset. It achieves the following results on the evaluation set: - Loss: 0.4746 - Accuracy: 0.8439 The model was fine-tuned to compare [google/fnet-base](https://huggingface.co/google/fnet-base) as introduced in [this paper](https://arxiv.org/abs/2105.03824) against [bert-base-cased](https://huggingface.co/bert-base-cased). ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure This model is trained using the [run_glue](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) script. The following command was used: ```bash #!/usr/bin/bash python ../run_glue.py \\n --model_name_or_path google/fnet-base \\n --task_name qnli \\n --do_train \\n --do_eval \\n --max_seq_length 512 \\n --per_device_train_batch_size 16 \\n --learning_rate 2e-5 \\n --num_train_epochs 3 \\n --output_dir fnet-base-finetuned-qnli \\n --push_to_hub \\n --hub_strategy all_checkpoints \\n --logging_strategy epoch \\n --save_strategy epoch \\n --evaluation_strategy epoch \\n``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:--------:| | 0.4597 | 1.0 | 6547 | 0.3713 | 0.8411 | | 0.3252 | 2.0 | 13094 | 0.3781 | 0.8420 | | 0.2243 | 3.0 | 19641 | 0.4746 | 0.8439 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer", "fnet-bert-base-comparison"], "datasets": ["glue"], "metrics": ["accuracy"], "model-index": [{"name": "fnet-base-finetuned-qnli", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE QNLI", "type": "glue", "args": "qnli"}, "metrics": [{"type": "accuracy", "value": 0.8438586857038257, "name": "Accuracy"}]}]}]}
text-classification
gchhablani/fnet-base-finetuned-qnli
[ "transformers", "pytorch", "tensorboard", "fnet", "text-classification", "generated_from_trainer", "fnet-bert-base-comparison", "en", "dataset:glue", "arxiv:2105.03824", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2105.03824" ]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #fnet #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
fnet-base-finetuned-qnli ======================== This model is a fine-tuned version of google/fnet-base on the GLUE QNLI dataset. It achieves the following results on the evaluation set: * Loss: 0.4746 * Accuracy: 0.8439 The model was fine-tuned to compare google/fnet-base as introduced in this paper against bert-base-cased. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ This model is trained using the run\_glue script. The following command was used: ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #fnet #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 88, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #fnet #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.1281575709581375, 0.15705668926239014, -0.0038979940582066774, 0.11255989223718643, 0.11293601244688034, -0.008617952466011047, 0.13138744235038757, 0.1563730537891388, -0.1094919964671135, 0.05590956658124924, 0.1562153697013855, 0.15063650906085968, 0.03187095746397972, 0.18714870512485504, -0.05448957160115242, -0.23206175863742828, 0.030974131077528, 0.07522018253803253, -0.05673399940133095, 0.13729923963546753, 0.09973176568746567, -0.11717205494642258, 0.09386496245861053, 0.041752949357032776, -0.20222944021224976, -0.015307755209505558, -0.0008101215353235602, -0.08269140124320984, 0.11289821565151215, 0.02265351265668869, 0.08919014781713486, 0.03242024779319763, 0.03147246316075325, -0.1466279774904251, 0.007479942869395018, 0.055549394339323044, 0.0026599117554724216, 0.11788346618413925, 0.03937623277306557, -0.014160647056996822, 0.07863719761371613, -0.09087078273296356, 0.046092916280031204, 0.026117689907550812, -0.10794810950756073, -0.2786095142364502, -0.08820846676826477, 0.07192650437355042, 0.042675234377384186, 0.0806850865483284, 0.001147752278484404, 0.17639178037643433, -0.007389090955257416, 0.11422055959701538, 0.2610849142074585, -0.3091604709625244, -0.061695314943790436, 0.024227019399404526, 0.009995708242058754, 0.06274503469467163, -0.08376740664243698, -0.031031707301735878, 0.04775812104344368, 0.04504694044589996, 0.17843511700630188, -0.011732942424714565, -0.007466572802513838, -0.025801893323659897, -0.14371579885482788, -0.07526436448097229, 0.19688530266284943, 0.04901329055428505, -0.048377830535173416, -0.06737475097179413, -0.08691056817770004, -0.16888527572155, -0.03213018923997879, -0.018070582300424576, 0.04411018267273903, -0.0378425307571888, -0.060805536806583405, -0.029689611867070198, -0.07623965293169022, -0.042504265904426575, -0.042007382959127426, 0.15460588037967682, 0.05135969817638397, 0.03275564685463905, -0.03648478910326958, 0.08528219908475876, -0.016176627948880196, -0.15604232251644135, -0.0033359206281602383, 0.008210553787648678, 0.032338954508304596, -0.01879490725696087, -0.03313892334699631, -0.09054049104452133, 0.012796161696314812, 0.12658193707466125, -0.09896866977214813, 0.0699896514415741, 0.014115668833255768, 0.05199545621871948, -0.0832819789648056, 0.18305599689483643, -0.03047999180853367, 0.014890378341078758, 0.025784319266676903, 0.0863259807229042, 0.045972537249326706, -0.02756713703274727, -0.10970760136842728, 0.026695188134908676, 0.13855181634426117, 0.006172229070216417, -0.03736800700426102, 0.07175908982753754, -0.04684898629784584, -0.04453456029295921, 0.056356143206357956, -0.11108183860778809, 0.012465346604585648, 0.0007881103083491325, -0.08858437836170197, -0.03560861945152283, 0.026957271620631218, -0.012466381303966045, -0.04220158979296684, 0.05419426038861275, -0.0943819209933281, 0.009732727892696857, -0.05831371992826462, -0.1146949976682663, 0.013262145221233368, -0.11782565712928772, 0.0005620947922579944, -0.11043663322925568, -0.13709686696529388, -0.008001538924872875, 0.05364624038338661, -0.021408166736364365, -0.0810130313038826, -0.055306028574705124, -0.09359939396381378, 0.02663268707692623, -0.015616829507052898, 0.05365283414721489, -0.06507189571857452, 0.08449108898639679, 0.049848102033138275, 0.07236256450414658, -0.03752796724438667, 0.04536004737019539, -0.07848503440618515, 0.0464148111641407, -0.20472228527069092, 0.07016371190547943, -0.064207062125206, 0.06198544800281525, -0.1068328395485878, -0.11787286400794983, 0.030897412449121475, -0.03552781045436859, 0.08240031450986862, 0.09914099425077438, -0.14732807874679565, -0.08291295915842056, 0.19005797803401947, -0.08365947753190994, -0.12725262343883514, 0.12337704747915268, -0.04553937166929245, 0.0029195500537753105, 0.05737419053912163, 0.2257438600063324, 0.0889558345079422, -0.0515265017747879, -0.02331758849322796, -0.0051168533973395824, 0.04387769103050232, -0.07846325635910034, 0.08052174746990204, 0.006332707591354847, 0.04367716237902641, 0.02254851721227169, -0.029947886243462563, 0.04004385322332382, -0.08367564529180527, -0.08474674820899963, -0.05332755297422409, -0.08056257665157318, 0.056404486298561096, 0.04767722263932228, 0.08231600373983383, -0.11374399065971375, -0.094481460750103, 0.02864118292927742, 0.09102724492549896, -0.08461901545524597, 0.04763771593570709, -0.09820594638586044, 0.12363352626562119, -0.06146226450800896, -0.004213334526866674, -0.18391123414039612, -0.009381342679262161, 0.0438222773373127, -0.023048149421811104, -0.0010736447293311357, -0.017531832680106163, 0.0634535625576973, 0.05774131044745445, -0.039399147033691406, -0.0408325120806694, -0.033703941851854324, -0.009450753219425678, -0.11244195699691772, -0.18618963658809662, -0.04849305376410484, -0.034880541265010834, 0.11852548271417618, -0.17111770808696747, 0.06201837584376335, 0.06472091376781464, 0.11162162572145462, 0.02296465076506138, -0.02818126603960991, -0.0076725417748093605, 0.041496098041534424, -0.039293959736824036, -0.0763016790151596, 0.06855007261037827, 0.03613912686705589, -0.0994340106844902, -0.021636441349983215, -0.11005594581365585, 0.18589626252651215, 0.12711931765079498, -0.015304340980947018, -0.046145275235176086, -0.0037261333782225847, -0.06693658232688904, -0.0252322256565094, 0.00010479353659320623, 0.02630312740802765, 0.18688923120498657, 0.007668614853173494, 0.17470361292362213, -0.10448402166366577, -0.054649028927087784, 0.04829766973853111, -0.027804838493466377, -0.012250205501914024, 0.11106427013874054, 0.000900426646694541, -0.11244355887174606, 0.14791768789291382, 0.12212676554918289, -0.06123341992497444, 0.1253422647714615, -0.06269760429859161, -0.042313165962696075, -0.03154568001627922, 0.0070209321565926075, 0.016009759157896042, 0.09681913256645203, -0.12338456511497498, -0.019060473889112473, 0.03688754513859749, 0.026065392419695854, 0.022046322003006935, -0.18368276953697205, 0.0016608438454568386, 0.046017080545425415, -0.06232096627354622, -0.0001330666127614677, -0.01134573481976986, -0.0014391003642231226, 0.09911268204450607, 0.018035314977169037, -0.08723576366901398, 0.04871227592229843, 0.008767826482653618, -0.07373929023742676, 0.20202793180942535, -0.10205624997615814, -0.19618737697601318, -0.11792904138565063, -0.0520332008600235, -0.09464605152606964, -0.00033909876947291195, 0.06213480234146118, -0.07916445285081863, -0.022073807194828987, -0.09520765393972397, -0.03123565763235092, -0.01737889274954796, 0.03542753681540489, 0.07313533127307892, -0.02203269489109516, 0.10872673243284225, -0.1269455999135971, -0.02306712418794632, -0.0358131118118763, 0.008673591539263725, 0.055851276963949203, 0.009536282159388065, 0.09724908322095871, 0.1170153021812439, -0.033818624913692474, 0.05197916179895401, -0.028282886371016502, 0.24284608662128448, -0.053264424204826355, -0.02976847253739834, 0.12733469903469086, -0.0040460084564983845, 0.08623170107603073, 0.08496475964784622, 0.04557865113019943, -0.08525460958480835, -0.01157840620726347, 0.00595470005646348, -0.03817666694521904, -0.21714363992214203, -0.036886055022478104, -0.04140350595116615, 0.02079310268163681, 0.10649195313453674, 0.04592718183994293, 0.043558333069086075, 0.057499635964632034, 0.023834481835365295, 0.04960625246167183, -0.02408022992312908, 0.09037080407142639, 0.1317465603351593, 0.04990197345614433, 0.13514135777950287, -0.04064502939581871, -0.03152943775057793, 0.039630502462387085, -0.002475862158462405, 0.19950714707374573, -0.02927609719336033, 0.1829756200313568, 0.04599745571613312, 0.1883370578289032, 0.011177031323313713, 0.06271716207265854, -0.024891022592782974, -0.006736312061548233, -0.009674341417849064, -0.0412435419857502, -0.05800659582018852, 0.0024435357190668583, -0.04474163055419922, 0.07293359190225601, -0.117747962474823, 0.040190309286117554, 0.06010587513446808, 0.29027169942855835, 0.022407572716474533, -0.37493225932121277, -0.11181744188070297, -0.016597270965576172, -0.031712714582681656, -0.04552937671542168, 0.010863134637475014, 0.11205030232667923, -0.09440278261899948, 0.06324154883623123, -0.08447496592998505, 0.09077057987451553, -0.07360748946666718, 0.03601958230137825, 0.05221237242221832, 0.09796657413244247, 0.006393694784492254, 0.06149907410144806, -0.26434561610221863, 0.25069257616996765, 0.018371766433119774, 0.05161698907613754, -0.06146470457315445, 0.015569853596389294, 0.02125987596809864, 0.06923633813858032, 0.07857105880975723, 0.0007500528590753675, -0.02767954021692276, -0.16410666704177856, -0.11515739560127258, 0.016607077792286873, 0.07020550966262817, -0.02121029421687126, 0.08431950211524963, -0.0010764591861516237, 0.004067974630743265, 0.04312531650066376, -0.008349082432687283, -0.027957666665315628, -0.09104249626398087, 0.016294151544570923, 0.06238500773906708, -0.04210955649614334, -0.08233670145273209, -0.1198863536119461, -0.08412427455186844, 0.1865774542093277, -0.008332050405442715, -0.0823029950261116, -0.12076611816883087, 0.06184208020567894, 0.06703545898199081, -0.09185026586055756, 0.042081426829099655, -0.017016833648085594, 0.12382151186466217, 0.003344601020216942, -0.07244402170181274, 0.09798984974622726, -0.04916444793343544, -0.16366566717624664, -0.03181470185518265, 0.13521257042884827, 0.034096188843250275, 0.05754973739385605, -0.012549868784844875, 0.019313739612698555, -0.026896705850958824, -0.0765208750963211, 0.03267578408122063, 0.002446891972795129, 0.09423114359378815, -0.015256565995514393, 0.0032036362681537867, 0.028266185894608498, -0.07761038839817047, 0.005427155178040266, 0.19771218299865723, 0.2606421709060669, -0.1043548434972763, 0.03651893138885498, 0.027820715680718422, -0.04207947105169296, -0.1520952433347702, 0.017168138176202774, 0.08070502430200577, 0.007420012727379799, -0.010656541213393211, -0.1764611452817917, 0.05012203007936478, 0.08809427171945572, -0.01788947917521, 0.08023256808519363, -0.29865652322769165, -0.11554376780986786, 0.11132355779409409, 0.12564530968666077, 0.09193845093250275, -0.13787555694580078, -0.04598912596702576, -0.0075193652883172035, -0.12548284232616425, 0.1158299669623375, -0.0956348180770874, 0.10974547266960144, -0.04320225492119789, 0.05354843661189079, 0.007337958551943302, -0.052445266395807266, 0.12752960622310638, 0.01908416859805584, 0.0783768892288208, -0.0478614941239357, 0.003130370983853936, 0.10798399150371552, -0.0758182629942894, 0.06092115119099617, -0.10174112766981125, 0.04936999827623367, -0.12913434207439423, -0.01087163481861353, -0.08215732127428055, 0.02880539558827877, -0.029542503878474236, -0.03560418263077736, -0.058059368282556534, 0.009000681340694427, 0.07482077181339264, -0.0021140656899660826, 0.17803633213043213, 0.047338198870420456, 0.12804824113845825, 0.2045678347349167, 0.07827051728963852, -0.11816922575235367, -0.10215629637241364, -0.0021673408336937428, -0.009365938603878021, 0.0566975399851799, -0.15991105139255524, 0.03947640210390091, 0.1441853940486908, 0.004068439826369286, 0.11697175353765488, 0.06757407635450363, -0.058885324746370316, -0.00002974093695229385, 0.04177243635058403, -0.17993171513080597, -0.10290152579545975, -0.014580353163182735, -0.012066712602972984, -0.13221070170402527, 0.07946693897247314, 0.11126816272735596, -0.07109289616346359, -0.018744582310318947, 0.0011641265591606498, -0.0008233633125200868, -0.025426305830478668, 0.18059362471103668, 0.0713382288813591, 0.07033462822437286, -0.10616692155599594, 0.09460733085870743, 0.0462157279253006, -0.0725565180182457, 0.04061013460159302, 0.06810374557971954, -0.11764568090438843, -0.022825488820672035, 0.037216611206531525, 0.14856918156147003, -0.027179433032870293, -0.05499974638223648, -0.17361967265605927, -0.1086450144648552, 0.08841122686862946, 0.13100364804267883, 0.10317601263523102, 0.017121080309152603, -0.04885329306125641, -0.009807179681956768, -0.11012768745422363, 0.09655274450778961, 0.05846559628844261, 0.0660039484500885, -0.16399246454238892, 0.11977937817573547, -0.0003578494652174413, 0.07169348001480103, -0.012192845344543457, 0.001448655384592712, -0.0843949168920517, 0.004430484492331743, -0.09232339262962341, 0.012558856047689915, -0.04001414775848389, 0.00048618309665471315, -0.022317882627248764, -0.04360825940966606, -0.0546182319521904, 0.03691510483622551, -0.10589302331209183, -0.03589541092514992, 0.030760591849684715, 0.02635170891880989, -0.11072052270174026, -0.031555552035570145, 0.0018842692952603102, -0.08543401211500168, 0.09491559118032455, 0.053661517798900604, -0.007116627413779497, 0.009966611862182617, -0.003835842479020357, 0.002103820675984025, 0.0713348463177681, 0.007812848314642906, 0.06204405426979065, -0.11863148212432861, -0.004516935441643, 0.004596107639372349, 0.001546198152936995, 0.019018597900867462, 0.12120135128498077, -0.12007837742567062, -0.014022485353052616, -0.022460926324129105, -0.022654950618743896, -0.0753025934100151, 0.05987093597650528, 0.10378355532884598, 0.041243866086006165, 0.19044536352157593, -0.06453124433755875, 0.015417633578181267, -0.2059982270002365, -0.0010957218473777175, 0.009144444018602371, -0.1515446901321411, -0.05387648567557335, -0.025656811892986298, 0.05819327011704445, -0.07224544882774353, 0.11503966152667999, 0.004568712320178747, -0.01583913527429104, 0.03687077388167381, -0.044159963726997375, -0.016699714586138725, 0.008686945773661137, 0.16097450256347656, 0.013965492136776447, -0.037191856652498245, 0.12147503346204758, 0.023390216752886772, 0.1060599610209465, 0.12157408893108368, 0.1685730367898941, 0.12814003229141235, 0.026453150436282158, 0.10737672448158264, 0.032515060156583786, -0.030444564297795296, -0.18784967064857483, 0.06720752269029617, -0.04578623175621033, 0.13226191699504852, 0.006823073141276836, 0.17648069560527802, 0.11036133766174316, -0.15301413834095, 0.049124132841825485, -0.021638398990035057, -0.09448863565921783, -0.10050928592681885, -0.09679119288921356, -0.07743744552135468, -0.15149177610874176, -0.01236912701278925, -0.11908751726150513, -0.001917796558700502, 0.06672867387533188, 0.0058901347219944, -0.021161947399377823, 0.13932976126670837, 0.023673847317695618, -0.0022931741550564766, 0.0803397074341774, -0.0035189997870475054, -0.05596965178847313, -0.07125981897115707, -0.07409729808568954, 0.013776594772934914, 0.013752772472798824, 0.05871286615729332, -0.02784240059554577, 0.009659827686846256, 0.051329080015420914, -0.02629757858812809, -0.10809190571308136, 0.009442382492125034, 0.021296720951795578, 0.054612502455711365, 0.020937973633408546, 0.013347321189939976, -0.011912370100617409, -0.017500825226306915, 0.17539910972118378, -0.06457335501909256, -0.022280963137745857, -0.10723160207271576, 0.20094625651836395, 0.036107197403907776, -0.03869583457708359, 0.040720485150814056, -0.06905907392501831, -0.020572103559970856, 0.18964605033397675, 0.20552918314933777, -0.026153136044740677, 0.0006357349921017885, -0.004643670748919249, -0.01797187142074108, -0.007700672838836908, 0.08245248347520828, 0.13317352533340454, -0.0074555822648108006, -0.06725732237100601, -0.026400936767458916, -0.06664355099201202, -0.004993689712136984, -0.048690151423215866, 0.06622577458620071, 0.027980800718069077, 0.012009805999696255, -0.046264056116342545, 0.022495536133646965, -0.042431484907865524, -0.07840681821107864, 0.030966809019446373, -0.20489253103733063, -0.15987305343151093, -0.008346200920641422, 0.03857562318444252, 0.0004496309265960008, 0.062375955283641815, -0.0009975176071748137, 0.012196631170809269, 0.0908537283539772, -0.02237413451075554, -0.08646247535943985, -0.07932375371456146, 0.09022057801485062, -0.17671364545822144, 0.2005920559167862, -0.04314102232456207, 0.049184780567884445, 0.1326240599155426, 0.04266604781150818, -0.10272050648927689, 0.0379687063395977, 0.03511279076337814, -0.008898447267711163, -0.004219701513648033, 0.09855833649635315, -0.012647976167500019, 0.07372915744781494, 0.03857356309890747, -0.09686969965696335, -0.039119016379117966, -0.07064490765333176, -0.0207537654787302, -0.040169861167669296, -0.052368659526109695, -0.05137869715690613, 0.11046379059553146, 0.19000554084777832, -0.049329470843076706, -0.017489463090896606, -0.06654652208089828, 0.005614506546407938, 0.07303186506032944, -0.011176830157637596, -0.04266827553510666, -0.253805011510849, 0.010820526629686356, 0.046582285314798355, -0.017032980918884277, -0.2275182157754898, -0.1029951274394989, -0.004221665672957897, -0.06776802986860275, -0.07437815517187119, 0.07514068484306335, 0.04294868931174278, 0.05498082563281059, -0.06696301698684692, 0.028071772307157516, -0.09679107367992401, 0.15699414908885956, -0.15248030424118042, -0.07556789368391037 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # fnet-base-finetuned-qqp This model is a fine-tuned version of [google/fnet-base](https://huggingface.co/google/fnet-base) on the GLUE QQP dataset. It achieves the following results on the evaluation set: - Loss: 0.3686 - Accuracy: 0.8847 - F1: 0.8466 - Combined Score: 0.8657 The model was fine-tuned to compare [google/fnet-base](https://huggingface.co/google/fnet-base) as introduced in [this paper](https://arxiv.org/abs/2105.03824) against [bert-base-cased](https://huggingface.co/bert-base-cased). ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure This model is trained using the [run_glue](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) script. The following command was used: ```bash #!/usr/bin/bash #!/usr/bin/bash python ../run_glue.py \\n --model_name_or_path google/fnet-base \\n --task_name qqp \\n --do_train \\n --do_eval \\n --max_seq_length 512 \\n --per_device_train_batch_size 16 \\n --learning_rate 2e-5 \\n --num_train_epochs 3 \\n --output_dir fnet-base-finetuned-qqp \\n --push_to_hub \\n --hub_strategy all_checkpoints \\n --logging_strategy epoch \\n --save_strategy epoch \\n --evaluation_strategy epoch \\n``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Combined Score | |:-------------:|:-----:|:-----:|:---------------:|:--------:|:------:|:--------------:| | 0.3484 | 1.0 | 22741 | 0.3014 | 0.8676 | 0.8297 | 0.8487 | | 0.2387 | 2.0 | 45482 | 0.3011 | 0.8801 | 0.8429 | 0.8615 | | 0.1739 | 3.0 | 68223 | 0.3686 | 0.8847 | 0.8466 | 0.8657 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
{"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer", "fnet-bert-base-comparison"], "datasets": ["glue"], "metrics": ["accuracy", "f1"], "model-index": [{"name": "fnet-base-finetuned-qqp", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "GLUE QQP", "type": "glue", "args": "qqp"}, "metrics": [{"type": "accuracy", "value": 0.8847390551570616, "name": "Accuracy"}, {"type": "f1", "value": 0.8466197090382463, "name": "F1"}]}]}]}
text-classification
gchhablani/fnet-base-finetuned-qqp
[ "transformers", "pytorch", "tensorboard", "fnet", "text-classification", "generated_from_trainer", "fnet-bert-base-comparison", "en", "dataset:glue", "arxiv:2105.03824", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2105.03824" ]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #fnet #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
fnet-base-finetuned-qqp ======================= This model is a fine-tuned version of google/fnet-base on the GLUE QQP dataset. It achieves the following results on the evaluation set: * Loss: 0.3686 * Accuracy: 0.8847 * F1: 0.8466 * Combined Score: 0.8657 The model was fine-tuned to compare google/fnet-base as introduced in this paper against bert-base-cased. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ This model is trained using the run\_glue script. The following command was used: ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.11.0.dev0 * Pytorch 1.9.0 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #fnet #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 88, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #fnet #text-classification #generated_from_trainer #fnet-bert-base-comparison #en #dataset-glue #arxiv-2105.03824 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.11.0.dev0\n* Pytorch 1.9.0\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.1281575709581375, 0.15705668926239014, -0.0038979940582066774, 0.11255989223718643, 0.11293601244688034, -0.008617952466011047, 0.13138744235038757, 0.1563730537891388, -0.1094919964671135, 0.05590956658124924, 0.1562153697013855, 0.15063650906085968, 0.03187095746397972, 0.18714870512485504, -0.05448957160115242, -0.23206175863742828, 0.030974131077528, 0.07522018253803253, -0.05673399940133095, 0.13729923963546753, 0.09973176568746567, -0.11717205494642258, 0.09386496245861053, 0.041752949357032776, -0.20222944021224976, -0.015307755209505558, -0.0008101215353235602, -0.08269140124320984, 0.11289821565151215, 0.02265351265668869, 0.08919014781713486, 0.03242024779319763, 0.03147246316075325, -0.1466279774904251, 0.007479942869395018, 0.055549394339323044, 0.0026599117554724216, 0.11788346618413925, 0.03937623277306557, -0.014160647056996822, 0.07863719761371613, -0.09087078273296356, 0.046092916280031204, 0.026117689907550812, -0.10794810950756073, -0.2786095142364502, -0.08820846676826477, 0.07192650437355042, 0.042675234377384186, 0.0806850865483284, 0.001147752278484404, 0.17639178037643433, -0.007389090955257416, 0.11422055959701538, 0.2610849142074585, -0.3091604709625244, -0.061695314943790436, 0.024227019399404526, 0.009995708242058754, 0.06274503469467163, -0.08376740664243698, -0.031031707301735878, 0.04775812104344368, 0.04504694044589996, 0.17843511700630188, -0.011732942424714565, -0.007466572802513838, -0.025801893323659897, -0.14371579885482788, -0.07526436448097229, 0.19688530266284943, 0.04901329055428505, -0.048377830535173416, -0.06737475097179413, -0.08691056817770004, -0.16888527572155, -0.03213018923997879, -0.018070582300424576, 0.04411018267273903, -0.0378425307571888, -0.060805536806583405, -0.029689611867070198, -0.07623965293169022, -0.042504265904426575, -0.042007382959127426, 0.15460588037967682, 0.05135969817638397, 0.03275564685463905, -0.03648478910326958, 0.08528219908475876, -0.016176627948880196, -0.15604232251644135, -0.0033359206281602383, 0.008210553787648678, 0.032338954508304596, -0.01879490725696087, -0.03313892334699631, -0.09054049104452133, 0.012796161696314812, 0.12658193707466125, -0.09896866977214813, 0.0699896514415741, 0.014115668833255768, 0.05199545621871948, -0.0832819789648056, 0.18305599689483643, -0.03047999180853367, 0.014890378341078758, 0.025784319266676903, 0.0863259807229042, 0.045972537249326706, -0.02756713703274727, -0.10970760136842728, 0.026695188134908676, 0.13855181634426117, 0.006172229070216417, -0.03736800700426102, 0.07175908982753754, -0.04684898629784584, -0.04453456029295921, 0.056356143206357956, -0.11108183860778809, 0.012465346604585648, 0.0007881103083491325, -0.08858437836170197, -0.03560861945152283, 0.026957271620631218, -0.012466381303966045, -0.04220158979296684, 0.05419426038861275, -0.0943819209933281, 0.009732727892696857, -0.05831371992826462, -0.1146949976682663, 0.013262145221233368, -0.11782565712928772, 0.0005620947922579944, -0.11043663322925568, -0.13709686696529388, -0.008001538924872875, 0.05364624038338661, -0.021408166736364365, -0.0810130313038826, -0.055306028574705124, -0.09359939396381378, 0.02663268707692623, -0.015616829507052898, 0.05365283414721489, -0.06507189571857452, 0.08449108898639679, 0.049848102033138275, 0.07236256450414658, -0.03752796724438667, 0.04536004737019539, -0.07848503440618515, 0.0464148111641407, -0.20472228527069092, 0.07016371190547943, -0.064207062125206, 0.06198544800281525, -0.1068328395485878, -0.11787286400794983, 0.030897412449121475, -0.03552781045436859, 0.08240031450986862, 0.09914099425077438, -0.14732807874679565, -0.08291295915842056, 0.19005797803401947, -0.08365947753190994, -0.12725262343883514, 0.12337704747915268, -0.04553937166929245, 0.0029195500537753105, 0.05737419053912163, 0.2257438600063324, 0.0889558345079422, -0.0515265017747879, -0.02331758849322796, -0.0051168533973395824, 0.04387769103050232, -0.07846325635910034, 0.08052174746990204, 0.006332707591354847, 0.04367716237902641, 0.02254851721227169, -0.029947886243462563, 0.04004385322332382, -0.08367564529180527, -0.08474674820899963, -0.05332755297422409, -0.08056257665157318, 0.056404486298561096, 0.04767722263932228, 0.08231600373983383, -0.11374399065971375, -0.094481460750103, 0.02864118292927742, 0.09102724492549896, -0.08461901545524597, 0.04763771593570709, -0.09820594638586044, 0.12363352626562119, -0.06146226450800896, -0.004213334526866674, -0.18391123414039612, -0.009381342679262161, 0.0438222773373127, -0.023048149421811104, -0.0010736447293311357, -0.017531832680106163, 0.0634535625576973, 0.05774131044745445, -0.039399147033691406, -0.0408325120806694, -0.033703941851854324, -0.009450753219425678, -0.11244195699691772, -0.18618963658809662, -0.04849305376410484, -0.034880541265010834, 0.11852548271417618, -0.17111770808696747, 0.06201837584376335, 0.06472091376781464, 0.11162162572145462, 0.02296465076506138, -0.02818126603960991, -0.0076725417748093605, 0.041496098041534424, -0.039293959736824036, -0.0763016790151596, 0.06855007261037827, 0.03613912686705589, -0.0994340106844902, -0.021636441349983215, -0.11005594581365585, 0.18589626252651215, 0.12711931765079498, -0.015304340980947018, -0.046145275235176086, -0.0037261333782225847, -0.06693658232688904, -0.0252322256565094, 0.00010479353659320623, 0.02630312740802765, 0.18688923120498657, 0.007668614853173494, 0.17470361292362213, -0.10448402166366577, -0.054649028927087784, 0.04829766973853111, -0.027804838493466377, -0.012250205501914024, 0.11106427013874054, 0.000900426646694541, -0.11244355887174606, 0.14791768789291382, 0.12212676554918289, -0.06123341992497444, 0.1253422647714615, -0.06269760429859161, -0.042313165962696075, -0.03154568001627922, 0.0070209321565926075, 0.016009759157896042, 0.09681913256645203, -0.12338456511497498, -0.019060473889112473, 0.03688754513859749, 0.026065392419695854, 0.022046322003006935, -0.18368276953697205, 0.0016608438454568386, 0.046017080545425415, -0.06232096627354622, -0.0001330666127614677, -0.01134573481976986, -0.0014391003642231226, 0.09911268204450607, 0.018035314977169037, -0.08723576366901398, 0.04871227592229843, 0.008767826482653618, -0.07373929023742676, 0.20202793180942535, -0.10205624997615814, -0.19618737697601318, -0.11792904138565063, -0.0520332008600235, -0.09464605152606964, -0.00033909876947291195, 0.06213480234146118, -0.07916445285081863, -0.022073807194828987, -0.09520765393972397, -0.03123565763235092, -0.01737889274954796, 0.03542753681540489, 0.07313533127307892, -0.02203269489109516, 0.10872673243284225, -0.1269455999135971, -0.02306712418794632, -0.0358131118118763, 0.008673591539263725, 0.055851276963949203, 0.009536282159388065, 0.09724908322095871, 0.1170153021812439, -0.033818624913692474, 0.05197916179895401, -0.028282886371016502, 0.24284608662128448, -0.053264424204826355, -0.02976847253739834, 0.12733469903469086, -0.0040460084564983845, 0.08623170107603073, 0.08496475964784622, 0.04557865113019943, -0.08525460958480835, -0.01157840620726347, 0.00595470005646348, -0.03817666694521904, -0.21714363992214203, -0.036886055022478104, -0.04140350595116615, 0.02079310268163681, 0.10649195313453674, 0.04592718183994293, 0.043558333069086075, 0.057499635964632034, 0.023834481835365295, 0.04960625246167183, -0.02408022992312908, 0.09037080407142639, 0.1317465603351593, 0.04990197345614433, 0.13514135777950287, -0.04064502939581871, -0.03152943775057793, 0.039630502462387085, -0.002475862158462405, 0.19950714707374573, -0.02927609719336033, 0.1829756200313568, 0.04599745571613312, 0.1883370578289032, 0.011177031323313713, 0.06271716207265854, -0.024891022592782974, -0.006736312061548233, -0.009674341417849064, -0.0412435419857502, -0.05800659582018852, 0.0024435357190668583, -0.04474163055419922, 0.07293359190225601, -0.117747962474823, 0.040190309286117554, 0.06010587513446808, 0.29027169942855835, 0.022407572716474533, -0.37493225932121277, -0.11181744188070297, -0.016597270965576172, -0.031712714582681656, -0.04552937671542168, 0.010863134637475014, 0.11205030232667923, -0.09440278261899948, 0.06324154883623123, -0.08447496592998505, 0.09077057987451553, -0.07360748946666718, 0.03601958230137825, 0.05221237242221832, 0.09796657413244247, 0.006393694784492254, 0.06149907410144806, -0.26434561610221863, 0.25069257616996765, 0.018371766433119774, 0.05161698907613754, -0.06146470457315445, 0.015569853596389294, 0.02125987596809864, 0.06923633813858032, 0.07857105880975723, 0.0007500528590753675, -0.02767954021692276, -0.16410666704177856, -0.11515739560127258, 0.016607077792286873, 0.07020550966262817, -0.02121029421687126, 0.08431950211524963, -0.0010764591861516237, 0.004067974630743265, 0.04312531650066376, -0.008349082432687283, -0.027957666665315628, -0.09104249626398087, 0.016294151544570923, 0.06238500773906708, -0.04210955649614334, -0.08233670145273209, -0.1198863536119461, -0.08412427455186844, 0.1865774542093277, -0.008332050405442715, -0.0823029950261116, -0.12076611816883087, 0.06184208020567894, 0.06703545898199081, -0.09185026586055756, 0.042081426829099655, -0.017016833648085594, 0.12382151186466217, 0.003344601020216942, -0.07244402170181274, 0.09798984974622726, -0.04916444793343544, -0.16366566717624664, -0.03181470185518265, 0.13521257042884827, 0.034096188843250275, 0.05754973739385605, -0.012549868784844875, 0.019313739612698555, -0.026896705850958824, -0.0765208750963211, 0.03267578408122063, 0.002446891972795129, 0.09423114359378815, -0.015256565995514393, 0.0032036362681537867, 0.028266185894608498, -0.07761038839817047, 0.005427155178040266, 0.19771218299865723, 0.2606421709060669, -0.1043548434972763, 0.03651893138885498, 0.027820715680718422, -0.04207947105169296, -0.1520952433347702, 0.017168138176202774, 0.08070502430200577, 0.007420012727379799, -0.010656541213393211, -0.1764611452817917, 0.05012203007936478, 0.08809427171945572, -0.01788947917521, 0.08023256808519363, -0.29865652322769165, -0.11554376780986786, 0.11132355779409409, 0.12564530968666077, 0.09193845093250275, -0.13787555694580078, -0.04598912596702576, -0.0075193652883172035, -0.12548284232616425, 0.1158299669623375, -0.0956348180770874, 0.10974547266960144, -0.04320225492119789, 0.05354843661189079, 0.007337958551943302, -0.052445266395807266, 0.12752960622310638, 0.01908416859805584, 0.0783768892288208, -0.0478614941239357, 0.003130370983853936, 0.10798399150371552, -0.0758182629942894, 0.06092115119099617, -0.10174112766981125, 0.04936999827623367, -0.12913434207439423, -0.01087163481861353, -0.08215732127428055, 0.02880539558827877, -0.029542503878474236, -0.03560418263077736, -0.058059368282556534, 0.009000681340694427, 0.07482077181339264, -0.0021140656899660826, 0.17803633213043213, 0.047338198870420456, 0.12804824113845825, 0.2045678347349167, 0.07827051728963852, -0.11816922575235367, -0.10215629637241364, -0.0021673408336937428, -0.009365938603878021, 0.0566975399851799, -0.15991105139255524, 0.03947640210390091, 0.1441853940486908, 0.004068439826369286, 0.11697175353765488, 0.06757407635450363, -0.058885324746370316, -0.00002974093695229385, 0.04177243635058403, -0.17993171513080597, -0.10290152579545975, -0.014580353163182735, -0.012066712602972984, -0.13221070170402527, 0.07946693897247314, 0.11126816272735596, -0.07109289616346359, -0.018744582310318947, 0.0011641265591606498, -0.0008233633125200868, -0.025426305830478668, 0.18059362471103668, 0.0713382288813591, 0.07033462822437286, -0.10616692155599594, 0.09460733085870743, 0.0462157279253006, -0.0725565180182457, 0.04061013460159302, 0.06810374557971954, -0.11764568090438843, -0.022825488820672035, 0.037216611206531525, 0.14856918156147003, -0.027179433032870293, -0.05499974638223648, -0.17361967265605927, -0.1086450144648552, 0.08841122686862946, 0.13100364804267883, 0.10317601263523102, 0.017121080309152603, -0.04885329306125641, -0.009807179681956768, -0.11012768745422363, 0.09655274450778961, 0.05846559628844261, 0.0660039484500885, -0.16399246454238892, 0.11977937817573547, -0.0003578494652174413, 0.07169348001480103, -0.012192845344543457, 0.001448655384592712, -0.0843949168920517, 0.004430484492331743, -0.09232339262962341, 0.012558856047689915, -0.04001414775848389, 0.00048618309665471315, -0.022317882627248764, -0.04360825940966606, -0.0546182319521904, 0.03691510483622551, -0.10589302331209183, -0.03589541092514992, 0.030760591849684715, 0.02635170891880989, -0.11072052270174026, -0.031555552035570145, 0.0018842692952603102, -0.08543401211500168, 0.09491559118032455, 0.053661517798900604, -0.007116627413779497, 0.009966611862182617, -0.003835842479020357, 0.002103820675984025, 0.0713348463177681, 0.007812848314642906, 0.06204405426979065, -0.11863148212432861, -0.004516935441643, 0.004596107639372349, 0.001546198152936995, 0.019018597900867462, 0.12120135128498077, -0.12007837742567062, -0.014022485353052616, -0.022460926324129105, -0.022654950618743896, -0.0753025934100151, 0.05987093597650528, 0.10378355532884598, 0.041243866086006165, 0.19044536352157593, -0.06453124433755875, 0.015417633578181267, -0.2059982270002365, -0.0010957218473777175, 0.009144444018602371, -0.1515446901321411, -0.05387648567557335, -0.025656811892986298, 0.05819327011704445, -0.07224544882774353, 0.11503966152667999, 0.004568712320178747, -0.01583913527429104, 0.03687077388167381, -0.044159963726997375, -0.016699714586138725, 0.008686945773661137, 0.16097450256347656, 0.013965492136776447, -0.037191856652498245, 0.12147503346204758, 0.023390216752886772, 0.1060599610209465, 0.12157408893108368, 0.1685730367898941, 0.12814003229141235, 0.026453150436282158, 0.10737672448158264, 0.032515060156583786, -0.030444564297795296, -0.18784967064857483, 0.06720752269029617, -0.04578623175621033, 0.13226191699504852, 0.006823073141276836, 0.17648069560527802, 0.11036133766174316, -0.15301413834095, 0.049124132841825485, -0.021638398990035057, -0.09448863565921783, -0.10050928592681885, -0.09679119288921356, -0.07743744552135468, -0.15149177610874176, -0.01236912701278925, -0.11908751726150513, -0.001917796558700502, 0.06672867387533188, 0.0058901347219944, -0.021161947399377823, 0.13932976126670837, 0.023673847317695618, -0.0022931741550564766, 0.0803397074341774, -0.0035189997870475054, -0.05596965178847313, -0.07125981897115707, -0.07409729808568954, 0.013776594772934914, 0.013752772472798824, 0.05871286615729332, -0.02784240059554577, 0.009659827686846256, 0.051329080015420914, -0.02629757858812809, -0.10809190571308136, 0.009442382492125034, 0.021296720951795578, 0.054612502455711365, 0.020937973633408546, 0.013347321189939976, -0.011912370100617409, -0.017500825226306915, 0.17539910972118378, -0.06457335501909256, -0.022280963137745857, -0.10723160207271576, 0.20094625651836395, 0.036107197403907776, -0.03869583457708359, 0.040720485150814056, -0.06905907392501831, -0.020572103559970856, 0.18964605033397675, 0.20552918314933777, -0.026153136044740677, 0.0006357349921017885, -0.004643670748919249, -0.01797187142074108, -0.007700672838836908, 0.08245248347520828, 0.13317352533340454, -0.0074555822648108006, -0.06725732237100601, -0.026400936767458916, -0.06664355099201202, -0.004993689712136984, -0.048690151423215866, 0.06622577458620071, 0.027980800718069077, 0.012009805999696255, -0.046264056116342545, 0.022495536133646965, -0.042431484907865524, -0.07840681821107864, 0.030966809019446373, -0.20489253103733063, -0.15987305343151093, -0.008346200920641422, 0.03857562318444252, 0.0004496309265960008, 0.062375955283641815, -0.0009975176071748137, 0.012196631170809269, 0.0908537283539772, -0.02237413451075554, -0.08646247535943985, -0.07932375371456146, 0.09022057801485062, -0.17671364545822144, 0.2005920559167862, -0.04314102232456207, 0.049184780567884445, 0.1326240599155426, 0.04266604781150818, -0.10272050648927689, 0.0379687063395977, 0.03511279076337814, -0.008898447267711163, -0.004219701513648033, 0.09855833649635315, -0.012647976167500019, 0.07372915744781494, 0.03857356309890747, -0.09686969965696335, -0.039119016379117966, -0.07064490765333176, -0.0207537654787302, -0.040169861167669296, -0.052368659526109695, -0.05137869715690613, 0.11046379059553146, 0.19000554084777832, -0.049329470843076706, -0.017489463090896606, -0.06654652208089828, 0.005614506546407938, 0.07303186506032944, -0.011176830157637596, -0.04266827553510666, -0.253805011510849, 0.010820526629686356, 0.046582285314798355, -0.017032980918884277, -0.2275182157754898, -0.1029951274394989, -0.004221665672957897, -0.06776802986860275, -0.07437815517187119, 0.07514068484306335, 0.04294868931174278, 0.05498082563281059, -0.06696301698684692, 0.028071772307157516, -0.09679107367992401, 0.15699414908885956, -0.15248030424118042, -0.07556789368391037 ]