sha
null
last_modified
null
library_name
stringclasses
154 values
text
stringlengths
1
900k
metadata
stringlengths
2
348k
pipeline_tag
stringclasses
45 values
id
stringlengths
5
122
tags
sequencelengths
1
1.84k
created_at
stringlengths
25
25
arxiv
sequencelengths
0
201
languages
sequencelengths
0
1.83k
tags_str
stringlengths
17
9.34k
text_str
stringlengths
0
389k
text_lists
sequencelengths
0
722
processed_texts
sequencelengths
1
723
tokens_length
sequencelengths
1
723
input_texts
sequencelengths
1
61
embeddings
sequencelengths
768
768
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base_seed42_original_amh-esp-eng_train This model is a fine-tuned version of [FacebookAI/xlm-roberta-base](https://huggingface.co/FacebookAI/xlm-roberta-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.0175 - Spearman Corr: 0.8510 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 1.76 | 200 | 0.0231 | 0.8199 | | 0.0378 | 3.52 | 400 | 0.0142 | 0.8523 | | 0.021 | 5.29 | 600 | 0.0142 | 0.8544 | | 0.0157 | 7.05 | 800 | 0.0144 | 0.8553 | | 0.0125 | 8.81 | 1000 | 0.0159 | 0.8538 | | 0.0104 | 10.57 | 1200 | 0.0156 | 0.8515 | | 0.0083 | 12.33 | 1400 | 0.0158 | 0.8503 | | 0.0067 | 14.1 | 1600 | 0.0143 | 0.8510 | | 0.0067 | 15.86 | 1800 | 0.0183 | 0.8493 | | 0.0059 | 17.62 | 2000 | 0.0175 | 0.8510 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
{"license": "mit", "tags": ["generated_from_trainer"], "base_model": "FacebookAI/xlm-roberta-base", "model-index": [{"name": "xlm-roberta-base_seed42_original_amh-esp-eng_train", "results": []}]}
text-classification
shanhy/xlm-roberta-base_seed42_original_amh-esp-eng_train
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-15T00:53:13+00:00
[]
[]
TAGS #transformers #safetensors #xlm-roberta #text-classification #generated_from_trainer #base_model-FacebookAI/xlm-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us
xlm-roberta-base\_seed42\_original\_amh-esp-eng\_train ====================================================== This model is a fine-tuned version of FacebookAI/xlm-roberta-base on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 0.0175 * Spearman Corr: 0.8510 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 32 * eval\_batch\_size: 128 * seed: 42 * gradient\_accumulation\_steps: 2 * total\_train\_batch\_size: 64 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 30 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.37.2 * Pytorch 2.1.0+cu121 * Datasets 2.17.0 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 128\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #safetensors #xlm-roberta #text-classification #generated_from_trainer #base_model-FacebookAI/xlm-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 128\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ 68, 141, 4, 33 ]
[ "passage: TAGS\n#transformers #safetensors #xlm-roberta #text-classification #generated_from_trainer #base_model-FacebookAI/xlm-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 128\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ -0.13196544349193573, 0.09891481697559357, -0.0018682656809687614, 0.08302310854196548, 0.16318348050117493, 0.016677498817443848, 0.12253726273775101, 0.11726276576519012, -0.11173776537179947, 0.05556543171405792, 0.10338494926691055, 0.08458799868822098, 0.0286884643137455, 0.15043699741363525, -0.045705877244472504, -0.28430095314979553, -0.00045704812509939075, 0.02558436244726181, -0.1376832276582718, 0.13619966804981232, 0.10879813879728317, -0.12514753639698029, 0.06264033913612366, 0.023766178637742996, -0.1919274777173996, 0.0014368330594152212, 0.017397576943039894, -0.07561030238866806, 0.13846375048160553, 0.03705544397234917, 0.13360853493213654, 0.03372970595955849, 0.10445985943078995, -0.17197804152965546, 0.008810426108539104, 0.06506016850471497, 0.01182118896394968, 0.09916599839925766, 0.08596121519804001, -0.006887015886604786, 0.107834093272686, -0.061852216720581055, 0.07219021767377853, 0.03824690356850624, -0.11962353438138962, -0.28946825861930847, -0.1038181334733963, 0.09671656042337418, 0.11488976329565048, 0.08843138813972473, -0.010587817057967186, 0.11949947476387024, -0.09038527309894562, 0.08671168982982635, 0.24624861776828766, -0.2762778699398041, -0.08599308133125305, -0.006691739894449711, 0.04454458877444267, 0.0192960724234581, -0.11702073365449905, -0.034725237637758255, 0.060181524604558945, 0.0283932164311409, 0.11969391256570816, 0.0011189340148121119, -0.01149663608521223, -0.007112774066627026, -0.1498728096485138, -0.04236593469977379, 0.13226667046546936, 0.057521168142557144, -0.049670055508613586, -0.04971322789788246, -0.05079459026455879, -0.17936183512210846, -0.03869696706533432, 0.004522996488958597, 0.01884404383599758, -0.06204994395375252, -0.1170351505279541, 0.02485233172774315, -0.0986606776714325, -0.0883607417345047, -0.00536000169813633, 0.19927476346492767, 0.05297479033470154, 0.0019048440735787153, -0.012764488346874714, 0.13149689137935638, 0.029031140729784966, -0.15885967016220093, 0.02181168459355831, 0.013129622675478458, -0.04196780547499657, -0.02643256075680256, -0.03564628213644028, -0.013651377521455288, 0.0005222551408223808, 0.16112670302391052, -0.10241930186748505, 0.02287430316209793, 0.027450762689113617, 0.03275188431143761, -0.09093765914440155, 0.17895354330539703, -0.08061987161636353, -0.024040358141064644, 0.004189989995211363, 0.1055867150425911, 0.01888435333967209, -0.0002917523088399321, -0.09347851574420929, 0.009325179271399975, 0.1090208888053894, 0.044567689299583435, -0.029269354417920113, 0.035916756838560104, -0.04596438258886337, -0.026745637878775597, 0.07028650492429733, -0.08116807043552399, 0.031144915148615837, 0.009348084218800068, -0.08677830547094345, -0.03827532380819321, 0.014872299507260323, 0.011736470274627209, 0.013896521180868149, 0.14781735837459564, -0.09783969074487686, 0.0013644210994243622, -0.10506504029035568, -0.11206164211034775, 0.01479119062423706, -0.02137015573680401, 0.0012705293484032154, -0.10630778968334198, -0.14520946145057678, -0.020766140893101692, 0.023043107241392136, -0.03370993584394455, -0.07425355166196823, -0.009458414278924465, -0.09016023576259613, 0.036381740123033524, -0.019144224002957344, 0.1338672935962677, -0.04918650910258293, 0.1320859044790268, 0.06343449652194977, 0.04678048938512802, 0.00831086840480566, 0.04423472657799721, -0.09216220676898956, 0.03793376311659813, -0.1888304054737091, 0.022821834310889244, -0.05996130406856537, 0.059156384319067, -0.08510183542966843, -0.1328919231891632, 0.017381513491272926, -0.017376312986016273, 0.09732269495725632, 0.11362158507108688, -0.13049037754535675, -0.10281094163656235, 0.16968834400177002, -0.109103724360466, -0.1255229115486145, 0.12405472993850708, -0.02942158840596676, -0.003954533953219652, 0.030629022046923637, 0.10945684462785721, 0.0842948853969574, -0.08237522840499878, -0.020055506378412247, -0.05470968782901764, 0.08616973459720612, -0.016709523275494576, 0.10127030313014984, 0.003251338843256235, -0.007391751743853092, 0.014485578052699566, -0.05618653446435928, 0.06764025241136551, -0.12030204385519028, -0.08058790117502213, -0.03696097806096077, -0.08276177942752838, 0.08453356474637985, 0.07146790623664856, 0.050437234342098236, -0.10332237184047699, -0.11862462013959885, 0.07324158400297165, 0.11601345986127853, -0.048777882009744644, 0.0124500198289752, -0.059240709990262985, 0.08453813195228577, -0.051279596984386444, -0.026758845895528793, -0.18091709911823273, -0.06499651074409485, 0.02301541157066822, -0.018301941454410553, -0.004097240045666695, -0.037253014743328094, 0.062303297221660614, 0.093818798661232, -0.08045177906751633, -0.0764354020357132, -0.10619605332612991, -0.0158742256462574, -0.09482879191637039, -0.22273369133472443, -0.07861017435789108, -0.017165008932352066, 0.15580084919929504, -0.21151766180992126, 0.04360150545835495, -0.010889439843595028, 0.14122705161571503, 0.029071001335978508, -0.01567329280078411, -0.03531797602772713, 0.09140531718730927, -0.03568165749311447, -0.05378113314509392, 0.04078692942857742, 0.015963202342391014, -0.08712584525346756, -0.01978384144604206, -0.10062315315008163, 0.19644492864608765, 0.12115684151649475, -0.019988611340522766, -0.10672228783369064, -0.01419133972376585, -0.09363941103219986, -0.04462401568889618, -0.054406121373176575, -0.004758825991302729, 0.10676517337560654, 0.019794635474681854, 0.1471862941980362, -0.08466091006994247, -0.05152713879942894, 0.03815930709242821, -0.021593807265162468, 0.012436249293386936, 0.10759704560041428, 0.09588126838207245, -0.09077505767345428, 0.14478042721748352, 0.11831695586442947, -0.08035369962453842, 0.14524731040000916, -0.05666378140449524, -0.08607003837823868, -0.03343934565782547, -0.006182708777487278, 0.024729762226343155, 0.14715492725372314, -0.08842822164297104, 0.0009005532483570278, 0.02143796533346176, -0.0009729673038236797, 0.018144162371754646, -0.21142855286598206, -0.032100383192300797, 0.020743396133184433, -0.035349272191524506, -0.023111259564757347, -0.014213657937943935, 0.028337886556982994, 0.11677493900060654, 0.01414300873875618, -0.0540250800549984, 0.013599103316664696, 0.009574293158948421, -0.07914607971906662, 0.2171662151813507, -0.07426396012306213, -0.17233231663703918, -0.15179471671581268, 0.004287849646061659, -0.05426809936761856, -0.009687787853181362, 0.04670441523194313, -0.09638029336929321, -0.024891173467040062, -0.037057165056467056, 0.04515697434544563, -0.007374031934887171, 0.0434957891702652, -0.015290717594325542, 0.027705008164048195, 0.0815771147608757, -0.10898098349571228, 0.01840367168188095, -0.030161090195178986, -0.05450989678502083, 0.03729647025465965, 0.050532273948192596, 0.1120103970170021, 0.15729740262031555, -0.018926111981272697, 0.00281600933521986, -0.03385520726442337, 0.1718938797712326, -0.09474305063486099, -0.039306409657001495, 0.1497810333967209, 0.01193019188940525, 0.05696384236216545, 0.13361965119838715, 0.06738590449094772, -0.07117169350385666, 0.026298534125089645, 0.04004959389567375, -0.01705203577876091, -0.2273423969745636, -0.04503382742404938, -0.037529271095991135, 0.008658994920551777, 0.10548876971006393, 0.0318632572889328, 0.0270606130361557, 0.0570688471198082, -0.039488404989242554, 0.029293619096279144, -0.04103964939713478, 0.08947612345218658, 0.06445353478193283, 0.050813522189855576, 0.14866364002227783, -0.026137124747037888, -0.056426919996738434, 0.03509410098195076, -0.045765411108732224, 0.2088230550289154, -0.03246336430311203, 0.08511096239089966, 0.05439717695116997, 0.18534918129444122, 0.022988246753811836, 0.08172706514596939, 0.02111090160906315, -0.040226854383945465, 0.02130485698580742, -0.0510917529463768, -0.01913459412753582, 0.024741031229496002, -0.020088329911231995, 0.09492947906255722, -0.15800602734088898, -0.017486002296209335, 0.04466555640101433, 0.3070462644100189, 0.057204265147447586, -0.3239973783493042, -0.1329151690006256, -0.007546214386820793, -0.05137051269412041, -0.04527734965085983, 0.00596288638189435, 0.1046595349907875, -0.1131599098443985, 0.046224016696214676, -0.08372066169977188, 0.08514859527349472, -0.011406857520341873, 0.006385982036590576, 0.09116256982088089, 0.08809847384691238, -0.012875889427959919, 0.06686434149742126, -0.2405211478471756, 0.302921861410141, 0.0031665891874581575, 0.067683145403862, -0.03094049170613289, 0.023741696029901505, 0.03205963596701622, 0.05544143542647362, 0.044757336378097534, -0.022030169144272804, -0.03817500174045563, -0.21632836759090424, -0.07639524340629578, 0.01098565012216568, 0.11032047122716904, -0.08038049191236496, 0.1349044144153595, -0.02970171719789505, -0.01409254688769579, 0.054343778640031815, -0.03862052410840988, -0.09632794559001923, -0.0757531076669693, 0.01526376698166132, -0.01612093485891819, 0.042920488864183426, -0.12476500868797302, -0.13458886742591858, -0.053820978850126266, 0.18230336904525757, -0.06271585822105408, -0.04775528237223625, -0.13164086639881134, 0.11616203933954239, 0.1310301423072815, -0.0800657570362091, 0.05902821570634842, -0.0004891243879683316, 0.13768382370471954, -0.00018232640286441892, -0.034991513937711716, 0.09990507364273071, -0.08413300663232803, -0.25101998448371887, -0.05981237813830376, 0.16267050802707672, 0.023493312299251556, 0.0548059456050396, -0.02138851024210453, 0.02654334157705307, -0.01098493579775095, -0.09118691831827164, 0.023614194244146347, -0.03138713166117668, 0.07060311734676361, 0.038307689130306244, -0.04493559151887894, -0.003535062773153186, -0.055559176951646805, -0.04866667836904526, 0.11938028037548065, 0.2810666561126709, -0.10598211735486984, -0.009600569494068623, 0.058872971683740616, -0.03153645992279053, -0.16587300598621368, 0.03787032142281532, 0.11307208985090256, 0.024927588179707527, 0.014607351273298264, -0.16745203733444214, 0.07618924230337143, 0.113561250269413, -0.04220382124185562, 0.08733001351356506, -0.29393360018730164, -0.13554498553276062, 0.08696893602609634, 0.12788578867912292, 0.024306975305080414, -0.17132867872714996, -0.043222617357969284, -0.020857524126768112, -0.09924142807722092, 0.11386924982070923, -0.07394159585237503, 0.10555082559585571, -0.034590329974889755, 0.04956836998462677, 0.013746642507612705, -0.060336288064718246, 0.1332203894853592, -0.03701593726873398, 0.08962203562259674, -0.019909681752324104, -0.017622459679841995, 0.08537176251411438, -0.04859213903546333, 0.011331639252603054, -0.03726105019450188, 0.046149592846632004, -0.09338995814323425, -0.01944100856781006, -0.11331600695848465, 0.05712530016899109, -0.06554371863603592, -0.06349298357963562, -0.030393438413739204, 0.05302911624312401, 0.0006406907923519611, -0.028255226090550423, 0.12629824876785278, 0.015722298994660378, 0.2150397002696991, 0.11728408187627792, 0.06625007092952728, -0.016186660155653954, -0.0641254261136055, -0.008585327304899693, -0.03099910356104374, 0.06869698315858841, -0.13333941996097565, 0.02077336050570011, 0.12685872614383698, 0.06321674585342407, 0.12267275899648666, 0.07356739789247513, -0.057909898459911346, 0.015538030304014683, 0.09509473294019699, -0.13818930089473724, -0.08300869166851044, -0.03791758790612221, 0.021259013563394547, -0.1530337780714035, 0.060089360922575, 0.10215576738119125, -0.0836058109998703, -0.03092457726597786, 0.005163066554814577, 0.000374901806935668, -0.030314097180962563, 0.21469973027706146, 0.07397905737161636, 0.09001398086547852, -0.09201814979314804, 0.06724780797958374, 0.06080635264515877, -0.10922487825155258, -0.013464605435729027, 0.08993840962648392, -0.06848926842212677, -0.02020818367600441, 0.025699980556964874, 0.1014053151011467, -0.07606237381696701, -0.054409973323345184, -0.1661984771490097, -0.13389240205287933, 0.06314565241336823, 0.1618947684764862, 0.07241392135620117, 0.024320410564541817, -0.007316235452890396, 0.03363948315382004, -0.13871777057647705, 0.12411312758922577, 0.07404549419879913, 0.09306581318378448, -0.15791285037994385, 0.1818198412656784, 0.014229292050004005, 0.03504512831568718, -0.012622985057532787, 0.027776600793004036, -0.11654958128929138, -0.0019876465667039156, -0.1256277859210968, -0.05021267756819725, -0.037451595067977905, -0.008649054914712906, -0.01665184088051319, -0.05679382011294365, -0.06017814204096794, 0.020587364211678505, -0.1020774319767952, -0.03487476333975792, 0.013034898787736893, 0.036619529128074646, -0.12464044243097305, -0.016756128519773483, 0.041986919939517975, -0.09662823379039764, 0.08624295145273209, 0.05239079147577286, 0.06865750998258591, 0.04205240309238434, -0.07263530790805817, 0.01565958932042122, 0.04319623485207558, -0.01246478408575058, 0.04708905145525932, -0.1416696459054947, 0.013452060520648956, -0.035300228744745255, 0.032127853482961655, 0.012330862693488598, 0.041268181055784225, -0.15590620040893555, -0.0006988015957176685, 0.0005412847967818379, -0.05495034530758858, -0.06016375124454498, 0.029517367482185364, 0.06915541738271713, 0.023343803361058235, 0.16834065318107605, -0.09421270340681076, 0.043014321476221085, -0.24118341505527496, -0.0035449780989438295, -0.04173633083701134, -0.08540837466716766, -0.0987994521856308, -0.019733915105462074, 0.06711504608392715, -0.04421355575323105, 0.09756482392549515, -0.014044692739844322, 0.109415702521801, 0.044441673904657364, -0.061472225934267044, 0.04127075895667076, 0.03979703038930893, 0.17852424085140228, 0.03312114626169205, -0.02493695169687271, 0.05475730076432228, 0.03807259723544121, 0.07197780162096024, 0.0731436088681221, 0.18023869395256042, 0.12355324625968933, 0.011792181059718132, 0.06446578353643417, 0.06573721021413803, -0.08827177435159683, -0.16074307262897491, 0.014717316254973412, -0.029363291338086128, 0.08556829392910004, -0.021288832649588585, 0.17704680562019348, 0.14232785999774933, -0.1777249574661255, 0.03781313821673393, -0.030358977615833282, -0.06662668287754059, -0.09023876488208771, 0.008646022528409958, -0.07632064819335938, -0.17277851700782776, 0.011035195551812649, -0.12461110204458237, 0.002976049669086933, 0.06567922979593277, 0.005048900842666626, -0.006358268670737743, 0.14241522550582886, 0.07726656645536423, 0.014594470150768757, 0.08281748741865158, 0.03125395253300667, -0.0042702932842075825, -0.047413360327482224, -0.07322124391794205, 0.003198332618921995, -0.03393085300922394, 0.03494373336434364, -0.06007290259003639, -0.10644083470106125, 0.055227603763341904, 0.03205983340740204, -0.10651516914367676, 0.03054053708910942, 0.003860602155327797, 0.08455248922109604, 0.06187829002737999, 0.0024589267559349537, 0.028186585754156113, -0.028640104457736015, 0.2566484212875366, -0.09727800637483597, -0.057189494371414185, -0.1263759881258011, 0.29044467210769653, 0.016411423683166504, -0.027846038341522217, 0.04205775260925293, -0.0899231880903244, -0.014586132019758224, 0.1568177491426468, 0.1485815942287445, -0.019990285858511925, -0.004054858349263668, 0.01543956995010376, -0.026464708149433136, -0.05437016114592552, 0.086168073117733, 0.1241421103477478, 0.09672901779413223, -0.08731590956449509, -0.03509433567523956, -0.056863438338041306, -0.02664414793252945, -0.015760263428092003, 0.08034801483154297, 0.015216142870485783, -0.0015402467688545585, -0.05069897323846817, 0.07896960526704788, -0.016800498589873314, -0.12112859636545181, 0.06400942802429199, -0.20127683877944946, -0.18462198972702026, -0.019730333238840103, 0.05812923237681389, 0.00019156295456923544, 0.07645424455404282, -0.004090806934982538, -0.019641485065221786, 0.08557789772748947, -0.008854112587869167, -0.05160798877477646, -0.13622045516967773, 0.09911761432886124, -0.056526068598032, 0.20981615781784058, -0.05306484177708626, 0.04587500914931297, 0.12786905467510223, 0.04194230958819389, -0.10196087509393692, 0.036407627165317535, 0.07802671194076538, -0.1121029257774353, 0.032378170639276505, 0.15671691298484802, -0.02063562534749508, 0.09771383553743362, 0.027820155024528503, -0.16775496304035187, 0.013787583447992802, -0.08647667616605759, -0.05391157418489456, -0.07212251424789429, 0.005070052575320005, -0.03500846400856972, 0.13139131665229797, 0.229573592543602, -0.060340479016304016, -0.010522725060582161, -0.06468646973371506, 0.025870291516184807, 0.07577784359455109, 0.09794671833515167, -0.017301272600889206, -0.2772675156593323, 0.03158432990312576, 0.061059098690748215, -0.004956540651619434, -0.28039345145225525, -0.08054531365633011, 0.04352137818932533, -0.07393883913755417, -0.08716678619384766, 0.09514938294887543, 0.04989435523748398, 0.05062716826796532, -0.041774995625019073, -0.09182899445295334, -0.08364211767911911, 0.17357149720191956, -0.16998212039470673, -0.08033598959445953 ]
null
null
peft
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Adapter This model is a fine-tuned version of [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 1 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 4 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2 - training_steps: 10 - mixed_precision_training: Native AMP ### Training results ### Framework versions - PEFT 0.8.2 - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
{"library_name": "peft", "tags": ["generated_from_trainer"], "base_model": "meta-llama/Llama-2-7b-hf", "model-index": [{"name": "Adapter", "results": []}]}
null
akkky02/Adapter
[ "peft", "safetensors", "generated_from_trainer", "base_model:meta-llama/Llama-2-7b-hf", "region:us" ]
2024-02-15T00:54:58+00:00
[]
[]
TAGS #peft #safetensors #generated_from_trainer #base_model-meta-llama/Llama-2-7b-hf #region-us
# Adapter This model is a fine-tuned version of meta-llama/Llama-2-7b-hf on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 1 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 4 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2 - training_steps: 10 - mixed_precision_training: Native AMP ### Training results ### Framework versions - PEFT 0.8.2 - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
[ "# Adapter\n\nThis model is a fine-tuned version of meta-llama/Llama-2-7b-hf on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 1\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 4\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2\n- training_steps: 10\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.2" ]
[ "TAGS\n#peft #safetensors #generated_from_trainer #base_model-meta-llama/Llama-2-7b-hf #region-us \n", "# Adapter\n\nThis model is a fine-tuned version of meta-llama/Llama-2-7b-hf on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 1\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 4\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2\n- training_steps: 10\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.2" ]
[ 39, 34, 6, 12, 8, 3, 139, 4, 39 ]
[ "passage: TAGS\n#peft #safetensors #generated_from_trainer #base_model-meta-llama/Llama-2-7b-hf #region-us \n# Adapter\n\nThis model is a fine-tuned version of meta-llama/Llama-2-7b-hf on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 1\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 4\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2\n- training_steps: 10\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.2" ]
[ -0.12487402558326721, 0.10333660989999771, -0.0033091697841882706, 0.06184389069676399, 0.11711011081933975, 0.021471703425049782, 0.10876346379518509, 0.13817696273326874, -0.011044046841561794, 0.10776624828577042, 0.09683068841695786, 0.0304058026522398, 0.07049813866615295, 0.18305432796478271, -0.020913897082209587, -0.2592708170413971, 0.02518879994750023, -0.056490473449230194, -0.0892280712723732, 0.08806058019399643, 0.10763003677129745, -0.09409355372190475, 0.06466154754161835, 0.00572290550917387, -0.08801846206188202, -0.010425089858472347, -0.032045744359493256, -0.05619470030069351, 0.10132003575563431, -0.006095458287745714, 0.07161620259284973, 0.041211992502212524, 0.11663264036178589, -0.22143566608428955, 0.00401340750977397, 0.06479215621948242, 0.05314910039305687, 0.09628599137067795, 0.07785561680793762, -0.01923501119017601, 0.062397491186857224, -0.15368495881557465, 0.10672125965356827, 0.03255442902445793, -0.07202945649623871, -0.16500616073608398, -0.09653982520103455, 0.06230711191892624, 0.12635739147663116, 0.0585370771586895, 0.01365047786384821, 0.09297426044940948, -0.09454654902219772, 0.04026797413825989, 0.21839860081672668, -0.2735937833786011, -0.05862700566649437, 0.0034544861409813166, 0.05979664623737335, 0.08760064840316772, -0.10800094902515411, 0.0011169328354299068, 0.018202748149633408, 0.021847473457455635, 0.11857552081346512, 0.016517268493771553, 0.02486826479434967, -0.0017896379576995969, -0.1280122846364975, -0.009802439250051975, 0.10307404398918152, 0.044673819094896317, -0.056538742035627365, -0.15254153311252594, -0.03819379210472107, -0.0952584519982338, -0.012530981563031673, -0.007722607348114252, 0.01733063906431198, -0.0565871000289917, -0.050764795392751694, -0.05966201052069664, -0.05165061727166176, -0.07169816642999649, 0.027372654527425766, 0.1721286177635193, 0.03762056306004524, 0.008637294173240662, 0.027455352246761322, 0.12371521443128586, 0.02758646570146084, -0.11881411820650101, -0.018406495451927185, -0.018109314143657684, -0.12644533812999725, -0.03683250769972801, -0.042628560215234756, -0.016626061871647835, 0.005337996408343315, 0.13965265452861786, -0.026954172179102898, 0.07958229631185532, 0.032458871603012085, 0.0017931148177012801, -0.023555148392915726, 0.11554068326950073, -0.06170772388577461, -0.03245462477207184, -0.009475132450461388, 0.13652434945106506, 0.02368753030896187, -0.03492356091737747, -0.07770726829767227, -0.024953138083219528, 0.07337649166584015, 0.08986146748065948, -0.040426600724458694, -0.0012256810441613197, -0.055879149585962296, -0.03197590634226799, 0.04515693709254265, -0.13514460623264313, 0.03260309249162674, 0.01113179326057434, -0.07392166554927826, -0.045700062066316605, -0.005831385031342506, 0.010446880012750626, -0.024423591792583466, 0.0863928273320198, -0.07253166288137436, -0.020372983068227768, -0.0529119037091732, -0.04567675665020943, 0.006682693958282471, -0.10080142319202423, -0.020617008209228516, -0.05760122835636139, -0.13180756568908691, -0.035213906317949295, 0.04112337902188301, -0.09373415261507034, -0.04596101865172386, -0.033409636467695236, -0.03337651491165161, 0.04644564166665077, -0.018222477287054062, 0.13192011415958405, -0.05370037257671356, 0.06934699416160583, -0.03673442080616951, 0.05296831950545311, 0.05692184343934059, 0.03649213910102844, -0.04822072014212608, 0.07068377733230591, -0.15319953858852386, 0.08871210366487503, -0.09323062002658844, 0.014784115366637707, -0.14535865187644958, -0.06906712055206299, 0.017671430483460426, -0.02934991754591465, 0.08122795820236206, 0.14037083089351654, -0.14747537672519684, -0.0138822291046381, 0.15085192024707794, -0.06759873777627945, -0.08426696062088013, 0.089119553565979, -0.035085998475551605, 0.012720335274934769, 0.021061142906546593, 0.17948319017887115, 0.09357128292322159, -0.13843558728694916, 0.023029258474707603, 0.022786106914281845, 0.07668767869472504, 0.026256553828716278, 0.09176268428564072, -0.03469033166766167, 0.005027120467275381, -0.00278841913677752, -0.08096087723970413, -0.00035464041866362095, -0.06615345925092697, -0.07143516838550568, -0.02964749000966549, -0.08161448687314987, 0.0576133206486702, -0.023431023582816124, 0.01683744043111801, -0.06026928126811981, -0.11740950495004654, 0.04879558086395264, 0.13781709969043732, -0.061478227376937866, 0.014356537722051144, -0.07501735538244247, 0.04325798898935318, -0.022006822749972343, -0.031244007870554924, -0.1660270243883133, -0.11434584110975266, 0.048762354999780655, -0.07328635454177856, 0.00036982589517720044, -0.03862289711833, 0.06243765726685524, 0.06828754395246506, -0.040486425161361694, -0.041314590722322464, -0.06429964303970337, -0.0007124505937099457, -0.09586995840072632, -0.20625099539756775, -0.07537413388490677, -0.04099550470709801, 0.21877267956733704, -0.22304517030715942, 0.0005169542855583131, 0.004926816560328007, 0.15395620465278625, 0.03384312987327576, -0.05687596648931503, 0.024555625393986702, 0.021637680009007454, 0.009724678471684456, -0.10870018601417542, 0.03963162377476692, -0.0032656339462846518, -0.14313481748104095, 0.02498113177716732, -0.15609388053417206, 0.010845583863556385, 0.04632239416241646, 0.15332791209220886, -0.09302067011594772, -0.11988168954849243, -0.05977468937635422, -0.04921533539891243, -0.06663041561841965, 0.010710115544497967, 0.17305070161819458, 0.05100054293870926, 0.11488877236843109, -0.077410988509655, -0.07447611540555954, 0.011372278444468975, 0.012878555804491043, -0.008094136603176594, 0.07670538127422333, 0.03204426169395447, -0.10659786313772202, 0.04825756326317787, 0.12621580064296722, -0.01782907545566559, 0.11747897416353226, -0.043457210063934326, -0.11039897799491882, -0.03160430118441582, 0.035969190299510956, 0.008879859000444412, 0.13897298276424408, -0.01856572926044464, 0.012641135603189468, 0.03408847749233246, 0.03197907656431198, 0.016757767647504807, -0.16368910670280457, -0.010049433447420597, 0.019180109724402428, -0.04156952351331711, -0.030258631333708763, -0.0199134424328804, 0.0280714500695467, 0.0803973376750946, 0.05837724357843399, -0.013097047805786133, 0.025605062022805214, -0.018841741606593132, -0.08208855986595154, 0.1810712367296219, -0.11668847501277924, -0.1353953778743744, -0.10972218215465546, 0.033338796347379684, -0.03576714172959328, -0.03046390227973461, -0.006421225145459175, -0.058030929416418076, -0.04563053697347641, -0.11129574477672577, -0.038103848695755005, -0.03901916742324829, 0.011423878371715546, 0.024395326152443886, 0.0074576325714588165, 0.09022063761949539, -0.09889530390501022, 0.013238552026450634, -0.011016763746738434, -0.04066196456551552, 0.011351699009537697, 0.05006321519613266, 0.06649014353752136, 0.12202394008636475, 0.013827430084347725, 0.02631910890340805, -0.043871376663446426, 0.1815834939479828, -0.09585995972156525, 0.00851974356919527, 0.10656781494617462, -0.008266497403383255, 0.06767085939645767, 0.12881767749786377, 0.02434053085744381, -0.0882461667060852, 0.031384147703647614, 0.04897012189030647, -0.022410651668906212, -0.2434547245502472, -0.030707167461514473, -0.018655454739928246, -0.0632651075720787, 0.12514877319335938, 0.050142157822847366, -0.038730595260858536, 0.04128521308302879, -0.022790351882576942, -0.03803037106990814, 0.003698514075949788, 0.07589966058731079, 0.06077122315764427, 0.04318168759346008, 0.09583871066570282, -0.00814195815473795, 0.0012808124301955104, 0.061551425606012344, 0.03989458829164505, 0.2576327621936798, -0.04923133924603462, 0.12019674479961395, 0.009241546504199505, 0.13548454642295837, -0.03136202692985535, 0.04629630222916603, 0.024858519434928894, -0.010927325114607811, -0.0019383588805794716, -0.06565124541521072, -0.023639660328626633, 0.05535171926021576, 0.033728066831827164, 0.020185599103569984, -0.11220105737447739, 0.06450188905000687, 0.012611566111445427, 0.2981871962547302, 0.06790006160736084, -0.2888491451740265, -0.06944084167480469, 0.015958890318870544, -0.041745807975530624, -0.08414854109287262, 0.011848918162286282, 0.14047643542289734, -0.13866713643074036, 0.07071918994188309, -0.08002747595310211, 0.07055117934942245, -0.053353242576122284, -0.018637381494045258, 0.044529564678668976, 0.14231400191783905, -0.022775961086153984, 0.07814819365739822, -0.1569720059633255, 0.20968325436115265, 0.018551988527178764, 0.0741274431347847, -0.06687091290950775, 0.024607166647911072, 0.013155723921954632, 0.06332425028085709, 0.11989552527666092, 0.015003323554992676, -0.08430825173854828, -0.1631813645362854, -0.13821791112422943, 0.023954013362526894, 0.14859461784362793, -0.07098385691642761, 0.08408931642770767, -0.050384510308504105, -0.0029418885242193937, 0.01792592741549015, -0.09247498959302902, -0.11324324458837509, -0.13973820209503174, 0.039958253502845764, -0.006441498175263405, 0.012643261812627316, -0.1070866733789444, -0.09631228446960449, 0.01887550763785839, 0.14948001503944397, -0.02373550646007061, -0.06155940145254135, -0.17220178246498108, 0.020529797300696373, 0.15796378254890442, -0.05234958976507187, 0.03324204310774803, 0.008793788962066174, 0.1506715714931488, 0.01979858987033367, -0.04794667288661003, 0.09413349628448486, -0.07119425386190414, -0.19827201962471008, -0.06334654241800308, 0.16140656173229218, 0.05907374247908592, 0.04998994618654251, -0.002235773019492626, 0.011685607023537159, 0.02361569181084633, -0.08134674280881882, 0.04434936121106148, 0.0804663822054863, 0.03359860181808472, 0.04697932302951813, -0.05349791422486305, 0.11982834339141846, -0.04165414720773697, -0.02900727465748787, 0.12122434377670288, 0.2555544972419739, -0.08999429643154144, 0.13546903431415558, 0.057812079787254333, -0.039925239980220795, -0.16708502173423767, 0.023984776809811592, 0.13469533622264862, 0.038923945277929306, 0.059106580913066864, -0.19108779728412628, 0.0787704810500145, 0.11434578895568848, -0.03179992362856865, 0.06743532419204712, -0.30467721819877625, -0.11677826941013336, 0.05639045313000679, 0.08708545565605164, 0.0017519117100164294, -0.11294090747833252, -0.058703478425741196, -0.026323866099119186, -0.06884926557540894, 0.07379067689180374, -0.07165330648422241, 0.10753174871206284, -0.012757210992276669, 0.07159756124019623, 0.03506499528884888, -0.03099469281733036, 0.16561296582221985, 0.005978884641081095, 0.05249708518385887, -0.032668571919202805, 0.045694682747125626, 0.02173296920955181, -0.10125935077667236, 0.031406138092279434, -0.08184325695037842, 0.060906294733285904, -0.14784285426139832, -0.020488372072577477, -0.06302622705698013, 0.05361911281943321, -0.04827962443232536, -0.055441904813051224, -0.023296363651752472, 0.06566495448350906, 0.07606867700815201, -0.024615932255983353, 0.06625713407993317, -0.011096900328993797, 0.0710734948515892, 0.10509398579597473, 0.10135029256343842, -0.02957979217171669, -0.08712628483772278, -0.002189381280913949, -0.018693579360842705, 0.05267990380525589, -0.1385222226381302, 0.03747241199016571, 0.10546563565731049, 0.04074975848197937, 0.10949555784463882, 0.02678357996046543, -0.08530497550964355, -0.007346786092966795, 0.046848367899656296, -0.07899383455514908, -0.11799003928899765, -0.0012871518265455961, 0.06338878720998764, -0.15965865552425385, 0.013794693164527416, 0.13288868963718414, -0.026550747454166412, -0.020316973328590393, -0.010144833475351334, 0.027230024337768555, 0.00016816203424241394, 0.16364449262619019, 0.0409405380487442, 0.08821390569210052, -0.07004843652248383, 0.12134215235710144, 0.06283379346132278, -0.08910884708166122, 0.0520738884806633, 0.07537186145782471, -0.10387580841779709, -0.0029729369562119246, 0.03195420280098915, 0.05101862922310829, -0.017004305496811867, -0.03808216005563736, -0.06785834580659866, -0.12256874144077301, 0.04559924826025963, 0.08032334595918655, 0.02036212384700775, -0.01884213276207447, -0.005443142261356115, 0.011706247925758362, -0.12368172407150269, 0.08831384778022766, 0.03087298572063446, 0.07701487839221954, -0.13525928556919098, 0.0799410417675972, -0.0022802369203418493, 0.04216885566711426, -0.015328903682529926, 0.005354979541152716, -0.08979179710149765, -0.020431965589523315, -0.1362224966287613, -0.009258033707737923, -0.05010287091135979, 0.000792269769590348, -0.009599906392395496, -0.04464436694979668, -0.019348936155438423, 0.029352329671382904, -0.06857296079397202, -0.06601129472255707, -0.023037292063236237, 0.04635385423898697, -0.11656972765922546, -0.019244881346821785, 0.028005894273519516, -0.12812462449073792, 0.11589335650205612, 0.05521126091480255, 0.029905008152127266, -0.0029347767122089863, -0.0561499297618866, 0.0015982199693098664, 0.004630551673471928, 0.0024884906597435474, 0.052210550755262375, -0.13789789378643036, -0.029677169397473335, -0.05610549822449684, 0.00903787836432457, 0.016399679705500603, 0.053151123225688934, -0.12186174839735031, -0.028161006048321724, -0.05021322891116142, -0.039564650505781174, -0.059348221868276596, 0.051063280552625656, 0.058384720236063004, 0.031092867255210876, 0.12276668846607208, -0.07479219883680344, 0.053351618349552155, -0.16264137625694275, -0.03834903612732887, -0.01534962747246027, 0.004243571776896715, -0.03066478669643402, -0.016610940918326378, 0.08746685087680817, -0.05365384370088577, 0.12355896830558777, -0.05694881081581116, 0.086154505610466, 0.037619609385728836, -0.05018860101699829, 0.01285294909030199, 0.02845124900341034, 0.16688495874404907, 0.06176221743226051, -0.011608956381678581, 0.10450123250484467, -0.02824881672859192, 0.04467594251036644, 0.0653952807188034, 0.16471239924430847, 0.14453107118606567, 0.026946550235152245, 0.0622788667678833, 0.06616542488336563, -0.1271066665649414, -0.15005820989608765, 0.13246077299118042, -0.02306443266570568, 0.09461529552936554, -0.03622468188405037, 0.17335884273052216, 0.10973543673753738, -0.18922261893749237, 0.021480683237314224, -0.02727459743618965, -0.11185373365879059, -0.11127147078514099, -0.060039032250642776, -0.07192151248455048, -0.13503201305866241, 0.007831040769815445, -0.09097742289304733, 0.029685789719223976, 0.106021948158741, 0.018328355625271797, 0.04603414982557297, 0.14862915873527527, -0.006226696539670229, 0.004746988881379366, 0.07910118997097015, 0.044913798570632935, 0.021376265212893486, -0.006374666467308998, -0.09412035346031189, 0.04320492595434189, -0.0180347952991724, 0.06748607009649277, -0.05413752794265747, 0.010342922993004322, 0.05099884420633316, 0.005543888546526432, -0.08386197686195374, 0.028227664530277252, 0.0023935078643262386, 0.02003871090710163, 0.04148321598768234, 0.05832860991358757, -0.012152367271482944, -0.06069038063287735, 0.28676456212997437, -0.08981846272945404, -0.04192027077078819, -0.13252700865268707, 0.22386258840560913, 0.012948050163686275, -0.00491938553750515, 0.04775199294090271, -0.10995906591415405, -0.023258617147803307, 0.11921069025993347, 0.13627701997756958, -0.09202519059181213, -0.01191547978669405, -0.01621069386601448, -0.017738526687026024, -0.05255865678191185, 0.12779252231121063, 0.08005930483341217, -0.008568254299461842, -0.053732119500637054, 0.0019165454432368279, -0.005109188612550497, -0.04551243036985397, -0.10147487372159958, 0.08841133117675781, 0.010337477549910545, 0.011776178143918514, -0.042631953954696655, 0.07373423129320145, 0.017926421016454697, -0.15972156822681427, 0.035643815994262695, -0.12694936990737915, -0.1925448477268219, -0.04755654186010361, 0.017003178596496582, -0.016595255583524704, 0.05345078557729721, -0.000257848238106817, -0.019862979650497437, 0.14723293483257294, 0.0017720300238579512, -0.02036195993423462, -0.08205965161323547, 0.0814182460308075, -0.06480415165424347, 0.21947309374809265, 0.0033555671107023954, 0.04054191708564758, 0.09703636169433594, 0.0278131365776062, -0.1520587056875229, 0.018884390592575073, 0.10427939146757126, -0.07349102944135666, 0.034558024257421494, 0.17427948117256165, -0.04414869099855423, 0.10819920152425766, 0.054085202515125275, -0.10663867741823196, -0.012708619236946106, -0.0651036947965622, -0.024278702214360237, -0.06396938115358353, -0.008612058125436306, -0.04273022711277008, 0.16105124354362488, 0.18678319454193115, -0.05675996467471123, -0.011063601821660995, -0.05298365652561188, 0.010309400968253613, 0.042210619896650314, 0.08793202042579651, -0.01456852164119482, -0.19092878699302673, 0.023863373324275017, 0.03595956414937973, 0.043500129133462906, -0.25992730259895325, -0.08665163815021515, 0.025918366387486458, -0.051219116896390915, -0.04018337279558182, 0.12604662775993347, 0.02448297291994095, 0.02016819268465042, -0.037240706384181976, -0.1105092391371727, -0.03549932688474655, 0.14155340194702148, -0.16348512470722198, -0.029380619525909424 ]
null
null
peft
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # zephyr_outputs This model is a fine-tuned version of [HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.3066 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 10 - total_train_batch_size: 40 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 1.7024 | 1.0 | 9 | 1.2276 | | 0.8726 | 2.0 | 18 | 1.1114 | | 0.4336 | 3.0 | 27 | 1.1273 | | 0.1907 | 4.0 | 36 | 1.3390 | | 0.0789 | 5.0 | 45 | 1.3066 | ### Framework versions - PEFT 0.8.2 - Transformers 4.38.0.dev0 - Pytorch 2.0.1+cu117 - Datasets 2.16.1 - Tokenizers 0.15.2
{"license": "mit", "library_name": "peft", "tags": ["trl", "sft", "generated_from_trainer"], "base_model": "HuggingFaceH4/zephyr-7b-beta", "model-index": [{"name": "zephyr_outputs", "results": []}]}
null
lvcalucioli/zephyr_outputs
[ "peft", "safetensors", "mistral", "trl", "sft", "generated_from_trainer", "base_model:HuggingFaceH4/zephyr-7b-beta", "license:mit", "4-bit", "region:us" ]
2024-02-15T00:56:22+00:00
[]
[]
TAGS #peft #safetensors #mistral #trl #sft #generated_from_trainer #base_model-HuggingFaceH4/zephyr-7b-beta #license-mit #4-bit #region-us
zephyr\_outputs =============== This model is a fine-tuned version of HuggingFaceH4/zephyr-7b-beta on the None dataset. It achieves the following results on the evaluation set: * Loss: 1.3066 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.0005 * train\_batch\_size: 4 * eval\_batch\_size: 4 * seed: 42 * gradient\_accumulation\_steps: 10 * total\_train\_batch\_size: 40 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 5 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * PEFT 0.8.2 * Transformers 4.38.0.dev0 * Pytorch 2.0.1+cu117 * Datasets 2.16.1 * Tokenizers 0.15.2
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* gradient\\_accumulation\\_steps: 10\n* total\\_train\\_batch\\_size: 40\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.38.0.dev0\n* Pytorch 2.0.1+cu117\n* Datasets 2.16.1\n* Tokenizers 0.15.2" ]
[ "TAGS\n#peft #safetensors #mistral #trl #sft #generated_from_trainer #base_model-HuggingFaceH4/zephyr-7b-beta #license-mit #4-bit #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* gradient\\_accumulation\\_steps: 10\n* total\\_train\\_batch\\_size: 40\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.38.0.dev0\n* Pytorch 2.0.1+cu117\n* Datasets 2.16.1\n* Tokenizers 0.15.2" ]
[ 55, 140, 4, 44 ]
[ "passage: TAGS\n#peft #safetensors #mistral #trl #sft #generated_from_trainer #base_model-HuggingFaceH4/zephyr-7b-beta #license-mit #4-bit #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* gradient\\_accumulation\\_steps: 10\n* total\\_train\\_batch\\_size: 40\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.38.0.dev0\n* Pytorch 2.0.1+cu117\n* Datasets 2.16.1\n* Tokenizers 0.15.2" ]
[ -0.14137278497219086, 0.07696202397346497, -0.00299333268776536, 0.0933670923113823, 0.1373407393693924, 0.0206989124417305, 0.10784021764993668, 0.1133255586028099, -0.060070738196372986, 0.09337880462408066, 0.11632444709539413, 0.06267353892326355, 0.048209287226200104, 0.18082021176815033, -0.03660834953188896, -0.25579366087913513, 0.0029886469710618258, 0.0070506855845451355, -0.05669432878494263, 0.13012948632240295, 0.08259182423353195, -0.1284806877374649, 0.05344786122441292, -0.014441440813243389, -0.1445557326078415, -0.01570841670036316, -0.008307608775794506, -0.027106374502182007, 0.1214238628745079, 0.007932680658996105, 0.1264805644750595, 0.027550572529435158, 0.1071501225233078, -0.1947975903749466, 0.014225577935576439, 0.06483966112136841, 0.011813097633421421, 0.08337589353322983, 0.08473140001296997, -0.0033361585810780525, 0.12918633222579956, -0.10061205178499222, 0.055223409086465836, 0.011147244833409786, -0.13808976113796234, -0.27185946702957153, -0.10896129161119461, 0.0755428597331047, 0.11277631670236588, 0.07440251857042313, -0.018333906307816505, 0.11163700371980667, -0.07910964637994766, 0.06922239810228348, 0.262894868850708, -0.26797205209732056, -0.07862264662981033, 0.01060077827423811, 0.01884087547659874, 0.0894421860575676, -0.11627800762653351, -0.05111231282353401, 0.05402272567152977, 0.04868798702955246, 0.12863250076770782, 0.0036715478636324406, 0.040514763444662094, 0.004322725348174572, -0.16171589493751526, -0.05493250489234924, 0.09514886140823364, 0.04728927090764046, -0.042540278285741806, -0.047066930681467056, -0.05898550897836685, -0.19230613112449646, -0.047167591750621796, 0.013906741514801979, 0.022639349102973938, -0.042872339487075806, -0.057546112686395645, 0.021419981494545937, -0.08349096775054932, -0.08836901187896729, 0.007471705786883831, 0.1548222154378891, 0.0572846457362175, 0.0018773908959701657, 0.012327625416219234, 0.13716371357440948, -0.008696852251887321, -0.15026229619979858, 0.016122156754136086, 0.011232716031372547, -0.026248889043927193, -0.030277471989393234, -0.04389040917158127, 0.013184862211346626, 0.0061301179230213165, 0.15603908896446228, -0.12713368237018585, 0.05638907849788666, 0.044888705015182495, 0.03487315773963928, -0.10937627404928207, 0.11665870994329453, -0.06094883382320404, -0.024026785045862198, 0.003739774227142334, 0.13317587971687317, 0.030972428619861603, -0.006737965624779463, -0.06733252108097076, 0.00007184177957242355, 0.1093694269657135, 0.025602255016565323, -0.05375971645116806, 0.014856004156172276, -0.061759624630212784, -0.008928034454584122, 0.08379597216844559, -0.09696613997220993, 0.031129982322454453, 0.03077131323516369, -0.0683078020811081, -0.061459705233573914, 0.005973513238132, 0.00019229401368647814, -0.008270145393908024, 0.14130139350891113, -0.0957401916384697, 0.013391876593232155, -0.07919187098741531, -0.11453624814748764, 0.017369607463479042, -0.04925883561372757, 0.0037874095141887665, -0.07979793101549149, -0.13028450310230255, -0.049052104353904724, 0.02345149964094162, -0.04741545394062996, -0.04535062238574028, -0.043957848101854324, -0.08952102065086365, 0.0198507197201252, -0.020907383412122726, 0.1172618493437767, -0.06760187447071075, 0.1374196708202362, 0.019807634875178337, 0.05690397694706917, -0.007217101752758026, 0.03876977786421776, -0.07660876214504242, 0.05202973261475563, -0.18420685827732086, 0.03135378658771515, -0.06738384813070297, 0.05402419716119766, -0.1022162213921547, -0.12195371091365814, -0.028078215196728706, -0.028379643335938454, 0.13464751839637756, 0.1286783516407013, -0.1447526067495346, -0.07294667512178421, 0.20159852504730225, -0.10015618801116943, -0.1033901572227478, 0.1332065463066101, -0.02662615291774273, -0.02734627015888691, 0.03973877429962158, 0.17236243188381195, 0.09392843395471573, -0.11129507422447205, 0.011560767889022827, -0.03484329581260681, 0.11103351414203644, -0.013546252623200417, 0.09159346669912338, -0.014230318367481232, -0.00525273010134697, 0.006120920181274414, -0.07541951537132263, 0.07544775307178497, -0.11198615282773972, -0.07346673309803009, -0.021269578486680984, -0.08327487856149673, 0.0681120902299881, 0.064696304500103, 0.03183302283287048, -0.10702109336853027, -0.093013234436512, 0.026903798803687096, 0.11343999952077866, -0.06361225992441177, 0.021836943924427032, -0.037108127027750015, 0.12362456321716309, -0.05896664038300514, -0.031470730900764465, -0.17938446998596191, -0.0346778929233551, 0.019847948104143143, -0.012812555767595768, -0.012245239689946175, -0.03849734738469124, 0.08160675317049026, 0.10430382937192917, -0.06414767354726791, -0.06341450661420822, -0.0868505984544754, -0.012391922995448112, -0.12163791060447693, -0.2062528282403946, -0.06653245538473129, -0.02619626559317112, 0.14813357591629028, -0.2275659739971161, 0.02951415628194809, 0.003572357352823019, 0.10111398994922638, 0.02409498766064644, -0.04424476996064186, -0.02699178084731102, 0.09271343797445297, -0.00527948746457696, -0.06646797060966492, 0.06791934370994568, 0.005566832143813372, -0.07255210727453232, -0.023783976212143898, -0.1298774778842926, 0.13531479239463806, 0.10913094133138657, 0.014822456054389477, -0.10096505284309387, -0.06807373464107513, -0.07808834314346313, -0.0379110649228096, -0.04264569282531738, 0.020260756835341454, 0.12689438462257385, 0.020872460678219795, 0.1276550590991974, -0.0889853909611702, -0.03955798223614693, 0.03089223802089691, -0.014824222773313522, 0.028940590098500252, 0.13978935778141022, 0.07315534353256226, -0.05672875791788101, 0.126443549990654, 0.14613714814186096, -0.06372381001710892, 0.10996359586715698, -0.06727681308984756, -0.10400155931711197, -0.02985547110438347, 0.03727514669299126, 0.014065581373870373, 0.13833484053611755, -0.062424588948488235, 0.0036304867826402187, 0.01342342421412468, 0.02261381596326828, 0.005854841321706772, -0.23526562750339508, -0.045047685503959656, 0.02940133586525917, -0.0449863001704216, -0.04460185021162033, -0.011351948603987694, 0.003410955425351858, 0.11186062544584274, 0.009256678633391857, -0.07298525422811508, -0.013716340996325016, -0.004869664087891579, -0.08206302672624588, 0.20014628767967224, -0.08137790858745575, -0.08779153972864151, -0.10395403206348419, -0.0047131809405982494, -0.02218722365796566, -0.009018989279866219, 0.05096150562167168, -0.07193493098020554, -0.0155014768242836, -0.07912331074476242, -0.00589956808835268, 0.012516268528997898, 0.02496211789548397, -0.006435567047446966, -0.019442301243543625, 0.0713728666305542, -0.08815903216600418, -0.000050574479246279225, -0.03591352701187134, -0.02811434678733349, 0.05855327844619751, 0.03611091524362564, 0.10383016616106033, 0.14712277054786682, 0.0021995846182107925, 0.006935004144906998, -0.03512922674417496, 0.24155326187610626, -0.06866463273763657, -0.028836432844400406, 0.11355303972959518, -0.00666392594575882, 0.07870498299598694, 0.12037185579538345, 0.0626353770494461, -0.09497196972370148, 0.018080081790685654, 0.02258335053920746, -0.037512604147195816, -0.1914018839597702, -0.05120014771819115, -0.054350610822439194, -0.023960687220096588, 0.1174556091427803, 0.026791634038090706, -0.028077127411961555, 0.03973091021180153, -0.013908234424889088, 0.01627153530716896, -0.002354803029447794, 0.06961946934461594, 0.040092822164297104, 0.03376017138361931, 0.10920681804418564, -0.024403821676969528, -0.031129248440265656, 0.03572005406022072, -0.02376905456185341, 0.2419380247592926, -0.043986160308122635, 0.09241484850645065, 0.05801974982023239, 0.2056213915348053, -0.0008964711450971663, 0.0844574049115181, -0.007109592203050852, -0.03262859955430031, -0.0013784707989543676, -0.054404012858867645, -0.03847205638885498, 0.022027453407645226, -0.030715957283973694, 0.06477485597133636, -0.14375358819961548, -0.03544533625245094, 0.03251975402235985, 0.2927888035774231, 0.08383060991764069, -0.3188922107219696, -0.0992329865694046, -0.005435876082628965, -0.013184924609959126, -0.030763451009988785, 0.006082249339669943, 0.14611947536468506, -0.06681425124406815, 0.03576723858714104, -0.0703173279762268, 0.08230502903461456, -0.012478143908083439, 0.015303713269531727, 0.059013642370700836, 0.10052469372749329, -0.021666986867785454, 0.039395589381456375, -0.2374979555606842, 0.2799341380596161, -0.002183842472732067, 0.07096339762210846, -0.03231320530176163, -0.01355326920747757, 0.03953249007463455, 0.039619412273168564, 0.06663073599338531, 0.0038629346527159214, -0.03692971542477608, -0.22188754379749298, -0.10503984242677689, 0.02425563894212246, 0.10422193259000778, -0.049637917429208755, 0.1180163323879242, -0.021396096795797348, 0.0023359444458037615, 0.037932414561510086, -0.007477379869669676, -0.05533035844564438, -0.05122233182191849, -0.004424037411808968, -0.01860075816512108, -0.006494439207017422, -0.0974804013967514, -0.11359043419361115, -0.06581790745258331, 0.08470487594604492, -0.05060455948114395, -0.045307379215955734, -0.12210769951343536, 0.08864164352416992, 0.11774789541959763, -0.08203787356615067, 0.03506172075867653, 0.02840636670589447, 0.07906334102153778, 0.005549590568989515, -0.029346592724323273, 0.0899660736322403, -0.05733209103345871, -0.21151410043239594, -0.051472436636686325, 0.15249407291412354, 0.051437705755233765, 0.0686100423336029, -0.034537579864263535, 0.03335631266236305, -0.007064639590680599, -0.09907892346382141, 0.03210674971342087, 0.015674397349357605, 0.059933047741651535, 0.031600624322891235, -0.031775422394275665, 0.03778727725148201, -0.06330696493387222, -0.021000228822231293, 0.11222601681947708, 0.3337109386920929, -0.08193352818489075, 0.02038354054093361, 0.04330896958708763, -0.03790731355547905, -0.18503399193286896, 0.007589184679090977, 0.10165271162986755, 0.020381202921271324, 0.04230257123708725, -0.14632220566272736, 0.04073421657085419, 0.10873594135046005, -0.023630185052752495, 0.13456888496875763, -0.32549989223480225, -0.12398670613765717, 0.07655972987413406, 0.15105362236499786, 0.06945537030696869, -0.1679595708847046, -0.044417791068553925, 0.01722472719848156, -0.1291123777627945, 0.06357133388519287, -0.08106600493192673, 0.0964743122458458, -0.031230006366968155, 0.03941933438181877, 0.022516880184412003, -0.059829555451869965, 0.15472210943698883, -0.02583674155175686, 0.09327193349599838, -0.03601974621415138, 0.01409192569553852, 0.018048280850052834, -0.06378158926963806, 0.006925210356712341, -0.06862291693687439, 0.03311757370829582, -0.0988282784819603, -0.011645524762570858, -0.10563655197620392, 0.015049817971885204, -0.04509050026535988, -0.04284201189875603, -0.029640667140483856, 0.050945330411195755, 0.04196327179670334, -0.009816951118409634, 0.11030755192041397, 0.0011333584552630782, 0.18744948506355286, 0.09925451129674911, 0.037939704954624176, -0.025036601349711418, -0.05265950784087181, -0.012297861278057098, -0.02245856635272503, 0.05235214903950691, -0.16084738075733185, 0.010478210635483265, 0.14863263070583344, 0.052556123584508896, 0.12961000204086304, 0.06691120564937592, -0.06421640515327454, 0.011476690880954266, 0.07734257727861404, -0.1364612728357315, -0.10927890241146088, -0.005738690495491028, -0.01114862784743309, -0.13459894061088562, 0.014169584959745407, 0.09118160605430603, -0.07845747470855713, -0.024852856993675232, -0.008937296457588673, 0.02514490857720375, -0.051088761538267136, 0.2287413477897644, 0.06298815459012985, 0.06594584882259369, -0.09921687096357346, 0.07043351233005524, 0.04607217013835907, -0.09346350282430649, 0.0012649956624954939, 0.09395099431276321, -0.08665042370557785, -0.02462892420589924, 0.05335787311196327, 0.10711196064949036, -0.00651408638805151, -0.031185057014226913, -0.12600156664848328, -0.12326701730489731, 0.07850756496191025, 0.12295699119567871, 0.06791596859693527, 0.015267294831573963, -0.0012325822608545423, 0.020312249660491943, -0.11071477085351944, 0.09997356683015823, 0.06964638084173203, 0.07714217156171799, -0.12258227169513702, 0.17100635170936584, -0.004409297369420528, 0.02549135498702526, -0.010612420737743378, 0.032094284892082214, -0.11404383182525635, 0.013578920625150204, -0.13363049924373627, -0.014191074296832085, -0.043213795870542526, -0.0019200185779482126, -0.01679098978638649, -0.058077339082956314, -0.043107207864522934, 0.026057086884975433, -0.11290928721427917, -0.041270479559898376, 0.0008687659283168614, 0.03815319389104843, -0.11504891514778137, -0.03467787057161331, 0.03597525507211685, -0.09730874747037888, 0.07662249356508255, 0.044749315828084946, 0.04569469392299652, 0.05360366404056549, -0.10378123074769974, 0.022059302777051926, 0.051588743925094604, -0.015278303064405918, 0.03796154633164406, -0.12137196958065033, -0.007548733148723841, -0.03775890916585922, 0.012523403391242027, 0.014415103942155838, 0.05774587765336037, -0.1304355263710022, -0.007107038050889969, -0.02490011416375637, -0.05555844306945801, -0.04712865501642227, 0.018400471657514572, 0.06524742394685745, 0.05037263408303261, 0.1418098658323288, -0.08927832543849945, 0.04498915374279022, -0.2392284870147705, -0.024671748280525208, -0.02484199032187462, -0.07749352604150772, -0.084147147834301, -0.030310073867440224, 0.08805524557828903, -0.037524838000535965, 0.05787104740738869, -0.02236003987491131, 0.05558665096759796, 0.027025550603866577, -0.03641989454627037, 0.01785370148718357, 0.03181414306163788, 0.18798565864562988, 0.021196816116571426, -0.04932210221886635, 0.026768380776047707, 0.040477167814970016, 0.0682629868388176, 0.11216197162866592, 0.18209388852119446, 0.15206089615821838, 0.03921658918261528, 0.056509796530008316, 0.03062579222023487, -0.08673494309186935, -0.13954032957553864, 0.048372503370046616, -0.016323063522577286, 0.07985515147447586, -0.019506661221385002, 0.22222544252872467, 0.11007917672395706, -0.20178386569023132, 0.03584969416260719, -0.041822005063295364, -0.07621406018733978, -0.0975618064403534, -0.03580215200781822, -0.060764387249946594, -0.15181799232959747, 0.0002975847164634615, -0.10589855909347534, 0.013723261654376984, 0.10473368316888809, 0.010115713812410831, 0.01159158069640398, 0.15186846256256104, 0.06930144876241684, 0.017866915091872215, 0.07018112391233444, 0.028082070872187614, -0.010431709699332714, -0.03265712410211563, -0.10000821948051453, 0.04490792006254196, -0.06783351302146912, 0.04940422251820564, -0.059905532747507095, -0.07552210241556168, 0.04995771497488022, 0.01860741153359413, -0.10231295228004456, 0.025285936892032623, 0.01464917790144682, 0.06464977562427521, 0.09746132791042328, 0.027114829048514366, 0.020929386839270592, -0.02702830731868744, 0.22642141580581665, -0.07112693041563034, -0.0445355549454689, -0.10681512951850891, 0.28925764560699463, 0.035125065594911575, -0.020384518429636955, 0.03881359100341797, -0.09465617686510086, 0.025444133207201958, 0.13563907146453857, 0.1313234269618988, -0.07251814007759094, 0.008033770136535168, 0.004243350587785244, -0.01159600168466568, -0.04259544983506203, 0.11128076165914536, 0.13465027511119843, 0.06241263076663017, -0.09321761876344681, -0.00760673126205802, -0.07110683619976044, -0.015708277001976967, -0.03140866383910179, 0.027732672169804573, 0.02924508787691593, 0.0022438527084887028, -0.055912084877491, 0.07566149532794952, -0.035189948976039886, -0.1166907474398613, 0.09196352958679199, -0.20076307654380798, -0.17140163481235504, -0.022751059383153915, 0.03395860269665718, 0.02587791346013546, 0.06865397095680237, -0.02839241363108158, 0.006346386391669512, 0.09323533624410629, -0.03218027576804161, -0.03480523079633713, -0.11490198224782944, 0.08127693086862564, -0.07927784323692322, 0.19507098197937012, -0.03547009825706482, 0.03938350826501846, 0.1118726059794426, 0.05907205119729042, -0.10416585952043533, 0.0666687861084938, 0.06662499904632568, -0.08860524743795395, 0.0035937107168138027, 0.12951143085956573, -0.0412793792784214, 0.05661194398999214, 0.04741509258747101, -0.11530349403619766, 0.003540988778695464, -0.03597434237599373, -0.04192514345049858, -0.04710030555725098, -0.013025007210671902, -0.04188520088791847, 0.14029067754745483, 0.20399972796440125, -0.04806214198470116, 0.008286256343126297, -0.05183962360024452, 0.013302515260875225, 0.05919056013226509, 0.10497722029685974, -0.017444606870412827, -0.23828674852848053, 0.03176133707165718, 0.05228722468018532, -0.006136499345302582, -0.26488691568374634, -0.07834336906671524, 0.0035031328443437815, -0.060916513204574585, -0.0886736586689949, 0.11686032265424728, 0.03685416653752327, 0.055621590465307236, -0.046749547123909, -0.1148144006729126, -0.06792119145393372, 0.17465762794017792, -0.13698257505893707, -0.07626383006572723 ]
null
null
diffusers
# RVC_JonahJameson <Gallery /> ## Download model [Download](/coversia21/RVC_JonahJameson/tree/main) them in the Files & versions tab.
{"license": "openrail", "tags": ["text-to-image", "stable-diffusion", "lora", "diffusers", "template:sd-lora"], "widget": [{"text": "-", "output": {"url": "images/1_8W7Z4aMRw_etKmtmUpmxWw.jpg"}}], "base_model": "h94/IP-Adapter-FaceID"}
text-to-image
coversia21/RVC_JonahJameson
[ "diffusers", "text-to-image", "stable-diffusion", "lora", "template:sd-lora", "base_model:h94/IP-Adapter-FaceID", "license:openrail", "region:us" ]
2024-02-15T00:58:09+00:00
[]
[]
TAGS #diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-h94/IP-Adapter-FaceID #license-openrail #region-us
# RVC_JonahJameson <Gallery /> ## Download model Download them in the Files & versions tab.
[ "# RVC_JonahJameson\n\n<Gallery />", "## Download model\n\n\nDownload them in the Files & versions tab." ]
[ "TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-h94/IP-Adapter-FaceID #license-openrail #region-us \n", "# RVC_JonahJameson\n\n<Gallery />", "## Download model\n\n\nDownload them in the Files & versions tab." ]
[ 58, 14, 14 ]
[ "passage: TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-h94/IP-Adapter-FaceID #license-openrail #region-us \n# RVC_JonahJameson\n\n<Gallery />## Download model\n\n\nDownload them in the Files & versions tab." ]
[ -0.06943126022815704, 0.05244957283139229, -0.0014449325390160084, 0.031445689499378204, 0.06236153841018677, 0.04276648536324501, 0.18257656693458557, 0.07835233956575394, 0.059335269033908844, -0.010468242689967155, 0.0958445817232132, 0.09755333513021469, 0.020097441971302032, 0.14329171180725098, -0.057362619787454605, -0.21288825571537018, 0.0254999827593565, 0.0014304106589406729, -0.02228572778403759, 0.050864435732364655, 0.06348266452550888, -0.009373027831315994, 0.1306639164686203, -0.05867549404501915, -0.010275754146277905, 0.029708286747336388, -0.02961747720837593, -0.046809930354356766, -0.0034579739440232515, 0.0405878908932209, -0.013715533539652824, 0.14263661205768585, 0.13291151821613312, -0.09177940338850021, 0.058977555483579636, 0.02707676775753498, -0.11231935024261475, 0.04605339840054512, -0.000293569901259616, -0.03994501009583473, 0.2071482241153717, 0.04365270957350731, -0.01551925577223301, 0.028959069401025772, -0.041498973965644836, -0.09723833203315735, -0.03563440218567848, 0.08485794067382812, 0.07107449322938919, -0.009835032746195793, 0.05381447449326515, 0.041387829929590225, -0.03131728246808052, 0.04239858686923981, 0.22921840846538544, -0.20150917768478394, -0.03070301003754139, 0.285409539937973, 0.06686502695083618, 0.16889789700508118, -0.04458211362361908, 0.07819905877113342, 0.07723060250282288, -0.04625500738620758, -0.013057010248303413, -0.031305305659770966, 0.003390934783965349, 0.005372365936636925, -0.03159942850470543, 0.022074447944760323, 0.3630547523498535, 0.06980347633361816, 0.008450087159872055, -0.07019700109958649, -0.06891903281211853, 0.0754823088645935, -0.07839285582304001, 0.1032985970377922, 0.036288440227508545, 0.038994841277599335, -0.10296308994293213, -0.10703124850988388, -0.07126370072364807, -0.09104279428720474, -0.05042121186852455, 0.0844750925898552, 0.008486811071634293, 0.09730271995067596, -0.033174317330121994, 0.10683827847242355, -0.059535346925258636, -0.15803031623363495, 0.007275452837347984, -0.1081082671880722, 0.1455090492963791, 0.09884455054998398, -0.013263597153127193, -0.045945681631565094, 0.07741985470056534, 0.12823665142059326, 0.10915389657020569, -0.0030339988879859447, -0.05397920683026314, 0.14670802652835846, -0.019329240545630455, -0.06740063428878784, -0.0605078861117363, -0.04881305992603302, 0.06543038785457611, 0.016321290284395218, 0.10309461504220963, -0.05763620883226395, -0.17721164226531982, 0.053191300481557846, -0.222009539604187, 0.014872122555971146, 0.04518996924161911, 0.03328903391957283, -0.03821619600057602, -0.027880113571882248, 0.23027223348617554, 0.03308285400271416, -0.04589204117655754, -0.026850983500480652, -0.08080525696277618, 0.10047219693660736, 0.12655475735664368, -0.02008034847676754, 0.07263751327991486, 0.0867246463894844, -0.07875314354896545, -0.051803141832351685, -0.012506331317126751, -0.03629359230399132, 0.03741711378097534, -0.14436489343643188, 0.04665205627679825, -0.1467904895544052, -0.22269850969314575, 0.04203284904360771, 0.04809718206524849, -0.057195305824279785, 0.008935097604990005, -0.005014124792069197, -0.03540113568305969, 0.04155948758125305, -0.042052268981933594, -0.055941298604011536, -0.10821566730737686, 0.06755848228931427, -0.09010139107704163, 0.14501015841960907, -0.1342965066432953, 0.024203289300203323, -0.054111141711473465, 0.0041953194886446, -0.1531180888414383, -0.008778268471360207, -0.1676713079214096, 0.08018075674772263, -0.05212607607245445, -0.05041837692260742, -0.0725797787308693, 0.02500378154218197, -0.03038901463150978, 0.18365657329559326, -0.09345704317092896, -0.06106320396065712, 0.1337062567472458, -0.16572710871696472, -0.09007804840803146, 0.01587691903114319, 0.01663978397846222, 0.027769675478339195, 0.01089727133512497, 0.14610369503498077, 0.027784772217273712, -0.2998066842556, 0.06558024883270264, 0.15314920246601105, -0.051506608724594116, -0.12792706489562988, 0.07902573049068451, 0.03103821538388729, 0.10741814225912094, 0.03047202154994011, -0.09209764748811722, 0.11107359081506729, -0.03562111780047417, 0.005758522544056177, -0.0127717899158597, -0.0895102396607399, 0.0060234032571315765, 0.062038205564022064, 0.03605400398373604, -0.01931372843682766, -0.03903961926698685, -0.03803057596087456, 0.11612850427627563, -0.014045631512999535, 0.014408782124519348, -0.051072634756565094, 0.1235978901386261, -0.06951121240854263, -0.021968603134155273, -0.013515869155526161, -0.03325039893388748, -0.019173482432961464, 0.09004272520542145, 0.02580966241657734, 0.1373376101255417, 0.07367968559265137, -0.03163595125079155, -0.08209408819675446, -0.002680859761312604, 0.0360797718167305, 0.0014960516709834337, 0.04194368049502373, -0.17354241013526917, 0.024789493530988693, -0.03672007471323013, 0.03980063274502754, -0.13863380253314972, 0.022350631654262543, 0.07899601012468338, 0.15121039748191833, 0.051790013909339905, 0.013081836514174938, 0.04463043436408043, -0.06466854363679886, -0.05802580714225769, -0.027393803000450134, 0.07895082980394363, 0.016949670389294624, -0.041449110954999924, 0.2028527706861496, -0.015184311196208, 0.17033764719963074, 0.20119455456733704, -0.06366261094808578, 0.024991216138005257, -0.04107152670621872, -0.007858357392251492, 0.01867597922682762, -0.01238664798438549, -0.0488557331264019, -0.1146477684378624, -0.008458071388304234, 0.13164736330509186, -0.08604644238948822, 0.051244158297777176, 0.05657215788960457, -0.02967517450451851, 0.001098208362236619, 0.05271980166435242, 0.24991226196289062, -0.06617193669080734, 0.0937940776348114, 0.1535511016845703, -0.03590182587504387, 0.21482519805431366, -0.03339662402868271, -0.07772237807512283, 0.057413116097450256, -0.01194440945982933, 0.012836386449635029, 0.19514136016368866, 0.0038955153431743383, -0.035448431968688965, 0.05733602121472359, -0.016866246238350868, 0.015489012002944946, -0.09369613230228424, -0.07615862786769867, -0.0005772691802121699, -0.031731851398944855, -0.004479355178773403, 0.10987859964370728, -0.10716751962900162, 0.058871734887361526, -0.07357732951641083, -0.10674764961004257, 0.008711502887308598, -0.04770065099000931, -0.05061952397227287, 0.059147682040929794, -0.10134812444448471, -0.06210686266422272, -0.15819662809371948, -0.1027657762169838, -0.08109020441770554, 0.0005824652034789324, 0.055840108543634415, -0.054387737065553665, -0.054964009672403336, -0.06264470517635345, -0.02473963238298893, 0.026048874482512474, -0.07504060864448547, -0.03422747179865837, 0.030244335532188416, -0.052319347858428955, -0.16441068053245544, 0.01806139387190342, -0.04088278487324715, 0.017277048900723457, 0.05513982102274895, -0.1370924860239029, 0.12700557708740234, 0.043212179094552994, 0.04500799998641014, 0.04367242380976677, 0.003649358870461583, 0.10029419511556625, -0.041782014071941376, 0.06276798993349075, 0.1698443591594696, 0.07326839864253998, 0.02182736247777939, 0.09322158992290497, 0.07786844670772552, -0.05869010090827942, 0.01930946297943592, -0.04071451723575592, -0.11770496517419815, -0.1426066756248474, -0.11595932394266129, -0.10329419374465942, 0.08969858288764954, 0.0752451941370964, 0.0426415279507637, 0.049870558083057404, 0.1207485944032669, -0.038538459688425064, -0.030652349814772606, 0.07122017443180084, 0.08088867366313934, 0.09866508096456528, -0.041567638516426086, 0.05872833728790283, -0.10460861027240753, 0.014972222968935966, 0.17176516354084015, 0.013108794577419758, 0.127618208527565, -0.013766582123935223, 0.05851699039340019, 0.060633063316345215, 0.0738125815987587, 0.1537647843360901, 0.09909893572330475, -0.027391895651817322, -0.04076875001192093, -0.02941274084150791, -0.07703103125095367, 0.041963063180446625, 0.05468800663948059, 0.0037473184056580067, -0.1133275106549263, 0.05401057377457619, -0.0631360411643982, -0.035257309675216675, 0.0221906416118145, 0.07339593768119812, -0.2444823533296585, 0.1006026640534401, 0.04803560674190521, 0.10099011659622192, -0.012752382084727287, 0.09937505424022675, 0.10120043903589249, -0.06096721440553665, 0.11720814555883408, 0.01573212631046772, 0.12016448378562927, -0.021710917353630066, -0.08006815612316132, 0.014460972510278225, -0.01214798167347908, -0.012513820081949234, 0.016366567462682724, -0.02037818543612957, 0.15366466343402863, 0.015598103404045105, -0.028710996732115746, 0.025220153853297234, -0.05066812038421631, 0.11538051068782806, 0.17092901468276978, 0.10829855501651764, 0.0176888108253479, 0.02439168281853199, -0.018854599446058273, -0.09842339903116226, 0.010377931408584118, 0.045835573226213455, -0.03760121762752533, -0.06236845999956131, 0.013111358508467674, -0.028295142576098442, 0.0028018478769809008, 0.06527142226696014, -0.05649634823203087, -0.13163834810256958, -0.0073404693976044655, 0.12638606131076813, 0.03576734662055969, -0.048968005925416946, -0.043014366179704666, -0.09741422533988953, 0.009906238876283169, -0.015308245085179806, -0.10411408543586731, -0.06037398800253868, -0.06655870378017426, 0.04669524356722832, 0.012530799955129623, 0.05858875811100006, -0.03156774118542671, 0.0009710068115964532, -0.06683854013681412, -0.1346198320388794, 0.02931688167154789, -0.10390446335077286, -0.11194232106208801, -0.09740272164344788, 0.12529653310775757, -0.050511427223682404, -0.0004756932030431926, -0.02563263289630413, 0.030287660658359528, -0.028899258002638817, -0.10464415699243546, 0.012398803606629372, 0.029484663158655167, -0.026393838226795197, 0.06251579523086548, -0.03818412125110626, -0.00216603628359735, 0.07646433264017105, -0.01841782219707966, -0.002949920017272234, 0.2385273575782776, -0.06366229802370071, 0.07863141596317291, 0.1820588856935501, -0.01952168717980385, -0.2265569418668747, -0.10077746957540512, -0.08686970919370651, -0.08858513832092285, 0.05717751756310463, -0.11032940447330475, 0.10792849957942963, 0.05723196640610695, -0.07039374858140945, 0.18848735094070435, -0.23653936386108398, -0.07859979569911957, 0.03473469987511635, 0.0676531046628952, 0.2550694942474365, -0.1619381606578827, -0.06191471219062805, -0.07219336181879044, -0.34818172454833984, 0.048219677060842514, -0.11209257692098618, 0.04640189930796623, -0.012147301807999611, 0.021151741966605186, -0.04878843203186989, -0.00856944639235735, 0.13365279138088226, -0.04314243420958519, 0.050436392426490784, -0.09786096215248108, 0.05631203576922417, 0.18808385729789734, -0.006397712510079145, 0.018481280654668808, -0.229384183883667, 0.05365009605884552, -0.17246761918067932, 0.0037408864591270685, 0.002124364487826824, 0.03825821354985237, -0.017314018681645393, -0.04566018283367157, -0.0830821543931961, 0.00969691202044487, -0.0012014161329716444, 0.013541847467422485, 0.17059578001499176, -0.026324214413762093, 0.05613701790571213, 0.172649085521698, -0.04582547768950462, -0.03796335309743881, -0.09211935102939606, -0.10997386276721954, -0.04741313308477402, 0.09577338397502899, -0.2377949208021164, -0.05122395232319832, 0.0711820125579834, 0.05356042459607124, 0.04408049210906029, 0.020747780799865723, 0.003894323483109474, 0.12946292757987976, 0.1330554187297821, -0.09719385206699371, 0.042058445513248444, -0.02526463009417057, 0.035752926021814346, 0.1012306734919548, 0.06470289081335068, 0.10750307142734528, -0.06521391123533249, 0.045824140310287476, 0.029858041554689407, 0.023073436692357063, -0.04711258038878441, 0.027800178155303, 0.1219300925731659, -0.004502979107201099, -0.10086511075496674, 0.11612547188997269, -0.03175728768110275, -0.042829375714063644, -0.09352415800094604, 0.06898579001426697, -0.13350260257720947, -0.05520976334810257, -0.0427369624376297, 0.06490430980920792, -0.13063499331474304, -0.026992036029696465, -0.0689905509352684, -0.030742008239030838, -0.011167623102664948, 0.07589918375015259, 0.06850610673427582, -0.02625473029911518, 0.03831586614251137, -0.0055511766113340855, 0.010918626561760902, 0.019289376214146614, -0.00896016601473093, 0.06442654877901077, -0.13416890799999237, -0.23586168885231018, 0.014093528501689434, -0.006633236072957516, -0.09101997315883636, -0.014707179740071297, -0.09606152027845383, 0.011362479068338871, -0.07464867830276489, 0.06448836624622345, -0.09207800030708313, -0.010007910430431366, -0.060603924095630646, -0.07734202593564987, -0.06481355428695679, 0.008943592198193073, -0.06294698268175125, 0.015982041135430336, 0.049322258681058884, 0.07335561513900757, -0.08668119460344315, -0.018153339624404907, 0.017361728474497795, -0.051957108080387115, 0.05902387201786041, 0.019887851551175117, -0.07028085738420486, 0.014051085337996483, -0.17038117349147797, -0.05752396211028099, 0.029656121507287025, 0.05921470746397972, -0.002925639972090721, 0.15608322620391846, 0.05429510399699211, -0.007688325829803944, -0.0004255404928699136, -0.026623448356986046, -0.05501677095890045, -0.10522064566612244, 0.05007217824459076, -0.10406769067049026, 0.01658272184431553, -0.04302152246236801, 0.017953569069504738, 0.18058517575263977, 0.08811968564987183, 0.1138572245836258, -0.05838005244731903, 0.015247312374413013, -0.056625936180353165, 0.007862747646868229, 0.058520637452602386, -0.10439562797546387, 0.009663174860179424, -0.0486055426299572, -0.033311937004327774, -0.028590448200702667, 0.26259446144104004, 0.04078681021928787, -0.19747038185596466, 0.016229944303631783, 0.08468642085790634, 0.030281726270914078, -0.02708606980741024, 0.2447715550661087, 0.03887694701552391, 0.06991104036569595, -0.15784679353237152, 0.045408740639686584, 0.08369778841733932, -0.08451168984174728, -0.09777303040027618, 0.1363571584224701, -0.007070288993418217, 0.019637687131762505, 0.0950126126408577, -0.033529702574014664, 0.01944904215633869, 0.030020100995898247, 0.07792992889881134, 0.11229599267244339, -0.02941054478287697, 0.01331748254597187, 0.15858185291290283, -0.014406256377696991, -0.045869067311286926, 0.07286301255226135, 0.02645430900156498, -0.08172546327114105, -0.17414820194244385, -0.06037397310137749, -0.3208220601081848, 0.08218646049499512, -0.025689905509352684, 0.016051363199949265, 0.12913796305656433, 0.08328811079263687, 0.00459769181907177, -0.042425837367773056, -0.08915697783231735, -0.10347810387611389, 0.09276948869228363, -0.007918554358184338, -0.07929874956607819, -0.07025101780891418, -0.028381425887346268, 0.08631806820631027, -0.08308111131191254, -0.06515200436115265, 0.03006436862051487, 0.05257953330874443, 0.02336886338889599, -0.015432113781571388, -0.054311130195856094, -0.026778310537338257, 0.030545903369784355, -0.012545772828161716, 0.21611374616622925, 0.010734056122601032, 0.043892111629247665, 0.006495289970189333, 0.15171490609645844, -0.024202177301049232, -0.0739196315407753, -0.06638546288013458, -0.0653144046664238, -0.07098549604415894, 0.07530606538057327, -0.027731696143746376, -0.07685770094394684, -0.022563839331269264, 0.21338440477848053, 0.23010589182376862, -0.07347102463245392, 0.034202493727207184, 0.016693688929080963, -0.0003502614563331008, 0.01658417098224163, 0.018364407122135162, 0.041954994201660156, 0.19324122369289398, -0.007513327989727259, -0.05336839705705643, -0.06406963616609573, -0.01783699356019497, -0.026822874322533607, -0.081122487783432, 0.0072245835326612, -0.14426133036613464, -0.0634111687541008, 0.09246346354484558, -0.1355895847082138, -0.055544737726449966, 0.09915808588266373, -0.12553276121616364, 0.03167304769158363, -0.09896986931562424, 0.09296415001153946, 0.10105518996715546, -0.015794329345226288, -0.09522788226604462, -0.028276173397898674, -0.006639229133725166, -0.021359939128160477, -0.14850345253944397, -0.10523145645856857, 0.00013450377446133643, -0.20871445536613464, 0.11891131103038788, -0.07738813757896423, 0.020313438028097153, 0.010363365523517132, 0.0015687012346461415, -0.03606106713414192, 0.0534757524728775, 0.020511629059910774, -0.1549564152956009, -0.09372735023498535, 0.03362589329481125, -0.04111894220113754, 0.09708303958177567, 0.015021990053355694, -0.06843272596597672, 0.02052328549325466, 0.15889346599578857, -0.07734276354312897, -0.10032612830400467, 0.005973944906145334, -0.08957918733358383, 0.09530527144670486, -0.011480724439024925, 0.0012383974390104413, -0.09755201637744904, -0.027121448889374733, 0.039024002850055695, 0.12420821934938431, -0.17495973408222198, 0.09790875017642975, -0.010919323191046715, -0.0913856029510498, 0.014155704528093338, 0.06143808737397194, -0.11212804913520813, 0.03892508149147034, -0.17263129353523254, -0.0006844925228506327, -0.0013592757750302553, 0.033402908593416214, 0.19469596445560455, -0.0009336582734249532, -0.014093669131398201, -0.15862654149532318, 0.06021393835544586, 0.030548229813575745, -0.09702280163764954, -0.0862378478050232 ]
null
null
peft
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.8.2
{"library_name": "peft", "base_model": "Viet-Mistral/Vistral-7B-Chat"}
null
longcule123/adapter-14-2
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Viet-Mistral/Vistral-7B-Chat", "region:us" ]
2024-02-15T01:02:15+00:00
[ "1910.09700" ]
[]
TAGS #peft #safetensors #arxiv-1910.09700 #base_model-Viet-Mistral/Vistral-7B-Chat #region-us
# Model Card for Model ID ## Model Details ### Model Description - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact ### Framework versions - PEFT 0.8.2
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ "TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-Viet-Mistral/Vistral-7B-Chat #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ 40, 6, 3, 54, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4, 11 ]
[ "passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-Viet-Mistral/Vistral-7B-Chat #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2" ]
[ -0.12140171229839325, 0.1996288150548935, -0.002685773652046919, 0.029247287660837173, 0.0854666605591774, 0.020799465477466583, 0.045453328639268875, 0.13058903813362122, 0.010872160084545612, 0.10369975864887238, 0.06859787553548813, 0.11098543554544449, 0.11239472031593323, 0.21304285526275635, 0.004450436215847731, -0.1745002120733261, 0.02866232395172119, -0.0897039845585823, 0.0044947424903512, 0.12609374523162842, 0.14543266594409943, -0.09981926530599594, 0.08291763067245483, -0.013929805718362331, -0.002055539283901453, -0.0383547805249691, -0.06647428125143051, -0.01653502695262432, 0.04440229758620262, 0.034578561782836914, 0.05856473743915558, -0.00970715843141079, 0.09121838212013245, -0.2568903863430023, 0.020260317251086235, 0.04398832470178604, 0.006071143317967653, 0.08759643882513046, 0.09654129296541214, -0.035593435168266296, 0.11848976463079453, -0.029184922575950623, 0.14039507508277893, 0.09117726981639862, -0.08722258359193802, -0.22106707096099854, -0.06760361045598984, 0.07795003801584244, 0.19125211238861084, 0.0810496062040329, -0.04232846200466156, 0.13710665702819824, -0.06996362656354904, 0.026338007301092148, 0.036162540316581726, -0.08454577624797821, -0.07017991691827774, 0.0677034929394722, 0.12779705226421356, 0.06350866705179214, -0.12449177354574203, -0.03513680770993233, 0.029265446588397026, 0.03784489631652832, 0.06407812982797623, 0.009263482876121998, 0.16365839540958405, 0.02701052464544773, -0.14242784678936005, -0.04820939525961876, 0.14901210367679596, 0.02045786939561367, -0.04289337247610092, -0.22641831636428833, -0.0055978368036448956, -0.09003156423568726, -0.025992443785071373, -0.05627164617180824, 0.03092438541352749, 0.012004666030406952, 0.12351442128419876, -0.04285154491662979, -0.09451472014188766, -0.021432537585496902, 0.09541884064674377, 0.038220636546611786, 0.023475531488656998, -0.020127376541495323, 0.00817085150629282, 0.12301617115736008, 0.08390626311302185, -0.13206104934215546, -0.0643700435757637, -0.0804099589586258, -0.048097070306539536, -0.04145583510398865, 0.04350230470299721, 0.03478552773594856, 0.06396618485450745, 0.25539758801460266, -0.022000029683113098, 0.06534954905509949, 0.06569833308458328, 0.014136167243123055, 0.049107424914836884, 0.10443467646837234, -0.03820338100194931, -0.16123418509960175, -0.011152511462569237, 0.09760474413633347, -0.005995761603116989, -0.028723053634166718, -0.04779423400759697, 0.03533463180065155, 0.03770200163125992, 0.11234533786773682, 0.11136460304260254, -0.014614148065447807, -0.07538284361362457, -0.0658780038356781, 0.21729516983032227, -0.15683338046073914, 0.0454099215567112, 0.025797449052333832, -0.008632282726466656, -0.043065816164016724, 0.005697535816580057, 0.01633268967270851, -0.028755255043506622, 0.06178741902112961, -0.06519939750432968, -0.04203398898243904, -0.1280553638935089, -0.02809559367597103, 0.029268385842442513, 0.013044212013483047, -0.041874825954437256, -0.041122741997241974, -0.0813145861029625, -0.1049538254737854, 0.11088553816080093, -0.057920925319194794, -0.05927994102239609, -0.03126302734017372, -0.09342576563358307, 0.02484290860593319, 0.02899276278913021, 0.07629222422838211, -0.02674579806625843, 0.045166417956352234, -0.012311708182096481, 0.0617121122777462, 0.07634655386209488, 0.028596004471182823, -0.07608471810817719, 0.06143266707658768, -0.20020845532417297, 0.08498112112283707, -0.08287771046161652, 0.03804438188672066, -0.16377763450145721, -0.008559136651456356, 0.020300207659602165, 0.024427078664302826, 0.033310942351818085, 0.1640872061252594, -0.223727747797966, -0.026182027533650398, 0.1536749005317688, -0.10926694422960281, -0.12663504481315613, 0.04130322113633156, -0.04045126959681511, 0.1753060668706894, 0.02905443124473095, 0.0027341332752257586, 0.10754626244306564, -0.1651504784822464, -0.026169566437602043, -0.021064262837171555, -0.0039011964108794928, 0.08618078380823135, 0.08508234471082687, -0.08516842871904373, 0.009165518917143345, 0.015477876178920269, -0.05846118927001953, -0.01406903751194477, -0.04116612672805786, -0.10679883509874344, 0.004673507064580917, -0.08388800173997879, 0.02230624295771122, -0.0015398967079818249, -0.0942300334572792, -0.006296060048043728, -0.1561613231897354, -0.053208786994218826, 0.08923131227493286, 0.0037225193809717894, -0.025666022673249245, -0.10913068056106567, 0.05110194906592369, -0.02986770309507847, -0.02128872647881508, -0.13891030848026276, -0.023833010345697403, 0.02044750563800335, -0.14259950816631317, -0.010823407210409641, -0.11183588206768036, 0.06778913736343384, 0.005426047835499048, -0.05040513351559639, -0.043891314417123795, 0.00005681530456058681, 0.0034284135326743126, -0.05229370296001434, -0.23952442407608032, -0.029328711330890656, -0.05353354290127754, 0.16000612080097198, -0.2212785929441452, 0.042650870978832245, 0.02844754233956337, 0.12341764569282532, 0.0025756764225661755, -0.06657850742340088, 0.021354427561163902, -0.0734001025557518, -0.02456078678369522, -0.07532016932964325, -0.003003099001944065, 0.000914686534088105, -0.028069064021110535, 0.01648116298019886, -0.1102515310049057, -0.05876988545060158, 0.10013826191425323, 0.06427323073148727, -0.15206581354141235, 0.0039567300118505955, -0.041764236986637115, -0.059390679001808167, -0.07272350043058395, -0.07224766910076141, 0.08919842541217804, 0.056044433265924454, 0.0368100181221962, -0.07057216763496399, -0.07209638506174088, 0.007629978470504284, -0.023844093084335327, -0.01380052138119936, 0.11541832238435745, 0.07140751928091049, -0.10882892459630966, 0.0898965373635292, 0.069275863468647, 0.030122097581624985, 0.08105552196502686, -0.028384028002619743, -0.1013713926076889, -0.032042428851127625, 0.053069405257701874, 0.010286478325724602, 0.17401395738124847, -0.07367897033691406, 0.055050693452358246, 0.04611266404390335, -0.04116055369377136, 0.04700435698032379, -0.08446896076202393, 0.01146949827671051, 0.003319385228678584, -0.014752005226910114, 0.03114120475947857, -0.021028390154242516, 0.00835016742348671, 0.07627564668655396, 0.052324797958135605, 0.028222380205988884, 0.02147953398525715, -0.036995451897382736, -0.1406952440738678, 0.17802509665489197, -0.0988021045923233, -0.24273580312728882, -0.15937398374080658, 0.06389971077442169, 0.05078127980232239, -0.015630539506673813, 0.022470748052001, -0.05795908346772194, -0.10539022088050842, -0.08473992347717285, 0.004364791326224804, 0.03551246225833893, -0.056366242468357086, -0.06758508086204529, 0.04414505511522293, 0.04482196271419525, -0.12203235924243927, 0.030440595000982285, 0.06563146412372589, -0.020631937310099602, -0.0013604846317321062, 0.05356726422905922, 0.08566778153181076, 0.18531592190265656, -0.004273437429219484, 0.0023420231882482767, 0.062050193548202515, 0.2840237021446228, -0.15735939145088196, 0.12300650775432587, 0.14043428003787994, -0.06530723720788956, 0.07276659458875656, 0.18698835372924805, 0.02757974900305271, -0.09697845578193665, 0.02588566765189171, 0.02651793137192726, -0.019485831260681152, -0.2667730152606964, -0.05803757160902023, -0.01689564436674118, -0.08576837927103043, 0.07342088967561722, 0.08853690326213837, 0.0829588770866394, 0.037367966026067734, -0.0652301013469696, -0.10410547256469727, 0.03347534313797951, 0.10536608099937439, -0.017498433589935303, 0.002891843905672431, 0.08134210109710693, -0.04220569133758545, 0.012078248895704746, 0.09143298119306564, -0.01966429315507412, 0.14899444580078125, 0.05574857443571091, 0.10264859348535538, 0.0792556032538414, 0.09858350455760956, -0.006479883100837469, 0.03284841403365135, 0.014936949126422405, 0.025222769007086754, 0.020120199769735336, -0.08479912579059601, 0.014658810570836067, 0.1074795126914978, 0.035428423434495926, 0.02606106735765934, 0.021256664767861366, -0.04265487939119339, 0.0454370342195034, 0.18824486434459686, 0.023020580410957336, -0.2065679132938385, -0.08585014939308167, 0.055482350289821625, -0.07993795722723007, -0.1533515900373459, -0.012806334532797337, 0.032344646751880646, -0.16468888521194458, 0.018896665424108505, -0.04089738801121712, 0.10386265069246292, -0.08831246197223663, -0.04042929410934448, 0.11345794796943665, 0.055337049067020416, -0.015935160219669342, 0.04677500203251839, -0.18564487993717194, 0.1079627126455307, 0.028686456382274628, 0.07856442034244537, -0.08634798973798752, 0.10170068591833115, 0.0008308578981086612, -0.011965720914304256, 0.1649085283279419, 0.005654322449117899, -0.049547333270311356, -0.07776187360286713, -0.09971585869789124, -0.005121064838021994, 0.07930052280426025, -0.1334974616765976, 0.07504528015851974, -0.03378620371222496, -0.028863485902547836, -0.008269481360912323, -0.08668395131826401, -0.13609841465950012, -0.16255056858062744, 0.05355598032474518, -0.10016095638275146, 0.023583268746733665, -0.08554266393184662, -0.0527990497648716, 0.007138650398701429, 0.18128862977027893, -0.22105713188648224, -0.10718805342912674, -0.1461164504289627, -0.11072908341884613, 0.16221396625041962, -0.040649473667144775, 0.08355574309825897, 0.0008875716011971235, 0.16423599421977997, 0.0109701594337821, -0.0167712289839983, 0.09075432270765305, -0.09449305385351181, -0.18709802627563477, -0.05307750776410103, 0.16303491592407227, 0.14550264179706573, 0.029919244349002838, -0.007599715143442154, 0.027966538444161415, -0.06217687577009201, -0.1191592589020729, 0.02584492601454258, 0.16889643669128418, 0.06025314703583717, -0.019540660083293915, -0.021393582224845886, -0.11318079382181168, -0.05972632020711899, -0.039860863238573074, -0.011405568569898605, 0.19359667599201202, -0.06896117329597473, 0.15481042861938477, 0.10825737565755844, -0.057591404765844345, -0.2098388373851776, 0.03888315707445145, 0.04977087676525116, 0.02127411589026451, 0.04138597100973129, -0.1863064169883728, 0.09062792360782623, -0.013610406778752804, -0.08054398000240326, 0.16408313810825348, -0.16853953897953033, -0.13511410355567932, 0.1044762060046196, 0.025110026821494102, -0.21466077864170074, -0.13316631317138672, -0.09946936368942261, -0.01880308985710144, -0.13308514654636383, 0.04869459196925163, 0.00564036937430501, 0.0035910343285650015, 0.021390864625573158, 0.012868966907262802, 0.033493030816316605, -0.05172407627105713, 0.2110792100429535, -0.036133285611867905, -0.00012843671720474958, -0.049878817051649094, -0.08089101314544678, 0.026608414947986603, -0.05181719362735748, 0.11327837407588959, -0.002463910961523652, 0.03293551504611969, -0.16258499026298523, -0.03897335007786751, -0.05402439832687378, 0.0348532572388649, -0.09134560078382492, -0.08267658203840256, -0.04306390881538391, 0.09412466734647751, 0.09511726349592209, -0.023314641788601875, 0.00009480959124630317, -0.08904723078012466, 0.06446263939142227, 0.20621053874492645, 0.1964423954486847, 0.06613453477621078, -0.054598335176706314, 0.024959085509181023, -0.032794494181871414, 0.046619441360235214, -0.20953883230686188, 0.042191650718450546, 0.059372443705797195, 0.01804354600608349, 0.06947237998247147, -0.0129817770794034, -0.15375259518623352, -0.07473690807819366, 0.08322246372699738, -0.057340461760759354, -0.1727190464735031, -0.029810788109898567, 0.024720804765820503, -0.2059432715177536, -0.04173646122217178, 0.033822666853666306, -0.01468494813889265, -0.03864968568086624, 0.02272002585232258, 0.08130394667387009, -0.023910703137516975, 0.09875325113534927, 0.0798734650015831, 0.09315463900566101, -0.10403429716825485, 0.05676693096756935, 0.07293252646923065, -0.041745129972696304, 0.030413631349802017, 0.11752652376890182, -0.04929342493414879, -0.04154818877577782, 0.07834380120038986, 0.10979950428009033, 0.013254201039671898, -0.054075997322797775, 0.010844049975275993, -0.049409545958042145, 0.05536507070064545, 0.0938677042722702, 0.030804138630628586, 0.005809402093291283, 0.06659463793039322, 0.03393374755978584, -0.08928819745779037, 0.11723754554986954, 0.059764400124549866, 0.019269829615950584, -0.05834539979696274, -0.042147595435380936, -0.015383705496788025, -0.016226164996623993, -0.019659316167235374, -0.0071244873106479645, -0.08376505225896835, -0.004261997994035482, -0.10972808301448822, 0.02174658142030239, -0.08058100938796997, 0.007632415276020765, 0.034754928201436996, -0.04893473535776138, 0.0034708515740931034, 0.0007130385492928326, -0.07215137034654617, -0.05719831958413124, -0.014260835014283657, 0.07996530085802078, -0.13224442303180695, 0.042190779000520706, 0.07592333853244781, -0.10856133699417114, 0.06985551118850708, -0.0032452407758682966, 0.008484702557325363, 0.0028293095529079437, -0.14569728076457977, 0.052403394132852554, -0.02580653876066208, -0.004745488986372948, 0.015609911642968655, -0.2000858187675476, -0.008540639653801918, -0.03470335528254509, -0.06450921297073364, 0.013455076143145561, -0.006125046871602535, -0.11855901032686234, 0.10617338865995407, 0.004685448482632637, -0.06008615344762802, -0.023406794294714928, 0.04029490426182747, 0.09822620451450348, -0.010627121664583683, 0.1329609900712967, -0.027220485731959343, 0.0733216255903244, -0.17580360174179077, -0.007634709123522043, -0.0125382449477911, 0.05296187847852707, -0.022741321474313736, -0.032708387821912766, 0.061248231679201126, -0.02180355042219162, 0.16758792102336884, -0.006902157329022884, 0.07010539621114731, 0.053951606154441833, 0.00940990261733532, 0.026810355484485626, 0.07991054654121399, 0.05916164442896843, -0.00938441976904869, 0.0016061511123552918, 0.038850944489240646, -0.006934515666216612, -0.051438603550195694, -0.16005392372608185, 0.05878189206123352, 0.16005302965641022, 0.05419100821018219, 0.026895418763160706, 0.017228612676262856, -0.11425723135471344, -0.07545158267021179, 0.1177670806646347, -0.02470051683485508, -0.03056151233613491, -0.06843115389347076, 0.188817098736763, 0.13671961426734924, -0.20005854964256287, 0.07108461856842041, -0.05932101234793663, -0.04659246280789375, -0.14023378491401672, -0.175625279545784, -0.058310866355895996, -0.05486767366528511, -0.02868063934147358, -0.057045865803956985, 0.05057992786169052, 0.03460799157619476, 0.001578762661665678, -0.021884800866246223, 0.10130146890878677, 0.0198769923299551, -0.027918711304664612, 0.04581545665860176, 0.060270968824625015, 0.0355856716632843, -0.09240631759166718, 0.010182301513850689, 0.0016920811031013727, 0.023191548883914948, 0.06728342920541763, 0.021998250856995583, -0.0660434141755104, 0.0257414560765028, -0.01975911855697632, -0.12305665761232376, 0.04115123301744461, -0.011946647427976131, -0.04131697490811348, 0.15035617351531982, 0.04134224355220795, 0.007536596152931452, -0.01823064126074314, 0.22907505929470062, -0.08023092150688171, -0.07742613554000854, -0.15012292563915253, 0.05612820386886597, -0.07369625568389893, 0.02989843115210533, 0.030745480209589005, -0.11943057924509048, 0.00962632242590189, 0.16664361953735352, 0.12330155074596405, -0.010571794584393501, 0.007678080350160599, 0.04456869512796402, 0.003708271309733391, -0.044946689158678055, 0.02043893374502659, 0.04892880842089653, 0.1815471202135086, -0.0724794864654541, 0.06229013204574585, -0.013828114606440067, -0.08458880335092545, -0.017309753224253654, 0.08938106149435043, -0.010677346028387547, -0.00007220871339086443, -0.06379398703575134, 0.1495051383972168, -0.08133602887392044, -0.21546892821788788, 0.060935042798519135, -0.05918857827782631, -0.13724619150161743, -0.043350450694561005, 0.044285181909799576, -0.021160870790481567, 0.004746788181364536, 0.06840398907661438, -0.04342907294631004, 0.19166730344295502, 0.029359783977270126, -0.0463566891849041, -0.09055039286613464, 0.06004186347126961, -0.15500584244728088, 0.27908793091773987, 0.02221224643290043, 0.0553024522960186, 0.10818982124328613, -0.02120942249894142, -0.1516295075416565, 0.005266777705401182, 0.10748716443777084, -0.07201637327671051, 0.06458400934934616, 0.16982518136501312, 0.00629363814368844, 0.12980881333351135, 0.06180741637945175, -0.051081154495477676, 0.03321162238717079, -0.08993881195783615, -0.04347948729991913, -0.11812365055084229, 0.08511599898338318, -0.0885571837425232, 0.15917131304740906, 0.11743192374706268, -0.07028013467788696, 0.007001075893640518, -0.019600598141551018, 0.08464017510414124, 0.010971790179610252, 0.1136329397559166, 0.009725179523229599, -0.19502851366996765, 0.03495243191719055, 0.008341696113348007, 0.10137444734573364, -0.18354253470897675, -0.05594315752387047, 0.038479823619127274, -0.01961253024637699, -0.07385113835334778, 0.12024649232625961, 0.034318290650844574, 0.029738325625658035, -0.03702659532427788, -0.02529929019510746, 0.008096402511000633, 0.1478181630373001, -0.11028417199850082, -0.01337889488786459 ]
null
null
nemo
# CHiME8 DASR NeMo Baseline Models - The model files in this repository are the models used in this paper [The CHiME-7 Challenge: System Description and Performance of NeMo Team’s DASR System](https://arxiv.org/pdf/2310.12378.pdf). - These models are needed to execute the CHiME8-DASR baseline [CHiME8-DASR-Baseline NeMo](https://github.com/chimechallenge/C8DASR-Baseline-NeMo/tree/main/scripts/chime8) - VAD, Diarization and ASR models are all based on [NVIDIA NeMo Conversational AI Toolkits](https://github.com/NVIDIA/NeMo). ## 1. Voice Activity Detection (VAD) Model: ### **[**MarbleNet_frame_VAD_chime7_Acrobat.nemo**](https://huggingface.co/chime-dasr/nemo_baseline_models/blob/main/MarbleNet_frame_VAD_chime7_Acrobat.nemo)** - This model is based on [NeMo MarbleNet VAD model](https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/main/asr/speech_classification/models.html#marblenet-vad). - For validation, we use dataset comprises the CHiME-6 development subset as well as 50 hours of simulated audio data. - The simulated data is generated using the [NeMo multi-speaker data simulator](https://github.com/NVIDIA/NeMo/blob/main/tutorials/tools/Multispeaker_Simulator.ipynb) on [VoxCeleb1&2 datasets](https://www.robots.ox.ac.uk/~vgg/data/voxceleb/vox1.html) - The multi-speaker data simulation results in a total of 2,000 hours of audio, of which approximately 30% is silence. - The Model training incorporates [SpecAugment](https://arxiv.org/abs/1904.08779) and noise augmentation through [MUSAN noise dataset](https://arxiv.org/abs/1510.08484). ## 2. Speaker Diarization Model: Multi-scale Diarization Decoder (MSDD-v2) ### **[**MSDD_v2_PALO_100ms_intrpl_3scales.nemo**](https://huggingface.co/chime-dasr/nemo_baseline_models/blob/main/MSDD_v2_PALO_100ms_intrpl_3scales.nemo)** Our DASR system is based on the speaker diarization system using the multi-scale diarization decoder (MSDD). - MSDD Reference: [Park et al. (2022)](https://arxiv.org/pdf/2203.15974.pdf) - MSDD-v2 speaker diarization system employs a multi-scale embedding approach and utilizes TitaNet speaker embedding extractor. - TitaNet Reference: [Koluguri et al. (2022)](https://arxiv.org/abs/2110.04410) - TitaNet Model is included in [MSDD-v2 .nemo checkpoint file](https://huggingface.co/chime-dasr/nemo_baseline_models/blob/main/MSDD_v2_PALO_100ms_intrpl_3scales.nemo). - Unlike the system that uses a multi-layer LSTM architecture, we employ a four-layer Transformer architecture with a hidden size of 384. - This neural model generates logit values indicating speaker existence. - Our diarization model is trained on approximately 3,000 hours of simulated audio mixture data from the same multi-speaker data simulator used in VAD model training, drawing from VoxCeleb1&2 and LibriSpeech datasets. - LibriSpeech Reference: [OpenSLR Download](https://www.openslr.org/12),[LibriSpeech, Panayotov et al. (2015)](https://ieeexplore.ieee.org/document/7178964) - MUSAN noise is also used for adding additive background noise, focusing on music and broadband noise. ## 3. Automatic Speech Recognition (ASR) model ### **[**FastConformerXL-RNNT-chime7-GSS-finetuned.nemo**](https://huggingface.co/chime-dasr/nemo_baseline_models/blob/main/FastConformerXL-RNNT-chime7-GSS-finetuned.nemo)** - This ASR model is based on [NeMo FastConformer XL model](https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/main/asr/models.html#fast-conformer). - Single-channel audio generated using a multi-channel front-end (Guided Source Separation, GSS) is transcribed using a 0.6B parameter Conformer-based transducer (RNNT) model. - Model Reference: [Gulati et al. (2020)](https://arxiv.org/abs/2005.08100) - The model was initialized using a publicly available NeMo checkpoint. - NeMo Checkpoint: [NGC Model Card: Conformer Transducer XL](https://catalog.ngc.nvidia.com/orgs/nvidia/teams/nemo/models/stt_en_conformer_transducer_xlarge) - This model was then fine-tuned on the CHiME-7 train and dev set, which includes the CHiME-6 and Mixer6 training subsets, after processing the data through the multi-channel ASR front-end, utilizing ground-truth diarization. - Fine-Tuning Details: - Fine-tuning Duration: 35,000 updates - Batch Size: 128 ## 4. Language Model for ASR Decoding: KenLM Model ### **[**ASR_LM_chime7_only.kenlm**](https://huggingface.co/chime-dasr/nemo_baseline_models/blob/main/ASR_LM_chime7_only.kenlm)** - This KenLM model is trained solely on CHiME7-DASR datasets (Mixer6, CHiME6, DipCo). - We apply a word-piece level N-gram language model using byte-pair-encoding (BPE) tokens. - This approach utilizes the SentencePiece and KenLM toolkits, based on the transcription of CHiME-7 train and dev sets. - SentencePiece: [Kudo and Richardson (2018)](https://arxiv.org/abs/1808.06226) - KenLM: [KenLM GitRepo](https://github.com/kpu/kenlm) - The token sets of our ASR and LM models were matched to ensure consistency. - To combine several N-gram models with equal weights, we used the OpenGrm library. - OpenGrm: [Roark et al. (2012)](https://aclanthology.org/P12-3011/) - MAES decoding was employed for the transducer, which accelerates the decoding process. - MAES Decoding: [Kim et al. (2020)](https://ieeexplore.ieee.org/document/9250505) - As expected, integrating the beam-search decoder with the language model significantly enhances the performance of the end-to-end model compared to its pure counterpart.
{"library_name": "nemo"}
null
chime-dasr/nemo_baseline_models
[ "nemo", "arxiv:2310.12378", "arxiv:1904.08779", "arxiv:1510.08484", "arxiv:2203.15974", "arxiv:2110.04410", "arxiv:2005.08100", "arxiv:1808.06226", "region:us" ]
2024-02-15T01:03:13+00:00
[ "2310.12378", "1904.08779", "1510.08484", "2203.15974", "2110.04410", "2005.08100", "1808.06226" ]
[]
TAGS #nemo #arxiv-2310.12378 #arxiv-1904.08779 #arxiv-1510.08484 #arxiv-2203.15974 #arxiv-2110.04410 #arxiv-2005.08100 #arxiv-1808.06226 #region-us
# CHiME8 DASR NeMo Baseline Models - The model files in this repository are the models used in this paper The CHiME-7 Challenge: System Description and Performance of NeMo Team’s DASR System. - These models are needed to execute the CHiME8-DASR baseline CHiME8-DASR-Baseline NeMo - VAD, Diarization and ASR models are all based on NVIDIA NeMo Conversational AI Toolkits. ## 1. Voice Activity Detection (VAD) Model: ### MarbleNet_frame_VAD_chime7_Acrobat.nemo - This model is based on NeMo MarbleNet VAD model. - For validation, we use dataset comprises the CHiME-6 development subset as well as 50 hours of simulated audio data. - The simulated data is generated using the NeMo multi-speaker data simulator on VoxCeleb1&2 datasets - The multi-speaker data simulation results in a total of 2,000 hours of audio, of which approximately 30% is silence. - The Model training incorporates SpecAugment and noise augmentation through MUSAN noise dataset. ## 2. Speaker Diarization Model: Multi-scale Diarization Decoder (MSDD-v2) ### MSDD_v2_PALO_100ms_intrpl_3scales.nemo Our DASR system is based on the speaker diarization system using the multi-scale diarization decoder (MSDD). - MSDD Reference: Park et al. (2022) - MSDD-v2 speaker diarization system employs a multi-scale embedding approach and utilizes TitaNet speaker embedding extractor. - TitaNet Reference: Koluguri et al. (2022) - TitaNet Model is included in MSDD-v2 .nemo checkpoint file. - Unlike the system that uses a multi-layer LSTM architecture, we employ a four-layer Transformer architecture with a hidden size of 384. - This neural model generates logit values indicating speaker existence. - Our diarization model is trained on approximately 3,000 hours of simulated audio mixture data from the same multi-speaker data simulator used in VAD model training, drawing from VoxCeleb1&2 and LibriSpeech datasets. - LibriSpeech Reference: OpenSLR Download,LibriSpeech, Panayotov et al. (2015) - MUSAN noise is also used for adding additive background noise, focusing on music and broadband noise. ## 3. Automatic Speech Recognition (ASR) model ### URL - This ASR model is based on NeMo FastConformer XL model. - Single-channel audio generated using a multi-channel front-end (Guided Source Separation, GSS) is transcribed using a 0.6B parameter Conformer-based transducer (RNNT) model. - Model Reference: Gulati et al. (2020) - The model was initialized using a publicly available NeMo checkpoint. - NeMo Checkpoint: NGC Model Card: Conformer Transducer XL - This model was then fine-tuned on the CHiME-7 train and dev set, which includes the CHiME-6 and Mixer6 training subsets, after processing the data through the multi-channel ASR front-end, utilizing ground-truth diarization. - Fine-Tuning Details: - Fine-tuning Duration: 35,000 updates - Batch Size: 128 ## 4. Language Model for ASR Decoding: KenLM Model ### ASR_LM_chime7_only.kenlm - This KenLM model is trained solely on CHiME7-DASR datasets (Mixer6, CHiME6, DipCo). - We apply a word-piece level N-gram language model using byte-pair-encoding (BPE) tokens. - This approach utilizes the SentencePiece and KenLM toolkits, based on the transcription of CHiME-7 train and dev sets. - SentencePiece: Kudo and Richardson (2018) - KenLM: KenLM GitRepo - The token sets of our ASR and LM models were matched to ensure consistency. - To combine several N-gram models with equal weights, we used the OpenGrm library. - OpenGrm: Roark et al. (2012) - MAES decoding was employed for the transducer, which accelerates the decoding process. - MAES Decoding: Kim et al. (2020) - As expected, integrating the beam-search decoder with the language model significantly enhances the performance of the end-to-end model compared to its pure counterpart.
[ "# CHiME8 DASR NeMo Baseline Models\n\n- The model files in this repository are the models used in this paper The CHiME-7 Challenge: System Description and Performance of\nNeMo Team’s DASR System.\n- These models are needed to execute the CHiME8-DASR baseline CHiME8-DASR-Baseline NeMo\n- VAD, Diarization and ASR models are all based on NVIDIA NeMo Conversational AI Toolkits.", "## 1. Voice Activity Detection (VAD) Model:", "### MarbleNet_frame_VAD_chime7_Acrobat.nemo \n- This model is based on NeMo MarbleNet VAD model.\n- For validation, we use dataset comprises the CHiME-6 development subset as well as 50 hours of simulated audio data.\n- The simulated data is generated using the NeMo multi-speaker data simulator\non VoxCeleb1&2 datasets\n- The multi-speaker data simulation results in a total of 2,000 hours of audio, of which approximately 30% is silence.\n- The Model training incorporates SpecAugment and noise augmentation through MUSAN noise dataset.", "## 2. Speaker Diarization Model: Multi-scale Diarization Decoder (MSDD-v2)", "### MSDD_v2_PALO_100ms_intrpl_3scales.nemo\n\n Our DASR system is based on the speaker diarization system using the multi-scale diarization decoder (MSDD).\n - MSDD Reference: Park et al. (2022)\n- MSDD-v2 speaker diarization system employs a multi-scale embedding approach and utilizes TitaNet speaker embedding extractor.\n - TitaNet Reference: Koluguri et al. (2022)\n - TitaNet Model is included in MSDD-v2 .nemo checkpoint file.\n- Unlike the system that uses a multi-layer LSTM architecture, we employ a four-layer Transformer architecture with a hidden size of 384.\n- This neural model generates logit values indicating speaker existence.\n- Our diarization model is trained on approximately 3,000 hours of simulated audio mixture data from the same multi-speaker data simulator used in VAD model training, drawing from VoxCeleb1&2 and LibriSpeech datasets.\n - LibriSpeech Reference: OpenSLR Download,LibriSpeech, Panayotov et al. (2015)\n- MUSAN noise is also used for adding additive background noise, focusing on music and broadband noise.", "## 3. Automatic Speech Recognition (ASR) model", "### URL\n- This ASR model is based on NeMo FastConformer XL model. \n- Single-channel audio generated using a multi-channel front-end (Guided Source Separation, GSS) is transcribed using a 0.6B parameter Conformer-based transducer (RNNT) model.\n - Model Reference: Gulati et al. (2020)\n- The model was initialized using a publicly available NeMo checkpoint.\n - NeMo Checkpoint: NGC Model Card: Conformer Transducer XL\n- This model was then fine-tuned on the CHiME-7 train and dev set, which includes the CHiME-6 and Mixer6 training subsets, after processing the data through the multi-channel ASR front-end, utilizing ground-truth diarization.\n - Fine-Tuning Details:\n - Fine-tuning Duration: 35,000 updates\n - Batch Size: 128", "## 4. Language Model for ASR Decoding: KenLM Model", "### ASR_LM_chime7_only.kenlm\n\n- This KenLM model is trained solely on CHiME7-DASR datasets (Mixer6, CHiME6, DipCo).\n- We apply a word-piece level N-gram language model using byte-pair-encoding (BPE) tokens.\n- This approach utilizes the SentencePiece and KenLM toolkits, based on the transcription of CHiME-7 train and dev sets.\n - SentencePiece: Kudo and Richardson (2018)\n - KenLM: KenLM GitRepo\n- The token sets of our ASR and LM models were matched to ensure consistency.\n- To combine several N-gram models with equal weights, we used the OpenGrm library.\n - OpenGrm: Roark et al. (2012)\n- MAES decoding was employed for the transducer, which accelerates the decoding process.\n - MAES Decoding: Kim et al. (2020)\n- As expected, integrating the beam-search decoder with the language model significantly enhances the performance of the end-to-end model compared to its pure counterpart." ]
[ "TAGS\n#nemo #arxiv-2310.12378 #arxiv-1904.08779 #arxiv-1510.08484 #arxiv-2203.15974 #arxiv-2110.04410 #arxiv-2005.08100 #arxiv-1808.06226 #region-us \n", "# CHiME8 DASR NeMo Baseline Models\n\n- The model files in this repository are the models used in this paper The CHiME-7 Challenge: System Description and Performance of\nNeMo Team’s DASR System.\n- These models are needed to execute the CHiME8-DASR baseline CHiME8-DASR-Baseline NeMo\n- VAD, Diarization and ASR models are all based on NVIDIA NeMo Conversational AI Toolkits.", "## 1. Voice Activity Detection (VAD) Model:", "### MarbleNet_frame_VAD_chime7_Acrobat.nemo \n- This model is based on NeMo MarbleNet VAD model.\n- For validation, we use dataset comprises the CHiME-6 development subset as well as 50 hours of simulated audio data.\n- The simulated data is generated using the NeMo multi-speaker data simulator\non VoxCeleb1&2 datasets\n- The multi-speaker data simulation results in a total of 2,000 hours of audio, of which approximately 30% is silence.\n- The Model training incorporates SpecAugment and noise augmentation through MUSAN noise dataset.", "## 2. Speaker Diarization Model: Multi-scale Diarization Decoder (MSDD-v2)", "### MSDD_v2_PALO_100ms_intrpl_3scales.nemo\n\n Our DASR system is based on the speaker diarization system using the multi-scale diarization decoder (MSDD).\n - MSDD Reference: Park et al. (2022)\n- MSDD-v2 speaker diarization system employs a multi-scale embedding approach and utilizes TitaNet speaker embedding extractor.\n - TitaNet Reference: Koluguri et al. (2022)\n - TitaNet Model is included in MSDD-v2 .nemo checkpoint file.\n- Unlike the system that uses a multi-layer LSTM architecture, we employ a four-layer Transformer architecture with a hidden size of 384.\n- This neural model generates logit values indicating speaker existence.\n- Our diarization model is trained on approximately 3,000 hours of simulated audio mixture data from the same multi-speaker data simulator used in VAD model training, drawing from VoxCeleb1&2 and LibriSpeech datasets.\n - LibriSpeech Reference: OpenSLR Download,LibriSpeech, Panayotov et al. (2015)\n- MUSAN noise is also used for adding additive background noise, focusing on music and broadband noise.", "## 3. Automatic Speech Recognition (ASR) model", "### URL\n- This ASR model is based on NeMo FastConformer XL model. \n- Single-channel audio generated using a multi-channel front-end (Guided Source Separation, GSS) is transcribed using a 0.6B parameter Conformer-based transducer (RNNT) model.\n - Model Reference: Gulati et al. (2020)\n- The model was initialized using a publicly available NeMo checkpoint.\n - NeMo Checkpoint: NGC Model Card: Conformer Transducer XL\n- This model was then fine-tuned on the CHiME-7 train and dev set, which includes the CHiME-6 and Mixer6 training subsets, after processing the data through the multi-channel ASR front-end, utilizing ground-truth diarization.\n - Fine-Tuning Details:\n - Fine-tuning Duration: 35,000 updates\n - Batch Size: 128", "## 4. Language Model for ASR Decoding: KenLM Model", "### ASR_LM_chime7_only.kenlm\n\n- This KenLM model is trained solely on CHiME7-DASR datasets (Mixer6, CHiME6, DipCo).\n- We apply a word-piece level N-gram language model using byte-pair-encoding (BPE) tokens.\n- This approach utilizes the SentencePiece and KenLM toolkits, based on the transcription of CHiME-7 train and dev sets.\n - SentencePiece: Kudo and Richardson (2018)\n - KenLM: KenLM GitRepo\n- The token sets of our ASR and LM models were matched to ensure consistency.\n- To combine several N-gram models with equal weights, we used the OpenGrm library.\n - OpenGrm: Roark et al. (2012)\n- MAES decoding was employed for the transducer, which accelerates the decoding process.\n - MAES Decoding: Kim et al. (2020)\n- As expected, integrating the beam-search decoder with the language model significantly enhances the performance of the end-to-end model compared to its pure counterpart." ]
[ 69, 106, 12, 145, 23, 296, 13, 202, 14, 266 ]
[ "passage: TAGS\n#nemo #arxiv-2310.12378 #arxiv-1904.08779 #arxiv-1510.08484 #arxiv-2203.15974 #arxiv-2110.04410 #arxiv-2005.08100 #arxiv-1808.06226 #region-us \n# CHiME8 DASR NeMo Baseline Models\n\n- The model files in this repository are the models used in this paper The CHiME-7 Challenge: System Description and Performance of\nNeMo Team’s DASR System.\n- These models are needed to execute the CHiME8-DASR baseline CHiME8-DASR-Baseline NeMo\n- VAD, Diarization and ASR models are all based on NVIDIA NeMo Conversational AI Toolkits.## 1. Voice Activity Detection (VAD) Model:### MarbleNet_frame_VAD_chime7_Acrobat.nemo \n- This model is based on NeMo MarbleNet VAD model.\n- For validation, we use dataset comprises the CHiME-6 development subset as well as 50 hours of simulated audio data.\n- The simulated data is generated using the NeMo multi-speaker data simulator\non VoxCeleb1&2 datasets\n- The multi-speaker data simulation results in a total of 2,000 hours of audio, of which approximately 30% is silence.\n- The Model training incorporates SpecAugment and noise augmentation through MUSAN noise dataset.## 2. Speaker Diarization Model: Multi-scale Diarization Decoder (MSDD-v2)" ]
[ -0.09301925450563431, 0.1557115912437439, -0.0018100047018378973, -0.0160148236900568, 0.04547710344195366, -0.026436185464262962, 0.15972641110420227, 0.04765421524643898, -0.14838001132011414, 0.12511518597602844, 0.01806117221713066, -0.08957774192094803, 0.08140423893928528, 0.17265966534614563, 0.05273652449250221, -0.14524446427822113, 0.05899139493703842, -0.06421715766191483, 0.06685683876276016, 0.09366529434919357, 0.09237994998693466, -0.057932715862989426, 0.025405049324035645, 0.07686670869588852, -0.11001983284950256, -0.015193268656730652, -0.00709491316229105, -0.023728497326374054, 0.0679110437631607, 0.0031283723656088114, 0.030019447207450867, 0.00010320339060854167, 0.10213000327348709, -0.22467872500419617, 0.008260319009423256, 0.05476254224777222, 0.03361845389008522, 0.0735325813293457, 0.04533815383911133, 0.009752443060278893, 0.24993661046028137, -0.016191672533750534, 0.007515602745115757, 0.09907770901918411, -0.08635441213846207, -0.11981489509344101, -0.099699467420578, 0.005992290563881397, 0.11636171489953995, 0.1145486906170845, -0.06287745386362076, 0.10378199815750122, -0.052921924740076065, 0.06699050962924957, 0.08460764586925507, -0.12938791513442993, -0.06001361086964607, 0.13405409455299377, 0.08863981068134308, 0.03444000706076622, -0.03982797637581825, 0.06348205357789993, 0.11427558213472366, 0.00011297812307020649, 0.09262527525424957, 0.04959225282073021, 0.013222684152424335, -0.030198264867067337, -0.15329274535179138, -0.06059621274471283, 0.12154804170131683, 0.043133631348609924, -0.04419315978884697, -0.07613958418369293, -0.03764072433114052, -0.12272314727306366, -0.05466162785887718, -0.008288051933050156, -0.010111786425113678, -0.06960572302341461, 0.02107188105583191, 0.011573860421776772, -0.05464065447449684, -0.11315977573394775, -0.017474941909313202, -0.09684941172599792, 0.023323075845837593, 0.04051532596349716, -0.07284822314977646, 0.023486563935875893, -0.161285400390625, -0.08463390916585922, 0.003305678255856037, -0.03149097412824631, -0.17702344059944153, -0.016174057498574257, -0.04231705516576767, -0.10552781075239182, -0.009980233386158943, 0.1309584528207779, 0.025028832256793976, 0.028029028326272964, 0.046263206750154495, -0.02103501372039318, 0.0768495574593544, 0.0006103536579757929, -0.07767271250486374, -0.16623803973197937, 0.0026022831443697214, 0.025681333616375923, 0.057069990783929825, -0.04285581782460213, -0.05421188101172447, 0.03917775675654411, 0.021778089925646782, 0.0032415238674730062, 0.02937532588839531, -0.09322772920131683, -0.06789363920688629, 0.001542419195175171, 0.13869021832942963, -0.037846002727746964, 0.016012052074074745, 0.022074086591601372, -0.059720415621995926, 0.05338543280959129, 0.0022156434133648872, 0.04329705238342285, -0.036388151347637177, -0.0025380710139870644, -0.09217322617769241, -0.02105049602687359, -0.09784071147441864, -0.08806832134723663, 0.10643378645181656, -0.05575598031282425, -0.05257255584001541, -0.05159154534339905, -0.07313176989555359, -0.04605112969875336, 0.06236007437109947, -0.10741511732339859, -0.05110270529985428, -0.08305322378873825, -0.056516267359256744, 0.02728816121816635, -0.011139060370624065, 0.09712015837430954, -0.011060908436775208, -0.005235371179878712, 0.13192574679851532, 0.11809743940830231, 0.004878602921962738, 0.03012387827038765, -0.008172908797860146, 0.058740053325891495, -0.015447571873664856, 0.05069267377257347, -0.10282002389431, -0.0500015988945961, -0.11833421140909195, -0.060632143169641495, -0.08449103683233261, -0.004610327072441578, 0.13128294050693512, 0.11715389043092728, -0.2822481691837311, -0.038725703954696655, 0.11791758984327316, -0.09913584589958191, -0.08736919611692429, 0.14602984488010406, -0.007897664792835712, 0.05171021446585655, 0.06073577702045441, 0.10616414994001389, 0.038364700973033905, -0.19554725289344788, -0.048845842480659485, -0.07605879008769989, 0.08439784497022629, -0.06790603697299957, 0.03883451968431473, 0.0131826875731349, 0.12567138671875, 0.006283967290073633, 0.049276743084192276, 0.05231790617108345, -0.0702282190322876, -0.07771768420934677, 0.016093235462903976, -0.08482684195041656, 0.04521969333291054, -0.011745215393602848, -0.013937893323600292, -0.02799615077674389, -0.06801807880401611, 0.05388430505990982, 0.1085486114025116, -0.026263408362865448, 0.035882413387298584, -0.06762683391571045, 0.08261793851852417, -0.12331237643957138, -0.018085939809679985, -0.16381417214870453, -0.051225967705249786, 0.00596714299172163, -0.1548130363225937, 0.061172157526016235, -0.009341484867036343, 0.02262995019555092, 0.05555069074034691, -0.005488018039613962, 0.037763308733701706, -0.08386988192796707, 0.04492558538913727, 0.0062195113860070705, -0.1071765348315239, -0.049450621008872986, -0.04101262614130974, 0.20520471036434174, -0.1626753956079483, -0.004299654625356197, -0.037098582834005356, 0.10606782883405685, -0.007644650060683489, -0.1067444235086441, 0.041596755385398865, 0.03479767218232155, -0.03075864538550377, -0.003787567839026451, 0.011571231298148632, 0.00135273567866534, -0.011033432558178902, 0.0507865846157074, -0.1411706954240799, -0.10606776177883148, 0.06857083737850189, 0.12460287660360336, -0.03202607110142708, 0.061980124562978745, 0.014158829115331173, -0.09055399149656296, -0.0578923262655735, -0.00655656773597002, 0.04390082508325577, 0.053462810814380646, 0.044836655259132385, -0.08540413528680801, -0.021233482286334038, 0.10480072349309921, -0.0688256025314331, -0.0009091176907531917, 0.06122591719031334, 0.09807205200195312, -0.03270731866359711, 0.044417329132556915, -0.029023898765444756, -0.09366112947463989, 0.12002485245466232, -0.005628046113997698, -0.1385001242160797, -0.07123695313930511, -0.07977893948554993, 0.008354926481842995, 0.15909343957901, -0.1032566949725151, 0.00966791994869709, 0.02929558791220188, 0.026738008484244347, 0.06528414785861969, -0.10567499697208405, 0.05052749812602997, 0.002593784825876355, -0.038255833089351654, -0.11723563820123672, 0.05005111172795296, -0.09815758466720581, 0.0313321053981781, -0.017941072583198547, -0.03996529430150986, -0.036510784178972244, -0.01674136519432068, -0.11541805416345596, 0.08086392283439636, -0.09286398440599442, -0.1310845911502838, -0.14299753308296204, 0.011262143962085247, -0.10548415780067444, -0.04017326980829239, -0.022430261597037315, -0.04129709303379059, -0.04810299724340439, -0.0803145170211792, -0.031139938160777092, -0.0429268442094326, -0.053765181452035904, 0.06339771300554276, -0.002829891163855791, 0.05183764547109604, -0.1029970794916153, 0.03266209363937378, -0.03638969734311104, 0.09819382429122925, -0.0376092791557312, -0.019685400649905205, 0.13419191539287567, 0.1814192831516266, -0.003583367681130767, 0.006867688149213791, 0.04450897127389908, 0.2725188136100769, -0.09038735181093216, 0.07191137969493866, 0.18667428195476532, 0.015447650104761124, 0.014169602654874325, 0.11294271796941757, 0.028720371425151825, -0.04570440575480461, 0.011052125133574009, 0.05624319240450859, -0.1011577844619751, -0.23851364850997925, -0.11197847872972488, -0.10176397860050201, -0.026218008249998093, -0.034408364444971085, -0.020093020051717758, 0.09268921613693237, -0.012720243073999882, -0.004787186626344919, -0.08702581375837326, 0.04965604096651077, 0.043206535279750824, 0.15436196327209473, -0.015099934302270412, 0.05959044024348259, -0.047310467809438705, 0.013875113800168037, 0.036217525601387024, 0.04392337426543236, 0.1846223622560501, 0.05347910895943642, 0.14503413438796997, 0.07253655046224594, -0.004302118439227343, 0.012126577086746693, 0.031808216124773026, -0.023864515125751495, -0.019754910841584206, 0.020282337442040443, -0.07310144603252411, -0.01751616597175598, 0.08421852439641953, 0.10207831114530563, -0.0424458310008049, -0.0011913958005607128, 0.04758129641413689, -0.0056164939887821674, 0.061380308121442795, 0.21115073561668396, -0.2826905846595764, -0.07800150662660599, 0.006332738324999809, 0.027508020401000977, -0.04468116909265518, 0.009534657932817936, 0.08913840353488922, -0.03616318851709366, 0.0756462812423706, -0.007843642495572567, 0.08743265271186829, 0.010660030879080296, -0.018207911401987076, -0.055983759462833405, 0.1367635428905487, -0.010224517434835434, 0.070603147149086, -0.16954903304576874, 0.1671202927827835, 0.006682869978249073, 0.09587087482213974, 0.0028702656272798777, -0.006170810665935278, -0.04827526956796646, -0.035639528185129166, 0.12059678137302399, 0.03902829438447952, -0.13772454857826233, 0.06336358189582825, -0.14777925610542297, -0.016859790310263634, 0.02768009528517723, 0.033342987298965454, 0.08156212419271469, 0.03585100546479225, -0.04545934498310089, 0.0149285439401865, 0.04889313504099846, -0.15782193839550018, -0.1342935711145401, 0.0417860709130764, 0.11532389372587204, -0.0787329375743866, -0.06124905124306679, -0.09454034268856049, -0.04862307384610176, 0.1559610813856125, 0.02324879914522171, -0.026132283732295036, -0.06971992552280426, -0.0011271265102550387, 0.21726082265377045, -0.021578524261713028, 0.10631855577230453, 0.004976275376975536, 0.1554504781961441, -0.0746060460805893, -0.031791720539331436, 0.05678439512848854, -0.08122691512107849, -0.08766696602106094, -0.05825212970376015, 0.17243692278862, 0.10414585471153259, 0.0587562657892704, 0.0047308108769357204, 0.056850798428058624, 0.009553046897053719, -0.034641996026039124, -0.008165592327713966, 0.15258671343326569, -0.0499364472925663, -0.05601521208882332, 0.05403980612754822, -0.14696760475635529, -0.035581525415182114, -0.06646010279655457, 0.07569388300180435, 0.2130689173936844, -0.025293270125985146, 0.13083139061927795, 0.1705823391675949, -0.05117395892739296, -0.1695214807987213, -0.018986022099852562, 0.11508645862340927, 0.07772760838270187, 0.06208082661032677, -0.09624502062797546, -0.014391531236469746, 0.005798629485070705, -0.0417485274374485, 0.007811983581632376, -0.29210734367370605, -0.08843684941530228, 0.019471989944577217, 0.007212779484689236, 0.10901729017496109, -0.01211476605385542, -0.016633495688438416, -0.07463988661766052, -0.1519651859998703, 0.10186443477869034, -0.028154971078038216, 0.07741331309080124, 0.04046408459544182, 0.06461574882268906, 0.040681224316358566, -0.06959402561187744, 0.1210457980632782, 0.11681609600782394, -0.0028736412059515715, 0.0006398989353328943, 0.07259276509284973, 0.08570444583892822, -0.08894707262516022, 0.10156670957803726, 0.0070074028335511684, 0.039166633039712906, 0.008623803034424782, -0.051269277930259705, -0.06484058499336243, -0.009284366853535175, -0.05239131301641464, -0.014869067817926407, -0.052937667816877365, 0.08461058139801025, 0.034799881279468536, -0.002602272666990757, 0.016140522435307503, -0.05176514387130737, 0.015090184286236763, 0.24435795843601227, 0.04912157729268074, 0.017622174695134163, -0.03255947306752205, 0.023625437170267105, -0.030185136944055557, 0.03290674462914467, -0.17102110385894775, 0.025988483801484108, 0.11685691773891449, 0.06076200678944588, 0.0561000294983387, -0.024748649448156357, -0.20352639257907867, -0.020995812490582466, 0.05673787370324135, -0.012329710647463799, -0.11702439934015274, -0.006449272856116295, -0.04909113422036171, -0.11218398064374924, 0.044802792370319366, 0.1181660145521164, -0.09756742417812347, 0.02000693790614605, 0.014246426522731781, 0.08288458734750748, -0.08225302398204803, 0.2316230833530426, -0.0023959516547620296, 0.06575765460729599, -0.08841823786497116, 0.0466659739613533, 0.05784665048122406, -0.015728428959846497, 0.026341833174228668, 0.030913708731532097, -0.0685604065656662, -0.012379862368106842, -0.06280233711004257, -0.0019871529657393694, -0.06799305975437164, -0.08969365060329437, -0.026219544932246208, -0.07751256972551346, -0.006204923614859581, 0.07521256059408188, 0.022098848596215248, 0.032444749027490616, -0.04593576863408089, -0.025433620437979698, -0.16801483929157257, 0.14475522935390472, 0.1033104807138443, 0.05758710578083992, -0.07692927122116089, 0.010424302890896797, 0.05420107766985893, 0.0006848787888884544, -0.03523207828402519, -0.013061737641692162, -0.04794517159461975, 0.062073253095149994, -0.14422568678855896, -0.07850095629692078, -0.09826456755399704, -0.013559844344854355, -0.029806481674313545, 0.036080822348594666, -0.06010293960571289, 0.06506665050983429, -0.02408420480787754, 0.0007191611803136766, -0.06699059158563614, 0.04758865386247635, -0.0901348739862442, 0.027572624385356903, 0.02048099786043167, -0.09040690213441849, 0.027540815994143486, 0.042871180921792984, 0.041255801916122437, 0.025128306820988655, -0.12911899387836456, 0.011794352903962135, 0.047851625829935074, 0.03884430229663849, 0.009122777730226517, -0.1984492838382721, -0.013236930593848228, 0.04153675213456154, 0.0107367392629385, -0.04530324041843414, -0.06900683790445328, -0.03547827526926994, -0.09155923873186111, -0.08460157364606857, -0.04488692805171013, 0.008421378210186958, 0.009331838227808475, 0.054945219308137894, 0.07758897542953491, 0.11925157159566879, 0.000937908363994211, 0.05500543490052223, -0.11746842414140701, -0.00424747820943594, -0.033350978046655655, 0.02743225172162056, -0.0843966081738472, -0.09320814162492752, 0.04835653677582741, -0.048359308391809464, 0.1360858529806137, -0.023554695770144463, 0.02774992026388645, 0.003168990835547447, -0.05346067622303963, -0.10748869925737381, 0.01761987805366516, 0.10415682196617126, 0.0496983677148819, 0.02388160303235054, 0.006893117446452379, 0.010883763432502747, 0.0035281379241496325, 0.1318276822566986, 0.08303060382604599, 0.016780667006969452, -0.020459776744246483, 0.07624073326587677, 0.0974041074514389, -0.07225699722766876, -0.06913483887910843, 0.21402621269226074, -0.123601533472538, 0.05430074408650398, -0.01871340721845627, 0.03193292021751404, 0.16033051908016205, -0.12434844672679901, 0.07166969031095505, -0.016602100804448128, -0.0992540642619133, -0.08332649618387222, -0.128025621175766, -0.07915057241916656, -0.10509391129016876, 0.01600836217403412, -0.08351946622133255, 0.02318810299038887, 0.028371717780828476, 0.038172733038663864, -0.04479120671749115, 0.11835554987192154, -0.018534626811742783, -0.04397348687052727, 0.00315618934109807, -0.03147989511489868, -0.022823624312877655, 0.05799601227045059, 0.0296824611723423, 0.11812612414360046, 0.042591556906700134, 0.10216546803712845, 0.02285991795361042, 0.03710085526108742, 0.10605835169553757, 0.050617706030607224, -0.07591331005096436, -0.029050638899207115, -0.02783113531768322, 0.08402995765209198, 0.1784915030002594, 0.09649133682250977, -0.04718480631709099, 0.006059438455849886, 0.03952351585030556, -0.04976874589920044, -0.12739647924900055, -0.1727130115032196, 0.22127608954906464, -0.011989835649728775, 0.03722856938838959, -0.01162747573107481, -0.05498315021395683, -0.08150853961706161, 0.0739249438047409, 0.13787125051021576, 0.007954037748277187, -0.08640152961015701, -0.01808907650411129, -0.024285761639475822, -0.12078391015529633, 0.00702672591432929, 0.007599622011184692, 0.25038912892341614, 0.002508777193725109, 0.043349117040634155, -0.0624496191740036, -0.054493363946676254, 0.0007929093553684652, 0.08439536392688751, 0.013243254274129868, -0.024429457262158394, 0.0005676352302543819, 0.13952825963497162, -0.126317098736763, -0.13346534967422485, -0.027755117043852806, -0.06856326013803482, -0.1397305577993393, 0.02873087488114834, 0.05465236306190491, 0.0368916317820549, 0.03935970366001129, -0.07018569111824036, 0.016517117619514465, 0.17303787171840668, -0.008769253268837929, -0.02518773265182972, -0.058616917580366135, 0.03747035935521126, -0.02579021453857422, 0.13259480893611908, -0.007129667326807976, 0.09557253867387772, 0.04964255914092064, 0.011719899252057076, -0.09970211237668991, 0.1290137618780136, 0.029817121103405952, -0.04868044704198837, 0.03530662879347801, 0.16387642920017242, -0.025127574801445007, 0.0955599993467331, 0.05818434804677963, -0.027370352298021317, 0.0032973242923617363, -0.06151294335722923, 0.028963277116417885, -0.0647093877196312, 0.08063316345214844, -0.04900138080120087, 0.14035871624946594, 0.02592732198536396, -0.07654072344303131, -0.013790108263492584, -0.016195019707083702, 0.05247480049729347, 0.00413617305457592, 0.14110207557678223, 0.027093209326267242, -0.23049113154411316, 0.04428384080529213, -0.08553311228752136, 0.020843300968408585, -0.22119902074337006, 0.006697134114801884, -0.02499605342745781, -0.035186801105737686, 0.047012001276016235, 0.07150428742170334, 0.10460695624351501, -0.00902819074690342, -0.06493712216615677, -0.04488462954759598, 0.0022901776246726513, 0.1046287938952446, -0.10473042726516724, -0.10312273353338242 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
null
dmusingu/phi2tokenizer
[ "transformers", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
2024-02-15T01:08:06+00:00
[ "1910.09700" ]
[]
TAGS #transformers #arxiv-1910.09700 #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #arxiv-1910.09700 #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 26, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.08389580249786377, 0.19830818474292755, -0.0013316317927092314, 0.02313883788883686, 0.11396584659814835, 0.01961737498641014, 0.053626976907253265, 0.14538456499576569, 0.0060051376931369305, 0.10656800121068954, 0.066679947078228, 0.09131570905447006, 0.09678101539611816, 0.20042605698108673, 0.04371999576687813, -0.17659740149974823, 0.010636410675942898, -0.06930278241634369, -0.010073255747556686, 0.11651819199323654, 0.141214057803154, -0.10151198506355286, 0.07627976685762405, -0.03319970890879631, -0.02870541252195835, -0.0070160143077373505, -0.07769215852022171, -0.05755697935819626, 0.07573003321886063, 0.054863471537828445, 0.04207949340343475, -0.0008347301045432687, 0.08447454124689102, -0.2674994468688965, 0.013753628358244896, 0.07452993094921112, 0.010659529827535152, 0.05990942195057869, 0.07833302766084671, -0.04036625102162361, 0.12881849706172943, -0.06320446729660034, 0.13035163283348083, 0.0906217098236084, -0.0681561604142189, -0.24378153681755066, -0.08239314705133438, 0.06505522131919861, 0.12533815205097198, 0.07694927603006363, -0.02823091857135296, 0.16422191262245178, -0.07247646898031235, 0.019290022552013397, 0.09481704235076904, -0.1151006743311882, -0.060644298791885376, 0.08318385481834412, 0.14101974666118622, 0.10340547561645508, -0.1255619376897812, -0.012289565056562424, 0.04275871813297272, 0.045979104936122894, 0.07389909774065018, 0.011339850723743439, 0.1143413558602333, 0.05629947781562805, -0.13526225090026855, -0.05700986459851265, 0.14547574520111084, 0.023872992023825645, -0.057064127177000046, -0.2138909548521042, -0.002902575535699725, -0.07730814069509506, -0.011685127392411232, -0.06846728920936584, 0.0291305985301733, -0.01194276288151741, 0.060226380825042725, -0.0496203787624836, -0.09797755628824234, -0.046314824372529984, 0.1015089675784111, 0.054820988327264786, 0.011354796588420868, -0.01489334274083376, 0.03576440364122391, 0.13432876765727997, 0.04213530570268631, -0.10012737661600113, -0.07065672427415848, -0.0701170489192009, -0.09620913118124008, -0.03947552293539047, 0.04272124543786049, 0.020167991518974304, 0.042202774435281754, 0.2283228635787964, 0.024096308276057243, 0.05459817871451378, 0.029667891561985016, 0.0026177873369306326, 0.03211980313062668, 0.1073630079627037, -0.041210614144802094, -0.188126802444458, -0.03292805701494217, 0.0931866466999054, -0.009821015410125256, -0.028658604249358177, -0.033444397151470184, 0.035014089196920395, 0.08379437029361725, 0.11821532249450684, 0.08875755965709686, -0.012828069739043713, -0.037612639367580414, -0.03493109717965126, 0.2115669697523117, -0.14141373336315155, 0.045799970626831055, -0.022097334265708923, -0.018195297569036484, -0.06905751675367355, 0.030103791505098343, 0.01831657998263836, -0.003142025787383318, 0.06966056674718857, -0.061253178864717484, -0.05794486775994301, -0.11518853157758713, -0.045523155480623245, 0.04711875319480896, -0.024105608463287354, -0.024469668045639992, -0.07765042781829834, -0.11219723522663116, -0.06417357176542282, 0.06612563133239746, -0.04156653955578804, -0.03974827378988266, 0.005308232270181179, -0.07131324708461761, 0.008387917652726173, 0.008993842639029026, 0.12122467905282974, -0.030063031241297722, 0.05833350867033005, -0.002476902212947607, 0.05916252359747887, 0.10643328726291656, 0.03227818012237549, -0.08492200076580048, 0.057466037571430206, -0.20633617043495178, 0.08371785283088684, -0.11420095711946487, 0.034276340156793594, -0.17048145830631256, -0.024183684960007668, 0.008447963744401932, 0.023597201332449913, 0.023726604878902435, 0.1338067352771759, -0.2097422182559967, -0.016196569427847862, 0.14133213460445404, -0.09649793803691864, -0.12422871589660645, 0.07990546524524689, -0.03459475561976433, 0.1747698187828064, 0.038475677371025085, -0.019652999937534332, 0.09909367561340332, -0.15559963881969452, -0.05852397903800011, -0.026064254343509674, -0.008927824907004833, 0.08823978155851364, 0.07542291283607483, -0.05844951793551445, 0.02285866066813469, 0.02562655322253704, -0.04727208614349365, -0.0268824752420187, -0.05256075784564018, -0.10127434879541397, -0.023140445351600647, -0.09642518311738968, 0.026515161618590355, 0.000058677000197349116, -0.07310442626476288, -0.028560271486639977, -0.17347893118858337, -0.02563360333442688, 0.10103316605091095, 0.004820956848561764, -0.007559072691947222, -0.08540112525224686, 0.022149885073304176, -0.05362366884946823, -0.006164622958749533, -0.16996455192565918, -0.03558015450835228, 0.051895126700401306, -0.14917676150798798, 0.015460150316357613, -0.07327745854854584, 0.07047311216592789, 0.02098717913031578, -0.05859505757689476, -0.03108096309006214, 0.0007694467785768211, 0.004292082041501999, -0.06229274719953537, -0.1903683841228485, -0.058886781334877014, -0.041500482708215714, 0.15720732510089874, -0.24841000139713287, 0.0300158578902483, 0.03247617185115814, 0.13185922801494598, 0.007058668415993452, -0.06344027817249298, 0.02096918225288391, -0.04676475748419762, -0.050621338188648224, -0.06898977607488632, -0.009901339188218117, -0.014539826661348343, -0.031393732875585556, 0.012980648316442966, -0.14970256388187408, -0.060514215379953384, 0.09452559798955917, 0.11224991828203201, -0.14555825293064117, 0.00204002158716321, -0.0460561066865921, -0.07002599537372589, -0.07487804442644119, -0.0761631652712822, 0.07739497721195221, 0.044650159776210785, 0.049250341951847076, -0.06317461282014847, -0.06234706938266754, 0.023210179060697556, 0.005524294450879097, -0.019023682922124863, 0.0948529988527298, 0.074309803545475, -0.09122881293296814, 0.07973480224609375, 0.08461450785398483, 0.04414684325456619, 0.086973637342453, 0.005991141777485609, -0.11396963149309158, -0.03062884695827961, 0.037754856050014496, 0.024159027263522148, 0.15351562201976776, -0.08692087233066559, 0.030462130904197693, 0.052177220582962036, -0.03854219615459442, 0.03157065063714981, -0.0923321321606636, 0.025362705811858177, 0.021495236083865166, -0.006555700208991766, 0.05864228308200836, -0.018769768998026848, -0.01403577346354723, 0.06336429715156555, 0.05677810311317444, 0.044270504266023636, 0.02595379762351513, -0.02093072421848774, -0.1278371512889862, 0.16537296772003174, -0.09028079360723495, -0.2540280222892761, -0.17074446380138397, 0.015454737469553947, 0.03706491366028786, -0.021728800609707832, 0.039588842540979385, -0.06286025792360306, -0.10237989574670792, -0.09417891502380371, 0.0029635571409016848, 0.023925531655550003, -0.058347854763269424, -0.0817074254155159, 0.060779985040426254, 0.04047083482146263, -0.13689260184764862, 0.0349188968539238, 0.06170675903558731, -0.03042641654610634, 0.0018567070364952087, 0.07321398705244064, 0.12743599712848663, 0.14838241040706635, -0.006730219814926386, -0.012446845881640911, 0.035035960376262665, 0.229813352227211, -0.1490442156791687, 0.10630457103252411, 0.14053207635879517, -0.021705523133277893, 0.06635113060474396, 0.1461038440465927, 0.023231739178299904, -0.07546708732843399, 0.04147516191005707, 0.04027445614337921, -0.04228919371962547, -0.2589097023010254, -0.05694316700100899, -0.00946022942662239, -0.07043391466140747, 0.09718906134366989, 0.09238530695438385, 0.11972260475158691, 0.0337289460003376, -0.05568677559494972, -0.025771914049983025, -0.003401360474526882, 0.114128477871418, -0.027640055865049362, -0.004564122296869755, 0.07965842634439468, -0.05878787487745285, 0.011684526689350605, 0.09941446036100388, 0.019347423687577248, 0.17601320147514343, 0.02533329278230667, 0.10681075602769852, 0.06725578010082245, 0.09347675740718842, -0.0015635732561349869, 0.034774236381053925, 0.05337131395936012, 0.022044572979211807, 0.010453542694449425, -0.09408048540353775, -0.012431944720447063, 0.13713060319423676, 0.019816776737570763, 0.009031654335558414, 0.008926562033593655, -0.01010479498654604, 0.03131420537829399, 0.20501568913459778, 0.0009575071162544191, -0.22537250816822052, -0.09500737488269806, 0.059459153562784195, -0.06931101530790329, -0.143676295876503, -0.02094252221286297, 0.030270220711827278, -0.17292405664920807, 0.016790566965937614, -0.0316389761865139, 0.09112390875816345, -0.07145322859287262, -0.028050832450389862, 0.06891903281211853, 0.07569212466478348, -0.012108199298381805, 0.07973295450210571, -0.19069278240203857, 0.12254468351602554, 0.03037673607468605, 0.08605273067951202, -0.11708726733922958, 0.07849059253931046, -0.0019813794642686844, -0.014807495288550854, 0.17999744415283203, -0.014062200672924519, -0.0586031936109066, -0.08878950774669647, -0.08704045414924622, -0.011727320961654186, 0.10361312329769135, -0.09322915226221085, 0.09586969763040543, -0.02775636687874794, -0.03705112263560295, 0.012418309226632118, -0.10469507426023483, -0.1636953055858612, -0.18679304420948029, 0.06244563311338425, -0.07802703976631165, 0.012347841635346413, -0.11227322369813919, -0.06334327906370163, -0.01575082167983055, 0.23160123825073242, -0.16648635268211365, -0.07049825042486191, -0.1498587429523468, -0.03997112438082695, 0.17463743686676025, -0.042160745710134506, 0.06849376112222672, -0.021383514627814293, 0.1873992383480072, -0.008081548847258091, -0.013158116489648819, 0.06569221615791321, -0.09637628495693207, -0.16879262030124664, -0.05748843029141426, 0.14160962402820587, 0.10863390564918518, 0.05731578543782234, -0.0038195757661014795, 0.013171887956559658, -0.03383830562233925, -0.09896382689476013, 0.013824623078107834, 0.13817466795444489, 0.0034514935687184334, 0.00682973163202405, -0.03995988517999649, -0.07027145475149155, -0.05825701728463173, -0.07912654429674149, 0.057147104293107986, 0.187900573015213, -0.09512355923652649, 0.1602867990732193, 0.12431421875953674, -0.06468851119279861, -0.2306901067495346, 0.03996593505144119, 0.04701630026102066, 0.007666614837944508, 0.022401191294193268, -0.19138796627521515, 0.09788824617862701, 0.0009011493530124426, -0.06807263940572739, 0.14616990089416504, -0.16564498841762543, -0.1461436152458191, 0.08002161979675293, 0.025075770914554596, -0.22560662031173706, -0.14821304380893707, -0.1037549376487732, -0.03735695406794548, -0.13707835972309113, 0.048581719398498535, 0.02614329755306244, 0.019834673032164574, 0.025222565978765488, 0.005338077899068594, 0.029657263308763504, -0.07272187620401382, 0.1870686560869217, -0.020297454670071602, 0.0072362530045211315, -0.050640691071748734, -0.04617878794670105, 0.09227550774812698, -0.06150037795305252, 0.11741586774587631, 0.018679620698094368, 0.018796883523464203, -0.1431548148393631, -0.049209367483854294, -0.060803934931755066, 0.04456847906112671, -0.07284719496965408, -0.09393193572759628, -0.04137463867664337, 0.08888561278581619, 0.07211937010288239, -0.032792408019304276, -0.0027768779546022415, -0.07569456845521927, 0.09405932575464249, 0.184477761387825, 0.17357055842876434, 0.009977072477340698, -0.07020942866802216, 0.024555526673793793, -0.042279548943042755, 0.03349342197179794, -0.24652716517448425, 0.03456863760948181, 0.066053606569767, 0.03803660348057747, 0.08509242534637451, -0.016836483031511307, -0.1781480610370636, -0.04086102172732353, 0.08498652279376984, -0.06206206604838371, -0.19876568019390106, -0.02703288197517395, 0.08424776047468185, -0.20383712649345398, -0.032998621463775635, 0.041543323546648026, -0.03834589570760727, -0.02396267279982567, -0.002415500348433852, 0.06396626681089401, -0.008327016606926918, 0.12156640738248825, 0.06747189164161682, 0.10266115516424179, -0.09284433722496033, 0.08920657634735107, 0.10416955500841141, -0.09140542894601822, 0.03545991703867912, 0.10264154523611069, -0.05670900270342827, -0.04460543021559715, 0.033935222774744034, 0.05925208330154419, -0.028357384726405144, -0.06409841030836105, -0.000502707262057811, -0.0359574519097805, 0.04993389546871185, 0.08058220148086548, 0.036113787442445755, -0.01202210783958435, 0.06544706225395203, 0.028145326301455498, -0.11693570017814636, 0.10949387401342392, 0.04405685141682625, 0.04509059712290764, -0.07182393968105316, -0.012280966155230999, 0.015999672934412956, 0.032540347427129745, -0.019734015688300133, -0.014576527290046215, -0.03146412968635559, -0.007561005651950836, -0.1553635597229004, -0.02064543403685093, -0.06516171246767044, 0.006067827809602022, 0.022207623347640038, -0.03830232471227646, -0.012014663778245449, 0.01381110493093729, -0.07979435473680496, -0.07571027427911758, -0.01700955256819725, 0.08539021760225296, -0.1381402313709259, 0.006627439055591822, 0.07182712107896805, -0.10980239510536194, 0.07347989827394485, -0.0048679932951927185, 0.017079560086131096, 0.010923396795988083, -0.11654401570558548, 0.04386281594634056, -0.005810429807752371, 0.01551580335944891, 0.022556742653250694, -0.171111062169075, 0.011553828604519367, -0.038553636521101, -0.03114982508122921, 0.011926400475203991, -0.025060230866074562, -0.11875922232866287, 0.08676479011774063, -0.028097305446863174, -0.037512701004743576, -0.03292486071586609, 0.06296087801456451, 0.08736220002174377, -0.011740099638700485, 0.09667140990495682, -0.025766119360923767, 0.04818311333656311, -0.1756584197282791, -0.01910574547946453, -0.050167568027973175, 0.02537350542843342, -0.01759655587375164, -0.0070639788173139095, 0.055272240191698074, -0.004191063344478607, 0.20991376042366028, -0.03921036794781685, 0.1548677533864975, 0.05199402943253517, -0.009925156831741333, 0.010884369723498821, 0.05032730847597122, 0.06423956155776978, 0.031145188957452774, 0.00853167474269867, 0.04660189896821976, -0.004552975296974182, -0.020357951521873474, -0.13699717819690704, 0.02791593410074711, 0.16117429733276367, 0.061918217688798904, 0.0392887257039547, 0.03704594820737839, -0.1422400325536728, -0.09538721293210983, 0.10306388139724731, -0.0331864058971405, 0.014331420883536339, -0.08317886292934418, 0.17621558904647827, 0.12328410148620605, -0.1574767529964447, 0.0577850341796875, -0.07234696298837662, -0.05066767707467079, -0.1024852767586708, -0.11832084506750107, -0.06293155997991562, -0.06027044355869293, -0.004747506696730852, -0.042489297688007355, 0.05734556168317795, 0.026751231402158737, -0.003270963439717889, -0.006759525276720524, 0.12665949761867523, -0.0249644722789526, -0.004145825747400522, 0.04152364656329155, 0.0326087586581707, 0.019319625571370125, -0.05872373282909393, 0.017997145652770996, 0.018602589145302773, 0.022180357947945595, 0.06835069507360458, 0.0260987039655447, -0.059317342936992645, 0.044286735355854034, 0.00319746439345181, -0.11313364654779434, 0.018146557733416557, -0.00002245741598017048, -0.05020225793123245, 0.13557326793670654, 0.04076748713850975, 0.01548024732619524, -0.029270920902490616, 0.24342355132102966, -0.07199113070964813, -0.08681939542293549, -0.13965600728988647, 0.11511493474245071, -0.023563209921121597, 0.03755274787545204, 0.016542524099349976, -0.12659503519535065, 0.011511262506246567, 0.18531471490859985, 0.12824349105358124, 0.012459068559110165, -0.007656481582671404, 0.05736639350652695, -0.0007639875984750688, -0.05985576659440994, 0.05051197111606598, 0.0664999932050705, 0.16097788512706757, -0.09069112688302994, 0.0652846097946167, -0.008405503816902637, -0.0831485390663147, -0.027498632669448853, 0.11705785244703293, -0.022675158455967903, 0.02148384228348732, -0.03778035193681717, 0.11204422265291214, -0.052532415837049484, -0.2719486355781555, 0.02952493168413639, -0.09503202140331268, -0.13993041217327118, -0.02591860294342041, 0.041448429226875305, -0.03349510580301285, 0.01577647216618061, 0.06254769116640091, -0.045389387756586075, 0.18837277591228485, 0.025987716391682625, -0.08679025620222092, -0.07755549252033234, 0.05874146893620491, -0.08695939928293228, 0.2789687216281891, 0.003863075515255332, 0.04782010242342949, 0.12108923494815826, -0.03053574077785015, -0.18664880096912384, 0.014769754372537136, 0.11989909410476685, -0.09114406257867813, 0.07780203968286514, 0.18139931559562683, -0.005561648402363062, 0.12649618089199066, 0.04705416411161423, -0.03877115994691849, 0.03976387158036232, -0.02721380814909935, -0.03821522742509842, -0.12209630757570267, 0.05661242455244064, -0.0612691193819046, 0.15957388281822205, 0.1158948540687561, -0.05964287370443344, 0.001120698289014399, -0.06126941740512848, 0.06300627440214157, 0.014774397015571594, 0.12115653604269028, 0.018452486023306847, -0.2023056596517563, 0.05087360367178917, -0.03283824771642685, 0.08166342973709106, -0.254973828792572, -0.08186668157577515, 0.07622263580560684, -0.019022729247808456, -0.04275642707943916, 0.12311509251594543, 0.06101066991686821, 0.03676839917898178, -0.03853875398635864, -0.08537755906581879, -0.01412904355674982, 0.15376435220241547, -0.14123432338237762, -0.029574336484074593 ]
null
null
peft
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.7.1
{"library_name": "peft", "base_model": "codeparrot/codeparrot"}
null
adalib/sfepy-cond-gen-codeparrot-prefix
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:codeparrot/codeparrot", "region:us" ]
2024-02-15T01:09:56+00:00
[ "1910.09700" ]
[]
TAGS #peft #safetensors #arxiv-1910.09700 #base_model-codeparrot/codeparrot #region-us
# Model Card for Model ID ## Model Details ### Model Description - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact ### Framework versions - PEFT 0.7.1
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.7.1" ]
[ "TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-codeparrot/codeparrot #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.7.1" ]
[ 35, 6, 3, 54, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4, 11 ]
[ "passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-codeparrot/codeparrot #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.7.1" ]
[ -0.10329236835241318, 0.1991094946861267, -0.0035496081691235304, 0.0320764034986496, 0.0945892333984375, 0.018722474575042725, 0.04824448376893997, 0.12525925040245056, -0.042864445596933365, 0.10972880572080612, 0.06623604893684387, 0.10955594480037689, 0.10282865911722183, 0.2034357488155365, 0.007155488710850477, -0.1959189474582672, 0.02770180068910122, -0.09508345276117325, -0.009301785379648209, 0.12406817823648453, 0.15033432841300964, -0.09707682579755783, 0.07696907222270966, -0.016842491924762726, -0.01898537389934063, -0.036891840398311615, -0.0783899649977684, -0.032675329595804214, 0.0423843115568161, 0.04027989134192467, 0.06320202350616455, -0.0010284704621881247, 0.08786033093929291, -0.2682170271873474, 0.01724882982671261, 0.046349991112947464, -0.005619754083454609, 0.08398384600877762, 0.09925679117441177, -0.04155835881829262, 0.11729296296834946, -0.03429543226957321, 0.1447087824344635, 0.07977385073900223, -0.09631329029798508, -0.20587898790836334, -0.0711367204785347, 0.07684213668107986, 0.1776040941476822, 0.07997488230466843, -0.044256117194890976, 0.13360482454299927, -0.09620600193738937, 0.018231961876153946, 0.04062868654727936, -0.0796104297041893, -0.07343902438879013, 0.06711498647928238, 0.10942767560482025, 0.053163494914770126, -0.13837280869483948, -0.03540616109967232, 0.024569101631641388, 0.04050806909799576, 0.0760534256696701, 0.019766200333833694, 0.1391512155532837, 0.02985544688999653, -0.1491759866476059, -0.04452425613999367, 0.12775851786136627, 0.026105333119630814, -0.03582258149981499, -0.22357040643692017, 0.0049335164949297905, -0.08705339580774307, -0.02808760292828083, -0.050879329442977905, 0.03442491590976715, 0.004562058951705694, 0.08835969120264053, -0.024290746077895164, -0.09167808294296265, -0.011855660937726498, 0.09715835005044937, 0.043023012578487396, 0.023720107972621918, -0.0279318206012249, 0.002449614927172661, 0.12286990135908127, 0.059708040207624435, -0.12925924360752106, -0.05743660032749176, -0.06734061986207962, -0.04403950646519661, -0.047377679497003555, 0.03163853660225868, 0.027819188311696053, 0.059605587273836136, 0.2507312297821045, -0.024732042104005814, 0.05848440155386925, 0.054291076958179474, 0.018252437934279442, 0.044734321534633636, 0.0977662056684494, -0.05256020277738571, -0.1497306376695633, -0.018936388194561005, 0.09864485263824463, -0.009359381161630154, -0.02077452838420868, -0.04808277264237404, 0.04318903386592865, 0.04737257957458496, 0.10851886123418808, 0.09459429979324341, -0.0035771166440099478, -0.07812892645597458, -0.04933774471282959, 0.21165241301059723, -0.15036462247371674, 0.042139485478401184, 0.022374821826815605, -0.014428223483264446, -0.06152331084012985, 0.007106795907020569, 0.018991848453879356, -0.024411004036664963, 0.09467687457799911, -0.06350524723529816, -0.03970067575573921, -0.12008386850357056, -0.020254241302609444, 0.037786852568387985, 0.012760885991156101, -0.02904166653752327, -0.029857967048883438, -0.05983944982290268, -0.09628355503082275, 0.10356172174215317, -0.06480321288108826, -0.0610169991850853, -0.03797745332121849, -0.09993628412485123, 0.020773757249116898, 0.03109990805387497, 0.10628335922956467, -0.02319858781993389, 0.04236413165926933, -0.013732589781284332, 0.06189176067709923, 0.08645255863666534, 0.036228246986866, -0.07309529185295105, 0.06289207190275192, -0.19817276298999786, 0.09132614731788635, -0.07704638689756393, 0.02864651195704937, -0.15782223641872406, -0.013861071318387985, 0.006248894613236189, 0.02022254467010498, 0.03507586196064949, 0.15412302315235138, -0.2035851627588272, -0.024532441049814224, 0.16218020021915436, -0.09985975176095963, -0.11708329617977142, 0.03965356573462486, -0.05680323764681816, 0.16186797618865967, 0.019650926813483238, -0.009927471168339252, 0.08559273183345795, -0.1474023312330246, -0.025457140058279037, -0.03177822753787041, -0.004023641347885132, 0.10740065574645996, 0.08121296018362045, -0.08140331506729126, 0.033494699746370316, 0.016097290441393852, -0.042079586535692215, -0.03229745477437973, -0.05116637796163559, -0.111270472407341, 0.0032467953860759735, -0.08391474932432175, 0.029999537393450737, -0.007374199107289314, -0.07686883956193924, -0.015222935006022453, -0.16282664239406586, -0.02819419465959072, 0.07947083562612534, 0.01443477813154459, -0.019197700545191765, -0.09322435408830643, 0.03170805424451828, -0.02483070269227028, -0.023633912205696106, -0.15835340321063995, -0.029325641691684723, 0.02153729274868965, -0.14053747057914734, 0.014486008323729038, -0.11843785643577576, 0.06604547053575516, 0.013066625222563744, -0.06505228579044342, -0.03554535657167435, -0.011915669776499271, 0.009852970018982887, -0.049514275044202805, -0.23495401442050934, -0.020242391154170036, -0.05747365579009056, 0.15570616722106934, -0.22586597502231598, 0.04150522127747536, 0.042241960763931274, 0.1306493878364563, 0.0035597607493400574, -0.0632866621017456, 0.02558288909494877, -0.06882250308990479, -0.02370600588619709, -0.0716695562005043, -0.005979429464787245, -0.002668083878234029, -0.03843466192483902, 0.009915197268128395, -0.11917467415332794, -0.05634033679962158, 0.10177865624427795, 0.05994154512882233, -0.17016547918319702, -0.015063108876347542, -0.04285911098122597, -0.06520961970090866, -0.08418621122837067, -0.06258635222911835, 0.1028047502040863, 0.05036497488617897, 0.037443552166223526, -0.07239457964897156, -0.07109133899211884, 0.00712712574750185, -0.023250464349985123, -0.02447434514760971, 0.11420147866010666, 0.06716551631689072, -0.11119965463876724, 0.09866861253976822, 0.07664984464645386, 0.029714545235037804, 0.081314817070961, -0.026460956782102585, -0.10333805531263351, -0.02987206168472767, 0.047711081802845, 0.012710303999483585, 0.1535472720861435, -0.07919952273368835, 0.04847134277224541, 0.04236757382750511, -0.03324655070900917, 0.0498337596654892, -0.09778207540512085, 0.012127798050642014, 0.005471101496368647, -0.010423904284834862, 0.019459696486592293, -0.02341809868812561, 0.012817231938242912, 0.08085740357637405, 0.05212326720356941, 0.0306490957736969, 0.03414905071258545, -0.03192135691642761, -0.12666679918766022, 0.180520698428154, -0.10406907647848129, -0.2341674566268921, -0.15631847083568573, 0.05662774667143822, 0.05440960079431534, -0.01599929668009281, 0.025788674131035805, -0.05863429233431816, -0.10053522139787674, -0.07932037115097046, -0.00039412869955413043, 0.02934226021170616, -0.06844586133956909, -0.0697588250041008, 0.05478376895189285, 0.04027644917368889, -0.11884526908397675, 0.03800419718027115, 0.06230232119560242, -0.01963106170296669, 0.00817993376404047, 0.05443860962986946, 0.08367152512073517, 0.1851641833782196, -0.009397785179316998, -0.005272079724818468, 0.05240032821893692, 0.27972644567489624, -0.15969666838645935, 0.11143927276134491, 0.11566600203514099, -0.0676729753613472, 0.079993836581707, 0.19276244938373566, 0.031111817806959152, -0.10343173891305923, 0.035727959126234055, 0.029416367411613464, -0.024281971156597137, -0.2687993347644806, -0.048981089144945145, -0.011912602931261063, -0.09494394809007645, 0.07778402417898178, 0.08550424128770828, 0.08882345259189606, 0.035288695245981216, -0.06378500908613205, -0.09618993103504181, 0.03923006355762482, 0.10156437009572983, -0.02127963677048683, 0.0045198858715593815, 0.08195331692695618, -0.028203364461660385, 0.0051971618086099625, 0.08904259651899338, -0.012480389326810837, 0.16496773064136505, 0.05493351072072983, 0.11151713877916336, 0.07972449064254761, 0.0957811251282692, -0.004405023064464331, 0.023521078750491142, 0.016099847853183746, 0.023084251210093498, 0.01289111003279686, -0.08184043318033218, 0.027636084705591202, 0.10693375021219254, 0.03995392471551895, 0.02337908186018467, 0.01462940126657486, -0.04491084814071655, 0.051204681396484375, 0.1860136240720749, 0.015520991757512093, -0.20027755200862885, -0.07545466721057892, 0.05730283632874489, -0.07728450745344162, -0.14053741097450256, -0.0176075492054224, 0.025777561590075493, -0.16871507465839386, 0.011927909217774868, -0.04473460465669632, 0.10090383887290955, -0.07111002504825592, -0.037563178688287735, 0.09582705795764923, 0.06932663172483444, -0.023215198889374733, 0.060129713267087936, -0.19539691507816315, 0.12666799128055573, 0.022838624194264412, 0.07298749685287476, -0.0895838588476181, 0.09916916489601135, -0.0012288594152778387, -0.011584814637899399, 0.1641952395439148, 0.001357188099063933, -0.07620725780725479, -0.06010977178812027, -0.08906254172325134, -0.014810220338404179, 0.10829322040081024, -0.13085880875587463, 0.06802394241094589, -0.0162927508354187, -0.030407987534999847, 0.004131825640797615, -0.08247561007738113, -0.12173120677471161, -0.1741877794265747, 0.05575304105877876, -0.10765613615512848, 0.03913215175271034, -0.09199798852205276, -0.0628759115934372, 0.007102193310856819, 0.1793217957019806, -0.18385601043701172, -0.09184932708740234, -0.14156349003314972, -0.09236942231655121, 0.16691496968269348, -0.03811681643128395, 0.08896732330322266, 0.0006371980998665094, 0.167173832654953, 0.013212242163717747, 0.0008053527562879026, 0.09846283495426178, -0.08838007599115372, -0.195245623588562, -0.060053374618291855, 0.16687048971652985, 0.1390485316514969, 0.03822357952594757, -0.008282543160021305, 0.028799647465348244, -0.049520693719387054, -0.10963472723960876, 0.02337036468088627, 0.130062997341156, 0.08295401930809021, -0.010147323831915855, -0.034744743257761, -0.10413122922182083, -0.06751642376184464, -0.05470778048038483, 0.004491682164371014, 0.19151616096496582, -0.07348428666591644, 0.15958790481090546, 0.12157517671585083, -0.05851850286126137, -0.2073076218366623, 0.04764602333307266, 0.05547701194882393, 0.010181306861341, 0.03388967365026474, -0.19430796802043915, 0.0895812064409256, -0.002211875980719924, -0.07187002897262573, 0.15506196022033691, -0.1751798391342163, -0.1419072151184082, 0.09645676612854004, 0.031242331489920616, -0.23179180920124054, -0.140110582113266, -0.10059773176908493, -0.020523587241768837, -0.11910218745470047, 0.06246551498770714, 0.004154243040829897, 0.013435576111078262, 0.03232221677899361, 0.022625509649515152, 0.0285483468323946, -0.05156245827674866, 0.20150147378444672, -0.021617839112877846, 0.0116057638078928, -0.05236828327178955, -0.09507346153259277, 0.03640175983309746, -0.050437383353710175, 0.09657271951436996, 0.0027802723925560713, 0.027580678462982178, -0.14331470429897308, -0.04037529602646828, -0.06119441241025925, 0.027803603559732437, -0.1000802218914032, -0.08828406035900116, -0.04697757214307785, 0.09533479064702988, 0.09622817486524582, -0.03175888583064079, 0.00981615949422121, -0.0842791348695755, 0.07180384546518326, 0.2082255482673645, 0.1852945238351822, 0.06956380605697632, -0.0607856884598732, 0.021769147366285324, -0.03267969191074371, 0.041049811989068985, -0.21977783739566803, 0.042970363050699234, 0.055601269006729126, 0.023724175989627838, 0.0868862122297287, -0.009627060033380985, -0.15081721544265747, -0.07427337765693665, 0.08015687763690948, -0.046447642147541046, -0.16000273823738098, -0.01915564388036728, 0.04663510248064995, -0.20957869291305542, -0.04341138154268265, 0.019573623314499855, -0.016891632229089737, -0.04244924709200859, 0.02295130304992199, 0.08179136365652084, -0.018676547333598137, 0.11014609783887863, 0.08830604702234268, 0.09282010793685913, -0.10115926712751389, 0.08067424595355988, 0.07793068140745163, -0.046769317239522934, 0.024064118042588234, 0.10837456583976746, -0.04830269142985344, -0.03586379438638687, 0.09512439370155334, 0.0946870818734169, 0.024960441514849663, -0.04589400440454483, 0.01367401797324419, -0.0520394966006279, 0.06197408214211464, 0.11672927439212799, 0.03141143545508385, -0.004231251776218414, 0.05709967017173767, 0.03799550607800484, -0.10058658570051193, 0.10852540284395218, 0.06427377462387085, 0.024908002465963364, -0.03996323049068451, -0.02511623315513134, -0.010927225463092327, -0.016850082203745842, -0.01710355468094349, -0.005908406805247068, -0.08984007686376572, -0.006262491457164288, -0.09741608798503876, 0.02910245582461357, -0.07385318726301193, 0.008803381584584713, 0.02880743332207203, -0.04584013670682907, 0.008316919207572937, 0.004094703122973442, -0.07695653289556503, -0.05422542989253998, -0.017286941409111023, 0.08707528561353683, -0.13027890026569366, 0.0341714583337307, 0.07813842594623566, -0.10886245220899582, 0.06746217608451843, -0.002395555842667818, 0.007175484672188759, 0.015258047729730606, -0.1680174022912979, 0.05064128711819649, -0.02307257056236267, -0.012751167640089989, 0.015237638726830482, -0.20685844123363495, -0.013277353718876839, -0.04568338766694069, -0.04917440563440323, 0.010579238645732403, -0.025905627757310867, -0.12403898686170578, 0.09993818402290344, -0.004685746040195227, -0.0767570436000824, -0.017627349123358727, 0.038374729454517365, 0.09996167570352554, -0.02670254558324814, 0.13602839410305023, -0.026359569281339645, 0.07776254415512085, -0.17095716297626495, -0.005339608062058687, -0.013490557670593262, 0.036647483706474304, -0.02355373091995716, -0.023600704967975616, 0.05970179662108421, -0.020983943715691566, 0.17308658361434937, -0.024301797151565552, 0.06980177015066147, 0.053884800523519516, 0.012619086541235447, 0.006580646615475416, 0.0890296995639801, 0.05994889512658119, -0.00224067154340446, -0.00451831566169858, 0.03458336740732193, -0.007273669820278883, -0.04424568638205528, -0.15858854353427887, 0.06816443055868149, 0.1642368882894516, 0.04969542473554611, 0.021598421037197113, 0.031035330146551132, -0.11431293189525604, -0.07415957748889923, 0.13565295934677124, -0.009394350461661816, -0.03296235576272011, -0.07538527995347977, 0.1824072301387787, 0.12872789800167084, -0.19966235756874084, 0.07852865755558014, -0.06783908605575562, -0.05593901127576828, -0.1232931911945343, -0.15602804720401764, -0.0648919865489006, -0.04340577498078346, -0.019624128937721252, -0.0639692172408104, 0.05613270401954651, 0.058813612908124924, 0.001425987109541893, -0.018078278750181198, 0.10630279779434204, 0.009662417694926262, -0.023386240005493164, 0.045132674276828766, 0.059963125735521317, 0.03376534953713417, -0.09632086008787155, 0.0096770990639925, -0.002340735634788871, 0.025244370102882385, 0.0662512481212616, 0.017782168462872505, -0.053059570491313934, 0.009303494356572628, -0.02173302322626114, -0.1160469576716423, 0.04477434605360031, -0.017120063304901123, -0.039619963616132736, 0.1449768990278244, 0.02904772013425827, 0.009236602112650871, -0.01978018693625927, 0.2358154058456421, -0.07457811385393143, -0.08457175642251968, -0.15541023015975952, 0.06076590716838837, -0.06565217673778534, 0.035359419882297516, 0.034571196883916855, -0.11856629699468613, 0.018512191250920296, 0.16088688373565674, 0.12946905195713043, -0.011538301594555378, 0.009198445826768875, 0.05150740593671799, 0.0027183787897229195, -0.03291062265634537, 0.017550481483340263, 0.054291024804115295, 0.14000749588012695, -0.07372044026851654, 0.06561611592769623, -0.01187864225357771, -0.07844166457653046, -0.021566620096564293, 0.10251908749341965, -0.0038683004677295685, 0.0015118604060262442, -0.06856850534677505, 0.14383848011493683, -0.084647536277771, -0.2348589301109314, 0.053169168531894684, -0.0701572448015213, -0.14949968457221985, -0.04537404328584671, 0.023992007598280907, -0.016428330913186073, 0.01799216866493225, 0.0795777291059494, -0.049674998968839645, 0.17259933054447174, 0.04154284670948982, -0.04437325522303581, -0.08154232054948807, 0.05796421319246292, -0.1414758861064911, 0.28577372431755066, 0.02069827914237976, 0.048000313341617584, 0.1059737354516983, -0.01667015440762043, -0.13934561610221863, 0.011938226409256458, 0.10723932832479477, -0.0637049525976181, 0.062433384358882904, 0.17652811110019684, -0.0022143926471471786, 0.12586885690689087, 0.0577392578125, -0.05767730250954628, 0.04133390262722969, -0.08329254388809204, -0.0484866127371788, -0.11132916063070297, 0.07939464598894119, -0.08110078424215317, 0.1582714021205902, 0.13221341371536255, -0.06497745960950851, -0.0005526122986339033, -0.02006004936993122, 0.08089762926101685, 0.0047617023810744286, 0.1109834760427475, 0.0031623311806470156, -0.1938789188861847, 0.03582930192351341, 0.013228845782577991, 0.10180895030498505, -0.2042633295059204, -0.06707853823900223, 0.050121136009693146, -0.018932392820715904, -0.07713542133569717, 0.11884412914514542, 0.04402478411793709, 0.03133934736251831, -0.03841261938214302, -0.04394432157278061, 0.0010597195941954851, 0.15094342827796936, -0.11137887090444565, -0.009867769666016102 ]
null
null
ml-agents
# **ppo** Agent playing **SnowballTarget** This is a trained model of a **ppo** agent playing **SnowballTarget** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: vpepe2003/ppo-SnowballTarget 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
{"library_name": "ml-agents", "tags": ["SnowballTarget", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-SnowballTarget"]}
reinforcement-learning
vpepe2003/ppo-SnowballTarget
[ "ml-agents", "tensorboard", "onnx", "SnowballTarget", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-SnowballTarget", "region:us" ]
2024-02-15T01:13:58+00:00
[]
[]
TAGS #ml-agents #tensorboard #onnx #SnowballTarget #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SnowballTarget #region-us
# ppo Agent playing SnowballTarget This is a trained model of a ppo agent playing SnowballTarget using the Unity ML-Agents Library. ## Usage (with ML-Agents) The Documentation: URL We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your browser: URL - A *longer tutorial* to understand how works ML-Agents: URL ### Resume the training ### Watch your Agent play You can watch your agent playing directly in your browser 1. If the environment is part of ML-Agents official environments, go to URL 2. Step 1: Find your model_id: vpepe2003/ppo-SnowballTarget 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play
[ "# ppo Agent playing SnowballTarget\n This is a trained model of a ppo agent playing SnowballTarget\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: vpepe2003/ppo-SnowballTarget\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play" ]
[ "TAGS\n#ml-agents #tensorboard #onnx #SnowballTarget #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SnowballTarget #region-us \n", "# ppo Agent playing SnowballTarget\n This is a trained model of a ppo agent playing SnowballTarget\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: vpepe2003/ppo-SnowballTarget\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play" ]
[ 50, 207 ]
[ "passage: TAGS\n#ml-agents #tensorboard #onnx #SnowballTarget #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SnowballTarget #region-us \n# ppo Agent playing SnowballTarget\n This is a trained model of a ppo agent playing SnowballTarget\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: vpepe2003/ppo-SnowballTarget\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play" ]
[ -0.04439889267086983, 0.049912337213754654, -0.003807131666690111, 0.10350304841995239, 0.17276470363140106, -0.015814581885933876, 0.12825283408164978, 0.10410739481449127, 0.11405503004789352, 0.060368914157152176, 0.09876507520675659, 0.09183704853057861, 0.06017263978719711, 0.1393967717885971, 0.09389042109251022, -0.2333889901638031, -0.04013935104012489, -0.09789051115512848, 0.011740473099052906, 0.08099237829446793, 0.04829489067196846, -0.04430742189288139, 0.03105415403842926, 0.05479073151946068, -0.012375447899103165, 0.001877451315522194, -0.07071779668331146, -0.04090152308344841, 0.07296263426542282, -0.03161272406578064, 0.01269887201488018, -0.053871579468250275, 0.10277482867240906, -0.1650945395231247, 0.023922165855765343, 0.05132848769426346, -0.017338555306196213, -0.01896302029490471, 0.1542959213256836, 0.02448205091059208, 0.09159449487924576, -0.11572563648223877, 0.1018763929605484, 0.07786574214696884, -0.050788022577762604, -0.03040768764913082, -0.07090898603200912, 0.05935535952448845, 0.2108878344297409, 0.13886861503124237, -0.003951001912355423, 0.09763488918542862, -0.027230344712734222, 0.059272218495607376, 0.17626598477363586, -0.2916005849838257, -0.06777605414390564, 0.16738690435886383, -0.0439586378633976, 0.03620297089219093, -0.00618718983605504, 0.047129642218351364, -0.012498414143919945, 0.030589129775762558, 0.0016045282827690244, 0.03409426659345627, 0.2646746039390564, 0.015753764659166336, -0.10524630546569824, -0.09529033303260803, 0.0013331403024494648, 0.030075080692768097, -0.04531341418623924, -0.1776294708251953, 0.003299141302704811, 0.10648646950721741, 0.010740037076175213, 0.025454986840486526, 0.06234362721443176, 0.009303842671215534, -0.08463209122419357, -0.14547188580036163, -0.042302705347537994, -0.06671007722616196, 0.11597026139497757, 0.08277010172605515, -0.02043319307267666, -0.004769916646182537, 0.03648645803332329, 0.08885723352432251, 0.10390873998403549, -0.034627314656972885, -0.03553019091486931, -0.02903798222541809, -0.14681802690029144, -0.012364516034722328, -0.029399383813142776, -0.019624417647719383, 0.026516560465097427, 0.14849282801151276, 0.17249685525894165, 0.03488928824663162, 0.039508819580078125, 0.027542119845747948, 0.0038056254852563143, 0.11737038940191269, 0.0426822155714035, -0.02771267294883728, 0.003652283689007163, 0.009011994116008282, 0.055326804518699646, -0.08130042999982834, -0.09061054140329361, 0.04101848229765892, -0.03164370357990265, 0.13506275415420532, 0.16029231250286102, -0.02187366411089897, -0.005869652610272169, -0.0362321101129055, 0.027937714010477066, -0.1446511596441269, 0.07424338161945343, 0.05803355947136879, -0.047366607934236526, -0.06323760747909546, -0.049662839621305466, 0.04272148758172989, -0.09555111825466156, 0.03350168839097023, 0.0012735265772789717, 0.0661223977804184, -0.002469439059495926, -0.04839840158820152, 0.0525270476937294, -0.11592737585306168, -0.01751389540731907, -0.17400577664375305, -0.1309453845024109, -0.08273766934871674, 0.02843782678246498, -0.03977731615304947, -0.11093488335609436, -0.09593846648931503, 0.04453174024820328, -0.07526736706495285, 0.023911992087960243, -0.014623391442000866, -0.06311451643705368, -0.029820704832673073, -0.11849825084209442, 0.063578762114048, 0.16816580295562744, 0.007619470823556185, -0.029468584805727005, 0.02340048737823963, -0.14744606614112854, 0.1433466374874115, -0.14740948379039764, 0.15734317898750305, -0.08529789745807648, 0.029832374304533005, 0.13006611168384552, -0.025232218205928802, 0.04696934297680855, 0.17751240730285645, -0.11887640506029129, -0.0731092095375061, 0.04769169166684151, -0.09618501365184784, -0.11986248195171356, 0.05453348532319069, 0.027674153447151184, 0.04964491352438927, 0.07138243317604065, 0.19096755981445312, 0.09816774725914001, -0.24983340501785278, 0.043363552540540695, 0.015424055978655815, -0.1361333429813385, -0.010392051190137863, 0.11941356211900711, -0.07511521130800247, -0.0069997264072299, -0.0339416079223156, -0.13658523559570312, 0.09284958988428116, -0.01177387684583664, -0.06102845445275307, 0.02772834524512291, -0.04992612451314926, -0.04872873052954674, -0.010275838896632195, 0.04068721830844879, -0.043224260210990906, -0.05375751852989197, -0.024916693568229675, 0.040121205151081085, -0.0011865461710840464, 0.07786134630441666, -0.026696542277932167, 0.12901188433170319, -0.006976486183702946, 0.012517069466412067, -0.10546591877937317, -0.14670750498771667, -0.009748381562530994, 0.021798167377710342, 0.06168495491147041, -0.07037268579006195, 0.10140793025493622, 0.07504085451364517, 0.04373408854007721, -0.07776704430580139, -0.070221446454525, 0.017934272065758705, -0.10612690448760986, -0.09921959787607193, -0.08290334045886993, -0.06097692251205444, 0.0860474705696106, -0.07306253910064697, 0.06324176490306854, -0.04813610762357712, 0.10090252757072449, -0.012120898813009262, -0.07094628363847733, 0.03470496088266373, -0.019053997471928596, 0.0457928329706192, -0.09395807236433029, 0.1013960912823677, 0.06675925850868225, -0.11534667015075684, 0.027979565784335136, 0.05924554914236069, -0.07913311570882797, 0.12648899853229523, 0.019463850185275078, -0.004019849468022585, -0.049483057111501694, -0.0637311339378357, 0.005197376478463411, -0.07488904893398285, 0.0032788163516670465, 0.19425490498542786, 0.12946929037570953, 0.08422670513391495, -0.04270714521408081, -0.06567375361919403, -0.031804345548152924, -0.04940011352300644, -0.06275977939367294, 0.12831729650497437, 0.040967561304569244, 0.007952353917062283, 0.03360868617892265, 0.011935165151953697, 0.09045687317848206, 0.1162833571434021, -0.0016064755618572235, -0.1226859837770462, 0.02285338193178177, 0.050398848950862885, 0.05314916744828224, 0.001684927032329142, 0.04631058871746063, -0.024190226569771767, -0.008766081184148788, -0.053415775299072266, -0.013227241113781929, -0.1056358739733696, -0.05748385563492775, 0.0652289092540741, -0.008809608407318592, 0.00861053541302681, -0.07159294933080673, -0.047804005444049835, 0.030375920236110687, 0.09617329388856888, 0.003588469000533223, 0.03341767191886902, -0.03733586519956589, -0.13351133465766907, 0.042717162519693375, -0.09560248255729675, -0.24289429187774658, -0.1046457588672638, -0.03534165397286415, -0.06134407967329025, 0.021022098138928413, 0.07497158646583557, -0.19310490787029266, -0.011680546216666698, -0.0920649915933609, 0.013425450772047043, -0.01889989525079727, -0.024640226736664772, 0.14397013187408447, 0.10401784628629684, -0.017751427367329597, -0.07631096988916397, 0.013951900415122509, 0.006737313233315945, -0.06730537861585617, -0.010909533128142357, 0.08434688299894333, 0.09350844472646713, 0.059004124253988266, 0.05926733836531639, 0.06281601637601852, -0.029429994523525238, 0.17099662125110626, -0.043137840926647186, 0.023054324090480804, 0.07696947455406189, -0.005731812212616205, 0.07227520644664764, 0.0003919210867024958, 0.02846631407737732, -0.002905979286879301, 0.007259448524564505, 0.007077640388160944, -0.07563012093305588, -0.214236319065094, -0.0842786356806755, 0.0015539317391812801, 0.17849363386631012, 0.16968224942684174, 0.09424608200788498, -0.11814949661493301, 0.02478238381445408, -0.0012255578767508268, -0.09634961187839508, 0.10459250211715698, 0.13096144795417786, -0.05209605023264885, -0.028606392443180084, 0.03435239568352699, -0.037211865186691284, 0.0626690611243248, 0.05367575213313103, -0.036472711712121964, 0.09636746346950531, 0.021306179463863373, -0.004393907729536295, -0.02035515196621418, -0.047393474727869034, -0.053058285266160965, 0.1419130265712738, 0.06290753185749054, 0.02645612694323063, 0.01606980338692665, -0.06749073415994644, -0.07794420421123505, 0.13299040496349335, 0.15939490497112274, -0.07214327156543732, -0.05695412680506706, 0.10147866606712341, 0.058641862124204636, 0.19945979118347168, -0.002561088651418686, -0.11719062179327011, -0.06041194126009941, -0.003692860249429941, -0.11803378909826279, 0.004608902148902416, 0.03840266540646553, -0.018890855833888054, -0.16754643619060516, 0.053641147911548615, 0.00425563333556056, 0.11116587370634079, 0.010936981998383999, -0.02598956972360611, 0.059420108795166016, 0.0077707841992378235, -0.026413362473249435, 0.051256828010082245, -0.14701761305332184, 0.034378260374069214, -0.005358757451176643, 0.09627066552639008, -0.05846956744790077, 0.029979413375258446, 0.07256167382001877, -0.04311668127775192, 0.16682952642440796, 0.04708149656653404, -0.015118398703634739, -0.11901576071977615, -0.16186672449111938, -0.05285417288541794, -0.03138944134116173, -0.10328014940023422, 0.06714294105768204, 0.03678945451974869, -0.019533870741724968, -0.09778768569231033, 0.011873756535351276, -0.05283432453870773, -0.10876956582069397, -0.04118747636675835, -0.09213145822286606, 0.07401678711175919, -0.05951050668954849, -0.07617615908384323, -0.09962309151887894, 0.1645018458366394, 0.07296931743621826, -0.11294091492891312, -0.11480977386236191, 0.01304116565734148, -0.05967115983366966, -0.029612800106406212, 0.086868055164814, 0.016042165458202362, 0.12004861980676651, -0.11136722564697266, -0.05044236034154892, -0.03054976835846901, -0.115926094353199, -0.1004403755068779, 0.03367765247821808, 0.17742595076560974, 0.03181007131934166, 0.0812879353761673, -0.00792812742292881, 0.09610220789909363, -0.01350976899266243, -0.07309939712285995, 0.12832756340503693, 0.10278724879026413, -0.04163403809070587, 0.04091760888695717, 0.03159418702125549, 0.06303171813488007, -0.11016447097063065, -0.013124278746545315, 0.22327043116092682, 0.28019335865974426, -0.07243723422288895, 0.20109272003173828, 0.02022855170071125, -0.04737080633640289, -0.16928432881832123, -0.0589495450258255, 0.02192763052880764, -0.053337499499320984, 0.10622058063745499, -0.1940634697675705, 0.08524646610021591, 0.010777327232062817, -0.007860255427658558, 0.028025364503264427, -0.14877745509147644, -0.08473508059978485, 0.029032031074166298, 0.09811859577894211, -0.07285894453525543, -0.07967051863670349, -0.07543997466564178, 0.00911183562129736, -0.07917997241020203, 0.0251882616430521, -0.09623318910598755, 0.05010443180799484, 0.009122335352003574, 0.03022541105747223, 0.05294404551386833, -0.05711977556347847, 0.12896032631397247, -0.030880529433488846, -0.06015218049287796, -0.06273084878921509, 0.04533303529024124, -0.024560583755373955, -0.08710303902626038, 0.06170625239610672, 0.003499547950923443, -0.020466504618525505, -0.19638752937316895, -0.048705361783504486, 0.008837164379656315, 0.04148681089282036, -0.039808813482522964, -0.08542324602603912, -0.021367903798818588, 0.04883788153529167, 0.0867818221449852, 0.017163073644042015, 0.11308714002370834, 0.007627409417182207, -0.013809598982334137, 0.07197991758584976, 0.027088522911071777, 0.024973012506961823, -0.15486736595630646, -0.06466183066368103, -0.06882050633430481, 0.01694786362349987, -0.0482020229101181, -0.01317096408456564, 0.05268671363592148, 0.05663503333926201, -0.014400750398635864, 0.05438785254955292, -0.09375910460948944, -0.01861260086297989, 0.02131059393286705, -0.1040828675031662, -0.10241839289665222, -0.0819472000002861, -0.1112980991601944, 0.026240704581141472, -0.07903856039047241, 0.09371311217546463, -0.05948198586702347, -0.006040521431714296, 0.012378808110952377, 0.03473612666130066, -0.0034899129532277584, 0.050429847091436386, 0.024858389049768448, 0.0338716134428978, -0.076564721763134, 0.12768876552581787, 0.00018708956486079842, -0.040120501071214676, 0.048346009105443954, 0.18887491524219513, -0.06671226769685745, -0.0724528357386589, -0.05619222670793533, 0.07075279206037521, 0.040167566388845444, -0.0174123365432024, -0.03882138058543205, -0.05595460534095764, 0.12628024816513062, -0.1579785794019699, 0.01732769049704075, -0.10993709415197372, 0.010009615682065487, 0.053764089941978455, -0.04177122190594673, 0.06997764855623245, -0.018297526985406876, -0.06129369139671326, -0.14589554071426392, 0.08161262422800064, 0.032911866903305054, 0.09035318344831467, -0.011273279786109924, -0.02841959334909916, -0.13559521734714508, 0.031566742807626724, -0.020666878670454025, -0.0040431213565170765, -0.15753604471683502, 0.016964925453066826, -0.002849607029929757, 0.036485470831394196, 0.030817708000540733, 0.07011096924543381, -0.04804741218686104, -0.09462463110685349, -0.05760747566819191, 0.062473662197589874, -0.08493445813655853, -0.015492663718760014, -0.02864011935889721, -0.07670464366674423, 0.0572299100458622, 0.07008442282676697, -0.01650749146938324, -0.03751889988780022, -0.061465997248888016, 0.013388310559093952, -0.00981742050498724, -0.04390721768140793, 0.05201111361384392, -0.11993372440338135, 0.021418794989585876, -0.06417661160230637, -0.12725254893302917, 0.03302133083343506, 0.12451978772878647, -0.07845497131347656, 0.051736410707235336, 0.06033722683787346, -0.0800190344452858, -0.06704964488744736, -0.015598664060235023, 0.06324461102485657, 0.057241473346948624, 0.1066289097070694, -0.07761025428771973, 0.1949520856142044, -0.11693723499774933, -0.030699174851179123, 0.013303917832672596, 0.0680723637342453, 0.029617005959153175, -0.08039536327123642, 0.044304970651865005, -0.018971489742398262, 0.06518574804067612, 0.08084764331579208, 0.013769801706075668, 0.05850149691104889, 0.051510442048311234, 0.1220434159040451, 0.017480183392763138, 0.07154817134141922, -0.01459385547786951, 0.031480953097343445, 0.10301093757152557, -0.00784341711550951, 0.07962087541818619, -0.07148469239473343, 0.06526653468608856, 0.0379079170525074, 0.0950070172548294, 0.07327516376972198, 0.059675879776477814, -0.09196990728378296, -0.18207429349422455, -0.05487143620848656, 0.0463142991065979, 0.05430727079510689, -0.04773616045713425, 0.1613515168428421, 0.13261792063713074, -0.1941465586423874, 0.006138294469565153, 0.0022727278992533684, 0.04067930206656456, -0.07487674802541733, -0.09003807604312897, 0.009953228756785393, -0.1464322805404663, 0.09594099968671799, -0.022355657070875168, -0.006990892346948385, -0.03656045347452164, 0.010325705632567406, 0.029902691021561623, 0.047721270471811295, -0.041943326592445374, -0.00810930784791708, 0.04391103237867355, -0.028702834621071815, 0.005818307399749756, -0.004232521168887615, -0.08136624842882156, -0.03763638064265251, -0.059385813772678375, -0.03443584218621254, 0.024983588606119156, -0.004868450108915567, 0.05598278343677521, 0.01946162059903145, -0.05807603895664215, 0.06726192682981491, -0.00008894111670088023, 0.005436551291495562, 0.21378350257873535, 0.09373034536838531, -0.04582633823156357, -0.04835307598114014, 0.1995811015367508, -0.03234390541911125, -0.05511699244379997, -0.08308549970388412, 0.10330899059772491, -0.04051796346902847, -0.03431115671992302, -0.03325990214943886, -0.15870897471904755, -0.062353044748306274, 0.17658643424510956, 0.12426155805587769, -0.008283818140625954, 0.00995112955570221, -0.06571287661790848, 0.003333514789119363, 0.02930682711303234, 0.10246095806360245, 0.07078471034765244, 0.05573917180299759, -0.10155672580003738, 0.011209698393940926, -0.07766894996166229, -0.10254649817943573, -0.19774098694324493, 0.04814191535115242, 0.029535556212067604, -0.02424588054418564, -0.016154533252120018, 0.12394275516271591, -0.09385097026824951, -0.0962488055229187, 0.10893499106168747, -0.02988174371421337, -0.07089414447546005, 0.009996742010116577, 0.03779611736536026, 0.005830450914800167, 0.11671634018421173, 0.08371727913618088, 0.041094619780778885, 0.05971981957554817, -0.015443498268723488, -0.09908518195152283, -0.005229398142546415, 0.054501354694366455, -0.12610776722431183, 0.24415548145771027, -0.021294137462973595, 0.004418717697262764, 0.08198809623718262, 0.06762892752885818, -0.18675166368484497, 0.009775658138096333, 0.055581118911504745, -0.17817506194114685, 0.015593589283525944, 0.0834864154458046, -0.043012574315071106, 0.026847492903470993, 0.04824228957295418, -0.04210304841399193, 0.001986941322684288, 0.1948210448026657, 0.039176441729068756, -0.03266752138733864, 0.0799291655421257, -0.14214113354682922, 0.10098541527986526, 0.09592614322900772, -0.06339462101459503, 0.0008535239030607045, -0.04089382663369179, 0.0071254814974963665, -0.00547598535194993, -0.02326262928545475, -0.011699257418513298, -0.12205637246370316, -0.01956801302731037, -0.035064175724983215, 0.023856744170188904, -0.1793336570262909, -0.13307024538516998, -0.058800823986530304, -0.08044157177209854, -0.051162220537662506, 0.08414264023303986, 0.0561566986143589, -0.0556475892663002, 0.019068434834480286, -0.085514597594738, 0.026086416095495224, 0.14460380375385284, -0.07263306528329849, -0.01253407634794712 ]
null
null
null
[phi-2](https://huggingface.co/microsoft/phi-2) quantized using imatrix data from [kalomaze's groups_merged.txt](https://github.com/ggerganov/llama.cpp/files/14194570/groups_merged.txt) Perplexities: Q8_0: 5.3886 Q4_0: 5.5526 IQ3_XXS: 6.0745 IQ2_XS: 7.2570 IQ2_XXS: 9.3666 IQ1_S: 18.7885
{}
null
ristew/phi-2-imatrix-gguf
[ "gguf", "region:us" ]
2024-02-15T01:22:09+00:00
[]
[]
TAGS #gguf #region-us
phi-2 quantized using imatrix data from kalomaze's groups_merged.txt Perplexities: Q8_0: 5.3886 Q4_0: 5.5526 IQ3_XXS: 6.0745 IQ2_XS: 7.2570 IQ2_XXS: 9.3666 IQ1_S: 18.7885
[]
[ "TAGS\n#gguf #region-us \n" ]
[ 9 ]
[ "passage: TAGS\n#gguf #region-us \n" ]
[ 0.030724648386240005, 0.026499787345528603, -0.010017825290560722, -0.05703527107834816, 0.08247160166501999, 0.07200847566127777, 0.01814177818596363, 0.020192064344882965, 0.2235025018453598, 0.017216520383954048, 0.1496623009443283, -0.031233953312039375, 0.006174509879201651, 0.05538657680153847, 0.039407629519701004, -0.19438467919826508, 0.058440499007701874, -0.02356063388288021, -0.020945189520716667, 0.01803453452885151, -0.05310691148042679, -0.04108472168445587, 0.022135348990559578, -0.07881014049053192, -0.15867982804775238, 0.0678698718547821, 0.017852067947387695, 0.0007025183876976371, 0.0820731669664383, 0.05882885307073593, 0.09657382220029831, -0.024203501641750336, -0.15220364928245544, -0.18796531856060028, 0.0366438589990139, -0.02974788099527359, -0.10282598435878754, 0.022019000723958015, 0.029453158378601074, -0.06967076659202576, 0.02238346077501774, 0.1427535116672516, -0.10206039994955063, 0.051592033356428146, -0.27165159583091736, -0.1715938150882721, -0.06585682183504105, -0.025845954194664955, -0.007345964200794697, 0.01241085771471262, -0.0010092189768329263, 0.047266922891139984, -0.20188692212104797, -0.005631127394735813, 0.09329266101121902, -0.25229454040527344, 0.02776304818689823, 0.21345718204975128, -0.010520953685045242, 0.09873088449239731, -0.05590669438242912, 0.14438565075397491, 0.03173782303929329, -0.019559340551495552, -0.1924813836812973, -0.070224329829216, -0.07177317887544632, 0.162109375, -0.0823177620768547, -0.11764442175626755, 0.24176421761512756, 0.009283576160669327, -0.026472626253962517, 0.15598991513252258, -0.029037300497293472, -0.009749599732458591, 0.04555726423859596, 0.01668328419327736, -0.010545015335083008, 0.1551385223865509, 0.17108163237571716, -0.08598228543996811, -0.10847756266593933, -0.030579885467886925, -0.2373785674571991, 0.2470305860042572, -0.01911027915775776, 0.12945520877838135, -0.20086053013801575, 0.018443629145622253, -0.3247532844543457, -0.0012029389617964625, -0.010316703468561172, -0.028618358075618744, -0.006935348734259605, 0.009301352314651012, -0.050316113978624344, 0.0739501491189003, 0.14580395817756653, 0.1393439620733261, -0.11465669423341751, 0.060509420931339264, -0.052172139286994934, 0.14876529574394226, 0.05827285721898079, 0.061183393001556396, 0.04079163819551468, 0.07037676870822906, -0.008353544399142265, -0.21633195877075195, -0.029873060062527657, -0.07057386636734009, -0.08445251733064651, -0.0130265261977911, -0.13896764814853668, 0.11386743932962418, -0.022273007780313492, -0.07913482189178467, -0.06810981780290604, 0.07626928389072418, 0.017650218680500984, -0.008536403998732567, -0.035703565925359726, -0.012481719255447388, 0.022218508645892143, -0.014872739091515541, -0.1519843488931656, 0.02295425534248352, 0.10455024242401123, 0.07257117331027985, -0.1489023119211197, -0.011344035156071186, -0.017298875376582146, 0.06959983706474304, 0.03884255141019821, -0.10402916371822357, 0.04283881187438965, -0.10747409611940384, -0.08414466679096222, 0.022628657519817352, -0.005062851123511791, -0.0418001152575016, 0.13524691760540009, 0.03997812792658806, 0.040150050073862076, -0.016940169036388397, -0.04259050637483597, -0.048133596777915955, -0.07602019608020782, 0.07334327697753906, 0.05418020859360695, 0.027240034192800522, -0.1915341019630432, 0.01154522504657507, -0.048245880752801895, 0.09175369143486023, -0.11856856942176819, 0.014575321227312088, -0.08105122298002243, 0.1604209989309311, 0.0349995456635952, 0.09055875241756439, -0.19562625885009766, 0.02605881541967392, -0.06191767752170563, 0.1854621320962906, -0.04451294615864754, -0.11786319315433502, 0.2698904871940613, -0.09105797111988068, -0.040079716593027115, 0.056803084909915924, 0.06560484319925308, -0.06272535026073456, 0.068723164498806, 0.4434472322463989, -0.06556011736392975, -0.07118581980466843, 0.05080527812242508, 0.17805561423301697, -0.1262815296649933, -0.09372174739837646, 0.09990617632865906, -0.1480535864830017, -0.211008220911026, 0.030864350497722626, 0.028955968096852303, 0.1494358479976654, -0.06205282360315323, -0.012456154450774193, 0.058214303106069565, -0.013022401370108128, 0.046677324920892715, 0.03563477098941803, 0.11109840869903564, -0.06493768095970154, 0.06851828098297119, -0.16232267022132874, 0.016065504401922226, 0.1209988072514534, -0.015012580901384354, -0.04126624017953873, 0.14286154508590698, -0.03809087723493576, 0.07199656218290329, -0.07730832695960999, -0.1804673671722412, 0.027612121775746346, 0.05621999502182007, 0.028122514486312866, 0.09176547825336456, 0.09526687115430832, -0.039257392287254333, 0.0013902259524911642, 0.0329861082136631, 0.061223939061164856, -0.007701692637056112, 0.015235940925776958, -0.015374142676591873, 0.12888981401920319, -0.07010363042354584, -0.04155188798904419, -0.09715848416090012, -0.00889967754483223, 0.2288777232170105, -0.01933911070227623, 0.02257734164595604, -0.06854789704084396, 0.033186767250299454, -0.0012386917369440198, 0.09506335854530334, -0.017756229266524315, 0.06063338369131088, -0.022011179476976395, -0.06201287358999252, 0.11652727425098419, -0.043086208403110504, 0.24556174874305725, 0.10792262107133865, -0.07513239979743958, -0.01741042546927929, -0.0871582105755806, -0.007020947523415089, 0.022898653522133827, 0.08814648538827896, -0.04863424599170685, 0.06471672654151917, -0.037898752838373184, -0.0013588295551016927, 0.018808960914611816, -0.008487841114401817, -0.030526969581842422, -0.04284367710351944, -0.08270563185214996, 0.09057542681694031, 0.0691855251789093, -0.13670015335083008, 0.17748047411441803, 0.2472171038389206, 0.1500423550605774, 0.2487964630126953, -0.06485911458730698, -0.014139159582555294, -0.02016172744333744, 0.03673918917775154, -0.020436765626072884, 0.13109654188156128, -0.18929845094680786, -0.032152432948350906, 0.02558354288339615, 0.029807843267917633, 0.10872193425893784, -0.1365325003862381, -0.1145850270986557, -0.0379912331700325, -0.047677598893642426, -0.08257206529378891, 0.07034620642662048, -0.12104500830173492, 0.03338077291846275, 0.07256745547056198, 0.0073080710135400295, 0.12201625853776932, 0.015417544171214104, -0.055278971791267395, 0.0998256728053093, -0.14543165266513824, -0.2384990155696869, -0.04642500355839729, -0.10990478098392487, 0.001206184271723032, 0.05318264663219452, 0.016633260995149612, -0.21265560388565063, -0.01741623878479004, 0.11141498386859894, 0.06650645285844803, -0.18111048638820648, 0.024138791486620903, 0.029385030269622803, -0.004455238115042448, -0.10212790220975876, -0.012687300331890583, -0.05387670546770096, -0.11039627343416214, -0.0691843032836914, 0.08163908869028091, -0.06936442852020264, 0.11164893209934235, 0.1582336574792862, 0.11141853034496307, 0.11249161511659622, -0.011774544604122639, 0.1976311057806015, -0.14119699597358704, -0.14489109814167023, 0.06405922025442123, -0.014498869888484478, 0.03640124574303627, 0.08232609927654266, 0.04930112138390541, -0.14269955456256866, -0.04848511889576912, -0.007545206230133772, -0.1497725397348404, -0.1323675513267517, -0.05164776369929314, -0.10658133774995804, 0.12379065901041031, -0.06248227879405022, 0.10150982439517975, 0.11162466555833817, 0.017522823065519333, 0.11151766777038574, -0.06246228888630867, -0.054680291563272476, -0.04807431995868683, 0.06297076493501663, -0.05410824716091156, -0.04205694422125816, -0.06721562892198563, -0.008002115413546562, 0.1349310278892517, 0.10885956883430481, 0.07581131905317307, 0.2265089601278305, 0.02780294418334961, 0.05355561524629593, 0.040789585560560226, 0.16015571355819702, 0.015284501947462559, -0.0046128155663609505, -0.08788388222455978, -0.014365277253091335, -0.0019687749445438385, -0.031080376356840134, -0.006052241660654545, 0.1340780407190323, -0.2559821307659149, 0.03235609456896782, -0.2989844083786011, 0.11946471780538559, -0.1565471589565277, 0.07426489144563675, 0.05220162868499756, 0.030080994591116905, 0.08841689676046371, 0.035069406032562256, -0.02871096506714821, 0.09149409085512161, 0.11694692075252533, -0.12628670036792755, 0.01540512777864933, 0.04918349161744118, 0.052707213908433914, -0.0142430504783988, 0.0931062400341034, -0.11024625599384308, -0.0737583339214325, -0.0024255106691271067, 0.07025767862796783, -0.2099330574274063, 0.23986183106899261, 0.03523903712630272, -0.10871971398591995, -0.021638909354805946, -0.0547538623213768, 0.03316742554306984, 0.08983159810304642, 0.1342458724975586, 0.11251148581504822, -0.11371640861034393, -0.12470904737710953, 0.029020745307207108, 0.03679748624563217, 0.1757190227508545, -0.09047917276620865, -0.14164063334465027, 0.001811441034078598, 0.05263577029109001, -0.053646381944417953, 0.07645093649625778, -0.05327983945608139, -0.0941789522767067, 0.03495060279965401, 0.04520740360021591, 0.00641082925722003, -0.019971303641796112, 0.08110581338405609, -0.02520396187901497, 0.085345059633255, -0.04878882318735123, 0.00847524031996727, -0.10202991217374802, -0.03634759038686752, 0.04376819357275963, -0.0722225159406662, 0.01614394783973694, -0.09818518906831741, -0.15651735663414001, -0.08556577563285828, -0.15303048491477966, 0.12497064471244812, -0.052672382444143295, 0.10244213044643402, -0.047614291310310364, 0.147609144449234, -0.013274060562252998, 0.030878636986017227, -0.05167607590556145, 0.028036773204803467, 0.011671020649373531, -0.14858771860599518, 0.20959575474262238, -0.1476162225008011, -0.023819662630558014, 0.16589532792568207, 0.05426561459898949, 0.1161220371723175, 0.04555299133062363, -0.0879630371928215, 0.23518426716327667, 0.2702784240245819, -0.0007818902959115803, 0.17838320136070251, 0.2352202981710434, -0.026693791151046753, -0.2436053603887558, -0.07260585576295853, -0.2063993662595749, -0.039628319442272186, 0.0004186074365861714, -0.282958060503006, 0.06042884290218353, 0.17210599780082703, -0.07570867985486984, 0.4319494664669037, -0.22352926433086395, 0.03153151646256447, 0.13982820510864258, -0.04242865741252899, 0.6181237101554871, -0.1820172369480133, -0.16550765931606293, 0.052592549473047256, -0.1248052790760994, 0.11609237641096115, -0.005267696920782328, 0.10048385709524155, -0.00011838242062367499, -0.02595684304833412, 0.03428659215569496, -0.0409976989030838, 0.23620888590812683, 0.018790103495121002, 0.045043930411338806, -0.09004033356904984, -0.1538960188627243, 0.10746775567531586, 0.02556895837187767, -0.10341835021972656, 0.03920651972293854, -0.06092366203665733, -0.10915451496839523, 0.011575369164347649, -0.08317004889249802, 0.03433287888765335, 0.09550272673368454, -0.050003789365291595, -0.0652989074587822, 0.024777809157967567, -0.16975140571594238, 0.028226720169186592, 0.1660151481628418, -0.08661750704050064, 0.17001861333847046, -0.04084239527583122, -0.0947834923863411, -0.15362800657749176, -0.020637191832065582, -0.07918675988912582, -0.01597081869840622, 0.10419487953186035, -0.11003783345222473, 0.006433290895074606, 0.09035904705524445, 0.002910176757723093, 0.07882846146821976, 0.09883374720811844, -0.08716033399105072, 0.05550702288746834, 0.1730797290802002, -0.21496161818504333, -0.1694899946451187, -0.04902869462966919, -0.1887752115726471, 0.2065081000328064, 0.03903897479176521, 0.04895683750510216, 0.16432031989097595, 0.015995748341083527, -0.010867753997445107, -0.020683420822024345, -0.11664224416017532, 0.00450828718021512, 0.04868127405643463, -0.005741522181779146, -0.11094820499420166, 0.13042977452278137, 0.05625306814908981, -0.010265284217894077, -0.04014173522591591, 0.1808832287788391, -0.06324239075183868, -0.06105973571538925, -0.29144585132598877, 0.07338178157806396, -0.10203809291124344, -0.033191971480846405, 0.08307401835918427, -0.024927617982029915, -0.0012370682088658214, 0.14441034197807312, 0.009444275870919228, 0.1295502781867981, 0.031338974833488464, 0.03218937665224075, 0.14084547758102417, -0.13805074989795685, -0.14429166913032532, -0.029582731425762177, -0.08434601873159409, -0.12847381830215454, -0.016780147328972816, 0.1751313954591751, -0.08363176882266998, -0.12467111647129059, -0.2756369411945343, 0.049299292266368866, -0.0641724020242691, -0.1138453483581543, -0.03101496584713459, -0.06544762849807739, 0.052310146391391754, -0.040101904422044754, 0.014005003497004509, -0.023109296336770058, -0.14451682567596436, 0.0458921417593956, 0.06695213168859482, 0.03172319754958153, -0.02931683138012886, 0.0015236766776069999, 0.15014788508415222, 0.026510147377848625, 0.16621503233909607, 0.22043149173259735, 0.061838917434215546, 0.20056213438510895, -0.2713247239589691, -0.10004157572984695, 0.10868333280086517, -0.07527677714824677, 0.021882841363549232, 0.13841275870800018, -0.01911449432373047, -0.0495067797601223, -0.03201347589492798, 0.08917038887739182, -0.017281996086239815, -0.08984966576099396, -0.04857974499464035, -0.003589637577533722, -0.18503929674625397, -0.0007536212215200067, -0.15319249033927917, 0.1420021951198578, 0.04460230842232704, -0.062356118112802505, 0.07465137541294098, 0.05997058004140854, 0.03977793827652931, 0.006764960940927267, 0.018739836290478706, -0.14650356769561768, 0.01704270951449871, -0.025170978158712387, -0.006106532644480467, 0.03402095288038254, 0.34655115008354187, -0.0466112419962883, -0.07675225287675858, -0.019784720614552498, 0.1001124382019043, 0.13863220810890198, -0.009452453814446926, 0.13600659370422363, 0.13898764550685883, -0.07470680773258209, -0.12456237524747849, 0.10025309771299362, -0.04034053534269333, -0.15969179570674896, 0.12802298367023468, -0.0435095950961113, -0.016280202195048332, 0.04011611267924309, -0.03383811563253403, -0.08241409808397293, 0.04869242012500763, -0.08193223923444748, -0.03468599542975426, -0.03921830281615257, -0.019609715789556503, -0.02835456281900406, 0.179523304104805, -0.03646359592676163, 0.07318142801523209, -0.02748848870396614, 0.010194642469286919, -0.10395175963640213, -0.1028568297624588, 0.05173351243138313, -0.12340104579925537, 0.07964924722909927, -0.03694985434412956, 0.030445387586951256, 0.22815105319023132, 0.02754553034901619, 0.015633730217814445, 0.13255921006202698, -0.00819331593811512, -0.0877854973077774, 0.03996758162975311, -0.044342756271362305, 0.021794743835926056, -0.030855976045131683, -0.07628626376390457, -0.0880078375339508, -0.10075201094150543, -0.049825526773929596, 0.03320961445569992, -0.030442843213677406, -0.05212388187646866, -0.14976045489311218, -0.02720625326037407, -0.07237301766872406, 0.11920249462127686, -0.09342960268259048, 0.08832328021526337, -0.012045936658978462, 0.0026839354541152716, 0.037163145840168, 0.1505078673362732, 0.010094218887388706, 0.10494716465473175, 0.006677085533738136, 0.09218452870845795, -0.06759306788444519, 0.14643312990665436, -0.12665413320064545, -0.02135086990892887, -0.03415476530790329, 0.2331210970878601, 0.20847657322883606, -0.11358945816755295, 0.009311644360423088, 0.03202449902892113, 0.04839635267853737, 0.185939759016037, 0.12599588930606842, 0.01761433109641075, 0.33329761028289795, -0.059357043355703354, -0.02227349951863289, 0.05721667781472206, -0.00022221643303055316, -0.06214975565671921, 0.0716261938214302, 0.08921460807323456, 0.013963594101369381, -0.1257423460483551, 0.11072274297475815, -0.21343208849430084, 0.15216094255447388, 0.07192383706569672, -0.18375952541828156, -0.009178245440125465, -0.05186039209365845, 0.008210902102291584, -0.027973614633083344, 0.13407447934150696, -0.07003656774759293, -0.1739543378353119, -0.19977876543998718, 0.060681428760290146, -0.35512542724609375, -0.20812080800533295, 0.06384200602769852, 0.1383514702320099, 0.10808566957712173, -0.06061858683824539, -0.013316533528268337, 0.006446295417845249, 0.01029437780380249, -0.019556531682610512, 0.028526417911052704, -0.008326482027769089, -0.05453765019774437, -0.25444141030311584, -0.006056090816855431, 0.0625600665807724, -0.15240277349948883, 0.05618175491690636, -0.017780732363462448, -0.008800189942121506, 0.13029517233371735, -0.021711476147174835, 0.03442413732409477, 0.00029493181500583887, -0.16273388266563416, 0.031801287084817886, 0.035038504749536514, 0.03614772483706474, -0.010639974847435951, -0.04227915778756142, -0.002239778870716691, 0.07848605513572693, -0.054354216903448105, -0.1438787877559662, 0.11021588742733002, -0.026462025940418243, 0.21526864171028137, -0.06517954170703888, -0.033111389726400375, 0.023098714649677277, -0.07031320035457611, 0.2018292248249054, -0.03690796345472336, 0.05650625377893448, 0.1586160659790039, 0.018734993413090706, 0.019857894629240036, -0.30062609910964966, 0.08813683688640594, -0.024517416954040527, 0.006894893944263458, -0.05270370468497276 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-finetuned-squad This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "bert-base-cased", "model-index": [{"name": "bert-finetuned-squad", "results": []}]}
question-answering
Peak1260/bert-finetuned-squad
[ "transformers", "tensorboard", "safetensors", "bert", "question-answering", "generated_from_trainer", "base_model:bert-base-cased", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2024-02-15T01:25:53+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #bert #question-answering #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #endpoints_compatible #region-us
# bert-finetuned-squad This model is a fine-tuned version of bert-base-cased on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
[ "# bert-finetuned-squad\n\nThis model is a fine-tuned version of bert-base-cased on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.2" ]
[ "TAGS\n#transformers #tensorboard #safetensors #bert #question-answering #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #endpoints_compatible #region-us \n", "# bert-finetuned-squad\n\nThis model is a fine-tuned version of bert-base-cased on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.2" ]
[ 60, 35, 6, 12, 8, 3, 103, 4, 33 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #bert #question-answering #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #endpoints_compatible #region-us \n# bert-finetuned-squad\n\nThis model is a fine-tuned version of bert-base-cased on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.2" ]
[ -0.0626351535320282, 0.06349974870681763, -0.0023771210107952356, 0.054038502275943756, 0.14659859240055084, 0.0202014297246933, 0.13172651827335358, 0.11050309985876083, -0.09484684467315674, 0.050080496817827225, 0.04347103089094162, 0.024793313816189766, 0.04500100389122963, 0.10279924422502518, -0.040943462401628494, -0.23241832852363586, 0.022829202935099602, 0.014582978561520576, -0.09761731326580048, 0.07703711092472076, 0.10796434432268143, -0.12149205058813095, 0.047987859696149826, 0.00970813911408186, -0.14694227278232574, 0.04365900903940201, -0.015008761547505856, -0.04717565327882767, 0.10612371563911438, 0.04258811101317406, 0.14830607175827026, 0.008187714032828808, 0.12048567086458206, -0.233250692486763, 0.013284264132380486, 0.09052932262420654, 0.02964795008301735, 0.0652635395526886, 0.056784577667713165, 0.021774083375930786, 0.06949657201766968, -0.10446500778198242, 0.12293750792741776, 0.020209090784192085, -0.0777631476521492, -0.21195341646671295, -0.08740760385990143, 0.02193804644048214, 0.1051797941327095, 0.06784109026193619, -0.0019075252348557115, 0.12885020673274994, -0.12137464433908463, 0.05377096310257912, 0.17989033460617065, -0.2773246765136719, -0.07259327173233032, 0.04142887145280838, 0.06232061609625816, 0.05568820610642433, -0.10720844566822052, -0.026770908385515213, 0.04498063400387764, 0.042791444808244705, 0.09501871466636658, -0.01531631126999855, -0.11926878243684769, 0.01096942275762558, -0.139540895819664, 0.011025520972907543, 0.09315509349107742, 0.04599865898489952, -0.04296383261680603, -0.09674439579248428, -0.035993702709674835, -0.0173293799161911, -0.04708664491772652, -0.032600920647382736, 0.04020747169852257, -0.03134194388985634, -0.06553614884614944, -0.05366462469100952, -0.08150224387645721, -0.08045835047960281, -0.022378534078598022, 0.11959962546825409, 0.04845806211233139, 0.004178166389465332, -0.0523882731795311, 0.08548489958047867, -0.05033800005912781, -0.09979114681482315, 0.009425445459783077, 0.0029356342274695635, -0.09634663909673691, -0.07788912951946259, -0.060572706162929535, -0.06672526895999908, 0.03540179878473282, 0.144344761967659, -0.03682560473680496, 0.08147840201854706, -0.04441814869642258, 0.011350047774612904, -0.0235584769397974, 0.11757354438304901, -0.031251389533281326, -0.014552532695233822, -0.001603359472937882, 0.08846733719110489, 0.0036015098448842764, 0.0023802134674042463, -0.08455023169517517, 0.00004387056833365932, 0.06342543661594391, 0.0475136935710907, -0.05328910052776337, 0.029311351478099823, -0.05233187600970268, -0.012801578268408775, -0.0172115545719862, -0.11451056599617004, 0.052782393991947174, 0.006666789297014475, -0.05089101940393448, -0.03674803301692009, 0.01525428332388401, 0.02431059256196022, 0.00032303406624123454, 0.10682182759046555, -0.06445055454969406, 0.002575164893642068, -0.10395312309265137, -0.10145235806703568, 0.02699451707303524, -0.047158583998680115, -0.00028866500360891223, -0.07354917377233505, -0.16222646832466125, -0.028593139722943306, 0.0530487485229969, -0.04338015243411064, -0.011369132436811924, -0.03805624693632126, -0.03800542280077934, 0.01437789760529995, -0.019031012430787086, 0.1705009639263153, -0.05325034633278847, 0.06699587404727936, -0.021701397374272346, 0.027903232723474503, -0.01185133308172226, 0.04180927202105522, -0.07395479828119278, 0.03009333461523056, -0.16190636157989502, 0.053450919687747955, -0.11242321133613586, 0.012424306012690067, -0.12335046380758286, -0.07940176874399185, 0.010893761180341244, 0.008402434177696705, 0.08605838567018509, 0.09367364645004272, -0.16884182393550873, -0.03698640689253807, 0.13600026071071625, -0.09385493397712708, -0.06609205901622772, 0.10539988428354263, -0.07094112038612366, 0.04243702441453934, 0.06578367203474045, 0.14971794188022614, 0.0717838853597641, -0.1400676667690277, 0.027748623862862587, -0.000967926811426878, 0.09221459925174713, 0.047699760645627975, 0.04627188295125961, -0.027211161330342293, -0.03465547040104866, -0.00046942135668359697, -0.05286065489053726, 0.02370985969901085, -0.0847461074590683, -0.0705610141158104, -0.036782845854759216, -0.08441439270973206, 0.05513174459338188, 0.02302834577858448, 0.03801527991890907, -0.08313535898923874, -0.10228437930345535, 0.1842222511768341, 0.11011616140604019, -0.06746245175600052, 0.01487654633820057, -0.09034719318151474, 0.02324450947344303, -0.031262386590242386, -0.015390435233712196, -0.19543471932411194, -0.12418006360530853, 0.030238065868616104, -0.05251884460449219, 0.044333215802907944, 0.04718517139554024, 0.07330933213233948, 0.05997847765684128, -0.07127439975738525, -0.007366594392806292, -0.08176840096712112, 0.018480997532606125, -0.11199255287647247, -0.21504409611225128, -0.05532152205705643, -0.04870268329977989, 0.15313999354839325, -0.24108804762363434, 0.016929324716329575, -0.013706785626709461, 0.14687441289424896, 0.05147673934698105, -0.02969682402908802, -0.027998849749565125, 0.08356752246618271, 0.005709488410502672, -0.07139848172664642, 0.043101727962493896, 0.00362464040517807, -0.10879407078027725, -0.05337706208229065, -0.12416478246450424, 0.04951661080121994, 0.08502870053052902, 0.008750362321734428, -0.09615259617567062, -0.05871731787919998, -0.04276357218623161, -0.037217218428850174, -0.09647045284509659, -0.004748283885419369, 0.1970035582780838, 0.009135439060628414, 0.1313830018043518, -0.06252285838127136, -0.04325341805815697, -0.008631257340312004, -0.003409863682463765, 0.009233377873897552, 0.08497599512338638, 0.06331750750541687, -0.1000773087143898, 0.09145647287368774, 0.14043734967708588, -0.0893128290772438, 0.1247052326798439, -0.06898362934589386, -0.085690438747406, -0.004528808873146772, 0.01741194725036621, -0.0013438909081742167, 0.1248251348733902, -0.11725401878356934, 0.012454932555556297, 0.021342214196920395, 0.023211855441331863, 0.03195527568459511, -0.1888953149318695, 0.0017084769206121564, 0.019095685333013535, -0.03210026025772095, -0.009602019563317299, -0.03784564137458801, 0.04710018262267113, 0.09659767895936966, 0.025160137563943863, -0.024390053004026413, 0.028936589136719704, -0.01727258414030075, -0.07824894785881042, 0.19472980499267578, -0.1330595165491104, -0.10100682824850082, -0.09929271787405014, -0.007308184634894133, -0.026045963168144226, -0.023746566846966743, 0.0314616896212101, -0.10386908054351807, -0.060790080577135086, -0.08875348418951035, 0.026292799040675163, -0.014013700187206268, -0.009036676958203316, 0.015903228893876076, 0.00847722589969635, 0.09268664568662643, -0.13014012575149536, 0.014070316217839718, -0.028385447338223457, -0.10781466960906982, 0.00555811170488596, 0.05837671458721161, 0.07445930689573288, 0.12847819924354553, -0.023281153291463852, -0.011952470988035202, -0.054212141782045364, 0.2074718028306961, -0.05863717198371887, -0.018631696701049805, 0.11055373400449753, -0.022567417472600937, 0.041955482214689255, 0.14337486028671265, 0.05639788135886192, -0.09512684494256973, 0.03943298012018204, 0.09092938899993896, -0.007159837521612644, -0.24677951633930206, -0.0356668159365654, -0.03569613769650459, -0.09581660479307175, 0.09577725827693939, 0.04621104151010513, -0.039915990084409714, 0.04631709307432175, -0.010035854764282703, 0.04895976930856705, 0.004256907384842634, 0.08063098043203354, 0.12873633205890656, 0.026379574090242386, 0.10316189378499985, -0.03264365345239639, -0.0386200025677681, 0.0619947724044323, 0.013319376856088638, 0.26561617851257324, 0.0029082116670906544, 0.07158004492521286, 0.06819714605808258, 0.12675006687641144, -0.005986168514937162, 0.013651341199874878, -0.00014948489842936397, -0.018601039424538612, -0.00725026847794652, -0.06394995003938675, -0.009941418655216694, 0.028224807232618332, -0.030165934935212135, 0.030965138226747513, -0.09665132313966751, -0.04283244535326958, 0.01629032753407955, 0.25870248675346375, 0.03586959093809128, -0.2416473776102066, -0.06611748784780502, 0.03326227143406868, -0.05520888790488243, -0.0663573145866394, 0.011581110768020153, 0.1336134672164917, -0.11332639306783676, 0.047725167125463486, -0.06118238344788551, 0.1042790561914444, -0.005568974185734987, 0.0007387851946987212, 0.0377180278301239, 0.12632648646831512, -0.013400617986917496, 0.0736008808016777, -0.24570424854755402, 0.2259187400341034, 0.02433323860168457, 0.11504196375608444, -0.04909200593829155, 0.03581841289997101, 0.024815650656819344, 0.06327223777770996, 0.0558878630399704, -0.021121807396411896, -0.07802660018205643, -0.19217504560947418, -0.03533013537526131, 0.05883469805121422, 0.1053692102432251, -0.0058393580839037895, 0.08270611613988876, -0.03756185993552208, 0.018145356327295303, 0.06458994001150131, -0.05488630384206772, -0.18782733380794525, -0.12274201214313507, -0.018143298104405403, 0.02190225198864937, -0.011513235978782177, -0.1124291941523552, -0.1013031154870987, -0.04631994292140007, 0.16161926090717316, 0.01629534550011158, -0.02837955392897129, -0.12556181848049164, 0.09086199849843979, 0.10417784750461578, -0.036565013229846954, 0.032827481627464294, 0.01773456297814846, 0.12216205149888992, 0.022557102143764496, -0.08203301578760147, 0.08235036581754684, -0.07566101849079132, -0.15665416419506073, -0.08211486786603928, 0.10150076448917389, 0.07372432202100754, 0.054491087794303894, 0.010877764783799648, 0.01807473786175251, 0.020203253254294395, -0.07364479452371597, -0.012126457877457142, 0.09798593819141388, 0.07803212106227875, 0.09101195633411407, -0.10958345979452133, -0.016092713922262192, -0.017829662188887596, -0.01579556241631508, 0.12146096676588058, 0.22849124670028687, -0.07718966901302338, 0.07249078154563904, 0.10638540238142014, -0.08216334879398346, -0.1911330223083496, 0.1028435081243515, 0.11044103652238846, 0.014629368670284748, 0.05952579900622368, -0.2104048728942871, 0.1373923271894455, 0.12098723649978638, -0.02427295409142971, 0.03020162135362625, -0.2997613549232483, -0.12836198508739471, 0.10926277935504913, 0.12650899589061737, 0.03523939475417137, -0.14197778701782227, -0.031758714467287064, -0.03585036098957062, -0.12458446621894836, 0.12400279194116592, -0.15279538929462433, 0.09149879962205887, 0.017928825691342354, 0.07797252386808395, 0.025845075026154518, -0.031722377985715866, 0.14801135659217834, -0.020162416622042656, 0.10147401690483093, -0.045756809413433075, 0.05089633911848068, 0.050551917403936386, -0.05103519186377525, 0.002634112723171711, -0.05043568089604378, 0.03886198624968529, -0.1209440529346466, -0.025685878470540047, -0.061050478368997574, 0.060148876160383224, -0.042117610573768616, -0.06355733424425125, -0.05403918772935867, 0.05352356657385826, 0.03979725390672684, -0.03162714093923569, 0.07989583909511566, -0.01815629005432129, 0.1600371152162552, 0.05475352331995964, 0.1356605440378189, -0.029108857735991478, -0.08210358768701553, 0.018579361960291862, -0.0372510701417923, 0.09251852333545685, -0.12433384358882904, 0.043050605803728104, 0.11424116045236588, 0.03173967823386192, 0.1558791995048523, 0.04694931581616402, -0.057469893246889114, 0.019246384501457214, 0.041224442422389984, -0.08617132902145386, -0.16028831899166107, 0.01921832002699375, 0.05977039411664009, -0.14330846071243286, 0.016158755868673325, 0.11151225864887238, -0.04346982389688492, -0.017927782610058784, -0.0010482992511242628, 0.011376814916729927, -0.03848477452993393, 0.17414119839668274, 0.0318097360432148, 0.050994183868169785, -0.08286263793706894, 0.10515038669109344, 0.08555009961128235, -0.11820162087678909, 0.04902256280183792, 0.04355047643184662, -0.062319789081811905, -0.02234034053981304, 0.09164650738239288, 0.2221025675535202, 0.002900870982557535, -0.05424705892801285, -0.07209830731153488, -0.12493113428354263, 0.03343741223216057, 0.15057995915412903, 0.04703885689377785, -0.020299233496189117, -0.021193426102399826, 0.050040680915117264, -0.10732634365558624, 0.07439127564430237, 0.027183782309293747, 0.07540509104728699, -0.10027534514665604, 0.10808634757995605, 0.009333116933703423, 0.008085554465651512, -0.01771860010921955, 0.020424339920282364, -0.12193065881729126, -0.010112407617270947, -0.18792405724525452, 0.006254600360989571, -0.026960114017128944, 0.01581365056335926, 0.012711024843156338, -0.04768962040543556, -0.03460667282342911, 0.03109298087656498, -0.0841427594423294, -0.03197425603866577, 0.020674189552664757, 0.07018938660621643, -0.1373666524887085, -0.01606280542910099, 0.02968611940741539, -0.08806298673152924, 0.06028575077652931, 0.02577284723520279, 0.028303664177656174, 0.050130754709243774, -0.17557811737060547, -0.033833231776952744, 0.02628490701317787, 0.02831597626209259, 0.060513563454151154, -0.0856848731637001, -0.02075131982564926, -0.01818194054067135, 0.06106989085674286, 0.013730241917073727, 0.06097390130162239, -0.11443782597780228, -0.04066574200987816, -0.04585647210478783, -0.076888307929039, -0.05193425714969635, 0.03162306919693947, 0.10020570456981659, 0.04896533861756325, 0.16675391793251038, -0.09952831268310547, 0.03400572016835213, -0.18201978504657745, -0.028461920097470284, -0.01794864609837532, -0.020469265058636665, -0.0771404430270195, -0.0455455482006073, 0.06978406012058258, -0.05067688599228859, 0.12186877429485321, -0.01840820536017418, 0.09198892116546631, 0.045698150992393494, -0.07488039880990982, 0.018699226900935173, 0.031184682622551918, 0.22507131099700928, 0.06406471878290176, -0.01644502580165863, 0.04186457023024559, -0.006902199238538742, 0.04929452762007713, 0.09082802385091782, 0.1696092039346695, 0.1897340565919876, -0.015094943344593048, 0.05103874206542969, 0.07374574989080429, -0.09224342554807663, -0.09965503960847855, 0.1240263283252716, -0.007879526354372501, 0.10947924852371216, -0.06523092091083527, 0.2296963334083557, 0.07010674476623535, -0.17488037049770355, 0.050358451902866364, -0.07806925475597382, -0.08950487524271011, -0.11266803741455078, -0.019690630957484245, -0.08353494852781296, -0.14298303425312042, 0.008755232207477093, -0.12382513284683228, 0.024576526135206223, 0.09757527709007263, 0.01316893007606268, 0.02871774137020111, 0.1399165689945221, -0.041760288178920746, 0.02275794930756092, 0.04925164207816124, 0.007152634207159281, 0.007997393608093262, -0.07579079270362854, -0.08378837257623672, 0.056162141263484955, 0.0012333289487287402, 0.04394959658384323, -0.03231234475970268, -0.0008517075912095606, 0.028534166514873505, -0.006668651010841131, -0.06939984858036041, 0.03502114489674568, 0.013454189524054527, 0.04443288594484329, 0.08176009356975555, 0.04610295593738556, 0.005319039802998304, -0.036966722458601, 0.27890869975090027, -0.07854106277227402, -0.09840981662273407, -0.14890821278095245, 0.22478032112121582, -0.004036687780171633, 0.008358627557754517, 0.043863337486982346, -0.10805326700210571, -0.005907649174332619, 0.1595521867275238, 0.12537088990211487, -0.07337973266839981, -0.005470559932291508, -0.004908227361738682, -0.02782849222421646, -0.08195878565311432, 0.1276509314775467, 0.09744963049888611, 0.022241905331611633, -0.06012153625488281, -0.0327179878950119, 0.002443439792841673, -0.0150130121037364, -0.08920463174581528, 0.042812153697013855, 0.023925716057419777, -0.004462741315364838, -0.03500613197684288, 0.08345053344964981, 0.031022457405924797, -0.21354278922080994, 0.03896607458591461, -0.13279227912425995, -0.1684979349374771, -0.04080396890640259, 0.0955069437623024, -0.013415209949016571, 0.039108019322156906, -0.0342266745865345, 0.011874650605022907, 0.10967787355184555, -0.027994569391012192, -0.005043831653892994, -0.14226599037647247, 0.12208429723978043, -0.05790996551513672, 0.2217032015323639, 0.005404379218816757, 0.07372995465993881, 0.11537814140319824, 0.03276059776544571, -0.10028285533189774, 0.04908449575304985, 0.06513281911611557, -0.11011875420808792, 0.01611950621008873, 0.13241416215896606, -0.05665956810116768, 0.12210933864116669, 0.04522630572319031, -0.1424448937177658, 0.02367076463997364, -0.07915899902582169, -0.06504205614328384, -0.07274679094552994, 0.006672010291367769, -0.08308223634958267, 0.14080338180065155, 0.19733883440494537, -0.024458112195134163, 0.00482425419613719, -0.06519466638565063, 0.02523444965481758, 0.04730762913823128, 0.0993104800581932, -0.04405133053660393, -0.20962342619895935, 0.027107825502753258, 0.0563543476164341, 0.019459228962659836, -0.2557174265384674, -0.09481098502874374, 0.046131815761327744, -0.020656397566199303, -0.05942300334572792, 0.09957673400640488, 0.093963623046875, 0.03521773964166641, -0.047501370310783386, -0.2310742437839508, -0.038381632417440414, 0.14272905886173248, -0.11307855695486069, -0.0415554977953434 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # mi-super-modelo This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.6143 - Accuracy: 0.225 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.6863 | 0.5 | 5 | 1.6243 | 0.225 | | 1.6154 | 1.0 | 10 | 1.6143 | 0.225 | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "bert-base-cased", "model-index": [{"name": "mi-super-modelo", "results": []}]}
text-classification
Hernan1970/mi-super-modelo
[ "transformers", "tensorboard", "safetensors", "bert", "text-classification", "generated_from_trainer", "base_model:bert-base-cased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-15T01:36:25+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
mi-super-modelo =============== This model is a fine-tuned version of bert-base-cased on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 1.6143 * Accuracy: 0.225 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 5e-05 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 1 ### Training results ### Framework versions * Transformers 4.35.2 * Pytorch 2.1.0+cu121 * Datasets 2.17.0 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ 67, 98, 4, 33 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ -0.09015761315822601, 0.09940165281295776, -0.002217498840764165, 0.10443039238452911, 0.14845909178256989, 0.021810904145240784, 0.1547500640153885, 0.11064709722995758, -0.07466781139373779, 0.04434923082590103, 0.12447204440832138, 0.12864921987056732, 0.014141353778541088, 0.1340656727552414, -0.06396052241325378, -0.22988569736480713, 0.009612617082893848, 0.030656777322292328, -0.06373316049575806, 0.11578419059515, 0.09288354218006134, -0.13012751936912537, 0.08910137414932251, -0.008386824280023575, -0.17668090760707855, 0.01740233600139618, 0.019704295322299004, -0.0521991103887558, 0.13538405299186707, 0.03521361202001572, 0.13967135548591614, 0.021988267078995705, 0.09033101052045822, -0.21194705367088318, 0.012127738445997238, 0.06318140774965286, -0.00893273577094078, 0.08196625113487244, 0.03525562211871147, 0.006450648419559002, 0.09316089004278183, -0.07966925948858261, 0.06629111617803574, 0.019888952374458313, -0.11809346824884415, -0.19987331330776215, -0.07456857711076736, 0.045184176415205, 0.09734154492616653, 0.07771283388137817, -0.011992166750133038, 0.12358204275369644, -0.0626552402973175, 0.09042464941740036, 0.21226301789283752, -0.32310187816619873, -0.06516525894403458, 0.05345531553030014, 0.030515138059854507, 0.08562616258859634, -0.10917175561189651, -0.01865510269999504, 0.06652399897575378, 0.026521094143390656, 0.1299317330121994, -0.027854889631271362, -0.07010691612958908, 0.007540024816989899, -0.14972291886806488, -0.016074158251285553, 0.1496792733669281, 0.051505230367183685, -0.039028823375701904, -0.05005514249205589, -0.061962325125932693, -0.15232078731060028, -0.04254809394478798, -0.024940647184848785, 0.04995272308588028, -0.019617730751633644, -0.05531402677297592, -0.0070569440722465515, -0.1087278202176094, -0.07930323481559753, -0.05625521019101143, 0.13144688308238983, 0.03945402428507805, 0.0005574580864049494, -0.01362527348101139, 0.09973486512899399, -0.03692048788070679, -0.12299849838018417, 0.021696919575333595, 0.02326720394194126, 0.008293411694467068, -0.058168984949588776, -0.05264725908637047, -0.06163351610302925, 0.03024865873157978, 0.13595205545425415, -0.050870031118392944, 0.052413903176784515, 0.00827560480684042, 0.0475141815841198, -0.10201013833284378, 0.15882553160190582, -0.03668052703142166, -0.03162522241473198, 0.017231011763215065, 0.07795605808496475, 0.039136942476034164, -0.0014890393940731883, -0.12688188254833221, 0.03041842207312584, 0.11091997474431992, 0.016971919685602188, -0.07608995586633682, 0.0795038640499115, -0.04835745319724083, 0.00310936919413507, 0.019778473302721977, -0.08708149194717407, 0.03160583972930908, 0.009020530618727207, -0.048713598400354385, -0.0748233050107956, 0.029386624693870544, 0.020906373858451843, 0.003405615920200944, 0.10589828342199326, -0.08579162508249283, 0.0097367437556386, -0.09320563822984695, -0.12235194444656372, 0.020746154710650444, -0.07666952162981033, 0.023502497002482414, -0.10998893529176712, -0.15820850431919098, -0.010554099455475807, 0.057936377823352814, -0.022096935659646988, -0.031499531120061874, -0.05755440890789032, -0.07463076710700989, 0.011876440607011318, -0.018066149204969406, 0.0977366641163826, -0.06626800447702408, 0.09448052197694778, 0.03351430594921112, 0.06525414437055588, -0.04770437255501747, 0.03833281993865967, -0.09638004004955292, 0.024883143603801727, -0.1811767965555191, 0.01239373255521059, -0.07222947478294373, 0.06358571350574493, -0.08214914053678513, -0.07680941373109818, -0.007914965972304344, 0.01391855999827385, 0.06999486684799194, 0.09051568061113358, -0.1611717790365219, -0.06468399614095688, 0.17430298030376434, -0.09890494495630264, -0.14243853092193604, 0.12289959192276001, -0.05898919701576233, 0.05525429546833038, 0.06614524871110916, 0.18653064966201782, 0.07319247722625732, -0.10152753442525864, 0.004650274757295847, 0.0015076363924890757, 0.06044926866889, -0.04873938485980034, 0.05968411639332771, 0.0037367886397987604, -0.008481893688440323, 0.018085956573486328, -0.046482328325510025, 0.04926978424191475, -0.08077820390462875, -0.08499391376972198, -0.038576871156692505, -0.1015431210398674, 0.0422392301261425, 0.05910111591219902, 0.06621081382036209, -0.11497509479522705, -0.08828133344650269, 0.07135268300771713, 0.07286989688873291, -0.07116153836250305, 0.017508242279291153, -0.0715346410870552, 0.08081936091184616, -0.05152818560600281, -0.017903488129377365, -0.1538083702325821, -0.0489889495074749, 0.01529446616768837, -0.0009248852729797363, 0.010421299375593662, 0.010825482197105885, 0.07248035818338394, 0.0738869458436966, -0.06807757914066315, -0.025149550288915634, -0.011117201298475266, 0.01886129379272461, -0.13185550272464752, -0.2089308351278305, -0.01405309233814478, -0.03335762396454811, 0.13568340241909027, -0.24554486572742462, 0.05366189777851105, 0.001906441175378859, 0.08036042004823685, 0.03434646502137184, -0.004079947713762522, -0.0464482344686985, 0.07556486874818802, -0.04180878400802612, -0.05320076271891594, 0.06200309470295906, 0.012151102535426617, -0.08482126891613007, -0.03771084547042847, -0.14574870467185974, 0.19369125366210938, 0.13455983996391296, -0.09520193189382553, -0.08489880710840225, -0.020306192338466644, -0.04279102012515068, -0.028536684811115265, -0.0462859645485878, 0.005099131725728512, 0.13661015033721924, -0.016947252675890923, 0.15117400884628296, -0.07821588963270187, -0.033779218792915344, 0.02560598775744438, -0.04444654658436775, 0.011944212950766087, 0.11099471151828766, 0.1177215501666069, -0.10820411145687103, 0.15628154575824738, 0.17304004728794098, -0.09332667291164398, 0.13532133400440216, -0.043603867292404175, -0.05636098235845566, -0.018608707934617996, 0.006651787552982569, 0.004182739183306694, 0.11856860667467117, -0.1386103481054306, -0.004667171277105808, 0.004897328559309244, 0.015456478111445904, 0.019264990463852882, -0.2266809493303299, -0.029025154188275337, 0.035806238651275635, -0.048772167414426804, 0.003119726898148656, -0.028580939397215843, -0.010730753652751446, 0.1003764271736145, -0.0016680997796356678, -0.08826681226491928, 0.039772845804691315, -0.006325630005449057, -0.09039987623691559, 0.21141086518764496, -0.08218229562044144, -0.13064654171466827, -0.13160686194896698, -0.07014437019824982, -0.032894317060709, 0.024127544835209846, 0.06873609870672226, -0.07847259193658829, -0.04016859084367752, -0.10708756744861603, -0.003855436109006405, 0.03648167848587036, 0.028972607105970383, 0.008238844573497772, 0.005796987563371658, 0.07848671078681946, -0.10359802097082138, -0.012987330555915833, -0.052822295576334, -0.05935932323336601, 0.03174804523587227, 0.02306213043630123, 0.11218155920505524, 0.14887170493602753, -0.026951109990477562, -0.00022402741888072342, -0.03714814782142639, 0.21810226142406464, -0.05850638076663017, -0.018004128709435463, 0.11826614290475845, -0.028090452775359154, 0.04435376450419426, 0.1415250152349472, 0.06374256312847137, -0.0940694510936737, 0.022891350090503693, 0.04135190695524216, -0.027036136016249657, -0.21240155398845673, -0.03639449551701546, -0.03625012934207916, 0.0000664740800857544, 0.09804322570562363, 0.036199141293764114, 0.025598792359232903, 0.0666297897696495, 0.03409915417432785, 0.07170633971691132, -0.005751233547925949, 0.06811205297708511, 0.11434672772884369, 0.03675801306962967, 0.12621080875396729, -0.0526769757270813, -0.058077458292245865, 0.03186590597033501, -0.001594046363607049, 0.20760001242160797, 0.02757384069263935, 0.12334167212247849, 0.06335020065307617, 0.15685221552848816, 0.005196072161197662, 0.06713493913412094, -0.012793400324881077, -0.050713300704956055, -0.008521794341504574, -0.04895905777812004, -0.02828424610197544, 0.03996187448501587, -0.09557926654815674, 0.056203220039606094, -0.11461750417947769, 0.009728193283081055, 0.05972889065742493, 0.24039550125598907, 0.04513793811202049, -0.3166147768497467, -0.08629325777292252, 0.023038093000650406, -0.030147196725010872, -0.0231489110738039, 0.034170977771282196, 0.1310756504535675, -0.051094185560941696, 0.027416102588176727, -0.0687294453382492, 0.08169586211442947, -0.0357351154088974, 0.04590949788689613, 0.06484175473451614, 0.08100341260433197, -0.011293882504105568, 0.07133091986179352, -0.2815931737422943, 0.2784622311592102, 0.011891269125044346, 0.075759656727314, -0.053158119320869446, 0.0010310186771675944, 0.03216962143778801, 0.09836027026176453, 0.07541889697313309, -0.022872373461723328, -0.07052116841077805, -0.19674791395664215, -0.05353608727455139, 0.03330092132091522, 0.09195225685834885, -0.030436616390943527, 0.09637612104415894, -0.03661596402525902, 0.0014207932399585843, 0.08981437981128693, -0.014745986089110374, -0.07914338260889053, -0.09819352626800537, -0.023552974686026573, 0.03620090335607529, -0.025153642520308495, -0.08511436730623245, -0.10363398492336273, -0.13746504485607147, 0.15897470712661743, -0.061407897621393204, -0.02005081996321678, -0.09363621473312378, 0.06800476461648941, 0.03999732807278633, -0.07482755929231644, 0.05123024061322212, 0.01172596774995327, 0.08658523112535477, 0.02403954043984413, -0.05687563866376877, 0.1287616789340973, -0.07485558837652206, -0.16658267378807068, -0.07799563556909561, 0.10089514404535294, 0.024533038958907127, 0.04681288078427315, 0.0025664118584245443, 0.004507578909397125, -0.012249757535755634, -0.07788068056106567, 0.019055400043725967, -0.0011498651001602411, 0.06647893786430359, 0.005657660774886608, -0.08654529601335526, -0.009787856601178646, -0.05373638868331909, -0.03329303488135338, 0.16722804307937622, 0.26887252926826477, -0.0924391821026802, 0.0027501648291945457, 0.06371024250984192, -0.07350728660821915, -0.2129034698009491, 0.04248690977692604, 0.03150835633277893, 0.0012388451723381877, 0.042494866997003555, -0.1467747837305069, 0.12197040766477585, 0.10222984105348587, -0.028392305597662926, 0.10799334943294525, -0.2833293378353119, -0.13245902955532074, 0.13858942687511444, 0.16309641301631927, 0.11181638389825821, -0.16146409511566162, -0.03016001358628273, -0.03003680519759655, -0.107781782746315, 0.10846210271120071, -0.1241891160607338, 0.10556800663471222, -0.004975308198481798, 0.04985843598842621, 0.0008875550120137632, -0.0584644190967083, 0.1198713481426239, -0.015014627017080784, 0.11428115516901016, -0.06393429636955261, -0.02553732320666313, 0.042306628078222275, -0.055278852581977844, 0.01671244204044342, -0.11327666789293289, 0.027039511129260063, -0.05633677542209625, -0.031080052256584167, -0.04765205830335617, 0.036333441734313965, -0.04061228781938553, -0.06350822746753693, -0.037082117050886154, 0.027116242796182632, 0.03793790563941002, -0.01236219983547926, 0.14330905675888062, 0.0026764641515910625, 0.16414004564285278, 0.1346205770969391, 0.083723284304142, -0.06791842728853226, -0.02015044167637825, -0.0008444359409622848, -0.03699110820889473, 0.0665038675069809, -0.15368326008319855, 0.0414353646337986, 0.12217312306165695, 0.005662091076374054, 0.15454208850860596, 0.0767735093832016, -0.02927430160343647, 0.005448286887258291, 0.0742219015955925, -0.16135193407535553, -0.10393426567316055, 0.000991012086160481, -0.030476855114102364, -0.11689391732215881, 0.0664457157254219, 0.11026910692453384, -0.07527332007884979, 0.0072439624927937984, -0.007518002763390541, 0.01455006655305624, -0.04291393235325813, 0.17702078819274902, 0.06983163952827454, 0.04968515411019325, -0.07537981122732162, 0.07535337656736374, 0.04262137413024902, -0.07753629982471466, 0.015957266092300415, 0.0407278910279274, -0.08743876963853836, -0.05291469767689705, 0.06499528139829636, 0.20220401883125305, -0.012420475482940674, -0.06379586458206177, -0.1422894299030304, -0.12202523648738861, 0.05600585788488388, 0.19779044389724731, 0.1013326570391655, 0.008504248224198818, -0.031146330758929253, 0.022593757137656212, -0.11516812443733215, 0.11470208317041397, 0.022291257977485657, 0.08211458474397659, -0.1497383862733841, 0.12559086084365845, 0.004100616555660963, 0.0023060685489326715, -0.026849431917071342, 0.0479857511818409, -0.12531781196594238, -0.005125884898006916, -0.12441294640302658, -0.007393726613372564, -0.026886457577347755, 0.008753543719649315, 0.011384980753064156, -0.0529012493789196, -0.06667019426822662, 0.011861109174787998, -0.10334586352109909, -0.012213276699185371, 0.03569402918219566, 0.06362321972846985, -0.12409414350986481, -0.03517964109778404, 0.023229124024510384, -0.06634417921304703, 0.06288248300552368, 0.016051752492785454, 0.023930663242936134, 0.05954477936029434, -0.1862882375717163, 0.03846414387226105, 0.07340697199106216, 0.011552832089364529, 0.04509330168366432, -0.09090662002563477, -0.009286590851843357, -0.00024529415532015264, 0.044354964047670364, 0.021183380857110023, 0.09018531441688538, -0.12808871269226074, 0.008816054090857506, -0.02757028490304947, -0.0696675106883049, -0.049322858452796936, 0.021156879141926765, 0.0884256362915039, -0.005605057813227177, 0.20360659062862396, -0.10119130462408066, 0.010651716962456703, -0.20261721312999725, 0.008927460759878159, -0.0001896684116218239, -0.10596857964992523, -0.11507272720336914, -0.060078077018260956, 0.046361178159713745, -0.05934962257742882, 0.15508216619491577, 0.02177431993186474, 0.019080346450209618, 0.03756354749202728, -0.049670860171318054, 0.03766300529241562, 0.03491191193461418, 0.22593122720718384, 0.02678450383245945, -0.04081840440630913, 0.01672079786658287, 0.03622029721736908, 0.11145514994859695, 0.07567618042230606, 0.17276336252689362, 0.168111190199852, -0.05362602695822716, 0.101323202252388, 0.04749523475766182, -0.05861610174179077, -0.12803277373313904, 0.06590881198644638, -0.05335947871208191, 0.09917302429676056, -0.028618700802326202, 0.2191939502954483, 0.09451963752508163, -0.15838059782981873, 0.01228067185729742, -0.06260021775960922, -0.07955073565244675, -0.11785911023616791, -0.05195650830864906, -0.09941284358501434, -0.15238288044929504, -0.0011915821814909577, -0.11262094974517822, 0.0013957471819594502, 0.10398545116186142, 0.006846359930932522, -0.013012412004172802, 0.1714066118001938, 0.006406668573617935, 0.040752820670604706, 0.051231976598501205, 0.004990288522094488, -0.03708909451961517, -0.11018657684326172, -0.08513810485601425, 0.0008090438786894083, -0.028154656291007996, 0.01797262392938137, -0.05046958476305008, -0.027112247422337532, 0.04092184081673622, -0.007962028495967388, -0.09523997455835342, 0.011356192640960217, 0.02109086699783802, 0.04795015975832939, 0.04859130457043648, 0.00868801400065422, 0.00814366526901722, 0.0041571869514882565, 0.2143542468547821, -0.07475148141384125, -0.06840631365776062, -0.10388786345720291, 0.2290145605802536, 0.032203156501054764, 0.022880077362060547, 0.008760429918766022, -0.0853312537074089, 0.03188841789960861, 0.2332543581724167, 0.19189393520355225, -0.07729238271713257, 0.005134772043675184, -0.0007411642000079155, -0.01171367522329092, -0.038377780467271805, 0.10100588202476501, 0.12286680191755295, 0.020741095766425133, -0.07418252527713776, -0.04974263906478882, -0.03410706669092178, -0.002807333366945386, -0.04180138558149338, 0.056387681514024734, 0.042380161583423615, 0.01358796562999487, -0.04952513426542282, 0.05407054349780083, -0.02645266242325306, -0.11239589750766754, 0.05953671410679817, -0.19774244725704193, -0.14658226072788239, -0.01060677133500576, 0.12204005569219589, -0.018530013039708138, 0.047909095883369446, -0.03167714551091194, -0.003979895729571581, 0.0718303695321083, -0.024695131927728653, -0.06814564764499664, -0.08224714547395706, 0.058747634291648865, -0.10204122215509415, 0.24126285314559937, -0.03762947767972946, 0.044041406363248825, 0.13268977403640747, 0.04327090084552765, -0.06877536326646805, 0.08759208768606186, 0.04449271783232689, -0.07870354503393173, 0.02336869016289711, 0.06368043273687363, -0.044481705874204636, 0.1195831298828125, 0.054379045963287354, -0.13371127843856812, 0.024072352796792984, -0.08117179572582245, -0.09294461458921432, -0.05350400134921074, -0.03044409677386284, -0.06232951954007149, 0.1336948573589325, 0.18822194635868073, -0.03009869158267975, 0.010325939394533634, -0.05015091225504875, 0.025403162464499474, 0.061279285699129105, 0.039351463317871094, -0.033158231526613235, -0.23624876141548157, 0.03285995498299599, 0.07785151898860931, -0.009712658822536469, -0.28543952107429504, -0.08836785703897476, -0.0014324705116450787, -0.044881295412778854, -0.10178390145301819, 0.07777399569749832, 0.1263887882232666, 0.05757888779044151, -0.0681326687335968, -0.11361874639987946, -0.07089872658252716, 0.15807732939720154, -0.12811248004436493, -0.10368068516254425 ]
null
null
transformers
# Uploaded model - **Developed by:** BarraHome - **License:** apache-2.0 - **Finetuned from model :** unsloth/mistral-7b-instruct-v0.2-bnb-4bit This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
{"language": ["en"], "license": "apache-2.0", "tags": ["text-generation-inference", "transformers", "unsloth", "mistral", "gguf"], "base_model": "unsloth/mistral-7b-instruct-v0.2-bnb-4bit"}
null
BarraHome/Wistral-7B-Instruct-v0.3-gguf
[ "transformers", "gguf", "mistral", "text-generation-inference", "unsloth", "en", "base_model:unsloth/mistral-7b-instruct-v0.2-bnb-4bit", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2024-02-15T01:37:37+00:00
[]
[ "en" ]
TAGS #transformers #gguf #mistral #text-generation-inference #unsloth #en #base_model-unsloth/mistral-7b-instruct-v0.2-bnb-4bit #license-apache-2.0 #endpoints_compatible #region-us
# Uploaded model - Developed by: BarraHome - License: apache-2.0 - Finetuned from model : unsloth/mistral-7b-instruct-v0.2-bnb-4bit This mistral model was trained 2x faster with Unsloth and Huggingface's TRL library. <img src="URL width="200"/>
[ "# Uploaded model\n\n- Developed by: BarraHome\n- License: apache-2.0\n- Finetuned from model : unsloth/mistral-7b-instruct-v0.2-bnb-4bit\n\nThis mistral model was trained 2x faster with Unsloth and Huggingface's TRL library.\n\n<img src=\"URL width=\"200\"/>" ]
[ "TAGS\n#transformers #gguf #mistral #text-generation-inference #unsloth #en #base_model-unsloth/mistral-7b-instruct-v0.2-bnb-4bit #license-apache-2.0 #endpoints_compatible #region-us \n", "# Uploaded model\n\n- Developed by: BarraHome\n- License: apache-2.0\n- Finetuned from model : unsloth/mistral-7b-instruct-v0.2-bnb-4bit\n\nThis mistral model was trained 2x faster with Unsloth and Huggingface's TRL library.\n\n<img src=\"URL width=\"200\"/>" ]
[ 70, 84 ]
[ "passage: TAGS\n#transformers #gguf #mistral #text-generation-inference #unsloth #en #base_model-unsloth/mistral-7b-instruct-v0.2-bnb-4bit #license-apache-2.0 #endpoints_compatible #region-us \n# Uploaded model\n\n- Developed by: BarraHome\n- License: apache-2.0\n- Finetuned from model : unsloth/mistral-7b-instruct-v0.2-bnb-4bit\n\nThis mistral model was trained 2x faster with Unsloth and Huggingface's TRL library.\n\n<img src=\"URL width=\"200\"/>" ]
[ -0.08675411343574524, 0.01683894917368889, -0.003272901987656951, 0.08940763771533966, 0.06491244584321976, 0.03651479259133339, 0.09864382445812225, 0.09573747217655182, 0.08010440319776535, -0.03783778101205826, 0.11853864043951035, 0.0596015527844429, 0.010935327969491482, 0.03134314715862274, -0.016323929652571678, -0.1505446434020996, 0.09356288611888885, -0.03538926690816879, -0.04361117258667946, 0.04144670069217682, 0.0714312270283699, 0.0010809076484292746, 0.09285992383956909, -0.062026187777519226, -0.04628913477063179, 0.019009247422218323, -0.02976168878376484, 0.0032359501346945763, 0.008583826944231987, 0.07659308612346649, -0.02350003272294998, 0.027681641280651093, 0.0515034943819046, -0.08966407179832458, 0.03604016453027725, 0.04237409308552742, -0.023324428126215935, 0.07426944375038147, -0.018672378733754158, 0.054074276238679886, 0.09583230316638947, -0.04067632183432579, -0.08845329284667969, 0.07175537198781967, -0.042198847979307175, -0.14987921714782715, -0.0697573721408844, 0.10802724957466125, 0.028067024424672127, 0.03982575237751007, 0.03572678565979004, 0.05428213253617287, -0.03791269287467003, 0.07164527475833893, 0.17286208271980286, -0.249709814786911, -0.0658288300037384, 0.1350189596414566, 0.04424847662448883, 0.045515693724155426, -0.0394115224480629, 0.004939265549182892, 0.036643922328948975, 0.00460635544732213, 0.024329254403710365, -0.08405502140522003, -0.040704816579818726, 0.04805044084787369, -0.10921024531126022, 0.010406161658465862, 0.20463204383850098, 0.05784571170806885, -0.051014430820941925, 0.058900270611047745, -0.1314111351966858, 0.06445883214473724, -0.05774697661399841, 0.08432278782129288, 0.05156417936086655, 0.10920292884111404, -0.028464244678616524, -0.111065573990345, -0.035293545573949814, -0.07159295678138733, -0.06901948899030685, 0.016078896820545197, 0.035119980573654175, 0.12082290649414062, -0.0646248534321785, 0.06689168512821198, -0.08077012002468109, -0.11584776639938354, -0.06640355288982391, -0.07262054830789566, 0.06808571517467499, 0.07019045948982239, -0.05305555835366249, 0.08202891051769257, 0.11749933660030365, 0.2595423758029938, 0.0716908723115921, 0.07015002518892288, 0.010045614093542099, 0.06859885901212692, -0.06807384639978409, 0.06014837697148323, -0.1207522377371788, -0.07404746115207672, 0.13565823435783386, 0.01758580282330513, 0.07028678804636002, -0.005569651257246733, -0.10704386979341507, -0.05823241174221039, 0.00504488218575716, 0.007779995910823345, 0.029242023825645447, 0.08776389062404633, -0.003613533917814493, -0.03409823775291443, 0.11901526898145676, -0.05638229846954346, -0.015509383752942085, 0.026557207107543945, -0.05000951141119003, 0.10414345562458038, 0.15539297461509705, -0.010391807183623314, -0.0380428284406662, -0.11386873573064804, -0.07090036571025848, 0.009281286969780922, -0.02757507562637329, -0.04324917122721672, 0.04207485169172287, -0.0382598415017128, 0.010016210377216339, -0.13718931376934052, -0.29741308093070984, 0.03789630904793739, 0.12877590954303741, -0.04268695041537285, 0.01908174529671669, -0.04162754863500595, -0.027329375967383385, 0.017941398546099663, -0.0328473336994648, -0.0012363468995317817, -0.07303812354803085, 0.019873913377523422, -0.04255600646138191, 0.08190721273422241, -0.2217114418745041, 0.022938445210456848, -0.10632824897766113, 0.03975459933280945, -0.10315385460853577, 0.08473284542560577, -0.07804688811302185, 0.11938246339559555, -0.1346333920955658, -0.003560172626748681, -0.01828094944357872, 0.0005231202812865376, 0.0791577473282814, 0.1330912858247757, -0.1610407531261444, 0.02311580441892147, 0.14127302169799805, -0.034187592566013336, -0.11872104555368423, 0.14068174362182617, 0.005966336000710726, 0.0800110474228859, 0.06200481206178665, 0.11491385847330093, 0.11704245954751968, -0.065682053565979, 0.07581786811351776, 0.1687072217464447, -0.0035402013454586267, -0.08591742813587189, 0.07141923159360886, 0.014236404560506344, -0.16646431386470795, 0.08364914357662201, -0.07122289389371872, 0.11976590752601624, 0.011853453703224659, -0.05074915662407875, -0.10717575252056122, -0.07387576252222061, -0.009209719486534595, -0.037915587425231934, 0.021738305687904358, 0.017283424735069275, -0.04860764369368553, 0.05330517888069153, 0.14327013492584229, -0.08243027329444885, 0.05402515083551407, -0.02540101855993271, 0.09753680229187012, -0.09622648358345032, 0.07233434170484543, -0.09346600621938705, 0.04223030433058739, -0.023239966481924057, 0.006401946302503347, 0.06777065992355347, 0.06331906467676163, 0.08223176002502441, -0.011553904972970486, -0.023065321147441864, 0.0025095778983086348, 0.10246320068836212, -0.015876881778240204, -0.06287563592195511, -0.13413530588150024, 0.0005296248127706349, -0.02288232557475567, 0.12300154566764832, -0.07861088216304779, 0.04793217405676842, -0.06146084517240524, 0.06243519112467766, -0.036786895245313644, 0.04441443458199501, 0.03286924585700035, -0.08913461118936539, -0.01050241943448782, -0.08380298316478729, 0.10330910980701447, 0.05932081863284111, -0.12384012341499329, 0.07878264039754868, -0.03894045203924179, 0.0658254623413086, 0.13480520248413086, -0.011643791571259499, 0.08575807511806488, -0.004396525211632252, -0.029683269560337067, -0.03495354577898979, 0.09689951688051224, -0.017402194440364838, 0.03769364580512047, 0.0026847815606743097, 0.12582515180110931, -0.09371422231197357, -0.009791379794478416, 0.0016285014571622014, -0.07651400566101074, 0.018793467432260513, 0.05723005533218384, 0.031683310866355896, -0.1734040230512619, 0.03990383818745613, 0.2651447355747223, -0.1304556280374527, 0.1341119408607483, -0.04524516314268112, -0.05403758957982063, -0.00933001283556223, 0.022974196821451187, -0.014537609182298183, 0.024369705468416214, -0.12280450016260147, 0.025942685082554817, 0.0470576211810112, -0.02391023188829422, 0.0564388744533062, -0.09230846166610718, 0.030226487666368484, -0.05230649933218956, -0.06200583279132843, -0.034654401242733, 0.06451678276062012, -0.08508674800395966, 0.03185312822461128, -0.015801219269633293, -0.08598461002111435, 0.04618782177567482, 0.028271973133087158, -0.06143732741475105, 0.13284210860729218, -0.1233680322766304, -0.06238545477390289, -0.20087353885173798, -0.04640598222613335, -0.13130250573158264, -0.007361900992691517, 0.075070321559906, -0.019373878836631775, -0.06460156291723251, -0.10572554171085358, -0.04262947291135788, 0.03751964494585991, 0.007492929231375456, 0.09207422286272049, 0.01522519625723362, 0.09512878209352493, -0.1355823129415512, -0.009658846072852612, 0.017184555530548096, -0.07885003089904785, 0.02867759019136429, -0.09479285031557083, 0.05212300270795822, 0.07051889598369598, 0.04101603105664253, -0.012384881265461445, 0.04670029133558273, 0.20170503854751587, 0.062043432146310806, 0.08873540163040161, 0.17208194732666016, -0.008656231686472893, 0.0909464880824089, 0.09983280301094055, 0.010053379461169243, -0.06230627000331879, 0.0002374094765400514, -0.03466562554240227, -0.04460477456450462, -0.16000591218471527, -0.0016522941878065467, -0.1064067929983139, 0.06067338213324547, 0.07308774441480637, 0.060167837888002396, -0.030401749536395073, 0.1391659677028656, -0.06589125841856003, 0.15135766565799713, 0.054264768958091736, 0.09263762831687927, 0.07548066228628159, -0.007332258392125368, 0.05936888977885246, -0.1080247312784195, 0.06403964012861252, 0.1452779471874237, 0.05955176427960396, 0.15409459173679352, -0.023581495508551598, 0.09169068187475204, 0.046612270176410675, 0.12900616228580475, 0.003593372879549861, 0.11613929271697998, -0.06506993621587753, 0.021364182233810425, -0.08027123659849167, -0.07413940131664276, -0.07630956172943115, 0.06502369046211243, -0.0875743106007576, 0.00017589377239346504, 0.0512007512152195, 0.09239271283149719, 0.09469372034072876, 0.1967184990644455, 0.06402544677257538, -0.22902549803256989, -0.09628675878047943, 0.09407218545675278, 0.05528188496828079, -0.03232433646917343, 0.05808621272444725, 0.0031369603238999844, 0.015937333926558495, 0.05011701211333275, -0.0374479740858078, 0.13894891738891602, 0.06988276541233063, 0.03982169181108475, 0.00716608390212059, 0.16740712523460388, 0.0516076385974884, 0.08601713925600052, -0.20808842778205872, 0.042601004242897034, 0.02353253774344921, 0.020716169849038124, -0.046202801167964935, 0.012954478152096272, 0.10617954283952713, 0.12342085689306259, 0.04900246486067772, 0.039727628231048584, -0.029523281380534172, 0.012453216128051281, -0.12535503506660461, 0.06930311769247055, -0.0047248066402971745, 0.003900136798620224, 0.032884303480386734, -0.09386655688285828, -0.029870087280869484, 0.015606972388923168, 0.06708518415689468, -0.11492013186216354, -0.09055889397859573, -0.0039351084269583225, 0.09333325922489166, -0.06086789071559906, -0.05374959856271744, 0.043236978352069855, -0.03162588179111481, 0.12744399905204773, -0.03654888644814491, -0.07218032330274582, -0.08799149096012115, -0.03801928088068962, 0.1471923440694809, -0.07706677168607712, 0.01847682148218155, -0.07274433225393295, -0.024821503087878227, 0.007675284054130316, -0.21026203036308289, 0.09715934097766876, -0.10310081392526627, -0.03851470723748207, 0.004944091197103262, 0.028461743146181107, -0.07730644196271896, -0.005475429818034172, 0.001705461647361517, -0.027578746899962425, -0.10745805501937866, -0.12530280649662018, -0.08463241159915924, 0.17143970727920532, -0.047955550253391266, 0.03980433568358421, -0.08359675109386444, -0.03786303102970123, 0.03150855004787445, 0.02558915875852108, 0.03400319814682007, 0.19683344662189484, -0.03678593412041664, 0.06775565445423126, 0.26108095049858093, -0.051259517669677734, -0.2681201100349426, -0.09391064196825027, -0.07513194531202316, -0.058164119720458984, -0.06305363029241562, -0.07874797284603119, 0.12648038566112518, 0.06021292507648468, -0.026518257334828377, 0.11805291473865509, -0.19756655395030975, -0.10187141597270966, 0.12388857454061508, 0.021014519035816193, 0.33598893880844116, -0.1275416612625122, -0.04998897761106491, -0.14925351738929749, -0.22212253510951996, -0.012439092621207237, -0.2546074092388153, 0.09491788595914841, -0.05538736283779144, 0.03358815610408783, -0.026097001507878304, -0.0374242328107357, 0.13399775326251984, 0.014315802603960037, 0.08423440158367157, -0.09371651709079742, 0.10037783533334732, 0.16493281722068787, -0.09886983782052994, 0.1729743927717209, -0.1763748973608017, 0.0968794971704483, -0.06079103797674179, 0.019711382687091827, -0.009206507354974747, -0.01378883421421051, 0.0041283611208200455, -0.022185413166880608, -0.050990115851163864, -0.01311104092746973, 0.03603115677833557, 0.004219209775328636, 0.15963329374790192, 0.03957191854715347, -0.0765395388007164, 0.20856472849845886, -0.008505186066031456, -0.12326834350824356, 0.023080307990312576, -0.048852477222681046, -0.045017778873443604, 0.10444606095552444, -0.24449360370635986, 0.05942244455218315, 0.0495293103158474, -0.037667643278837204, 0.0645674467086792, 0.0318630188703537, 0.01770043559372425, 0.005476042628288269, 0.030840426683425903, -0.12328227609395981, -0.09029514342546463, -0.03013894148170948, 0.007265204098075628, -0.08370617032051086, 0.09620974212884903, 0.1792910099029541, -0.05012465640902519, 0.012265567667782307, 0.017214521765708923, 0.045464515686035156, -0.10723298788070679, 0.05913105979561806, 0.0611543282866478, -0.021742770448327065, -0.11025956273078918, 0.16495762765407562, -0.027505401521921158, 0.050198256969451904, 0.006340602412819862, 0.07756485790014267, -0.15294097363948822, -0.11233875155448914, -0.02517702616751194, 0.11088112741708755, -0.11863667517900467, -0.021241813898086548, -0.05793095752596855, -0.023365462198853493, 0.05573144182562828, 0.061026643961668015, 0.06157573312520981, 0.00327630084939301, -0.03627776354551315, -0.014302250929176807, 0.010526401922106743, 0.028202397748827934, 0.03672716021537781, 0.05009095370769501, -0.13382764160633087, -0.06372376531362534, -0.04316357895731926, 0.03644726425409317, -0.03669825196266174, 0.00822498183697462, -0.14955639839172363, -0.0045658862218260765, -0.3704322874546051, 0.07348527759313583, -0.0857921689748764, 0.0485745333135128, -0.005439548287540674, -0.04172232002019882, -0.05030437186360359, 0.08903684467077255, -0.057541314512491226, -0.040556564927101135, -0.017326753586530685, 0.03415180370211601, -0.08770520240068436, -0.04339345544576645, 0.011562583968043327, -0.05165911465883255, 0.06494191288948059, 0.10635967552661896, -0.09711547940969467, 0.03072100132703781, -0.18112115561962128, -0.0788857713341713, 0.025296511128544807, 0.01623857207596302, 0.0010096434270963073, 0.055752839893102646, -0.017844308167696, 0.01924111880362034, 0.05656599625945091, -0.0446193553507328, 0.10989080369472504, -0.035602666437625885, -0.026069918647408485, -0.08346962183713913, 0.02293614111840725, -0.06278341263532639, -0.03182792291045189, 0.11765320599079132, 0.13694046437740326, 0.14036186039447784, -0.037063151597976685, -0.03749287873506546, -0.1382530927658081, -0.006191232707351446, 0.03889854997396469, -0.10425756126642227, -0.08056282997131348, -0.12258397042751312, -0.0006454741815105081, -0.042287081480026245, 0.11681294441223145, -0.04883919283747673, -0.035187821835279465, -0.03221452981233597, 0.0570310577750206, -0.009834888391196728, -0.02610824629664421, 0.24944527447223663, 0.025204678997397423, 0.05213044583797455, -0.10362540930509567, 0.02233477495610714, 0.10367557406425476, 0.0665445327758789, -0.033656880259513855, 0.09134224057197571, -0.0036672749556601048, 0.1820632666349411, -0.0016865934012457728, 0.05995919555425644, 0.004058873746544123, 0.09460962563753128, 0.017524762079119682, 0.08828059583902359, -0.07254040241241455, 0.10554292052984238, 0.16533470153808594, -0.07068187743425369, -0.02110847271978855, -0.02026640996336937, -0.04160476475954056, -0.11158563941717148, -0.19339922070503235, -0.09164351969957352, -0.17389141023159027, 0.0048205191269516945, -0.056006502360105515, 0.020462457090616226, 0.04260044917464256, 0.010681611485779285, 0.052895691245794296, 0.02114601992070675, -0.04575579613447189, -0.06759212911128998, 0.04992321506142616, -0.032890915870666504, -0.10757148265838623, 0.11574847996234894, -0.0429764986038208, 0.10396513342857361, -0.05226317048072815, 0.0055188448168337345, 0.039296090602874756, 0.1158481314778328, 0.06772539764642715, -0.054478924721479416, -0.06988323479890823, -0.05751284584403038, 0.07912522554397583, -0.032485611736774445, 0.11001747101545334, 0.058005742728710175, -0.03099212795495987, 0.04817763715982437, 0.15537633001804352, -0.103085458278656, -0.13398785889148712, -0.1204381138086319, 0.07208631187677383, -0.07140597701072693, 0.01680902950465679, -0.03769192472100258, -0.03636549040675163, 0.026248585432767868, 0.22640115022659302, 0.17008863389492035, -0.11893700063228607, -0.011437692679464817, 0.0011186229530721903, 0.00789870135486126, -0.0331297367811203, 0.14378701150417328, 0.12464046478271484, -0.0065673720091581345, -0.027075888589024544, -0.0411953330039978, 0.01084635965526104, -0.04556144401431084, -0.13746041059494019, -0.015689712017774582, -0.11123433709144592, -0.050394412130117416, -0.005257047712802887, -0.009890048764646053, -0.11488472670316696, -0.0812324807047844, -0.042963989078998566, 0.030964385718107224, -0.015056746080517769, -0.0928197056055069, 0.06388314813375473, 0.07870370149612427, 0.002710758475586772, -0.09333661198616028, 0.06143396720290184, 0.18861739337444305, -0.06932500749826431, -0.14919616281986237, -0.05772629752755165, 0.03707500919699669, 0.021693432703614235, 0.096107617020607, 0.022018900141119957, 0.014911269769072533, 0.07446934282779694, -0.005015179514884949, -0.1593298465013504, 0.054211780428886414, -0.03938549384474754, -0.03764606639742851, -0.027982603758573532, -0.008608952164649963, -0.09737072885036469, 0.052216317504644394, 0.03187326341867447, 0.018888922408223152, -0.04031196981668472, 0.13054780662059784, -0.040919650346040726, -0.07498334348201752, -0.022878792136907578, -0.09962502866983414, 0.11177625507116318, 0.06127462536096573, -0.053374044597148895, -0.053013186901807785, -0.08652359247207642, 0.056068431586027145, 0.016003839671611786, -0.08430780470371246, -0.010253862477838993, 0.0020482325926423073, -0.030132073909044266, 0.04681249335408211, 0.05749025195837021, -0.14012490212917328, -0.04076363146305084, -0.08344270288944244, -0.004431870300322771, -0.06399209052324295, 0.12027354538440704, 0.09523942321538925, 0.04190758615732193, -0.011619864962995052, -0.21355657279491425, -0.015023868530988693, 0.03790748119354248, -0.05815742164850235, -0.10924430936574936 ]
null
null
null
# Lora of marblehead/マーブルヘッド/马布尔黑德 (Azur Lane) ## What Is This? This is the LoRA model of waifu marblehead/マーブルヘッド/马布尔黑德 (Azur Lane). ## How Is It Trained? * This model is trained with [HCP-Diffusion](https://github.com/7eu7d7/HCP-Diffusion). * The [auto-training framework](https://github.com/deepghs/cyberharem) is maintained by [DeepGHS Team](https://huggingface.co/deepghs). * The base model used for training is [deepghs/animefull-latest](https://huggingface.co/deepghs/animefull-latest). * Dataset used for training is the `stage3-p480-800` in [CyberHarem/marblehead_azurlane](https://huggingface.co/datasets/CyberHarem/marblehead_azurlane), which contains 177 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 1800 steps, 40 checkpoints were saved and evaluated. * **Trigger word is `marblehead_azurlane`.** * Pruned core tags for this waifu are `blonde_hair, blue_eyes, breasts, hair_ornament, multicolored_hair, large_breasts, hairclip, pink_hair, two-tone_hair, hair_between_eyes, bangs, sidelocks`. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. ## How to Use It? ### If You Are Using A1111 WebUI v1.7+ **Just use it like the classic LoRA**. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 630, you need to download [`630/marblehead_azurlane.pt`](https://huggingface.co/CyberHarem/marblehead_azurlane/resolve/main/630/marblehead_azurlane.pt) as the embedding and [`630/marblehead_azurlane.safetensors`](https://huggingface.co/CyberHarem/marblehead_azurlane/resolve/main/630/marblehead_azurlane.safetensors) for loading Lora. By using both files together, you can generate images for the desired characters. ## Which Step Should I Use? We selected 5 good steps for you to choose. The best one is step 630. 1600 images (1.72 GiB) were generated for auto-testing. ![Metrics Plot](metrics_plot.png) The base model used for generating preview images is [Meina/MeinaMix_V11](https://huggingface.co/Meina/MeinaMix_V11). Here are the preview of the recommended steps: | Step | Epoch | CCIP | AI Corrupt | Bikini Plus | Score | Download | pattern_0_0 | pattern_0_1 | pattern_1 | pattern_2_0 | pattern_2_1 | portrait_0 | portrait_1 | portrait_2 | full_body_0 | full_body_1 | profile_0 | profile_1 | free_0 | free_1 | shorts | maid_0 | maid_1 | miko | yukata | suit | china | bikini_0 | bikini_1 | bikini_2 | sit | squat | kneel | jump | crossed_arms | angry | smile | cry | grin | n_lie_0 | n_lie_1 | n_stand_0 | n_stand_1 | n_stand_2 | n_sex_0 | n_sex_1 | |-------:|--------:|:----------|:-------------|:--------------|:----------|:------------------------------------------------------------------------------------------------------------|:----------------------------------------------|:----------------------------------------------|:------------------------------------------|:----------------------------------------------|:----------------------------------------------|:--------------------------------------------|:--------------------------------------------|:--------------------------------------------|:----------------------------------------------|:----------------------------------------------|:------------------------------------------|:------------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:--------------------------------|:------------------------------------|:--------------------------------|:----------------------------------|:----------------------------------------|:----------------------------------------|:----------------------------------------|:------------------------------|:----------------------------------|:----------------------------------|:--------------------------------|:------------------------------------------------|:----------------------------------|:----------------------------------|:------------------------------|:--------------------------------|:--------------------------------------|:--------------------------------------|:------------------------------------------|:------------------------------------------|:------------------------------------------|:--------------------------------------|:--------------------------------------| | 630 | 15 | **0.982** | 0.967 | **0.841** | **0.729** | [Download](https://huggingface.co/CyberHarem/marblehead_azurlane/resolve/main/630/marblehead_azurlane.zip) | ![pattern_0_0](630/previews/pattern_0_0.png) | ![pattern_0_1](630/previews/pattern_0_1.png) | ![pattern_1](630/previews/pattern_1.png) | ![pattern_2_0](630/previews/pattern_2_0.png) | ![pattern_2_1](630/previews/pattern_2_1.png) | ![portrait_0](630/previews/portrait_0.png) | ![portrait_1](630/previews/portrait_1.png) | ![portrait_2](630/previews/portrait_2.png) | ![full_body_0](630/previews/full_body_0.png) | ![full_body_1](630/previews/full_body_1.png) | ![profile_0](630/previews/profile_0.png) | ![profile_1](630/previews/profile_1.png) | ![free_0](630/previews/free_0.png) | ![free_1](630/previews/free_1.png) | ![shorts](630/previews/shorts.png) | ![maid_0](630/previews/maid_0.png) | ![maid_1](630/previews/maid_1.png) | ![miko](630/previews/miko.png) | ![yukata](630/previews/yukata.png) | ![suit](630/previews/suit.png) | ![china](630/previews/china.png) | ![bikini_0](630/previews/bikini_0.png) | ![bikini_1](630/previews/bikini_1.png) | ![bikini_2](630/previews/bikini_2.png) | ![sit](630/previews/sit.png) | ![squat](630/previews/squat.png) | ![kneel](630/previews/kneel.png) | ![jump](630/previews/jump.png) | ![crossed_arms](630/previews/crossed_arms.png) | ![angry](630/previews/angry.png) | ![smile](630/previews/smile.png) | ![cry](630/previews/cry.png) | ![grin](630/previews/grin.png) | ![n_lie_0](630/previews/n_lie_0.png) | ![n_lie_1](630/previews/n_lie_1.png) | ![n_stand_0](630/previews/n_stand_0.png) | ![n_stand_1](630/previews/n_stand_1.png) | ![n_stand_2](630/previews/n_stand_2.png) | ![n_sex_0](630/previews/n_sex_0.png) | ![n_sex_1](630/previews/n_sex_1.png) | | 405 | 10 | 0.982 | **0.987** | 0.840 | 0.727 | [Download](https://huggingface.co/CyberHarem/marblehead_azurlane/resolve/main/405/marblehead_azurlane.zip) | ![pattern_0_0](405/previews/pattern_0_0.png) | ![pattern_0_1](405/previews/pattern_0_1.png) | ![pattern_1](405/previews/pattern_1.png) | ![pattern_2_0](405/previews/pattern_2_0.png) | ![pattern_2_1](405/previews/pattern_2_1.png) | ![portrait_0](405/previews/portrait_0.png) | ![portrait_1](405/previews/portrait_1.png) | ![portrait_2](405/previews/portrait_2.png) | ![full_body_0](405/previews/full_body_0.png) | ![full_body_1](405/previews/full_body_1.png) | ![profile_0](405/previews/profile_0.png) | ![profile_1](405/previews/profile_1.png) | ![free_0](405/previews/free_0.png) | ![free_1](405/previews/free_1.png) | ![shorts](405/previews/shorts.png) | ![maid_0](405/previews/maid_0.png) | ![maid_1](405/previews/maid_1.png) | ![miko](405/previews/miko.png) | ![yukata](405/previews/yukata.png) | ![suit](405/previews/suit.png) | ![china](405/previews/china.png) | ![bikini_0](405/previews/bikini_0.png) | ![bikini_1](405/previews/bikini_1.png) | ![bikini_2](405/previews/bikini_2.png) | ![sit](405/previews/sit.png) | ![squat](405/previews/squat.png) | ![kneel](405/previews/kneel.png) | ![jump](405/previews/jump.png) | ![crossed_arms](405/previews/crossed_arms.png) | ![angry](405/previews/angry.png) | ![smile](405/previews/smile.png) | ![cry](405/previews/cry.png) | ![grin](405/previews/grin.png) | ![n_lie_0](405/previews/n_lie_0.png) | ![n_lie_1](405/previews/n_lie_1.png) | ![n_stand_0](405/previews/n_stand_0.png) | ![n_stand_1](405/previews/n_stand_1.png) | ![n_stand_2](405/previews/n_stand_2.png) | ![n_sex_0](405/previews/n_sex_0.png) | ![n_sex_1](405/previews/n_sex_1.png) | | 1215 | 28 | 0.972 | 0.973 | 0.835 | 0.708 | [Download](https://huggingface.co/CyberHarem/marblehead_azurlane/resolve/main/1215/marblehead_azurlane.zip) | ![pattern_0_0](1215/previews/pattern_0_0.png) | ![pattern_0_1](1215/previews/pattern_0_1.png) | ![pattern_1](1215/previews/pattern_1.png) | ![pattern_2_0](1215/previews/pattern_2_0.png) | ![pattern_2_1](1215/previews/pattern_2_1.png) | ![portrait_0](1215/previews/portrait_0.png) | ![portrait_1](1215/previews/portrait_1.png) | ![portrait_2](1215/previews/portrait_2.png) | ![full_body_0](1215/previews/full_body_0.png) | ![full_body_1](1215/previews/full_body_1.png) | ![profile_0](1215/previews/profile_0.png) | ![profile_1](1215/previews/profile_1.png) | ![free_0](1215/previews/free_0.png) | ![free_1](1215/previews/free_1.png) | ![shorts](1215/previews/shorts.png) | ![maid_0](1215/previews/maid_0.png) | ![maid_1](1215/previews/maid_1.png) | ![miko](1215/previews/miko.png) | ![yukata](1215/previews/yukata.png) | ![suit](1215/previews/suit.png) | ![china](1215/previews/china.png) | ![bikini_0](1215/previews/bikini_0.png) | ![bikini_1](1215/previews/bikini_1.png) | ![bikini_2](1215/previews/bikini_2.png) | ![sit](1215/previews/sit.png) | ![squat](1215/previews/squat.png) | ![kneel](1215/previews/kneel.png) | ![jump](1215/previews/jump.png) | ![crossed_arms](1215/previews/crossed_arms.png) | ![angry](1215/previews/angry.png) | ![smile](1215/previews/smile.png) | ![cry](1215/previews/cry.png) | ![grin](1215/previews/grin.png) | ![n_lie_0](1215/previews/n_lie_0.png) | ![n_lie_1](1215/previews/n_lie_1.png) | ![n_stand_0](1215/previews/n_stand_0.png) | ![n_stand_1](1215/previews/n_stand_1.png) | ![n_stand_2](1215/previews/n_stand_2.png) | ![n_sex_0](1215/previews/n_sex_0.png) | ![n_sex_1](1215/previews/n_sex_1.png) | | 1305 | 30 | 0.976 | 0.986 | 0.831 | 0.707 | [Download](https://huggingface.co/CyberHarem/marblehead_azurlane/resolve/main/1305/marblehead_azurlane.zip) | ![pattern_0_0](1305/previews/pattern_0_0.png) | ![pattern_0_1](1305/previews/pattern_0_1.png) | ![pattern_1](1305/previews/pattern_1.png) | ![pattern_2_0](1305/previews/pattern_2_0.png) | ![pattern_2_1](1305/previews/pattern_2_1.png) | ![portrait_0](1305/previews/portrait_0.png) | ![portrait_1](1305/previews/portrait_1.png) | ![portrait_2](1305/previews/portrait_2.png) | ![full_body_0](1305/previews/full_body_0.png) | ![full_body_1](1305/previews/full_body_1.png) | ![profile_0](1305/previews/profile_0.png) | ![profile_1](1305/previews/profile_1.png) | ![free_0](1305/previews/free_0.png) | ![free_1](1305/previews/free_1.png) | ![shorts](1305/previews/shorts.png) | ![maid_0](1305/previews/maid_0.png) | ![maid_1](1305/previews/maid_1.png) | ![miko](1305/previews/miko.png) | ![yukata](1305/previews/yukata.png) | ![suit](1305/previews/suit.png) | ![china](1305/previews/china.png) | ![bikini_0](1305/previews/bikini_0.png) | ![bikini_1](1305/previews/bikini_1.png) | ![bikini_2](1305/previews/bikini_2.png) | ![sit](1305/previews/sit.png) | ![squat](1305/previews/squat.png) | ![kneel](1305/previews/kneel.png) | ![jump](1305/previews/jump.png) | ![crossed_arms](1305/previews/crossed_arms.png) | ![angry](1305/previews/angry.png) | ![smile](1305/previews/smile.png) | ![cry](1305/previews/cry.png) | ![grin](1305/previews/grin.png) | ![n_lie_0](1305/previews/n_lie_0.png) | ![n_lie_1](1305/previews/n_lie_1.png) | ![n_stand_0](1305/previews/n_stand_0.png) | ![n_stand_1](1305/previews/n_stand_1.png) | ![n_stand_2](1305/previews/n_stand_2.png) | ![n_sex_0](1305/previews/n_sex_0.png) | ![n_sex_1](1305/previews/n_sex_1.png) | | 720 | 17 | 0.973 | 0.982 | 0.829 | 0.701 | [Download](https://huggingface.co/CyberHarem/marblehead_azurlane/resolve/main/720/marblehead_azurlane.zip) | ![pattern_0_0](720/previews/pattern_0_0.png) | ![pattern_0_1](720/previews/pattern_0_1.png) | ![pattern_1](720/previews/pattern_1.png) | ![pattern_2_0](720/previews/pattern_2_0.png) | ![pattern_2_1](720/previews/pattern_2_1.png) | ![portrait_0](720/previews/portrait_0.png) | ![portrait_1](720/previews/portrait_1.png) | ![portrait_2](720/previews/portrait_2.png) | ![full_body_0](720/previews/full_body_0.png) | ![full_body_1](720/previews/full_body_1.png) | ![profile_0](720/previews/profile_0.png) | ![profile_1](720/previews/profile_1.png) | ![free_0](720/previews/free_0.png) | ![free_1](720/previews/free_1.png) | ![shorts](720/previews/shorts.png) | ![maid_0](720/previews/maid_0.png) | ![maid_1](720/previews/maid_1.png) | ![miko](720/previews/miko.png) | ![yukata](720/previews/yukata.png) | ![suit](720/previews/suit.png) | ![china](720/previews/china.png) | ![bikini_0](720/previews/bikini_0.png) | ![bikini_1](720/previews/bikini_1.png) | ![bikini_2](720/previews/bikini_2.png) | ![sit](720/previews/sit.png) | ![squat](720/previews/squat.png) | ![kneel](720/previews/kneel.png) | ![jump](720/previews/jump.png) | ![crossed_arms](720/previews/crossed_arms.png) | ![angry](720/previews/angry.png) | ![smile](720/previews/smile.png) | ![cry](720/previews/cry.png) | ![grin](720/previews/grin.png) | ![n_lie_0](720/previews/n_lie_0.png) | ![n_lie_1](720/previews/n_lie_1.png) | ![n_stand_0](720/previews/n_stand_0.png) | ![n_stand_1](720/previews/n_stand_1.png) | ![n_stand_2](720/previews/n_stand_2.png) | ![n_sex_0](720/previews/n_sex_0.png) | ![n_sex_1](720/previews/n_sex_1.png) | ## Anything Else? Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. ## All Steps We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * [Steps From 1395 to 1800](all/0.md) * [Steps From 945 to 1350](all/1.md) * [Steps From 495 to 900](all/2.md) * [Steps From 45 to 450](all/3.md)
{"license": "mit", "tags": ["art", "not-for-all-audiences"], "datasets": ["CyberHarem/marblehead_azurlane"], "pipeline_tag": "text-to-image"}
text-to-image
CyberHarem/marblehead_azurlane
[ "art", "not-for-all-audiences", "text-to-image", "dataset:CyberHarem/marblehead_azurlane", "license:mit", "region:us" ]
2024-02-15T01:39:43+00:00
[]
[]
TAGS #art #not-for-all-audiences #text-to-image #dataset-CyberHarem/marblehead_azurlane #license-mit #region-us
Lora of marblehead/マーブルヘッド/马布尔黑德 (Azur Lane) ============================================ What Is This? ------------- This is the LoRA model of waifu marblehead/マーブルヘッド/马布尔黑德 (Azur Lane). How Is It Trained? ------------------ * This model is trained with HCP-Diffusion. * The auto-training framework is maintained by DeepGHS Team. * The base model used for training is deepghs/animefull-latest. * Dataset used for training is the 'stage3-p480-800' in CyberHarem/marblehead\_azurlane, which contains 177 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 1800 steps, 40 checkpoints were saved and evaluated. * Trigger word is 'marblehead\_azurlane'. * Pruned core tags for this waifu are 'blonde\_hair, blue\_eyes, breasts, hair\_ornament, multicolored\_hair, large\_breasts, hairclip, pink\_hair, two-tone\_hair, hair\_between\_eyes, bangs, sidelocks'. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. How to Use It? -------------- ### If You Are Using A1111 WebUI v1.7+ Just use it like the classic LoRA. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 630, you need to download '630/marblehead\_azurlane.pt' as the embedding and '630/marblehead\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters. Which Step Should I Use? ------------------------ We selected 5 good steps for you to choose. The best one is step 630. 1600 images (1.72 GiB) were generated for auto-testing. !Metrics Plot The base model used for generating preview images is Meina/MeinaMix\_V11. Here are the preview of the recommended steps: Anything Else? -------------- Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. All Steps --------- We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * Steps From 1395 to 1800 * Steps From 945 to 1350 * Steps From 495 to 900 * Steps From 45 to 450
[ "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 630, you need to download '630/marblehead\\_azurlane.pt' as the embedding and '630/marblehead\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 630.\n\n\n1600 images (1.72 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 1395 to 1800\n* Steps From 945 to 1350\n* Steps From 495 to 900\n* Steps From 45 to 450" ]
[ "TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/marblehead_azurlane #license-mit #region-us \n", "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 630, you need to download '630/marblehead\\_azurlane.pt' as the embedding and '630/marblehead\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 630.\n\n\n1600 images (1.72 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 1395 to 1800\n* Steps From 945 to 1350\n* Steps From 495 to 900\n* Steps From 45 to 450" ]
[ 45, 38, 470 ]
[ "passage: TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/marblehead_azurlane #license-mit #region-us \n### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file." ]
[ 0.005611651577055454, -0.003102501854300499, -0.004094063304364681, 0.08927373588085175, 0.07779097557067871, 0.08434705436229706, 0.2164371758699417, 0.0779314935207367, 0.1314191222190857, -0.0553281269967556, 0.09500580281019211, 0.07650057971477509, -0.00848967395722866, 0.027626018971204758, -0.04162229225039482, -0.15504474937915802, -0.07252649962902069, -0.034184832125902176, -0.009655528701841831, 0.014254990965127945, 0.06581319868564606, 0.003297241171821952, 0.09730031341314316, -0.05791318789124489, -0.05320879444479942, 0.04850701615214348, -0.022232508286833763, -0.03542651981115341, 0.038749564439058304, 0.0855504721403122, 0.10539966076612473, 0.014032541774213314, 0.0582641176879406, -0.14637771248817444, 0.07145245373249054, -0.0075097051449120045, -0.09880442917346954, 0.00021781558461952955, 0.019090304151177406, -0.038842953741550446, 0.09477468580007553, 0.013948556035757065, -0.10310826450586319, 0.04610701650381088, -0.1311943680047989, -0.03396261855959892, -0.060163047164678574, 0.02691439352929592, 0.13891804218292236, 0.04260998219251633, 0.02204793132841587, 0.049623411148786545, -0.051527515053749084, 0.07845582067966461, 0.12863413989543915, -0.1467132419347763, -0.07656972855329514, 0.11176427453756332, 0.010297540575265884, 0.15735234320163727, -0.08866556733846664, 0.10429082810878754, 0.06368470937013626, -0.04257700964808464, -0.14161185920238495, -0.09694703668355942, -0.20527617633342743, -0.002022411208599806, 0.00683244364336133, 0.03358926624059677, 0.4113456606864929, 0.0640043318271637, 0.04248479753732681, 0.0554841049015522, -0.06654620915651321, 0.05301003158092499, -0.101410873234272, 0.13999538123607635, 0.04926485940814018, 0.09244079887866974, -0.04735257849097252, -0.11275609582662582, -0.11185199022293091, -0.07198811322450638, -0.07184041291475296, -0.005473101045936346, 0.019283002242445946, 0.11856188625097275, -0.19952526688575745, 0.010493612848222256, -0.07803715765476227, -0.1189500018954277, 0.020775189623236656, -0.09914511442184448, 0.16160523891448975, 0.06781773269176483, -0.018381379544734955, 0.009549305774271488, 0.2526322305202484, 0.11837247014045715, 0.2295387238264084, 0.06180384010076523, -0.1062999740242958, 0.13670365512371063, 0.03702741861343384, -0.09394269436597824, -0.002953434130176902, -0.11050919443368912, 0.14576183259487152, -0.036984048783779144, 0.11510899662971497, -0.057994283735752106, -0.11399375647306442, 0.014889780431985855, -0.10875429958105087, 0.06803525239229202, 0.05514950677752495, 0.0015912052476778626, -0.04550142213702202, 0.05392547696828842, 0.04026840627193451, -0.037768151611089706, -0.007608095183968544, -0.012122144922614098, -0.04486894980072975, 0.045970603823661804, 0.11106469482183456, 0.027866650372743607, 0.06598444283008575, -0.002353840973228216, -0.026811659336090088, -0.007756002247333527, -0.042702529579401016, 0.009013527072966099, 0.04444952309131622, 0.03517754003405571, 0.08762889355421066, -0.14908410608768463, -0.07614938169717789, -0.00925363041460514, 0.056428324431180954, 0.00874935369938612, 0.08118686825037003, -0.006211032159626484, 0.05710915848612785, 0.009517240338027477, -0.020001590251922607, 0.06746718287467957, -0.0973999872803688, 0.08555923402309418, -0.016113756224513054, 0.09636291116476059, -0.20727358758449554, -0.004488123115152121, -0.02853107452392578, 0.010579385794699192, 0.03831074759364128, -0.011383737437427044, -0.1066979244351387, 0.1144731268286705, -0.014229892753064632, 0.0733361765742302, -0.11996978521347046, 0.04278630390763283, 0.02966330759227276, 0.0844842791557312, -0.10136362910270691, 0.004344338085502386, 0.1306982934474945, -0.12710896134376526, -0.16297276318073273, 0.09700187295675278, -0.020695557817816734, 0.02066480927169323, 0.04488270729780197, 0.15777930617332458, 0.1446622610092163, -0.20209302008152008, -0.009442918933928013, 0.05928001180291176, -0.027972472831606865, -0.08875632286071777, -0.01156108919531107, 0.08532287180423737, 0.030097240582108498, 0.04132606089115143, -0.03748537600040436, 0.11953465640544891, -0.03184807300567627, -0.07397210597991943, -0.037588704377412796, -0.07733294367790222, -0.0658588707447052, 0.048736296594142914, -0.017042510211467743, -0.05187937989830971, 0.008000025525689125, -0.15991954505443573, 0.15669526159763336, 0.024706708267331123, 0.023473046720027924, -0.06457500159740448, 0.12269475311040878, 0.0004112987662665546, 0.00817124079912901, -0.005112656392157078, -0.06440190225839615, -0.106041818857193, 0.23361600935459137, 0.10180477797985077, 0.09032915532588959, 0.06229126453399658, -0.049558125436306, -0.06886649131774902, 0.025185950100421906, 0.014887111261487007, -0.03494405746459961, 0.016465721651911736, -0.10443715006113052, 0.057509563863277435, -0.016570666804909706, 0.03842947632074356, -0.019881270825862885, -0.02422478049993515, 0.08821598440408707, 0.0029526979196816683, -0.009508872404694557, 0.08043354749679565, 0.04745474457740784, -0.029439378529787064, -0.05780396610498428, -0.001383772469125688, 0.07876720279455185, -0.005647291429340839, -0.07373128086328506, 0.010889062657952309, -0.009226027876138687, 0.03237205371260643, 0.2047792673110962, -0.20061896741390228, 0.04750150814652443, -0.008501271717250347, 0.04612832888960838, 0.034554820507764816, -0.0019772497471421957, -0.036525774747133255, 0.0454283244907856, -0.007469814270734787, 0.06452250480651855, -0.018392620608210564, 0.06855125725269318, -0.015746790915727615, -0.12668956816196442, -0.01824987307190895, -0.03503766283392906, 0.18847280740737915, -0.1665026843547821, 0.055551983416080475, 0.19367428123950958, -0.1214493066072464, 0.1353122740983963, 0.0008286600932478905, -0.005928272381424904, 0.010698522441089153, 0.03176074102520943, 0.003037400543689728, 0.11610621213912964, -0.08068432658910751, -0.03388485312461853, 0.02118007466197014, -0.08695048093795776, 0.03190994635224342, -0.12661267817020416, -0.10476461052894592, -0.07298129051923752, -0.027386657893657684, -0.03460989519953728, 0.0327472947537899, -0.0515279620885849, 0.07676135748624802, -0.07977675646543503, -0.06387436389923096, -0.02506067603826523, -0.08298036456108093, 0.01823282614350319, 0.0021702779922634363, -0.07226942479610443, -0.10792573541402817, -0.1151396632194519, -0.08513333648443222, -0.1391305774450302, -0.014684746041893959, 0.06710169464349747, -0.10646561533212662, -0.03553846478462219, 0.024392317980527878, -0.04283221438527107, 0.09146115183830261, -0.08067288994789124, 0.020833345130085945, 0.05103248730301857, -0.049208685755729675, -0.16591958701610565, -0.006205551791936159, -0.07290119677782059, -0.05443953350186348, 0.14973250031471252, -0.1481851488351822, 0.1740914285182953, -0.02566421963274479, 0.06930344551801682, 0.06878028810024261, 0.02373030036687851, 0.12070082873106003, -0.11066250503063202, 0.08938145637512207, 0.18556024134159088, 0.04931160435080528, 0.08357905596494675, 0.11821097135543823, 0.08583009243011475, -0.1135503500699997, 0.033362776041030884, 0.07653027027845383, -0.09951148927211761, -0.08522558957338333, -0.050350528210401535, -0.12166307121515274, -0.07360512763261795, 0.05192150920629501, 0.06267929822206497, 0.01838589832186699, 0.12119565159082413, -0.0456213504076004, -0.016026286408305168, 0.1101902574300766, 0.0443083792924881, 0.0825776532292366, 0.010495380498468876, 0.054101791232824326, -0.14292676746845245, -0.037562645971775055, 0.16575109958648682, 0.2286510467529297, 0.24505415558815002, 0.03083210438489914, 0.07145790010690689, 0.12361828982830048, 0.06623784452676773, 0.098662368953228, 0.049900367856025696, 0.018732978031039238, 0.012551801279187202, -0.07427076250314713, -0.05615765601396561, 0.023641260340809822, 0.003970942460000515, -0.026273557916283607, -0.1449851393699646, 0.10486703366041183, 0.016349593177437782, 0.07687962800264359, 0.16532208025455475, 0.028853213414549828, -0.102948397397995, 0.16604115068912506, 0.09621042758226395, 0.09396011382341385, -0.07385409623384476, 0.13126079738140106, 0.04604395478963852, -0.004066163208335638, 0.15583443641662598, 0.029582656919956207, 0.14792804419994354, -0.028636131435632706, -0.08108267188072205, -0.07613416016101837, -0.04767478629946709, 0.005368162877857685, 0.02052261307835579, -0.21051378548145294, 0.09939940273761749, 0.06251119077205658, 0.017985880374908447, 0.0001980347151402384, -0.04448606073856354, 0.18992865085601807, 0.15792669355869293, 0.09099548310041428, 0.02176274172961712, -0.02814219519495964, 0.0011792985023930669, -0.07753920555114746, 0.05870716646313667, 0.021602462977170944, 0.06385260820388794, -0.042732708156108856, -0.10013405978679657, -0.02385416068136692, -0.0017472775653004646, 0.017267949879169464, -0.08658292144536972, -0.11996955424547195, -0.04390353336930275, 0.2442772090435028, -0.08286160975694656, 0.04640701413154602, 0.05081944540143013, 0.012393992394208908, -0.037021275609731674, 0.020154686644673347, -0.028900036588311195, -0.01792619749903679, -0.04021279513835907, -0.00798122864216566, 0.004554790910333395, -0.04151232913136482, -0.062268201261758804, -0.01875687763094902, -0.11180063337087631, -0.10232369601726532, 0.00859174132347107, -0.04590277001261711, 0.021788714453577995, -0.031737398356199265, 0.02317929081618786, -0.10026799887418747, -0.034699149429798126, 0.023517724126577377, 0.045101601630449295, -0.07599000632762909, -0.12594196200370789, 0.009126625955104828, -0.002627810463309288, -0.052141327410936356, 0.03201933950185776, -0.10673461854457855, -0.09511637687683105, -0.05056805536150932, -0.031563226133584976, 0.12874959409236908, 0.23286405205726624, -0.029129352420568466, -0.01067547220736742, 0.14748577773571014, -0.10148665308952332, -0.3218301236629486, -0.15175803005695343, -0.16028651595115662, -0.09561324119567871, 0.03381315991282463, -0.07644589990377426, 0.03857244551181793, 0.07576323300600052, -0.045223668217659, 0.22537319362163544, -0.18822510540485382, -0.1007823720574379, 0.08118771761655807, 0.1066521406173706, 0.3156842291355133, -0.2495952844619751, 0.016350695863366127, -0.10584913194179535, -0.034128714352846146, 0.005956597626209259, -0.0837673619389534, 0.11364684998989105, 0.035862717777490616, 0.08032383024692535, -0.008109519258141518, -0.011528294533491135, 0.146105095744133, -0.07798227667808533, 0.1335935741662979, -0.10849808156490326, -0.11426171660423279, 0.19305039942264557, -0.039061881601810455, -0.007524522021412849, -0.2010762244462967, -0.03980712965130806, -0.009509528055787086, 0.04521208256483078, -0.011316030286252499, 0.04869646206498146, -0.010960583575069904, -0.006497535854578018, -0.1288205236196518, -0.017238816246390343, -0.05273851752281189, 0.054758209735155106, 0.20347630977630615, -0.07490991055965424, -0.06476718932390213, 0.03571365773677826, -0.00041041497024707496, 0.08271285146474838, -0.01215537078678608, -0.05487465113401413, -0.04953073710203171, 0.0902324840426445, -0.20274120569229126, 0.054354228079319, 0.010567859746515751, 0.008266650140285492, 0.0169236958026886, 0.013318716548383236, 0.01874549873173237, 0.11481408774852753, 0.1758970022201538, -0.012213828973472118, -0.05532156303524971, -0.01785215735435486, 0.02450866997241974, 0.12066560238599777, -0.03385799378156662, 0.10274028033018112, 0.03884687274694443, 0.03277619183063507, 0.006575687322765589, 0.05787518247961998, -0.07902365177869797, -0.10086620599031448, 0.10832025855779648, -0.04455535113811493, -0.08004777878522873, 0.08498834818601608, 0.04574546217918396, 0.05470900237560272, -0.008102711290121078, 0.0561099648475647, 0.01762893609702587, -0.12529657781124115, 0.02654694952070713, 0.20440290868282318, -0.03510061651468277, -0.05840816721320152, -0.06716486811637878, 0.003900367533788085, -0.12340519577264786, 0.07994858920574188, 0.03555787727236748, -0.03734728321433067, 0.11182186752557755, -0.04292277991771698, -0.02776738628745079, 0.014503344893455505, -0.08414949476718903, 0.031246952712535858, -0.1438794881105423, -0.21699035167694092, 0.051861073821783066, -0.010303844697773457, -0.06156131625175476, -0.09232711791992188, -0.08108507841825485, 0.06114879623055458, -0.1326569765806198, 0.14626404643058777, -0.0664295181632042, 0.059273697435855865, -0.04043989256024361, -0.04784848168492317, -0.11040833592414856, -0.0161623265594244, -0.052442923188209534, -0.01595066301524639, 0.061568234115839005, 0.02283504791557789, -0.12200481444597244, -0.11764005571603775, 0.060424190014600754, -0.008450393564999104, -0.0026371327694505453, 0.006104428321123123, -0.0704970732331276, 0.025914017111063004, -0.20734693109989166, -0.06904510408639908, 0.08185124397277832, 0.04577034339308739, -0.0858943834900856, 0.12045330554246902, 0.04697759076952934, -0.03509997949004173, 0.053244367241859436, 0.003154047532007098, 0.1709977090358734, -0.07239498943090439, 0.035991404205560684, -0.11854781955480576, -0.16080160439014435, -0.028371945023536682, 0.03588900715112686, 0.2360391914844513, 0.093897745013237, 0.10108233988285065, -0.04619836434721947, 0.01408863440155983, -0.02668340690433979, 0.07595262676477432, 0.01939030922949314, -0.10141859203577042, -0.03470245376229286, -0.1634538769721985, -0.06059949845075607, -0.06265285611152649, 0.15068113803863525, 0.03504906967282295, -0.157227024435997, 0.00428333505988121, 0.1161789745092392, -0.195809468626976, -0.011191016994416714, 0.143422931432724, -0.0540115088224411, 0.026034832000732422, -0.15254640579223633, 0.032233331352472305, 0.07885510474443436, -0.023256082087755203, 0.01730145886540413, 0.1286315768957138, 0.01697433367371559, 0.005452148150652647, 0.04423487186431885, -0.033862970769405365, 0.08041813224554062, -0.04878274351358414, 0.06044349819421768, 0.005921509116888046, -0.05465472862124443, -0.1124056801199913, 0.18955329060554504, -0.016431258991360664, 0.0060841175727546215, -0.0628613606095314, 0.00012281228555366397, -0.1128031313419342, -0.10206872224807739, -0.0695105791091919, -0.11765655130147934, 0.06994904577732086, -0.05695119500160217, 0.011667751707136631, -0.0052089388482272625, 0.01675351895391941, -0.06984754651784897, 0.004552491940557957, -0.17926691472530365, -0.05041951313614845, 0.02840171940624714, -0.02056204527616501, -0.030259238556027412, -0.054797105491161346, -0.03750401362776756, 0.026739463210105896, -0.07342465966939926, -0.06525203585624695, 0.05998700112104416, 0.06699075549840927, 0.05721447989344597, -0.17005503177642822, -0.09786710888147354, -0.071253702044487, 0.038323450833559036, 0.06869355589151382, 0.17698116600513458, 0.038865674287080765, -0.011105447076261044, 0.04241402447223663, 0.12967731058597565, 0.018127107992768288, -0.06932482868432999, -0.05823241174221039, -0.130996972322464, -0.14817045629024506, -0.013212445192039013, -0.061662010848522186, -0.02143324352800846, 0.028999952599406242, 0.22945939004421234, 0.18929296731948853, -0.14519886672496796, 0.040106188505887985, -0.06486187130212784, 0.03579358011484146, -0.033338937908411026, 0.16584251821041107, 0.04736374318599701, 0.15164698660373688, -0.03142806887626648, -0.031717341393232346, -0.06362447142601013, 0.022731833159923553, -0.10365651547908783, 0.025210388004779816, -0.0068429564125835896, -0.061167825013399124, -0.04939694330096245, 0.10819357633590698, -0.09228625148534775, 0.08700141310691833, 0.1647244691848755, -0.1558879017829895, -0.0226817037910223, -0.04363778606057167, 0.03933743014931679, 0.0894184410572052, 0.023754889145493507, -0.08597365021705627, -0.029439115896821022, 0.012454845011234283, 0.020006323233246803, -0.17028065025806427, -0.11003775894641876, -0.002211405197158456, -0.13930656015872955, 0.13169080018997192, -0.009905164130032063, 0.005263305269181728, 0.04042001813650131, -0.06330737471580505, -0.004144015721976757, 0.16540147364139557, 0.015146751888096333, -0.02316308207809925, -0.027935465797781944, -0.05798682942986488, -0.10609333217144012, 0.06323223561048508, 0.08920900523662567, 0.0717543214559555, -0.0021900974679738283, 0.14546559751033783, -0.0351899228990078, -0.04103538766503334, 0.1481592208147049, -0.1766006350517273, 0.08928535133600235, -0.02132611908018589, -0.010957217775285244, -0.069100521504879, -0.04007213935256004, 0.031798262149095535, 0.0833088606595993, -0.14738035202026367, -0.05294986814260483, 0.06626128405332565, -0.0893554762005806, 0.03929318115115166, 0.030480733141303062, -0.10413428395986557, 0.0032084970735013485, -0.12873846292495728, -0.0006810461054556072, -0.10423356294631958, 0.04179525002837181, 0.20501038432121277, -0.037991009652614594, 0.00805798638612032, -0.15407070517539978, 0.047857996076345444, -0.02450699359178543, -0.04955369606614113, -0.06902317702770233 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
longcule123/adapter-14-2_merged
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T01:44:06+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #mistral #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #mistral #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 60, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #mistral #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.04571164771914482, 0.1637648642063141, -0.005522117950022221, 0.017756497487425804, 0.09821303188800812, 0.01318030059337616, 0.06541220843791962, 0.1127115860581398, -0.017605241388082504, 0.1127321794629097, 0.030432263389229774, 0.09820804744958878, 0.1134178638458252, 0.14702944457530975, -0.003594378475099802, -0.22472713887691498, 0.052083637565374374, -0.12124937027692795, -0.03241228312253952, 0.1181139275431633, 0.14941681921482086, -0.09871039539575577, 0.07234785705804825, -0.030714161694049835, -0.01334790326654911, -0.03167412802577019, -0.05947697162628174, -0.045681875199079514, 0.046136777848005295, 0.0657167062163353, 0.06853367388248444, 0.007354621775448322, 0.08972878009080887, -0.2669793367385864, 0.019881360232830048, 0.06918594241142273, -0.0025153355672955513, 0.07059336453676224, 0.06344282627105713, -0.07033728063106537, 0.10271385312080383, -0.051166124641895294, 0.1467856466770172, 0.08377711474895477, -0.09116126596927643, -0.18892322480678558, -0.08764564990997314, 0.0990586131811142, 0.17651304602622986, 0.04750865325331688, -0.024397386237978935, 0.09895956516265869, -0.0878119245171547, 0.015860557556152344, 0.052259236574172974, -0.07261253148317337, -0.05407591536641121, 0.061004482209682465, 0.07816638052463531, 0.06616047024726868, -0.12551534175872803, -0.02998468652367592, 0.005221198312938213, 0.011705057695508003, 0.07518111169338226, 0.01836656779050827, 0.15222862362861633, 0.03479425609111786, -0.12653809785842896, -0.04834689199924469, 0.0983143299818039, 0.03359128534793854, -0.043975554406642914, -0.247073233127594, -0.031072303652763367, -0.026882093399763107, -0.030029185116291046, -0.038772210478782654, 0.04153512790799141, -0.006745535880327225, 0.08434242010116577, -0.0040448750369250774, -0.07344388216733932, -0.03874153643846512, 0.06087949126958847, 0.0669754296541214, 0.029331250116229057, -0.013996441848576069, 0.010876164771616459, 0.11490162461996078, 0.10806918889284134, -0.12199585139751434, -0.05589085817337036, -0.06492951512336731, -0.08786392956972122, -0.04284887760877609, 0.033410828560590744, 0.03509693965315819, 0.05435176193714142, 0.2536843419075012, 0.009815474040806293, 0.06126174330711365, 0.03745805472135544, 0.007310505956411362, 0.059651583433151245, 0.10812553018331528, -0.05987109988927841, -0.10409316420555115, -0.02881651371717453, 0.08857584744691849, 0.006609630770981312, -0.03354408219456673, -0.05052083358168602, 0.05901389569044113, 0.021856583654880524, 0.11749778687953949, 0.08884359151124954, 0.00984770804643631, -0.07126569002866745, -0.06146538630127907, 0.19450126588344574, -0.16384615004062653, 0.04264351725578308, 0.03702449053525925, -0.039683789014816284, -0.0003956064465455711, 0.011445282027125359, 0.01843930408358574, -0.023893611505627632, 0.09238249063491821, -0.05498874559998512, -0.04001082479953766, -0.1106586754322052, -0.0339570976793766, 0.034455835819244385, 0.010122774168848991, -0.03529255837202072, -0.03252722695469856, -0.08346389979124069, -0.07506290078163147, 0.09339368343353271, -0.07379438728094101, -0.04854428768157959, -0.018830472603440285, -0.0752616599202156, 0.02326788194477558, 0.02032634988427162, 0.07736726850271225, -0.023358777165412903, 0.04288764297962189, -0.054010841995477676, 0.05824148654937744, 0.11001134663820267, 0.035365406423807144, -0.05824809893965721, 0.06025301292538643, -0.2382364422082901, 0.09637492895126343, -0.07412451505661011, 0.05830197036266327, -0.15449334681034088, -0.02627694234251976, 0.04870045557618141, 0.0076532382518053055, -0.009597796015441418, 0.13436771929264069, -0.21578943729400635, -0.026375943794846535, 0.16865074634552002, -0.10160042345523834, -0.06946627050638199, 0.05867103114724159, -0.049256108701229095, 0.10817171633243561, 0.03891118988394737, -0.025492025539278984, 0.06244310364127159, -0.12527504563331604, 0.007147894706577063, -0.04992884770035744, -0.016554534435272217, 0.1592475026845932, 0.07294736802577972, -0.07235062122344971, 0.07110220938920975, 0.025814544409513474, -0.027441376820206642, -0.04532165080308914, -0.016039686277508736, -0.10585595667362213, 0.014911207370460033, -0.061168964952230453, 0.01876060478389263, -0.020111115649342537, -0.08977947384119034, -0.028080428019165993, -0.1748371720314026, -0.026230180636048317, 0.085477814078331, -0.007464459165930748, -0.018854627385735512, -0.11770102381706238, 0.008567224256694317, 0.044854406267404556, 0.006109896115958691, -0.13499478995800018, -0.04764661565423012, 0.027907660230994225, -0.16220368444919586, 0.033779170364141464, -0.05184612050652504, 0.05056280270218849, 0.026674345135688782, -0.029802238568663597, -0.025906935334205627, 0.022987615317106247, 0.006545235402882099, -0.011514187790453434, -0.24465326964855194, -0.026841215789318085, -0.026506783440709114, 0.166712686419487, -0.20777921378612518, 0.03577128052711487, 0.08057375997304916, 0.15318496525287628, 0.011457439512014389, -0.04087435454130173, 0.005527274217456579, -0.06868630647659302, -0.025992877781391144, -0.05823420733213425, -0.002480053110048175, -0.03337050974369049, -0.04843711107969284, 0.04469521716237068, -0.1662919819355011, -0.03491327911615372, 0.09593124687671661, 0.06427760422229767, -0.13986408710479736, -0.023568401113152504, -0.03526119887828827, -0.049809779971838, -0.047768235206604004, -0.06002878025174141, 0.11181395500898361, 0.058611296117305756, 0.04419868439435959, -0.059296321123838425, -0.07637067884206772, -0.0028071242850273848, -0.014342374168336391, -0.01986078731715679, 0.097631074488163, 0.06816094368696213, -0.1381729394197464, 0.09227006882429123, 0.09810956567525864, 0.07738673686981201, 0.09273158758878708, -0.02444581687450409, -0.08119411021471024, -0.0471174530684948, 0.03257923200726509, 0.018235107883810997, 0.1276484578847885, -0.027872784063220024, 0.04268912971019745, 0.0421174094080925, -0.018595336005091667, 0.013991083949804306, -0.08597505837678909, 0.033884208649396896, 0.02703946642577648, -0.0159194003790617, 0.04745442420244217, -0.037611253559589386, 0.024539871141314507, 0.08754327148199081, 0.04615016281604767, 0.033831849694252014, 0.015717241913080215, -0.05243339762091637, -0.10873834043741226, 0.1642032116651535, -0.12759798765182495, -0.22238075733184814, -0.13922695815563202, 0.003997850697487593, 0.036267586052417755, -0.01646288111805916, 0.002834152430295944, -0.060960907489061356, -0.12132686376571655, -0.08726011961698532, 0.015815909951925278, 0.050406474620103836, -0.0912260189652443, -0.060087788850069046, 0.056193675845861435, 0.037736181169748306, -0.14546552300453186, 0.01776101253926754, 0.04850281774997711, -0.09700650721788406, -0.004754792433232069, 0.07885372638702393, 0.06784981489181519, 0.17673011124134064, 0.018112216144800186, -0.021776698529720306, 0.031116241589188576, 0.20988549292087555, -0.13491620123386383, 0.11005933582782745, 0.13349974155426025, -0.09236859530210495, 0.08153878152370453, 0.20252206921577454, 0.04006611555814743, -0.09986240416765213, 0.032548144459724426, 0.02142537757754326, -0.027797512710094452, -0.2441972941160202, -0.07161470502614975, -0.004515932407230139, -0.06051458790898323, 0.07499068230390549, 0.09190185368061066, 0.08272628486156464, 0.011750337667763233, -0.09449771046638489, -0.08492138236761093, 0.06362129002809525, 0.10420511662960052, 0.02181125245988369, -0.009744768962264061, 0.09036174416542053, -0.03286943957209587, 0.01948373205959797, 0.08554471284151077, 0.0038120283279567957, 0.18320275843143463, 0.051725953817367554, 0.19073979556560516, 0.07944851368665695, 0.06951095163822174, 0.012023290619254112, 0.011227634735405445, 0.018135491758584976, 0.03228217363357544, -0.003646562807261944, -0.08350840210914612, -0.02080707624554634, 0.1153142973780632, 0.0672341138124466, 0.012952476739883423, 0.01729460060596466, -0.04021955281496048, 0.08128432929515839, 0.18377035856246948, -0.0093126455321908, -0.177269846200943, -0.06024068966507912, 0.07718996703624725, -0.09723462164402008, -0.09738315641880035, -0.01454379502683878, 0.030975129455327988, -0.1702532023191452, 0.025819219648838043, -0.023134231567382812, 0.11114585399627686, -0.13745717704296112, -0.020040949806571007, 0.07143081724643707, 0.07336213439702988, 0.004178736824542284, 0.055973317474126816, -0.16574905812740326, 0.1074945405125618, 0.007851972244679928, 0.06788748502731323, -0.0949488952755928, 0.10003086179494858, -0.002759356750175357, -0.016956903040409088, 0.13766175508499146, 0.003847390878945589, -0.0742180123925209, -0.07706846296787262, -0.08544620126485825, -0.010016623884439468, 0.12665624916553497, -0.13990990817546844, 0.08602021634578705, -0.03789570555090904, -0.04160536453127861, -0.0009961887262761593, -0.09994571655988693, -0.11771732568740845, -0.18694964051246643, 0.060274846851825714, -0.13818500936031342, 0.030693015083670616, -0.1080726683139801, -0.033236145973205566, -0.03044886700809002, 0.18898600339889526, -0.23496590554714203, -0.07289838045835495, -0.14654842019081116, -0.10314314812421799, 0.14515270292758942, -0.05135014280676842, 0.0824703797698021, -0.007518251892179251, 0.16955603659152985, 0.01909777894616127, -0.024870775640010834, 0.09702518582344055, -0.09090493619441986, -0.19369281828403473, -0.07736486196517944, 0.1553725302219391, 0.13563397526741028, 0.03274888917803764, -0.0031351360958069563, 0.03731042891740799, -0.016484085470438004, -0.119691863656044, 0.016338739544153214, 0.17828133702278137, 0.06005066633224487, 0.02449444867670536, -0.025351086631417274, -0.12034450471401215, -0.07065033912658691, -0.028268499299883842, 0.030481377616524696, 0.1794593334197998, -0.06955225765705109, 0.18364831805229187, 0.147920161485672, -0.05845186114311218, -0.20284810662269592, 0.01105605997145176, 0.03317207098007202, -0.00011460785754024982, 0.025185899809002876, -0.19945523142814636, 0.08448769152164459, 0.004838644526898861, -0.0498092919588089, 0.1281348466873169, -0.17351724207401276, -0.14425379037857056, 0.07726620137691498, 0.03829115256667137, -0.1926836371421814, -0.12892304360866547, -0.09138946235179901, -0.04540696740150452, -0.18867050111293793, 0.09461917728185654, 0.031194355338811874, 0.009373899549245834, 0.030387504026293755, 0.030604345723986626, 0.01938873715698719, -0.04181704297661781, 0.1860174536705017, -0.023930367082357407, 0.028327496722340584, -0.08596936613321304, -0.07190530747175217, 0.0391114242374897, -0.05227291211485863, 0.07252339273691177, -0.023452037945389748, 0.00719826715067029, -0.09769386798143387, -0.04156304895877838, -0.03843177855014801, 0.01581472158432007, -0.09648153930902481, -0.08523351699113846, -0.04445706307888031, 0.09780744463205338, 0.09553340077400208, -0.03473082184791565, -0.024805041030049324, -0.07508285343647003, 0.04805302992463112, 0.19605006277561188, 0.17889533936977386, 0.03904116898775101, -0.07846304774284363, -0.0033101453445851803, -0.010484009049832821, 0.04490501061081886, -0.20383046567440033, 0.06269704550504684, 0.05393069609999657, 0.019165942445397377, 0.11697915196418762, -0.01937638409435749, -0.15321338176727295, -0.07137971371412277, 0.062210626900196075, -0.05747547000646591, -0.19925202429294586, 0.008424095809459686, 0.062047190964221954, -0.16446428000926971, -0.045800499618053436, 0.046785544604063034, -0.004990153945982456, -0.03839265555143356, 0.022938871756196022, 0.09231305122375488, 0.0029900665394961834, 0.07426668703556061, 0.052022483199834824, 0.0835016593337059, -0.1060708537697792, 0.07922257483005524, 0.08730976283550262, -0.08381073921918869, 0.022620677947998047, 0.10530175268650055, -0.061487648636102676, -0.03560204058885574, 0.017662353813648224, 0.08361397683620453, 0.018624287098646164, -0.03893670439720154, 0.014383325353264809, -0.1065717563033104, 0.059272702783346176, 0.08645539730787277, 0.03302672877907753, 0.01618802361190319, 0.034192394465208054, 0.04655340686440468, -0.06840039044618607, 0.122025266289711, 0.032824426889419556, 0.017204686999320984, -0.035474274307489395, -0.04102595895528793, 0.01851540431380272, -0.03368416428565979, -0.005532157141715288, -0.03097093477845192, -0.07835554331541061, -0.015077406540513039, -0.16520504653453827, -0.009829589165747166, -0.05936548113822937, 0.012285472825169563, 0.031714752316474915, -0.034721489995718, 0.008415459655225277, 0.009580436162650585, -0.07713334262371063, -0.06541574746370316, -0.01965213567018509, 0.0961783304810524, -0.1606777459383011, 0.022340767085552216, 0.08350874483585358, -0.12098895758390427, 0.09293801337480545, 0.01664864458143711, -0.00869405921548605, 0.02654755860567093, -0.1516905426979065, 0.03389517217874527, -0.03324367105960846, 0.009356614202260971, 0.04251125827431679, -0.2180858999490738, -0.0012979574967175722, -0.034122150391340256, -0.06511902064085007, -0.008563618175685406, -0.035606082528829575, -0.1133907288312912, 0.10431582480669022, 0.007158213295042515, -0.08918852359056473, -0.031932637095451355, 0.02896781638264656, 0.08660420775413513, -0.02103978954255581, 0.1533614844083786, -0.008595003746449947, 0.07452014833688736, -0.16158120334148407, -0.019116591662168503, -0.0044966633431613445, 0.021838920190930367, -0.020337330177426338, -0.011089952662587166, 0.043057333678007126, -0.02310733124613762, 0.1769370436668396, -0.034001484513282776, 0.02080564945936203, 0.06879838556051254, 0.02382824197411537, -0.03270673379302025, 0.10420172661542892, 0.04176081717014313, 0.020029285922646523, 0.016749408096075058, 0.0014026050921529531, -0.04661702737212181, -0.03435906395316124, -0.1965997964143753, 0.07266207784414291, 0.15759599208831787, 0.09697116911411285, -0.019108884036540985, 0.07821404188871384, -0.0993313267827034, -0.10917975008487701, 0.12915705144405365, -0.04755320027470589, -0.004375945311039686, -0.07154709100723267, 0.13273866474628448, 0.14712604880332947, -0.18722544610500336, 0.07334931939840317, -0.07133730500936508, -0.04749078303575516, -0.10922681540250778, -0.194550022482872, -0.05630992352962494, -0.049111537635326385, -0.015855323523283005, -0.04727233946323395, 0.07431400567293167, 0.05443255603313446, 0.007043207995593548, -0.0018872307846322656, 0.06250270456075668, -0.02979675866663456, -0.004455813206732273, 0.033084239810705185, 0.06524696946144104, 0.012280851602554321, -0.028982065618038177, 0.017169395461678505, -0.009704679250717163, 0.04565926641225815, 0.06593092530965805, 0.0490880124270916, -0.02946917712688446, 0.01301988959312439, -0.040264759212732315, -0.10370729863643646, 0.044506072998046875, -0.02268853597342968, -0.081757090985775, 0.15341326594352722, 0.023376943543553352, 0.008703592233359814, -0.018961627036333084, 0.23797030746936798, -0.07337556779384613, -0.09915944188833237, -0.14910556375980377, 0.10603363811969757, -0.037726908922195435, 0.05897798761725426, 0.04798928648233414, -0.10144850611686707, 0.018896711990237236, 0.1251462697982788, 0.16306589543819427, -0.03724272549152374, 0.020064668729901314, 0.030806828290224075, 0.005520908627659082, -0.035788439214229584, 0.04845234379172325, 0.06755134463310242, 0.16263099014759064, -0.046816933900117874, 0.09447267651557922, 0.0011601726291701198, -0.09597980976104736, -0.03777771443128586, 0.10832508653402328, -0.014584118500351906, 0.018404638394713402, -0.059979453682899475, 0.11911186575889587, -0.06456011533737183, -0.2371375411748886, 0.062140509486198425, -0.06866546720266342, -0.13664314150810242, -0.023452885448932648, 0.08483598381280899, -0.011404541321098804, 0.028394777327775955, 0.07356005162000656, -0.07185159623622894, 0.20126941800117493, 0.03666449710726738, -0.05399559810757637, -0.054549336433410645, 0.0827551931142807, -0.09896446764469147, 0.27000707387924194, 0.015913790091872215, 0.048061735928058624, 0.1041264757514, -0.008932216092944145, -0.13759581744670868, 0.019727399572730064, 0.0954047441482544, -0.10358903557062149, 0.041838936507701874, 0.19829733669757843, -0.0014832824235782027, 0.1230277270078659, 0.07854447513818741, -0.07668869197368622, 0.0473078191280365, -0.08185897022485733, -0.06852826476097107, -0.0918748751282692, 0.10061057657003403, -0.07712632417678833, 0.14169210195541382, 0.13906599581241608, -0.05018797889351845, 0.011615060269832611, -0.031394075602293015, 0.04402702674269676, 0.0006254917825572193, 0.10420145094394684, 0.002576707163825631, -0.18477243185043335, 0.02472778968513012, 0.006634650751948357, 0.10846512019634247, -0.15925930440425873, -0.09642539173364639, 0.03936212509870529, 0.004935122560709715, -0.06595125794410706, 0.1294470727443695, 0.055943287909030914, 0.043614063411951065, -0.039108045399188995, -0.036952149122953415, -0.006302761845290661, 0.13504701852798462, -0.1053730770945549, 0.002390247769653797 ]
null
null
null
# Lora of hermann_kunne/ヘルマン・キュンネ/Z19 (Azur Lane) ## What Is This? This is the LoRA model of waifu hermann_kunne/ヘルマン・キュンネ/Z19 (Azur Lane). ## How Is It Trained? * This model is trained with [HCP-Diffusion](https://github.com/7eu7d7/HCP-Diffusion). * The [auto-training framework](https://github.com/deepghs/cyberharem) is maintained by [DeepGHS Team](https://huggingface.co/deepghs). * The base model used for training is [deepghs/animefull-latest](https://huggingface.co/deepghs/animefull-latest). * Dataset used for training is the `stage3-p480-800` in [CyberHarem/hermann_kunne_azurlane](https://huggingface.co/datasets/CyberHarem/hermann_kunne_azurlane), which contains 68 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 800 steps, 40 checkpoints were saved and evaluated. * **Trigger word is `hermann_kunne_azurlane`.** * Pruned core tags for this waifu are `black_hair, long_hair, hat, bangs`. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. ## How to Use It? ### If You Are Using A1111 WebUI v1.7+ **Just use it like the classic LoRA**. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 700, you need to download [`700/hermann_kunne_azurlane.pt`](https://huggingface.co/CyberHarem/hermann_kunne_azurlane/resolve/main/700/hermann_kunne_azurlane.pt) as the embedding and [`700/hermann_kunne_azurlane.safetensors`](https://huggingface.co/CyberHarem/hermann_kunne_azurlane/resolve/main/700/hermann_kunne_azurlane.safetensors) for loading Lora. By using both files together, you can generate images for the desired characters. ## Which Step Should I Use? We selected 5 good steps for you to choose. The best one is step 700. 1440 images (1.50 GiB) were generated for auto-testing. ![Metrics Plot](metrics_plot.png) The base model used for generating preview images is [Meina/MeinaMix_V11](https://huggingface.co/Meina/MeinaMix_V11). Here are the preview of the recommended steps: | Step | Epoch | CCIP | AI Corrupt | Bikini Plus | Score | Download | pattern_0 | portrait_0 | portrait_1 | portrait_2 | full_body_0 | full_body_1 | profile_0 | profile_1 | free_0 | free_1 | shorts | maid_0 | maid_1 | miko | yukata | suit | china | bikini_0 | bikini_1 | bikini_2 | sit | squat | kneel | jump | crossed_arms | angry | smile | cry | grin | n_lie_0 | n_lie_1 | n_stand_0 | n_stand_1 | n_stand_2 | n_sex_0 | n_sex_1 | |-------:|--------:|:----------|:-------------|:--------------|:----------|:-----------------------------------------------------------------------------------------------------------------|:-----------------------------------------|:-------------------------------------------|:-------------------------------------------|:-------------------------------------------|:---------------------------------------------|:---------------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-------------------------------|:-----------------------------------|:-------------------------------|:---------------------------------|:---------------------------------------|:---------------------------------------|:---------------------------------------|:-----------------------------|:---------------------------------|:---------------------------------|:-------------------------------|:-----------------------------------------------|:---------------------------------|:---------------------------------|:-----------------------------|:-------------------------------|:-------------------------------------|:-------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------------|:-------------------------------------|:-------------------------------------| | 700 | 42 | **0.858** | 0.982 | **0.854** | **0.819** | [Download](https://huggingface.co/CyberHarem/hermann_kunne_azurlane/resolve/main/700/hermann_kunne_azurlane.zip) | ![pattern_0](700/previews/pattern_0.png) | ![portrait_0](700/previews/portrait_0.png) | ![portrait_1](700/previews/portrait_1.png) | ![portrait_2](700/previews/portrait_2.png) | ![full_body_0](700/previews/full_body_0.png) | ![full_body_1](700/previews/full_body_1.png) | ![profile_0](700/previews/profile_0.png) | ![profile_1](700/previews/profile_1.png) | ![free_0](700/previews/free_0.png) | ![free_1](700/previews/free_1.png) | ![shorts](700/previews/shorts.png) | ![maid_0](700/previews/maid_0.png) | ![maid_1](700/previews/maid_1.png) | ![miko](700/previews/miko.png) | ![yukata](700/previews/yukata.png) | ![suit](700/previews/suit.png) | ![china](700/previews/china.png) | ![bikini_0](700/previews/bikini_0.png) | ![bikini_1](700/previews/bikini_1.png) | ![bikini_2](700/previews/bikini_2.png) | ![sit](700/previews/sit.png) | ![squat](700/previews/squat.png) | ![kneel](700/previews/kneel.png) | ![jump](700/previews/jump.png) | ![crossed_arms](700/previews/crossed_arms.png) | ![angry](700/previews/angry.png) | ![smile](700/previews/smile.png) | ![cry](700/previews/cry.png) | ![grin](700/previews/grin.png) | ![n_lie_0](700/previews/n_lie_0.png) | ![n_lie_1](700/previews/n_lie_1.png) | ![n_stand_0](700/previews/n_stand_0.png) | ![n_stand_1](700/previews/n_stand_1.png) | ![n_stand_2](700/previews/n_stand_2.png) | ![n_sex_0](700/previews/n_sex_0.png) | ![n_sex_1](700/previews/n_sex_1.png) | | 780 | 46 | 0.838 | **0.986** | 0.846 | 0.788 | [Download](https://huggingface.co/CyberHarem/hermann_kunne_azurlane/resolve/main/780/hermann_kunne_azurlane.zip) | ![pattern_0](780/previews/pattern_0.png) | ![portrait_0](780/previews/portrait_0.png) | ![portrait_1](780/previews/portrait_1.png) | ![portrait_2](780/previews/portrait_2.png) | ![full_body_0](780/previews/full_body_0.png) | ![full_body_1](780/previews/full_body_1.png) | ![profile_0](780/previews/profile_0.png) | ![profile_1](780/previews/profile_1.png) | ![free_0](780/previews/free_0.png) | ![free_1](780/previews/free_1.png) | ![shorts](780/previews/shorts.png) | ![maid_0](780/previews/maid_0.png) | ![maid_1](780/previews/maid_1.png) | ![miko](780/previews/miko.png) | ![yukata](780/previews/yukata.png) | ![suit](780/previews/suit.png) | ![china](780/previews/china.png) | ![bikini_0](780/previews/bikini_0.png) | ![bikini_1](780/previews/bikini_1.png) | ![bikini_2](780/previews/bikini_2.png) | ![sit](780/previews/sit.png) | ![squat](780/previews/squat.png) | ![kneel](780/previews/kneel.png) | ![jump](780/previews/jump.png) | ![crossed_arms](780/previews/crossed_arms.png) | ![angry](780/previews/angry.png) | ![smile](780/previews/smile.png) | ![cry](780/previews/cry.png) | ![grin](780/previews/grin.png) | ![n_lie_0](780/previews/n_lie_0.png) | ![n_lie_1](780/previews/n_lie_1.png) | ![n_stand_0](780/previews/n_stand_0.png) | ![n_stand_1](780/previews/n_stand_1.png) | ![n_stand_2](780/previews/n_stand_2.png) | ![n_sex_0](780/previews/n_sex_0.png) | ![n_sex_1](780/previews/n_sex_1.png) | | 620 | 37 | 0.850 | 0.962 | 0.836 | 0.774 | [Download](https://huggingface.co/CyberHarem/hermann_kunne_azurlane/resolve/main/620/hermann_kunne_azurlane.zip) | ![pattern_0](620/previews/pattern_0.png) | ![portrait_0](620/previews/portrait_0.png) | ![portrait_1](620/previews/portrait_1.png) | ![portrait_2](620/previews/portrait_2.png) | ![full_body_0](620/previews/full_body_0.png) | ![full_body_1](620/previews/full_body_1.png) | ![profile_0](620/previews/profile_0.png) | ![profile_1](620/previews/profile_1.png) | ![free_0](620/previews/free_0.png) | ![free_1](620/previews/free_1.png) | ![shorts](620/previews/shorts.png) | ![maid_0](620/previews/maid_0.png) | ![maid_1](620/previews/maid_1.png) | ![miko](620/previews/miko.png) | ![yukata](620/previews/yukata.png) | ![suit](620/previews/suit.png) | ![china](620/previews/china.png) | ![bikini_0](620/previews/bikini_0.png) | ![bikini_1](620/previews/bikini_1.png) | ![bikini_2](620/previews/bikini_2.png) | ![sit](620/previews/sit.png) | ![squat](620/previews/squat.png) | ![kneel](620/previews/kneel.png) | ![jump](620/previews/jump.png) | ![crossed_arms](620/previews/crossed_arms.png) | ![angry](620/previews/angry.png) | ![smile](620/previews/smile.png) | ![cry](620/previews/cry.png) | ![grin](620/previews/grin.png) | ![n_lie_0](620/previews/n_lie_0.png) | ![n_lie_1](620/previews/n_lie_1.png) | ![n_stand_0](620/previews/n_stand_0.png) | ![n_stand_1](620/previews/n_stand_1.png) | ![n_stand_2](620/previews/n_stand_2.png) | ![n_sex_0](620/previews/n_sex_0.png) | ![n_sex_1](620/previews/n_sex_1.png) | | 560 | 33 | 0.799 | 0.979 | 0.847 | 0.759 | [Download](https://huggingface.co/CyberHarem/hermann_kunne_azurlane/resolve/main/560/hermann_kunne_azurlane.zip) | ![pattern_0](560/previews/pattern_0.png) | ![portrait_0](560/previews/portrait_0.png) | ![portrait_1](560/previews/portrait_1.png) | ![portrait_2](560/previews/portrait_2.png) | ![full_body_0](560/previews/full_body_0.png) | ![full_body_1](560/previews/full_body_1.png) | ![profile_0](560/previews/profile_0.png) | ![profile_1](560/previews/profile_1.png) | ![free_0](560/previews/free_0.png) | ![free_1](560/previews/free_1.png) | ![shorts](560/previews/shorts.png) | ![maid_0](560/previews/maid_0.png) | ![maid_1](560/previews/maid_1.png) | ![miko](560/previews/miko.png) | ![yukata](560/previews/yukata.png) | ![suit](560/previews/suit.png) | ![china](560/previews/china.png) | ![bikini_0](560/previews/bikini_0.png) | ![bikini_1](560/previews/bikini_1.png) | ![bikini_2](560/previews/bikini_2.png) | ![sit](560/previews/sit.png) | ![squat](560/previews/squat.png) | ![kneel](560/previews/kneel.png) | ![jump](560/previews/jump.png) | ![crossed_arms](560/previews/crossed_arms.png) | ![angry](560/previews/angry.png) | ![smile](560/previews/smile.png) | ![cry](560/previews/cry.png) | ![grin](560/previews/grin.png) | ![n_lie_0](560/previews/n_lie_0.png) | ![n_lie_1](560/previews/n_lie_1.png) | ![n_stand_0](560/previews/n_stand_0.png) | ![n_stand_1](560/previews/n_stand_1.png) | ![n_stand_2](560/previews/n_stand_2.png) | ![n_sex_0](560/previews/n_sex_0.png) | ![n_sex_1](560/previews/n_sex_1.png) | | 740 | 44 | 0.822 | 0.951 | 0.833 | 0.744 | [Download](https://huggingface.co/CyberHarem/hermann_kunne_azurlane/resolve/main/740/hermann_kunne_azurlane.zip) | ![pattern_0](740/previews/pattern_0.png) | ![portrait_0](740/previews/portrait_0.png) | ![portrait_1](740/previews/portrait_1.png) | ![portrait_2](740/previews/portrait_2.png) | ![full_body_0](740/previews/full_body_0.png) | ![full_body_1](740/previews/full_body_1.png) | ![profile_0](740/previews/profile_0.png) | ![profile_1](740/previews/profile_1.png) | ![free_0](740/previews/free_0.png) | ![free_1](740/previews/free_1.png) | ![shorts](740/previews/shorts.png) | ![maid_0](740/previews/maid_0.png) | ![maid_1](740/previews/maid_1.png) | ![miko](740/previews/miko.png) | ![yukata](740/previews/yukata.png) | ![suit](740/previews/suit.png) | ![china](740/previews/china.png) | ![bikini_0](740/previews/bikini_0.png) | ![bikini_1](740/previews/bikini_1.png) | ![bikini_2](740/previews/bikini_2.png) | ![sit](740/previews/sit.png) | ![squat](740/previews/squat.png) | ![kneel](740/previews/kneel.png) | ![jump](740/previews/jump.png) | ![crossed_arms](740/previews/crossed_arms.png) | ![angry](740/previews/angry.png) | ![smile](740/previews/smile.png) | ![cry](740/previews/cry.png) | ![grin](740/previews/grin.png) | ![n_lie_0](740/previews/n_lie_0.png) | ![n_lie_1](740/previews/n_lie_1.png) | ![n_stand_0](740/previews/n_stand_0.png) | ![n_stand_1](740/previews/n_stand_1.png) | ![n_stand_2](740/previews/n_stand_2.png) | ![n_sex_0](740/previews/n_sex_0.png) | ![n_sex_1](740/previews/n_sex_1.png) | ## Anything Else? Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. ## All Steps We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * [Steps From 620 to 800](all/0.md) * [Steps From 420 to 600](all/1.md) * [Steps From 220 to 400](all/2.md) * [Steps From 20 to 200](all/3.md)
{"license": "mit", "tags": ["art", "not-for-all-audiences"], "datasets": ["CyberHarem/hermann_kunne_azurlane"], "pipeline_tag": "text-to-image"}
text-to-image
CyberHarem/hermann_kunne_azurlane
[ "art", "not-for-all-audiences", "text-to-image", "dataset:CyberHarem/hermann_kunne_azurlane", "license:mit", "region:us" ]
2024-02-15T01:44:44+00:00
[]
[]
TAGS #art #not-for-all-audiences #text-to-image #dataset-CyberHarem/hermann_kunne_azurlane #license-mit #region-us
Lora of hermann\_kunne/ヘルマン・キュンネ/Z19 (Azur Lane) ================================================ What Is This? ------------- This is the LoRA model of waifu hermann\_kunne/ヘルマン・キュンネ/Z19 (Azur Lane). How Is It Trained? ------------------ * This model is trained with HCP-Diffusion. * The auto-training framework is maintained by DeepGHS Team. * The base model used for training is deepghs/animefull-latest. * Dataset used for training is the 'stage3-p480-800' in CyberHarem/hermann\_kunne\_azurlane, which contains 68 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 800 steps, 40 checkpoints were saved and evaluated. * Trigger word is 'hermann\_kunne\_azurlane'. * Pruned core tags for this waifu are 'black\_hair, long\_hair, hat, bangs'. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. How to Use It? -------------- ### If You Are Using A1111 WebUI v1.7+ Just use it like the classic LoRA. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 700, you need to download '700/hermann\_kunne\_azurlane.pt' as the embedding and '700/hermann\_kunne\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters. Which Step Should I Use? ------------------------ We selected 5 good steps for you to choose. The best one is step 700. 1440 images (1.50 GiB) were generated for auto-testing. !Metrics Plot The base model used for generating preview images is Meina/MeinaMix\_V11. Here are the preview of the recommended steps: Anything Else? -------------- Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. All Steps --------- We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * Steps From 620 to 800 * Steps From 420 to 600 * Steps From 220 to 400 * Steps From 20 to 200
[ "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 700, you need to download '700/hermann\\_kunne\\_azurlane.pt' as the embedding and '700/hermann\\_kunne\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 700.\n\n\n1440 images (1.50 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 620 to 800\n* Steps From 420 to 600\n* Steps From 220 to 400\n* Steps From 20 to 200" ]
[ "TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/hermann_kunne_azurlane #license-mit #region-us \n", "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 700, you need to download '700/hermann\\_kunne\\_azurlane.pt' as the embedding and '700/hermann\\_kunne\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 700.\n\n\n1440 images (1.50 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 620 to 800\n* Steps From 420 to 600\n* Steps From 220 to 400\n* Steps From 20 to 200" ]
[ 47, 38, 473 ]
[ "passage: TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/hermann_kunne_azurlane #license-mit #region-us \n### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file." ]
[ 0.012291156686842442, -0.003746920032426715, -0.0034078112803399563, 0.0890384167432785, 0.06972639262676239, 0.06849721074104309, 0.23492613434791565, 0.07894853502511978, 0.14509499073028564, -0.07123377919197083, 0.09847396612167358, 0.06540002673864365, -0.01606076769530773, 0.021843455731868744, -0.03308902680873871, -0.15742379426956177, -0.056930188089609146, -0.025782357901334763, 0.007859259843826294, 0.009525056928396225, 0.07670311629772186, 0.006184544414281845, 0.10869637876749039, -0.04999382421374321, -0.02895388752222061, 0.050644584000110626, -0.027651995420455933, -0.035929836332798004, 0.03542588651180267, 0.07196124643087387, 0.11323728412389755, 0.014418192207813263, 0.06492871046066284, -0.1565932035446167, 0.06493302434682846, -0.009909635409712791, -0.10629967600107193, -0.006687026470899582, 0.01327525731176138, -0.036098312586545944, 0.1472652703523636, 0.020984675735235214, -0.10865424573421478, 0.03356911987066269, -0.1294587403535843, -0.02607450820505619, -0.044347185641527176, 0.04342588409781456, 0.1306077241897583, 0.06580864638090134, 0.023583032190799713, 0.0595572292804718, -0.047129712998867035, 0.087971031665802, 0.12817732989788055, -0.13350476324558258, -0.06711538881063461, 0.10930168628692627, 0.008905485272407532, 0.12500184774398804, -0.09765935689210892, 0.09809134900569916, 0.07634307444095612, -0.04727498069405556, -0.1460789144039154, -0.0987512394785881, -0.20876027643680573, -0.01201204489916563, 0.01029687374830246, 0.02056526206433773, 0.41102564334869385, 0.06719429045915604, 0.03899392858147621, 0.06685423105955124, -0.07366213947534561, 0.04066243767738342, -0.08317406475543976, 0.13046056032180786, 0.03927312046289444, 0.08430184423923492, -0.049874261021614075, -0.10215389728546143, -0.10912632942199707, -0.05855640023946762, -0.07087975740432739, -0.01532727386802435, 0.02195916324853897, 0.11570566147565842, -0.18839381635189056, 0.009440237656235695, -0.04964759200811386, -0.14236991107463837, 0.013532531447708607, -0.09783700853586197, 0.1773461401462555, 0.0696103572845459, -0.0064260344952344894, 0.0038961644750088453, 0.24405211210250854, 0.11862815171480179, 0.19013604521751404, 0.051909398287534714, -0.10354146361351013, 0.12507545948028564, 0.035467229783535004, -0.08709798753261566, -0.01054314523935318, -0.11083581298589706, 0.1339450478553772, -0.05262213200330734, 0.10845600068569183, -0.06394437700510025, -0.10997933149337769, 0.02084173448383808, -0.09335153549909592, 0.06313806027173996, 0.033667802810668945, 0.016472579911351204, -0.03960627689957619, 0.04476935416460037, 0.025908855721354485, -0.026134058833122253, -0.009789426811039448, -0.010843120515346527, -0.05230524390935898, 0.050202108919620514, 0.10406967997550964, 0.03569817543029785, 0.058716293424367905, 0.020764455199241638, -0.020188523456454277, 0.00486844964325428, -0.04570436477661133, 0.0065697873942554, 0.05226706340909004, 0.03608277812600136, 0.09111813455820084, -0.16782698035240173, -0.06736548990011215, -0.012274399399757385, 0.06088303402066231, 0.009576804004609585, 0.10031731426715851, -0.021139055490493774, 0.043388787657022476, 0.011024169623851776, -0.03094036877155304, 0.019173329696059227, -0.10939333587884903, 0.07363780587911606, -0.03455547243356705, 0.09157681465148926, -0.20482462644577026, -0.008497270755469799, -0.06020025908946991, 0.016092075034976006, 0.06768271327018738, -0.011075243353843689, -0.11030442267656326, 0.11205029487609863, -0.005673827137798071, 0.0654519572854042, -0.10376335680484772, 0.037829939275979996, 0.025021148845553398, 0.08338190615177155, -0.10143882036209106, 0.010193207301199436, 0.09215494990348816, -0.13773296773433685, -0.16458027064800262, 0.08391159772872925, -0.031915098428726196, 0.03611231595277786, 0.05259520187973976, 0.16951246559619904, 0.1779268980026245, -0.19816304743289948, -0.03242134302854538, 0.05964112654328346, -0.013604114763438702, -0.07851888984441757, -0.011917363852262497, 0.11152912676334381, 0.0018864281009882689, 0.0372985303401947, -0.04653327912092209, 0.10914042592048645, -0.026224764063954353, -0.08528263866901398, -0.03047148510813713, -0.07179675996303558, -0.07006102800369263, 0.05161423608660698, -0.005426543299108744, -0.058573462069034576, 0.020164959132671356, -0.14438924193382263, 0.1614341139793396, 0.007685406133532524, 0.028717348352074623, -0.057947710156440735, 0.11231344938278198, 0.011504951864480972, 0.004427031148225069, 0.006850259378552437, -0.07510019093751907, -0.10372909903526306, 0.22248288989067078, 0.08951693773269653, 0.11557505279779434, 0.0635417252779007, -0.04512343183159828, -0.07069370150566101, 0.01562411431223154, 0.0009434068924747407, -0.03541133925318718, 0.025448473170399666, -0.10661027580499649, 0.05168839916586876, -0.017073137685656548, 0.024509960785508156, -0.0021814636420458555, -0.03789101168513298, 0.07207758724689484, 0.022790394723415375, -0.011168565601110458, 0.08426328003406525, 0.03782620280981064, -0.01656152494251728, -0.06486067920923233, 0.004425338003784418, 0.08048655092716217, -0.014525511302053928, -0.06845346838235855, 0.014937391504645348, 0.009927807375788689, 0.029248755425214767, 0.19509945809841156, -0.21825826168060303, 0.04552562162280083, 0.019030116498470306, 0.04173891618847847, 0.03730613738298416, -0.003319747280329466, -0.030108816921710968, 0.04588966816663742, -0.028207845985889435, 0.059441398829221725, -0.018658936023712158, 0.06681989878416061, -0.02984524331986904, -0.13711808621883392, -0.01634187251329422, -0.013303040526807308, 0.16319771111011505, -0.16576595604419708, 0.06406041234731674, 0.18744589388370514, -0.12811703979969025, 0.14470088481903076, 0.00543685257434845, -0.007970000617206097, 0.008683524094522, 0.03901847079396248, -0.0031962196808308363, 0.11039240658283234, -0.09249764680862427, -0.021909428760409355, 0.0269231628626585, -0.07415829598903656, 0.031263455748558044, -0.11526276916265488, -0.10795190185308456, -0.07573018968105316, -0.03639020025730133, -0.02205469086766243, 0.02954452484846115, -0.05369826778769493, 0.07954192906618118, -0.0851740837097168, -0.09296772629022598, -0.022921063005924225, -0.08838162571191788, 0.023500951007008553, 0.015041339211165905, -0.04813646897673607, -0.13748010993003845, -0.12232650816440582, -0.08901172876358032, -0.1657264083623886, -0.007678444962948561, 0.06801160424947739, -0.10184689611196518, -0.051940735429525375, 0.005863219033926725, -0.04348857328295708, 0.10028807073831558, -0.07827585190534592, 0.014080413617193699, 0.0500212237238884, -0.03728418052196503, -0.17328116297721863, -0.007633950095623732, -0.06484590470790863, -0.04873761534690857, 0.14482584595680237, -0.13638776540756226, 0.18725775182247162, -0.04142236337065697, 0.050310809165239334, 0.05999568849802017, 0.03919263929128647, 0.13607007265090942, -0.10743879526853561, 0.06719060987234116, 0.1741786152124405, 0.03587309271097183, 0.07782840728759766, 0.12807302176952362, 0.0772821307182312, -0.10717969387769699, 0.03776148334145546, 0.07714438438415527, -0.1016213670372963, -0.0733373612165451, -0.059065647423267365, -0.10951132327318192, -0.03488713875412941, 0.06863710284233093, 0.05464741215109825, 0.04397730529308319, 0.12925665080547333, -0.057763487100601196, -0.011440221220254898, 0.0896696001291275, 0.051527757197618484, 0.06983613967895508, 0.015498622320592403, 0.05574984848499298, -0.14300879836082458, -0.0587930791079998, 0.1560894101858139, 0.2127230167388916, 0.2242249846458435, 0.016055775806307793, 0.054713163524866104, 0.1257137805223465, 0.09198303520679474, 0.08820798248052597, 0.04924144223332405, 0.0025478489696979523, 0.013577919453382492, -0.0702996626496315, -0.05988823249936104, 0.03141951188445091, 0.01115353312343359, -0.05620745196938515, -0.13844870030879974, 0.09808675199747086, -0.0028229004237800837, 0.08254153281450272, 0.13051775097846985, 0.060059841722249985, -0.10832562297582626, 0.15732957422733307, 0.09751251339912415, 0.08057212829589844, -0.06293907761573792, 0.13650228083133698, 0.05371217429637909, -0.020547015592455864, 0.16358180344104767, 0.027093136683106422, 0.14886566996574402, -0.043206050992012024, -0.07175970822572708, -0.07355764508247375, -0.05961419269442558, 0.012100983411073685, 0.04096740484237671, -0.21366089582443237, 0.10990328341722488, 0.046438366174697876, 0.00655394745990634, -0.01823233813047409, -0.06443661451339722, 0.17448405921459198, 0.14154523611068726, 0.08893796056509018, 0.02638963609933853, -0.045516811311244965, -0.008779584430158138, -0.08018200844526291, 0.05393474921584129, 0.020982254296541214, 0.0637827217578888, -0.03505280986428261, -0.09340415894985199, -0.019630659371614456, -0.0005728572141379118, 0.033093057572841644, -0.08041314035654068, -0.10987367480993271, -0.0492996908724308, 0.25508663058280945, -0.05910983681678772, 0.05519053712487221, 0.0485781766474247, 0.029228638857603073, -0.038926009088754654, 0.05510019510984421, -0.024467241019010544, -0.014983825385570526, -0.045906003564596176, 0.00001177459580503637, -0.0031611237209290266, -0.05742806941270828, -0.05882865935564041, -0.03680664300918579, -0.10391771048307419, -0.11117367446422577, 0.006010291166603565, -0.04110337421298027, 0.014067057520151138, -0.010153309442102909, 0.025149090215563774, -0.07994351536035538, -0.03275285288691521, 0.02117866277694702, 0.038522087037563324, -0.07798190414905548, -0.1339668482542038, -0.012627660296857357, -0.021992594003677368, -0.05455700308084488, 0.022511500865221024, -0.10649649053812027, -0.07782342284917831, -0.05897895246744156, -0.017236441373825073, 0.12243449687957764, 0.22884945571422577, -0.025814613327383995, 0.0028430605307221413, 0.1498338133096695, -0.0890055000782013, -0.3293430209159851, -0.14698277413845062, -0.16523891687393188, -0.10757101327180862, 0.029735349118709564, -0.058211568742990494, 0.05328226089477539, 0.08359060436487198, -0.03906066343188286, 0.2079939842224121, -0.18269772827625275, -0.10436288267374039, 0.08012571185827255, 0.10021474212408066, 0.32777103781700134, -0.2288505882024765, 0.013041168451309204, -0.11523302644491196, -0.056519925594329834, 0.019793083891272545, -0.09916858375072479, 0.11856129765510559, 0.03525589406490326, 0.06824663281440735, -0.007527212146669626, -0.009418155997991562, 0.14387473464012146, -0.05788194388151169, 0.14360009133815765, -0.12582054734230042, -0.09549357742071152, 0.19173896312713623, -0.03536482900381088, 0.020523222163319588, -0.21631044149398804, -0.030230509117245674, -0.02203114703297615, 0.03564296290278435, -0.018898215144872665, 0.07196331769227982, -0.007112293969839811, -0.028429251164197922, -0.14329084753990173, -0.013573896139860153, -0.031207801774144173, 0.05919418856501579, 0.2518880069255829, -0.05853872001171112, -0.06970474869012833, 0.03493918851017952, -0.0022205349523574114, 0.08909180760383606, 0.018228786066174507, -0.054483599960803986, -0.04948766157031059, 0.09573441743850708, -0.19785092771053314, 0.05294175073504448, 0.013776813633739948, -0.0038103959523141384, 0.01391022652387619, 0.009929115884006023, 0.02231549471616745, 0.11789502203464508, 0.17108915746212006, -0.02384704351425171, -0.04495924338698387, -0.02685699053108692, -0.006696767173707485, 0.13269443809986115, -0.007494466379284859, 0.12336544692516327, 0.009392283856868744, 0.04926110431551933, 0.009478956460952759, 0.0503760389983654, -0.08653893321752548, -0.0939977839589119, 0.09905029833316803, -0.0473245233297348, -0.08139318227767944, 0.09442596137523651, 0.04471476003527641, 0.06442990154027939, 0.00665435753762722, 0.04061055928468704, 0.007938424125313759, -0.12354370951652527, 0.003965970128774643, 0.2307519167661667, -0.08114311844110489, -0.06170322000980377, -0.07578965276479721, 0.01865093968808651, -0.12411388754844666, 0.06873854994773865, 0.04265693575143814, -0.030257217586040497, 0.11646486073732376, -0.043761011213064194, -0.0346079021692276, 0.01521964929997921, -0.05658663064241409, 0.04207305610179901, -0.1451244205236435, -0.20474877953529358, 0.0393182635307312, 0.0015420899726450443, -0.06241944059729576, -0.08772540092468262, -0.08836345374584198, 0.06168827414512634, -0.16143299639225006, 0.13253217935562134, -0.07489144057035446, 0.061049290001392365, -0.039567138999700546, -0.05188864842057228, -0.1076427772641182, -0.01361941359937191, -0.050830498337745667, -0.029061250388622284, 0.0576816201210022, 0.012407450005412102, -0.12432157248258591, -0.11524180322885513, 0.06447913497686386, -0.00496675306931138, -0.007333323359489441, 0.018599821254611015, -0.07131841778755188, 0.029102923348546028, -0.22483259439468384, -0.0663253515958786, 0.08771810680627823, 0.03948933258652687, -0.08000148087739944, 0.11597093194723129, 0.036210186779499054, -0.01851920783519745, 0.055187568068504333, 0.006029059179127216, 0.1646750569343567, -0.07385262101888657, 0.029330406337976456, -0.11945618689060211, -0.17799566686153412, -0.026980357244610786, 0.03381847217679024, 0.22568334639072418, 0.08522528409957886, 0.1373455673456192, -0.047418925911188126, 0.030142586678266525, -0.010418216697871685, 0.06949470192193985, 0.030586590990424156, -0.10920208692550659, -0.04424717277288437, -0.17249971628189087, -0.06587886065244675, -0.0695248395204544, 0.1516755074262619, 0.03606251999735832, -0.15215027332305908, -0.0032879174686968327, 0.10075408220291138, -0.16783441603183746, -0.010598273947834969, 0.17082248628139496, -0.04867987334728241, 0.02539229206740856, -0.16256959736347198, 0.03822553530335426, 0.07562730461359024, 0.0028078313916921616, 0.02138340286910534, 0.13658876717090607, -0.0002550820936448872, -0.002943232888355851, 0.04224032908678055, -0.019130660220980644, 0.06585104763507843, -0.0770287811756134, 0.06848125904798508, 0.005848011467605829, -0.048745207488536835, -0.11373285949230194, 0.1915249526500702, -0.02772369794547558, 0.011646909639239311, -0.048444077372550964, -0.002778133377432823, -0.10130880028009415, -0.11794594675302505, -0.08041505515575409, -0.1388130784034729, 0.08183538168668747, -0.06454929709434509, 0.0065811919048428535, 0.004471060819923878, 0.017891699448227882, -0.08121603727340698, 0.03281882777810097, -0.1898227035999298, -0.04601322486996651, 0.021920448169112206, -0.015467735007405281, -0.03101595677435398, -0.04082470014691353, -0.03458334505558014, 0.012523477897047997, -0.07255135476589203, -0.0755910649895668, 0.05370328575372696, 0.08830767869949341, 0.0459178164601326, -0.15584750473499298, -0.1091223657131195, -0.07091308385133743, 0.040964122861623764, 0.07856139540672302, 0.1918373703956604, 0.04142990708351135, -0.007814212702214718, 0.042996518313884735, 0.12714973092079163, 0.013673229143023491, -0.08330101519823074, -0.07382294535636902, -0.12811337411403656, -0.14030121266841888, -0.012363934889435768, -0.06265787035226822, -0.025271598249673843, 0.027876505628228188, 0.22584395110607147, 0.19783049821853638, -0.14133764803409576, 0.045628875494003296, -0.07863029837608337, 0.04235062375664711, -0.02975456602871418, 0.15768471360206604, 0.042111918330192566, 0.15423402190208435, -0.032396722584962845, -0.04785889759659767, -0.05756962299346924, 0.019093239679932594, -0.09899184107780457, 0.03605484962463379, -0.01658215932548046, -0.08133199065923691, -0.07484624534845352, 0.0983295738697052, -0.11768888682126999, 0.06321835517883301, 0.18466554582118988, -0.1398426741361618, -0.016972627490758896, -0.036960091441869736, 0.049714092165231705, 0.10351917892694473, 0.02754610776901245, -0.0793742910027504, -0.022963210940361023, 0.0020027540158480406, 0.029891764745116234, -0.17578929662704468, -0.1155494973063469, -0.0027088129427284002, -0.10384891927242279, 0.13110075891017914, -0.009498952887952328, 0.006045997608453035, 0.04035788029432297, -0.06467707455158234, -0.0047409492544829845, 0.17010259628295898, 0.02665841206908226, -0.03623194992542267, -0.03529456630349159, -0.062283676117658615, -0.10658619552850723, 0.08482838422060013, 0.09616433829069138, 0.04427092894911766, -0.008906448259949684, 0.1901608109474182, -0.016645947471261024, -0.05519014596939087, 0.13198022544384003, -0.1728641390800476, 0.0959412008523941, 0.005728099960833788, -0.011936472728848457, -0.07483912259340286, -0.04554193466901779, 0.0417434498667717, 0.07428132742643356, -0.18268442153930664, -0.04536203294992447, 0.06313684582710266, -0.09715813398361206, 0.06322001665830612, 0.03933476284146309, -0.103919617831707, 0.010328885167837143, -0.11091068387031555, -0.009971444495022297, -0.0902109444141388, 0.044166140258312225, 0.18730804324150085, -0.04501074180006981, 0.015408197417855263, -0.1297532320022583, 0.05941782146692276, -0.0325593538582325, -0.035462651401758194, -0.08492711186408997 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-large_robust_stream_speaker_s2 This model was trained from scratch on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 12 - total_train_batch_size: 96 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 50 - num_epochs: 20 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.36.2 - Pytorch 1.13.1+cu117 - Datasets 2.16.1 - Tokenizers 0.15.0
{"tags": ["generated_from_trainer"], "model-index": [{"name": "wav2vec2-large_robust_stream_speaker_s2", "results": []}]}
null
apirbadian/wav2vec2-large_robust_stream_speaker_s2
[ "transformers", "pytorch", "tensorboard", "safetensors", "wav2vec2", "generated_from_trainer", "endpoints_compatible", "region:us" ]
2024-02-15T01:49:07+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #safetensors #wav2vec2 #generated_from_trainer #endpoints_compatible #region-us
# wav2vec2-large_robust_stream_speaker_s2 This model was trained from scratch on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 12 - total_train_batch_size: 96 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 50 - num_epochs: 20 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.36.2 - Pytorch 1.13.1+cu117 - Datasets 2.16.1 - Tokenizers 0.15.0
[ "# wav2vec2-large_robust_stream_speaker_s2\n\nThis model was trained from scratch on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 12\n- total_train_batch_size: 96\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 50\n- num_epochs: 20\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.36.2\n- Pytorch 1.13.1+cu117\n- Datasets 2.16.1\n- Tokenizers 0.15.0" ]
[ "TAGS\n#transformers #pytorch #tensorboard #safetensors #wav2vec2 #generated_from_trainer #endpoints_compatible #region-us \n", "# wav2vec2-large_robust_stream_speaker_s2\n\nThis model was trained from scratch on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 12\n- total_train_batch_size: 96\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 50\n- num_epochs: 20\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.36.2\n- Pytorch 1.13.1+cu117\n- Datasets 2.16.1\n- Tokenizers 0.15.0" ]
[ 43, 35, 6, 12, 8, 3, 141, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #safetensors #wav2vec2 #generated_from_trainer #endpoints_compatible #region-us \n# wav2vec2-large_robust_stream_speaker_s2\n\nThis model was trained from scratch on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 12\n- total_train_batch_size: 96\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 50\n- num_epochs: 20\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.36.2\n- Pytorch 1.13.1+cu117\n- Datasets 2.16.1\n- Tokenizers 0.15.0" ]
[ -0.11438428610563278, 0.15456737577915192, -0.003222317900508642, 0.0713658332824707, 0.12639057636260986, 0.02950625866651535, 0.10691829025745392, 0.12076784670352936, -0.028479881584644318, 0.1076040267944336, 0.10019237548112869, 0.036153536289930344, 0.05960060656070709, 0.15346397459506989, -0.016522753983736038, -0.23918417096138, 0.025014355778694153, -0.04113146290183067, -0.11296404898166656, 0.09276082366704941, 0.08413094282150269, -0.10453387349843979, 0.058948636054992676, 0.007803452201187611, -0.11570743471384048, -0.00619992520660162, -0.048312027007341385, -0.05154317244887352, 0.08282860368490219, -0.0016060960479080677, 0.08409896492958069, 0.01579863205552101, 0.11971601098775864, -0.20421016216278076, 0.00235400116071105, 0.08169568330049515, 0.04573828727006912, 0.09963111579418182, 0.10258990526199341, 0.0026165961753576994, 0.05505843460559845, -0.16879042983055115, 0.09582289308309555, 0.02667771652340889, -0.07030586153268814, -0.1401423066854477, -0.08313523977994919, 0.0712742805480957, 0.10207988321781158, 0.07487351447343826, 0.009446242824196815, 0.11035293340682983, -0.0830884650349617, 0.05971091613173485, 0.21426254510879517, -0.26787760853767395, -0.05902205407619476, 0.021966055035591125, 0.05780372768640518, 0.043422263115644455, -0.12099234759807587, -0.018826056271791458, 0.025445178151130676, 0.007790861651301384, 0.09574462473392487, 0.021754590794444084, -0.01479675155133009, 0.013251631520688534, -0.13509847223758698, -0.01326081994920969, 0.07486991584300995, 0.04558517038822174, -0.03334137424826622, -0.1585128754377365, -0.06429189443588257, -0.15624938905239105, -0.01786159910261631, -0.011371888220310211, 0.009980457834899426, -0.0562748946249485, -0.05510884150862694, -0.010198996402323246, -0.0422079935669899, -0.08325225114822388, 0.02530398778617382, 0.14354221522808075, 0.05390497297048569, 0.007360804360359907, 0.000657576834782958, 0.12368607521057129, 0.05468197539448738, -0.1200360506772995, -0.005005670711398125, -0.0208301804959774, -0.15315771102905273, -0.029679784551262856, -0.016021093353629112, 0.015210860408842564, 0.00597778894007206, 0.11962827295064926, -0.0018303850665688515, 0.07633675634860992, 0.0559900626540184, -0.004222538787871599, -0.020245539024472237, 0.13631249964237213, -0.09501451253890991, -0.04209938645362854, -0.031020348891615868, 0.11397267132997513, -0.004140172153711319, -0.014288816601037979, -0.06083345040678978, -0.038486726582050323, 0.0794561579823494, 0.08745629340410233, -0.034030620008707047, 0.024703193455934525, -0.04877925664186478, -0.025680668652057648, 0.036883946508169174, -0.1302817463874817, 0.049728598445653915, 0.015237992629408836, -0.08579793572425842, 0.012796848081052303, 0.01266059186309576, -0.02449146658182144, -0.04167106747627258, 0.09657689183950424, -0.08505558222532272, -0.029735222458839417, -0.07510755956172943, -0.05674665421247482, 0.009066859260201454, -0.0986161157488823, -0.015489722602069378, -0.0373021699488163, -0.15828919410705566, -0.029587453231215477, 0.03853355720639229, -0.07850363850593567, -0.04515860974788666, -0.02992565929889679, -0.054849524050951004, 0.04332011193037033, -0.0011110976338386536, 0.14872579276561737, -0.0407063290476799, 0.06410268694162369, -0.042740605771541595, 0.049654360860586166, 0.09112287312746048, 0.03985020890831947, -0.0680876299738884, 0.06446485966444016, -0.10251130908727646, 0.07890856266021729, -0.10165029019117355, 0.016783544793725014, -0.13951756060123444, -0.0872240737080574, 0.01529486570507288, -0.026700302958488464, 0.07384241372346878, 0.1445728987455368, -0.16835954785346985, -0.0075461831875145435, 0.14672429859638214, -0.04980012774467468, -0.06010181084275246, 0.08634092658758163, -0.033866044133901596, -0.029492966830730438, 0.02502928301692009, 0.10659734904766083, 0.12290951609611511, -0.14651337265968323, -0.0018621293129399419, 0.02127206139266491, 0.057313472032547, 0.021029984578490257, 0.06579641252756119, -0.009970759972929955, 0.009527981281280518, 0.009107508696615696, -0.054970044642686844, 0.016782520338892937, -0.08207724243402481, -0.07226529717445374, -0.05297287181019783, -0.089707151055336, 0.0469491146504879, -0.012677939608693123, 0.014602772891521454, -0.04079398512840271, -0.136494979262352, 0.03407760709524155, 0.14603373408317566, -0.06779557466506958, -0.007570638321340084, -0.0739152580499649, 0.011283266358077526, -0.03339676931500435, -0.021156130358576775, -0.17196343839168549, -0.13099637627601624, 0.04203573614358902, -0.10888884961605072, -0.001266272971406579, 0.013363833539187908, 0.07200923562049866, 0.07288390398025513, -0.018334921449422836, -0.049340710043907166, -0.05185869708657265, -0.006450493820011616, -0.08640589565038681, -0.18498799204826355, -0.0696377158164978, -0.032026905566453934, 0.13765887916088104, -0.21991965174674988, -0.005773815326392651, -0.022528158500790596, 0.14764274656772614, 0.04153423011302948, -0.06181703880429268, 0.033200375735759735, 0.02125737816095352, 0.002416694536805153, -0.1141280084848404, 0.019215639680624008, -0.014168967492878437, -0.10206878930330276, 0.00337866204790771, -0.14203545451164246, 0.026662567630410194, 0.051481328904628754, 0.1417611688375473, -0.10183274000883102, -0.05429608374834061, -0.0707416906952858, -0.06510478258132935, -0.06454276293516159, 0.010137569159269333, 0.14902468025684357, 0.04646933078765869, 0.11297190189361572, -0.05388839170336723, -0.08080584555864334, 0.013875149190425873, 0.03438520058989525, -0.021394651383161545, 0.11076770722866058, 0.07179320603609085, -0.06889313459396362, 0.06952007859945297, 0.0880308672785759, -0.0263811182230711, 0.12449139356613159, -0.050535570830106735, -0.11093205213546753, -0.022891975939273834, 0.021306591108441353, 0.003732850542291999, 0.1254493147134781, -0.07056097686290741, 0.01811300218105316, 0.046996086835861206, 0.02098335139453411, 0.022352561354637146, -0.15477290749549866, -0.011314711533486843, 0.025724302977323532, -0.0389229878783226, -0.011539163067936897, 0.002658551326021552, 0.04085033759474754, 0.07773315906524658, 0.017847638577222824, -0.02772892266511917, 0.01149824634194374, -0.021517928689718246, -0.06445231288671494, 0.18166007101535797, -0.11041149497032166, -0.1243620216846466, -0.12491243332624435, 0.02777092345058918, -0.03289080038666725, -0.040282391011714935, -0.020226741209626198, -0.06612355262041092, -0.07298847287893295, -0.08466833084821701, -0.008953563868999481, -0.03325895220041275, 0.022819122299551964, 0.049377527087926865, 0.011360127478837967, 0.07938802987337112, -0.1179359182715416, 0.029774971306324005, -0.0055576348677277565, -0.051745738834142685, 0.004316581878811121, 0.07386938482522964, 0.07701658457517624, 0.1354074627161026, 0.007620191667228937, 0.016462570056319237, -0.012527602724730968, 0.2124505639076233, -0.08186979591846466, 0.004031686577945948, 0.09078918397426605, -0.006780172698199749, 0.06507730484008789, 0.11631881445646286, 0.034698545932769775, -0.08112287521362305, 0.02987419255077839, 0.06036698445677757, -0.014449688605964184, -0.25340837240219116, -0.04128853231668472, -0.025121405720710754, -0.04009576514363289, 0.13089267909526825, 0.05524785444140434, -0.05605282634496689, 0.05350295081734657, -0.02128877118229866, 0.013968272134661674, -0.022084981203079224, 0.0681353211402893, 0.022137710824608803, 0.048973772674798965, 0.08233027905225754, -0.016529135406017303, -0.03469438478350639, 0.04518460854887962, 0.020753221586346626, 0.27672281861305237, -0.04041874408721924, 0.15563331544399261, 0.005853472743183374, 0.13721352815628052, -0.002432603621855378, 0.07920613139867783, 0.022185655310750008, -0.00293294177390635, 0.019279256463050842, -0.07115097343921661, 0.0013671285705640912, 0.029382528737187386, 0.0075460197404026985, 0.034735701978206635, -0.09076845645904541, 0.07001928240060806, 0.0014482580590993166, 0.28407904505729675, 0.05969475582242012, -0.2932240068912506, -0.05659002810716629, 0.002717374125495553, -0.04238323122262955, -0.09702470898628235, 0.019818177446722984, 0.12810006737709045, -0.12184761464595795, 0.0457967109978199, -0.0806339755654335, 0.0810244008898735, -0.05505076423287392, -0.02281174249947071, 0.07167140394449234, 0.14766435325145721, -0.011841028928756714, 0.08833114802837372, -0.15146249532699585, 0.20708556473255157, 0.00877284910529852, 0.08060437440872192, -0.07365840673446655, 0.02806880883872509, -0.00005480749678099528, 0.01671433448791504, 0.12247413396835327, -0.0011882263934239745, -0.0654059424996376, -0.15591268241405487, -0.15224751830101013, 0.0129640968516469, 0.13304032385349274, -0.040388889610767365, 0.09542013704776764, -0.05459759011864662, -0.028287870809435844, 0.03799299895763397, -0.11808716505765915, -0.1359088122844696, -0.15066514909267426, 0.04064606502652168, 0.0034553539007902145, -0.007282060571014881, -0.08948401361703873, -0.10836686939001083, -0.026640545576810837, 0.16205114126205444, -0.04957637935876846, -0.05100400745868683, -0.1488027721643448, 0.020525718107819557, 0.19944483041763306, -0.05189864709973335, 0.03593242168426514, 0.006761929020285606, 0.15842017531394958, 0.03196658194065094, -0.08089668303728104, 0.07316574454307556, -0.08007989078760147, -0.19792886078357697, -0.06022263318300247, 0.14436252415180206, 0.06784713268280029, 0.034570179879665375, -0.01686733588576317, 0.011940229684114456, -0.006608901545405388, -0.09286841005086899, 0.030275937169790268, 0.09060940891504288, 0.027998171746730804, 0.0718468502163887, -0.04473923146724701, 0.056074827909469604, -0.0244932658970356, -0.013369472697377205, 0.09862041473388672, 0.22126223146915436, -0.09695113450288773, 0.10794518887996674, 0.056130364537239075, -0.05124802142381668, -0.1786218136548996, 0.02948305942118168, 0.12105102837085724, 0.03763893246650696, 0.03754222020506859, -0.18021677434444427, 0.10362472385168076, 0.09437016397714615, -0.017810378223657608, 0.08728604763746262, -0.339937299489975, -0.12729868292808533, 0.09179573506116867, 0.07666525989770889, 0.005271339789032936, -0.11374302208423615, -0.0470280684530735, -0.01671382412314415, -0.0945051833987236, 0.10786110162734985, -0.029964232817292213, 0.10823490470647812, -0.028019702062010765, 0.08147458732128143, 0.03945493325591087, -0.0581415556371212, 0.15415635704994202, 0.02216963842511177, 0.07503671199083328, -0.03181828185915947, 0.035287465900182724, 0.02996021881699562, -0.08574288338422775, 0.04003871977329254, -0.058361973613500595, 0.07338359206914902, -0.141632080078125, -0.017168613150715828, -0.05943458899855614, 0.04980472847819328, -0.045752350240945816, -0.05110766738653183, -0.03159250691533089, 0.050542738288640976, 0.06837309151887894, -0.032235872000455856, 0.056860752403736115, 0.04116462171077728, 0.047448769211769104, 0.09089136868715286, 0.0745648443698883, 0.014994361437857151, -0.12974004447460175, -0.0007893700385466218, -0.008115830831229687, 0.056735530495643616, -0.08986631035804749, 0.022793233394622803, 0.13026726245880127, 0.05081512778997421, 0.1029755100607872, 0.03563304618000984, -0.07672514021396637, -0.027429305016994476, 0.03936445713043213, -0.12074501067399979, -0.14594173431396484, -0.004846304189413786, 0.014048621989786625, -0.14103974401950836, 0.003854454029351473, 0.10025116801261902, -0.047580599784851074, -0.009943148121237755, -0.006245247554033995, 0.03037775307893753, 0.010910416021943092, 0.19448667764663696, 0.039258770644664764, 0.088833287358284, -0.0865059643983841, 0.11799678206443787, 0.06600336730480194, -0.08831063657999039, 0.042455971240997314, 0.06668635457754135, -0.08060889691114426, 0.0015055927215144038, 0.07204807549715042, 0.10321062803268433, -0.009636293165385723, -0.03072735294699669, -0.04994937777519226, -0.10656122118234634, 0.059975091367959976, 0.07052555680274963, 0.022066602483391762, 0.01618056930601597, -0.02329268865287304, 0.011610174551606178, -0.13653163611888885, 0.1071200966835022, 0.08133768290281296, 0.0698157474398613, -0.1612655222415924, 0.09939208626747131, -0.011853331699967384, 0.042466290295124054, -0.014244768768548965, 0.024094372987747192, -0.096869558095932, -0.024654073640704155, -0.1285267472267151, 0.008569616824388504, -0.04208831116557121, 0.0014103951398283243, -0.011856804601848125, -0.030499758198857307, -0.022278249263763428, 0.03027745895087719, -0.06404192000627518, -0.0608600415289402, -0.010023658163845539, 0.038415439426898956, -0.12541969120502472, -0.007103609386831522, 0.02641225978732109, -0.1232827678322792, 0.09005162119865417, 0.049834154546260834, 0.031209783628582954, 0.011620607227087021, -0.10676280409097672, -0.005442080553621054, 0.013734606094658375, 0.01132921501994133, 0.06539437174797058, -0.12075250595808029, -0.010929693467915058, -0.05918237194418907, 0.008198296651244164, -0.008939297869801521, 0.008685899898409843, -0.13679902255535126, -0.040104351937770844, -0.048270244151353836, -0.015424782410264015, -0.06367340683937073, 0.03956333175301552, 0.08418411016464233, 0.025968194007873535, 0.14284566044807434, -0.061341721564531326, 0.0605202242732048, -0.17803804576396942, -0.03078094683587551, -0.00894281454384327, -0.007765406742691994, -0.014925246126949787, 0.002598324790596962, 0.09167839586734772, -0.0532853789627552, 0.13610556721687317, -0.05114482343196869, 0.05103490129113197, 0.017864271998405457, -0.055019572377204895, -0.023853087797760963, 0.044342584908008575, 0.17088031768798828, 0.05434791371226311, -0.004332308657467365, 0.06986453384160995, -0.02664950117468834, 0.05281246453523636, 0.07054206728935242, 0.16619336605072021, 0.12891785800457, -0.013223553076386452, 0.0604412816464901, 0.06392139941453934, -0.14471301436424255, -0.14532588422298431, 0.12146856635808945, -0.05734924599528313, 0.1185038834810257, -0.055150724947452545, 0.1747611165046692, 0.07420559227466583, -0.1555631160736084, 0.035420265048742294, -0.03765324503183365, -0.11288338899612427, -0.1160544827580452, -0.0673723965883255, -0.0677112340927124, -0.11390909552574158, 0.020469140261411667, -0.09621817618608475, 0.0412067212164402, 0.06969374418258667, 0.035028718411922455, 0.027677681297063828, 0.15427976846694946, 0.006877178326249123, 0.01805948093533516, 0.06938920170068741, 0.055465880781412125, 0.004038501530885696, -0.004937176126986742, -0.06288469582796097, 0.04347294196486473, -0.013791080564260483, 0.06846370548009872, -0.04783687740564346, -0.007755483500659466, 0.060701608657836914, 0.00935895275324583, -0.07231287658214569, 0.018115414306521416, -0.004070221446454525, 0.01882237382233143, 0.04198558256030083, 0.05747875198721886, 0.006773913279175758, -0.07420400530099869, 0.2698356509208679, -0.08496079593896866, -0.05093250051140785, -0.12647348642349243, 0.16806036233901978, 0.014036827720701694, -0.0009056783746927977, 0.0704641044139862, -0.09661737084388733, -0.029444092884659767, 0.14483942091464996, 0.13402315974235535, -0.05880577489733696, -0.015859834849834442, -0.010833614505827427, -0.01774195209145546, -0.05488836392760277, 0.12779326736927032, 0.09682590514421463, 0.05396418645977974, -0.05234416946768761, -0.0016404403140768409, -0.015157231129705906, -0.04786693677306175, -0.09785380214452744, 0.07332777231931686, 0.006637228187173605, 0.010000882670283318, -0.041213709861040115, 0.06609903275966644, -0.026740865781903267, -0.17932169139385223, 0.0281570665538311, -0.12354522943496704, -0.19379065930843353, -0.048745181411504745, 0.061329808086156845, -0.009932998567819595, 0.05196377635002136, -0.005197447724640369, -0.01962941512465477, 0.14864599704742432, 0.0014656786806881428, -0.043136946856975555, -0.08240525424480438, 0.05990519002079964, -0.032382551580667496, 0.20754100382328033, 0.019625237211585045, 0.04749860614538193, 0.09035015851259232, 0.012429948896169662, -0.17900139093399048, 0.0039482577703893185, 0.08661308139562607, -0.09321341663599014, 0.04004286229610443, 0.1863265037536621, -0.03800394386053085, 0.06822311878204346, 0.02235768735408783, -0.09325271844863892, -0.02677239663898945, -0.05748370662331581, -0.000004137692940275883, -0.07594464719295502, 0.0044267307966947556, -0.026432927697896957, 0.1642974317073822, 0.21617929637432098, -0.0522853322327137, -0.003389514284208417, -0.05408094823360443, 0.020268145948648453, 0.022695891559123993, 0.08623094111680984, -0.0005735765444114804, -0.1972970962524414, 0.028761163353919983, 0.028506312519311905, 0.057238705456256866, -0.2068047672510147, -0.07527538388967514, 0.019682960584759712, -0.05770203098654747, -0.05238508805632591, 0.12463798373937607, 0.042343173176050186, 0.037112362682819366, -0.03388996422290802, -0.0653253048658371, -0.03674507141113281, 0.1304062455892563, -0.1649680733680725, -0.041626375168561935 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
devashat/244-test
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T01:54:17+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 57, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.05622259899973869, 0.16002345085144043, -0.004987028427422047, 0.023115945979952812, 0.0962471067905426, 0.011845538392663002, 0.06785304099321365, 0.11496778577566147, -0.020396295934915543, 0.11142492294311523, 0.03292480856180191, 0.0972127765417099, 0.11474913358688354, 0.16215258836746216, 0.004439093638211489, -0.23455148935317993, 0.04782992601394653, -0.12695099413394928, -0.033447545021772385, 0.11785799264907837, 0.14491069316864014, -0.10402194410562515, 0.07766910642385483, -0.030544815585017204, -0.009361269883811474, -0.03290390968322754, -0.06365230679512024, -0.05152205005288124, 0.05037128925323486, 0.06932847946882248, 0.06591591984033585, 0.007509593386203051, 0.09122733771800995, -0.2655104100704193, 0.02280162274837494, 0.07630051672458649, -0.0015554219717159867, 0.07497020810842514, 0.048351652920246124, -0.08209776133298874, 0.0788840726017952, -0.05696587264537811, 0.14718368649482727, 0.08216129243373871, -0.08924587815999985, -0.1965435892343521, -0.08464295417070389, 0.10284840315580368, 0.18357418477535248, 0.05158785358071327, -0.024141347035765648, 0.10476154088973999, -0.08419200032949448, 0.008797040209174156, 0.06024181470274925, -0.06443428993225098, -0.05412506312131882, 0.06934051215648651, 0.07975570857524872, 0.07967228442430496, -0.13025140762329102, -0.014651902951300144, 0.011243549175560474, 0.007594773545861244, 0.08504551649093628, 0.022028017789125443, 0.14595499634742737, 0.04393624886870384, -0.13030564785003662, -0.044304780662059784, 0.09771761298179626, 0.04345165938138962, -0.053857799619436264, -0.2537047266960144, -0.024983759969472885, -0.03927002474665642, -0.03094942681491375, -0.038562554866075516, 0.04431856796145439, -0.011080716736614704, 0.08032315224409103, -0.01118796318769455, -0.08149448037147522, -0.041395120322704315, 0.06544242054224014, 0.062143467366695404, 0.026896316558122635, -0.01158317644149065, 0.00973866879940033, 0.1224486380815506, 0.10907839238643646, -0.12763150036334991, -0.05768941715359688, -0.06755511462688446, -0.08307720720767975, -0.04300352931022644, 0.03337155282497406, 0.044020529836416245, 0.04436098039150238, 0.2466370165348053, 0.01108562108129263, 0.05453123152256012, 0.045806169509887695, 0.010608446784317493, 0.06787561625242233, 0.11606968939304352, -0.062306761741638184, -0.09178462624549866, -0.029058339074254036, 0.09215214103460312, 0.006741520017385483, -0.042814407497644424, -0.060904473066329956, 0.06479041278362274, 0.012608112767338753, 0.12110785394906998, 0.08444269746541977, 0.0026690615341067314, -0.07305197417736053, -0.06963318586349487, 0.18848419189453125, -0.1598394364118576, 0.047875016927719116, 0.031182926148176193, -0.038971830159425735, -0.0014042917173355818, 0.008752269670367241, 0.02394084818661213, -0.020246321335434914, 0.08923295140266418, -0.05574449151754379, -0.03784004598855972, -0.11079790443181992, -0.03252100944519043, 0.030985163524746895, 0.0051483530551195145, -0.027043871581554413, -0.033837489783763885, -0.09040277451276779, -0.059588029980659485, 0.0922931432723999, -0.07471107691526413, -0.04984431713819504, -0.013726521283388138, -0.07691634446382523, 0.023329194635152817, 0.016799474135041237, 0.08357251435518265, -0.02157396264374256, 0.0384126678109169, -0.0560205839574337, 0.0631464347243309, 0.11269522458314896, 0.029363946989178658, -0.053069718182086945, 0.05750001594424248, -0.24315528571605682, 0.10326608270406723, -0.07320205867290497, 0.050549428910017014, -0.15059062838554382, -0.026000602170825005, 0.044471126049757004, 0.00805877335369587, -0.013138634152710438, 0.14088952541351318, -0.21621745824813843, -0.0323486253619194, 0.16741067171096802, -0.0939871072769165, -0.07602590322494507, 0.059108685702085495, -0.05233629792928696, 0.10869261622428894, 0.04351044446229935, -0.02232111617922783, 0.060673557221889496, -0.14475463330745697, -0.01067100279033184, -0.04139741137623787, -0.02402937039732933, 0.16397778689861298, 0.07567544281482697, -0.06286642700433731, 0.08052356541156769, 0.024165838956832886, -0.017831770703196526, -0.04484899342060089, -0.023361295461654663, -0.10819391161203384, 0.009856974706053734, -0.06032416597008705, 0.02424289658665657, -0.025761527940630913, -0.09367526322603226, -0.02868773601949215, -0.1802000105381012, -0.009223134256899357, 0.0881323292851448, -0.011722641065716743, -0.021903391927480698, -0.12039245665073395, 0.011948852799832821, 0.031212422996759415, 0.002984174294397235, -0.13029038906097412, -0.05838731303811073, 0.027675874531269073, -0.16422230005264282, 0.03272955119609833, -0.05597274377942085, 0.05056252330541611, 0.03445037454366684, -0.03187771514058113, -0.033117350190877914, 0.009550533257424831, 0.006354342680424452, -0.010578392073512077, -0.2502359449863434, -0.02440580166876316, -0.0219739843159914, 0.17386503517627716, -0.21793730556964874, 0.04213962331414223, 0.07686693966388702, 0.14929872751235962, 0.006240781396627426, -0.038500864058732986, 0.010139784775674343, -0.08222103863954544, -0.030560437589883804, -0.0643099993467331, -0.012082485482096672, -0.03717579320073128, -0.05608142167329788, 0.05165567249059677, -0.16133594512939453, -0.028727244585752487, 0.1057019829750061, 0.06860516220331192, -0.14001330733299255, -0.019125886261463165, -0.04171464592218399, -0.043496038764715195, -0.05877087265253067, -0.0552728995680809, 0.1185101792216301, 0.05596614256501198, 0.04696191847324371, -0.06956122815608978, -0.07775315642356873, 0.007865429855883121, -0.017090093344449997, -0.017978519201278687, 0.08920905739068985, 0.07311701774597168, -0.12023317068815231, 0.09247473627328873, 0.10194233059883118, 0.09365488588809967, 0.108615942299366, -0.017981963232159615, -0.08929306268692017, -0.04584396257996559, 0.02045595459640026, 0.013332244008779526, 0.14797501266002655, -0.01403066236525774, 0.056954506784677505, 0.03922648727893829, -0.01123172789812088, 0.012020308524370193, -0.09384570270776749, 0.027314940467476845, 0.034342724829912186, -0.020308034494519234, 0.03796098753809929, -0.04001156985759735, 0.019826533272862434, 0.08712323755025864, 0.04676510766148567, 0.04415108636021614, 0.011758276261389256, -0.04233846068382263, -0.10904491692781448, 0.173858180642128, -0.12615609169006348, -0.24583272635936737, -0.14115718007087708, 0.0015609683468937874, 0.04152948409318924, -0.009671499952673912, 0.003867273684591055, -0.07054664939641953, -0.11710625886917114, -0.0934595838189125, 0.018713686615228653, 0.04491026699542999, -0.07426843047142029, -0.0596279613673687, 0.059872306883335114, 0.03894329443573952, -0.14430272579193115, 0.022237464785575867, 0.047419775277376175, -0.09032250195741653, -0.006925572175532579, 0.08398029953241348, 0.06729988008737564, 0.17764869332313538, 0.009659109637141228, -0.021044570952653885, 0.03080335259437561, 0.21258224546909332, -0.14283664524555206, 0.11252175271511078, 0.14021345973014832, -0.09024007618427277, 0.08099348843097687, 0.1948828399181366, 0.039186809211969376, -0.10478170961141586, 0.03259138762950897, 0.02489176020026207, -0.028939135372638702, -0.25018003582954407, -0.0680207833647728, 0.002590036718174815, -0.04892077296972275, 0.07092583924531937, 0.0918794497847557, 0.09946957975625992, 0.015428726561367512, -0.09732488542795181, -0.08017807453870773, 0.0468163788318634, 0.10640767961740494, 0.0070237633772194386, -0.01532268337905407, 0.08905128389596939, -0.03260866180062294, 0.018378758803009987, 0.0954233929514885, 0.00412675691768527, 0.17459604144096375, 0.05586163327097893, 0.17767499387264252, 0.07751350849866867, 0.06634163856506348, 0.019167855381965637, 0.0069374511949718, 0.02067388966679573, 0.017508454620838165, -0.004214957356452942, -0.08522020280361176, -0.00457410141825676, 0.12029227614402771, 0.06321834027767181, 0.024303704500198364, 0.0137604009360075, -0.03941800817847252, 0.08438141644001007, 0.17332784831523895, 0.0020201504230499268, -0.18486954271793365, -0.07240456342697144, 0.07921045273542404, -0.0910051167011261, -0.10552998632192612, -0.03353073075413704, 0.03346012532711029, -0.1747758537530899, 0.02097497321665287, -0.017018353566527367, 0.10809773951768875, -0.13855572044849396, -0.018670624122023582, 0.06328251957893372, 0.07232730835676193, -0.0028869258239865303, 0.06308864802122116, -0.153975248336792, 0.1050168052315712, 0.016289174556732178, 0.06754438579082489, -0.09747608006000519, 0.10138221830129623, -0.006303760688751936, -0.007241528946906328, 0.13875643908977509, 0.010596190579235554, -0.05694379657506943, -0.08987913280725479, -0.10555228590965271, -0.008462639525532722, 0.12933635711669922, -0.15157614648342133, 0.0847775787115097, -0.028662750497460365, -0.043171048164367676, 0.0024383023846894503, -0.1199452206492424, -0.1302652359008789, -0.1875755488872528, 0.058235347270965576, -0.1366453617811203, 0.039557021111249924, -0.10582595318555832, -0.04340389743447304, -0.028466427698731422, 0.2041483372449875, -0.2317875325679779, -0.0682469978928566, -0.1541893482208252, -0.08429346233606339, 0.14446710050106049, -0.04730919376015663, 0.08914490789175034, -0.0013825427740812302, 0.19013537466526031, 0.024473950266838074, -0.02387205697596073, 0.10308998823165894, -0.09543927758932114, -0.19450686872005463, -0.08603953570127487, 0.15582145750522614, 0.13931062817573547, 0.03702725097537041, -0.004593946039676666, 0.029260434210300446, -0.020000332966446877, -0.12535293400287628, 0.025526588782668114, 0.1793687790632248, 0.07859015464782715, 0.023437971249222755, -0.025896867737174034, -0.10993997752666473, -0.06524094194173813, -0.0335373692214489, 0.02718053013086319, 0.18264614045619965, -0.07421271502971649, 0.1900695115327835, 0.13626199960708618, -0.05445687845349312, -0.1955246478319168, 0.018216576427221298, 0.040417760610580444, 0.010847307741641998, 0.03138056397438049, -0.2078717201948166, 0.09027513861656189, 0.0014845491386950016, -0.05172133818268776, 0.141556978225708, -0.174949511885643, -0.1512570083141327, 0.06491631269454956, 0.0364508256316185, -0.19348180294036865, -0.117862768471241, -0.08817066252231598, -0.046907443553209305, -0.17498233914375305, 0.10519181191921234, 0.016932250931859016, 0.009516867808997631, 0.03492651879787445, 0.02640140987932682, 0.011080757714807987, -0.03873949125409126, 0.19461296498775482, -0.02505207620561123, 0.029532426968216896, -0.08079101145267487, -0.06136554479598999, 0.0607450045645237, -0.05577658861875534, 0.07896649837493896, -0.020188091322779655, 0.012835816480219364, -0.1100873053073883, -0.0468425452709198, -0.027396185323596, 0.017321845516562462, -0.09195652604103088, -0.09473495930433273, -0.05146971344947815, 0.09373841434717178, 0.08845265954732895, -0.036603908985853195, -0.04043547809123993, -0.07348548620939255, 0.0325477197766304, 0.17183002829551697, 0.17659065127372742, 0.038550034165382385, -0.08084331452846527, -0.005880105309188366, -0.01188716571778059, 0.04436201974749565, -0.22519725561141968, 0.06208868324756622, 0.04557957127690315, 0.015879612416028976, 0.11362850666046143, -0.018783990293741226, -0.16298477351665497, -0.06594224274158478, 0.06143777072429657, -0.06664001196622849, -0.18599680066108704, 0.0032026967965066433, 0.058006007224321365, -0.1646854728460312, -0.037671029567718506, 0.042260222136974335, -0.0045668939128518105, -0.04300284758210182, 0.01627597212791443, 0.08071378618478775, 0.005054219625890255, 0.07112491130828857, 0.05733523517847061, 0.0842885971069336, -0.10417009145021439, 0.07519911974668503, 0.08007751405239105, -0.08229218423366547, 0.031453702598810196, 0.08910130709409714, -0.061817802488803864, -0.03069761022925377, 0.032593827694654465, 0.07753410935401917, 0.019773589447140694, -0.041717879474163055, 0.008655321784317493, -0.09745000302791595, 0.06339588761329651, 0.09504765272140503, 0.03549657016992569, 0.014742289669811726, 0.034356739372015, 0.04988397657871246, -0.07460241764783859, 0.11766603589057922, 0.022336218506097794, 0.01780087500810623, -0.044981084764003754, -0.05459042266011238, 0.032110098749399185, -0.022974027320742607, -0.010163158178329468, -0.03885438293218613, -0.07015778869390488, -0.018130742013454437, -0.15929651260375977, -0.014899281784892082, -0.04085385054349899, 0.007158880587667227, 0.02551902085542679, -0.03834335505962372, 0.007963370531797409, 0.012195355258882046, -0.07085035741329193, -0.061454467475414276, -0.022903166711330414, 0.09224231541156769, -0.16436699032783508, 0.025155464187264442, 0.08285263180732727, -0.12099926173686981, 0.09775067120790482, 0.021939631551504135, 0.0031351554207503796, 0.028338242322206497, -0.1542527824640274, 0.04096807911992073, -0.024365095421671867, 0.01272035762667656, 0.04409142583608627, -0.22033950686454773, 0.001463581225834787, -0.03818526118993759, -0.05954346805810928, -0.010227864608168602, -0.033079732209444046, -0.11291328817605972, 0.09883669763803482, 0.008058897219598293, -0.08219768106937408, -0.030809206888079643, 0.03451729565858841, 0.08243680745363235, -0.02608415111899376, 0.15152283012866974, 0.0016822130419313908, 0.07172226905822754, -0.17519205808639526, -0.021702464669942856, -0.011611736379563808, 0.02207101881504059, -0.014536668546497822, -0.015496513806283474, 0.042471300810575485, -0.02421419881284237, 0.19108575582504272, -0.026401294395327568, 0.038726791739463806, 0.06405707448720932, 0.01593620702624321, -0.014801506884396076, 0.10957890748977661, 0.05975057929754257, 0.02399693801999092, 0.022115202620625496, 0.007329683285206556, -0.039842452853918076, -0.014149460941553116, -0.19538825750350952, 0.06474217027425766, 0.1377464383840561, 0.08781574666500092, -0.01322576031088829, 0.07683692127466202, -0.10024392604827881, -0.12397097796201706, 0.11215250939130783, -0.06283260136842728, -0.007701667957007885, -0.06531554460525513, 0.13346771895885468, 0.14944057166576385, -0.18992236256599426, 0.06835456937551498, -0.06228158622980118, -0.05332518368959427, -0.11744599789381027, -0.1957325041294098, -0.055616896599531174, -0.056456826627254486, -0.014700124971568584, -0.048795297741889954, 0.07307228446006775, 0.05693497136235237, 0.012962869368493557, 0.003600025549530983, 0.0766802653670311, -0.015357231721282005, 0.0008028073934838176, 0.03077360987663269, 0.06600049883127213, 0.013312965631484985, -0.02929985709488392, 0.020537450909614563, -0.007275243755429983, 0.04005419462919235, 0.06378308683633804, 0.038119763135910034, -0.02801438421010971, 0.01591232419013977, -0.03770609200000763, -0.10940317064523697, 0.0409080907702446, -0.028551526367664337, -0.08112191408872604, 0.13721226155757904, 0.02428387477993965, 0.005870606284588575, -0.02180131897330284, 0.24582624435424805, -0.07231455296278, -0.09001907706260681, -0.1473579704761505, 0.10211005061864853, -0.04095151647925377, 0.06560079753398895, 0.04110138490796089, -0.10732010751962662, 0.013498948886990547, 0.12688814103603363, 0.15896959602832794, -0.044884394854307175, 0.020156091079115868, 0.03252736106514931, 0.003683826420456171, -0.04006262496113777, 0.05253688618540764, 0.0694650411605835, 0.14883354306221008, -0.04907030612230301, 0.08928520232439041, 0.005485867150127888, -0.10256236046552658, -0.03822692111134529, 0.11808354407548904, -0.017866896465420723, 0.018703164532780647, -0.057248231023550034, 0.11889533698558807, -0.059861693531274796, -0.23005777597427368, 0.06317704170942307, -0.0720362737774849, -0.14286935329437256, -0.021647587418556213, 0.07456772774457932, -0.017636949196457863, 0.02658887766301632, 0.07326807081699371, -0.07681973278522491, 0.19899281859397888, 0.038975972682237625, -0.05729197710752487, -0.05658522993326187, 0.0789351835846901, -0.114089734852314, 0.2792985737323761, 0.01164181251078844, 0.04984506592154503, 0.10365619510412216, -0.016686614602804184, -0.13768579065799713, 0.015234606340527534, 0.09244892746210098, -0.09004336595535278, 0.03869183734059334, 0.2132277488708496, -0.002569539239630103, 0.1152428612112999, 0.07714667171239853, -0.07265080511569977, 0.04592108353972435, -0.1130065843462944, -0.0718315914273262, -0.086885966360569, 0.09441597014665604, -0.07240451127290726, 0.14123490452766418, 0.12318195402622223, -0.053516924381256104, 0.010368985123932362, -0.031209774315357208, 0.04651070013642311, 0.007842876948416233, 0.10365527868270874, 0.010769560933113098, -0.18099099397659302, 0.022656621411442757, 0.018202748149633408, 0.10856854915618896, -0.17241089046001434, -0.09672945737838745, 0.04725200682878494, 0.001958663808181882, -0.059874359518289566, 0.1282012164592743, 0.057909298688173294, 0.04923510178923607, -0.043742597103118896, -0.017267800867557526, -0.009560109116137028, 0.13584671914577484, -0.10737434774637222, -0.0021453071385622025 ]
null
null
stable-baselines3
# **DQN** Agent playing **SpaceInvadersNoFrameskip-v4** This is a trained model of a **DQN** agent playing **SpaceInvadersNoFrameskip-v4** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib Install the RL Zoo (with SB3 and SB3-Contrib): ```bash pip install rl_zoo3 ``` ``` # Download model and save it into the logs/ folder python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga Fhermin -f logs/ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do: ``` python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga Fhermin -f logs/ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` ## Training (with the RL Zoo) ``` python -m rl_zoo3.train --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ # Upload the model and generate video (when possible) python -m rl_zoo3.push_to_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ -orga Fhermin ``` ## Hyperparameters ```python OrderedDict([('batch_size', 64), ('buffer_size', 100000), ('env_wrapper', ['stable_baselines3.common.atari_wrappers.AtariWrapper']), ('exploration_final_eps', 0.01), ('exploration_fraction', 0.1), ('frame_stack', 6), ('gradient_steps', 1), ('learning_rate', 0.0001), ('learning_starts', 100000), ('n_timesteps', 1000000.0), ('optimize_memory_usage', False), ('policy', 'CnnPolicy'), ('target_update_interval', 1000), ('train_freq', 5), ('normalize', False)]) ``` # Environment Arguments ```python {'render_mode': 'rgb_array'} ```
{"library_name": "stable-baselines3", "tags": ["SpaceInvadersNoFrameskip-v4", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "DQN", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "SpaceInvadersNoFrameskip-v4", "type": "SpaceInvadersNoFrameskip-v4"}, "metrics": [{"type": "mean_reward", "value": "583.50 +/- 217.27", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
Fhermin/dqn-SpaceInvadersNoFrameskip-v4
[ "stable-baselines3", "SpaceInvadersNoFrameskip-v4", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
2024-02-15T01:55:23+00:00
[]
[]
TAGS #stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
# DQN Agent playing SpaceInvadersNoFrameskip-v4 This is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4 using the stable-baselines3 library and the RL Zoo. The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: URL SB3: URL SB3 Contrib: URL Install the RL Zoo (with SB3 and SB3-Contrib): If you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do: ## Training (with the RL Zoo) ## Hyperparameters # Environment Arguments
[ "# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.", "## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:", "## Training (with the RL Zoo)", "## Hyperparameters", "# Environment Arguments" ]
[ "TAGS\n#stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n", "# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.", "## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:", "## Training (with the RL Zoo)", "## Hyperparameters", "# Environment Arguments" ]
[ 43, 90, 73, 9, 5, 7 ]
[ "passage: TAGS\n#stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:## Training (with the RL Zoo)## Hyperparameters# Environment Arguments" ]
[ 0.043572068214416504, 0.2414778620004654, -0.0026879787910729647, 0.012635791674256325, 0.05784223601222038, 0.0030472534708678722, 0.08585051447153091, 0.10650663822889328, 0.024212315678596497, -0.001382096204906702, 0.003954293206334114, 0.17533031105995178, 0.03632635250687599, 0.13125447928905487, -0.018073517829179764, -0.2066594809293747, -0.013479253277182579, -0.06247470900416374, -0.07153085619211197, 0.036099132150411606, 0.07206681370735168, -0.030116932466626167, 0.036061208695173264, -0.051406677812337875, -0.057161085307598114, 0.036824777722358704, -0.03157254680991173, 0.007067287806421518, 0.15158706903457642, -0.1222257912158966, 0.12329676002264023, 0.020955175161361694, 0.1896144151687622, -0.12332789599895477, 0.0339222252368927, 0.08982209116220474, -0.036988191306591034, 0.013221588917076588, 0.00975361280143261, -0.052562564611434937, 0.1590864509344101, -0.09371145814657211, 0.07146181166172028, 0.010926910676062107, -0.07592244446277618, -0.1774153709411621, -0.09356249868869781, 0.07947742193937302, 0.0617753230035305, 0.005319166928529739, 0.03726791962981224, 0.11306490749120712, -0.020991774275898933, 0.06488905102014542, 0.11562903225421906, -0.17549200356006622, 0.013578375801444054, 0.17859570682048798, 0.003242473118007183, 0.15767055749893188, -0.05546637624502182, 0.019877681508660316, 0.02752300351858139, 0.04758313298225403, 0.06873945891857147, -0.08186400681734085, -0.1364826112985611, -0.056155186146497726, -0.15456219017505646, -0.03352400287985802, 0.05195203423500061, -0.011860138736665249, -0.05783402919769287, -0.010724928230047226, -0.04010869935154915, 0.0008851495804265141, -0.028637725859880447, 0.01805497519671917, 0.07031578570604324, -0.01226285845041275, 0.02092539705336094, -0.08391954004764557, -0.0390290804207325, -0.038563769310712814, -0.018022390082478523, 0.12054917961359024, 0.08285853266716003, 0.0266572255641222, -0.04135355353355408, 0.10274127870798111, -0.07091585546731949, -0.05454207584261894, 0.04555258899927139, -0.03786851093173027, -0.10615779459476471, 0.02120024710893631, -0.05905991420149803, 0.026879185810685158, 0.09943640232086182, 0.18048083782196045, -0.09862488508224487, 0.012620617635548115, -0.03430783003568649, 0.08121664822101593, -0.03196052461862564, 0.03197542577981949, -0.0840383991599083, -0.016251085326075554, 0.17835216224193573, 0.0030782297253608704, 0.022272996604442596, 0.002074616262689233, -0.049819961190223694, -0.02881433069705963, -0.017756454646587372, 0.06631895154714584, 0.07032092660665512, 0.010587303899228573, -0.0037596761249005795, -0.027667716145515442, -0.036921944469213486, -0.05629328638315201, -0.04952820762991905, 0.018803736194968224, -0.04712437093257904, -0.047942135483026505, 0.06027210131287575, -0.005624116864055395, 0.11337806284427643, -0.025607796385884285, 0.026316547766327858, -0.019410157576203346, -0.07494441419839859, -0.13221681118011475, -0.0304415225982666, 0.0691632330417633, 0.04371757060289383, -0.22497159242630005, -0.16994807124137878, -0.008539012633264065, 0.017946386709809303, -0.018741264939308167, -0.11334165185689926, 0.02453240379691124, -0.007166135590523481, -0.049758363515138626, -0.01601579785346985, 0.10474669933319092, -0.020438622683286667, 0.018010856583714485, -0.05593825876712799, 0.16603368520736694, -0.14290283620357513, 0.031004127115011215, -0.08706212788820267, 0.023509707301855087, -0.21286657452583313, 0.041208744049072266, -0.177636057138443, 0.04863585904240608, -0.08500861376523972, 0.02327173389494419, 0.021320728585124016, 0.01968831568956375, 0.08580207824707031, 0.10143322497606277, -0.23631145060062408, 0.05405791476368904, 0.07900930196046829, -0.022739801555871964, -0.04218491166830063, 0.06798892468214035, -0.06558530032634735, 0.1382148116827011, 0.046505436301231384, 0.24831900000572205, 0.10361487418413162, -0.2036508023738861, 0.061786454170942307, 0.0578593946993351, -0.08880111575126648, -0.004730981774628162, -0.020022382959723473, 0.11598580330610275, -0.01114928349852562, 0.03338807821273804, -0.12186288088560104, 0.1456439197063446, 0.02738998830318451, -0.0165485180914402, -0.04454165697097778, -0.1614885926246643, 0.10309953987598419, -0.015504824928939342, 0.09532155096530914, -0.042415786534547806, 0.0001161050095106475, -0.011168917641043663, 0.18012429773807526, -0.043841805309057236, 0.0007168867159634829, 0.07871408760547638, 0.10895700752735138, 0.028009075671434402, -0.020230965688824654, -0.20380273461341858, -0.0423048660159111, 0.02367858961224556, 0.044489551335573196, 0.2190362960100174, 0.19936694204807281, 0.07770156860351562, -0.022313760593533516, -0.025487221777439117, -0.003248062450438738, -0.05106664076447487, 0.03467361256480217, -0.027858436107635498, -0.024532482028007507, 0.06065356358885765, -0.09305168688297272, 0.02817818708717823, -0.13112716376781464, 0.06307920068502426, -0.17345242202281952, 0.06863926351070404, 0.021998396143317223, -0.005436043255031109, 0.024577690288424492, -0.011292695067822933, -0.034188106656074524, -0.06233125180006027, 0.07110602408647537, 0.06098933145403862, 0.014702376909554005, 0.0021991983521729708, -0.0683600977063179, -0.13828523457050323, 0.08231553435325623, -0.04042381793260574, -0.14305958151817322, 0.06392676383256912, 0.011172642931342125, 0.04875864461064339, -0.05975872278213501, 0.016254881396889687, 0.22900153696537018, 0.05321883037686348, 0.09785865992307663, -0.04092191904783249, -0.022525805979967117, -0.06617844104766846, -0.06677833944559097, 0.09694591909646988, 0.10812206566333771, 0.060318704694509506, -0.0030071530491113663, 0.07626225054264069, 0.10942911356687546, -0.1035122498869896, -0.0651884600520134, 0.03220061957836151, -0.05973697826266289, 0.019652515649795532, 0.049140311777591705, 0.02971293032169342, 0.08619047701358795, 0.1833551675081253, 0.008245792239904404, 0.0386311337351799, -0.025997694581747055, 0.026109617203474045, -0.15547916293144226, -0.03145433962345123, 0.04308181628584862, 0.00886955764144659, -0.07408110797405243, 0.04994636029005051, 0.051439400762319565, 0.13607151806354523, -0.08217083662748337, -0.13170577585697174, -0.059745315462350845, -0.03804200142621994, -0.04239124804735184, 0.14975430071353912, -0.08507520705461502, -0.19221234321594238, -0.017164425924420357, -0.15751953423023224, -0.02518727444112301, -0.005179801490157843, 0.002318724524229765, -0.08325926214456558, 0.017780914902687073, 0.010001576505601406, -0.03129372000694275, -0.0684933215379715, -0.06596160680055618, -0.05786636844277382, 0.09124112874269485, 0.06932931393384933, -0.12240120023488998, -0.00961651187390089, -0.03742414712905884, -0.020465577021241188, 0.04516167193651199, 0.08452648669481277, -0.007267598994076252, 0.07773483544588089, -0.13209199905395508, -0.06962883472442627, 0.02834828943014145, 0.2766247093677521, 0.02882981114089489, 0.004668009467422962, 0.17051753401756287, -0.03629542142152786, 0.04912714660167694, 0.16181479394435883, 0.030781643465161324, -0.14196757972240448, 0.07090470939874649, -0.011341600678861141, -0.09542687982320786, -0.1706860214471817, -0.10215658694505692, -0.037867411971092224, -0.05015881359577179, 0.05638284236192703, 0.004951419774442911, -0.04476970434188843, 0.05910305306315422, 0.08782228082418442, -0.017004497349262238, -0.06151578947901726, 0.11129767447710037, 0.032263003289699554, -0.030136963352560997, 0.08078382909297943, -0.042354047298431396, -0.04206389561295509, 0.0032403599470853806, 0.22643887996673584, 0.0937788337469101, -0.01775507442653179, -0.042567066848278046, 0.019317636266350746, 0.05095715448260307, 0.03613382205367088, 0.11312435567378998, -0.06975842267274857, -0.06826137751340866, -0.035185977816581726, 0.027829548344016075, -0.02945687249302864, 0.08205190300941467, 0.0630207508802414, 0.005563626065850258, -0.04653681069612503, -0.07972332090139389, -0.04849022626876831, 0.08408913016319275, -0.027642227709293365, -0.10093270242214203, 0.09321888536214828, 0.048575710505247116, 0.0016974330646917224, 0.03055831417441368, 0.027994604781270027, 0.01462269201874733, -0.07982148975133896, -0.06775744259357452, 0.011468625627458096, 0.07076629996299744, -0.06822766363620758, -0.027886953204870224, -0.19817815721035004, 0.14578363299369812, 0.010630400851368904, 0.04118429124355316, -0.13048617541790009, 0.1209396943449974, -0.023116756230592728, -0.026430301368236542, 0.013811616227030754, 0.0014643745962530375, 0.08203291147947311, -0.04806509613990784, 0.15762180089950562, 0.009528410620987415, -0.28092408180236816, -0.1418946087360382, -0.08416824042797089, -0.051183976233005524, -0.022873088717460632, 0.014752174727618694, 0.0642135739326477, 0.01516205258667469, 0.003868846921250224, -0.013076163828372955, 0.03185269236564636, -0.09826882928609848, -0.06493937969207764, -0.04839126765727997, -0.02250157669186592, -0.06525848805904388, -0.05647949501872063, -0.0006809153710491955, -0.17226077616214752, 0.12522587180137634, 0.11787347495555878, -0.06451737880706787, -0.041814323514699936, -0.06554657220840454, 0.046191465109586716, -0.07571537792682648, 0.0469326451420784, 0.003414976177737117, 0.019198855385184288, -0.06806991249322891, -0.17922484874725342, 0.016097763553261757, -0.10899919271469116, 0.03772687539458275, -0.05070559307932854, 0.020257100462913513, 0.08594245463609695, 0.17520126700401306, 0.05856714025139809, 0.01460097823292017, -0.07239776104688644, -0.07543374598026276, -0.0017121878918260336, -0.06344114243984222, 0.05762333422899246, -0.009151889942586422, -0.20333483815193176, 0.02763226442039013, -0.11414948850870132, 0.06860900670289993, 0.3310066759586334, 0.3324824273586273, -0.10698744654655457, 0.1177443116903305, 0.04819539934396744, -0.042202454060316086, -0.21051374077796936, -0.002244179602712393, 0.012272895313799381, 0.024992236867547035, 0.13725964725017548, -0.12924811244010925, 0.05453680083155632, 0.0794181227684021, -0.024458877742290497, 0.01456840243190527, -0.09078162908554077, -0.10816970467567444, 0.20847418904304504, 0.14226987957954407, 0.04421741142868996, -0.09421348571777344, 0.08391669392585754, 0.004295284394174814, 0.08375877887010574, 0.2107764035463333, -0.052112679928541183, 0.10695768147706985, 0.005195184610784054, 0.19852910935878754, 0.0328996516764164, -0.023768596351146698, 0.10834760218858719, -0.009801650419831276, 0.07911337912082672, 0.03985166177153587, -0.007676942739635706, 0.010487722232937813, -0.04522453248500824, 0.014148596674203873, -0.028376007452607155, 0.010284217074513435, -0.2274095118045807, 0.0582297146320343, -0.06368855386972427, 0.04604509472846985, 0.008256820961833, -0.0999874547123909, -0.03583388403058052, 0.06431841105222702, 0.08014573156833649, 0.01975327916443348, 0.0436067171394825, -0.03867863491177559, 0.11051398515701294, 0.20660489797592163, -0.009811338968575, 0.17751595377922058, -0.0615963339805603, 0.01464168168604374, -0.023011628538370132, -0.04223164543509483, -0.1462583988904953, -0.035259708762168884, 0.03498423472046852, 0.057734888046979904, 0.015203364193439484, 0.049647457897663116, -0.05656236410140991, 0.08498423546552658, 0.021687336266040802, -0.041541360318660736, 0.033579520881175995, 0.08835696429014206, 0.12415177375078201, 0.010754258371889591, -0.030121933668851852, 0.06147436052560806, -0.08128108084201813, -0.09446098655462265, -0.004497923422604799, -0.029991207644343376, -0.1083834245800972, 0.11353230476379395, 0.16914646327495575, 0.039594944566488266, -0.057076629251241684, 0.10688766092061996, -0.02768099494278431, 0.10047874599695206, 0.009198128245770931, 0.06507332623004913, -0.014091075398027897, -0.03691792115569115, 0.10611724853515625, -0.05442855879664421, -0.01637818105518818, 0.07645545154809952, -0.06522727757692337, -0.023877469822764397, -0.0801999643445015, 0.06034626066684723, 0.09222240000963211, -0.16854619979858398, -0.0639432892203331, -0.032122284173965454, -0.08628080040216446, 0.013965039514005184, 0.012447911314666271, 0.0710059329867363, -0.08589600026607513, 0.06316167116165161, -0.024337708950042725, 0.015639442950487137, -0.03689891844987869, 0.019222697243094444, -0.19525384902954102, -0.002140450058504939, -0.11280795186758041, -0.00348020251840353, -0.002931603929027915, 0.04463808611035347, -0.04961875081062317, -0.029358822852373123, -0.0030675032176077366, 0.044366419315338135, -0.16609135270118713, 0.002798673929646611, -0.011639905162155628, 0.03210212290287018, -0.0002893915225286037, -0.0983390137553215, 0.014195028692483902, -0.04294256120920181, -0.04198618605732918, 0.04925514757633209, 0.009436776861548424, 0.06470516324043274, -0.2795179784297943, -0.14905457198619843, 0.030816160142421722, 0.0683867484331131, 0.05483196675777435, -0.1830425262451172, 0.03568267077207565, -0.08042316138744354, -0.02253127470612526, -0.037770628929138184, 0.018491698428988457, -0.0539514496922493, 0.0018174031283706427, -0.04225044324994087, -0.023033907637000084, -0.028055014088749886, -0.07556360960006714, 0.0826747715473175, 0.12462522834539413, 0.07555580884218216, -0.03807181864976883, 0.09595896303653717, -0.10009756684303284, -0.04657831788063049, -0.04052736237645149, -0.036951083689928055, 0.017965637147426605, -0.0870552659034729, 0.048530060797929764, 0.05188591405749321, 0.18719671666622162, -0.08520494401454926, -0.058800119906663895, -0.014255574904382229, 0.0746525228023529, 0.07849094271659851, 0.005095830652862787, 0.17779210209846497, -0.045693784952163696, 0.05693846940994263, 0.021304311230778694, 0.046699028462171555, 0.10497613251209259, -0.023569339886307716, 0.14490213990211487, 0.21171095967292786, -0.037196725606918335, -0.11048602312803268, 0.043668005615472794, 0.01745123788714409, -0.002401199424639344, 0.05968761444091797, 0.11983796209096909, -0.050589341670274734, -0.10903856158256531, 0.23442286252975464, 0.054169271141290665, -0.11218088120222092, 0.09546315670013428, 0.039532262831926346, -0.015890996903181076, -0.1301896870136261, 0.010444961488246918, -0.0013640925753861666, -0.11233190447092056, 0.03386834263801575, -0.06087532266974449, -0.025547027587890625, 0.11809267848730087, 0.008789865300059319, 0.03317064419388771, -0.04139537364244461, -0.03756232187151909, -0.04352104663848877, -0.04273213446140289, -0.012549578212201595, -0.02991986647248268, -0.030186517164111137, -0.07621737569570541, -0.007770835887640715, -0.012012424878776073, 0.030795488506555557, -0.015285328030586243, -0.02503054589033127, -0.021192016080021858, -0.06697061657905579, -0.0026312144473195076, -0.008178025484085083, 0.015549594536423683, 0.010121971368789673, 0.2358063906431198, 0.07042546570301056, -0.10260069370269775, -0.01036880537867546, 0.22197756171226501, -0.03853277862071991, -0.06528383493423462, -0.07849395275115967, 0.25128230452537537, -0.10482002794742584, 0.051095426082611084, -0.005819917656481266, -0.06550488620996475, -0.07153836637735367, 0.2309868484735489, 0.13502730429172516, -0.1677926480770111, 0.06329060345888138, -0.0368385910987854, -0.009490780532360077, -0.14286863803863525, 0.16013580560684204, 0.1865294873714447, 0.09480160474777222, -0.12259847670793533, 0.0023130534682422876, -0.03518044203519821, -0.018328361213207245, -0.1660851687192917, -0.004593863617628813, -0.029364850372076035, -0.0427238829433918, -0.050771355628967285, 0.029773715883493423, -0.15205919742584229, -0.0927426889538765, -0.1916799396276474, -0.11482496559619904, -0.12386849522590637, -0.04549141973257065, -0.11142764985561371, -0.0019938007462769747, 0.02257080189883709, -0.0641874223947525, 0.021061956882476807, -0.0212461706250906, -0.05887424945831299, 0.015386379323899746, -0.08395619690418243, 0.0674985870718956, 0.06488548219203949, 0.15327942371368408, -0.0790991559624672, 0.025424562394618988, 0.07090727984905243, -0.057595450431108475, -0.10164349526166916, 0.06067253649234772, 0.015708057209849358, -0.1972588747739792, 0.007548294495791197, 0.17712996900081635, -0.10420889407396317, 0.09745754301548004, 0.048501528799533844, -0.012951982207596302, 0.0867827981710434, -0.024721821770071983, -0.016682926565408707, -0.04852180927991867, -0.011212974786758423, -0.10143939405679703, 0.09892100840806961, 0.0876845121383667, -0.0517118014395237, 0.07436849176883698, -0.09508965909481049, -0.04068392515182495, 0.13103286921977997, -0.010057874955236912, -0.08450483530759811, -0.11667824536561966, -0.04081142693758011, 0.09684515744447708, -0.018041390925645828, -0.20185889303684235, -0.11639472097158432, -0.11752668023109436, -0.00014377340266946703, -0.03563340753316879, 0.061800602823495865, 0.02430674433708191, -0.02556120604276657, -0.008150683715939522, -0.17615078389644623, -0.06614746153354645, 0.13479791581630707, -0.10176112502813339, -0.07456064969301224 ]
null
null
null
# Lora of unzen/雲仙/云仙 (Azur Lane) ## What Is This? This is the LoRA model of waifu unzen/雲仙/云仙 (Azur Lane). ## How Is It Trained? * This model is trained with [HCP-Diffusion](https://github.com/7eu7d7/HCP-Diffusion). * The [auto-training framework](https://github.com/deepghs/cyberharem) is maintained by [DeepGHS Team](https://huggingface.co/deepghs). * The base model used for training is [deepghs/animefull-latest](https://huggingface.co/deepghs/animefull-latest). * Dataset used for training is the `stage3-p480-800` in [CyberHarem/unzen_azurlane](https://huggingface.co/datasets/CyberHarem/unzen_azurlane), which contains 341 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 11, resolution is 720x720, clustering into 20 buckets. * Trained for 3440 steps, 40 checkpoints were saved and evaluated. * **Trigger word is `unzen_azurlane`.** * Pruned core tags for this waifu are `breasts, long_hair, large_breasts, white_hair, purple_eyes, hair_over_one_eye, multicolored_hair, bangs, streaked_hair, very_long_hair, ponytail`. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. ## How to Use It? ### If You Are Using A1111 WebUI v1.7+ **Just use it like the classic LoRA**. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 2150, you need to download [`2150/unzen_azurlane.pt`](https://huggingface.co/CyberHarem/unzen_azurlane/resolve/main/2150/unzen_azurlane.pt) as the embedding and [`2150/unzen_azurlane.safetensors`](https://huggingface.co/CyberHarem/unzen_azurlane/resolve/main/2150/unzen_azurlane.safetensors) for loading Lora. By using both files together, you can generate images for the desired characters. ## Which Step Should I Use? We selected 5 good steps for you to choose. The best one is step 2150. 1640 images (1.77 GiB) were generated for auto-testing. ![Metrics Plot](metrics_plot.png) The base model used for generating preview images is [Meina/MeinaMix_V11](https://huggingface.co/Meina/MeinaMix_V11). Here are the preview of the recommended steps: | Step | Epoch | CCIP | AI Corrupt | Bikini Plus | Score | Download | pattern_0_0 | pattern_0_1 | pattern_1 | pattern_2_0 | pattern_2_1 | pattern_2_2 | portrait_0 | portrait_1 | portrait_2 | full_body_0 | full_body_1 | profile_0 | profile_1 | free_0 | free_1 | shorts | maid_0 | maid_1 | miko | yukata | suit | china | bikini_0 | bikini_1 | bikini_2 | sit | squat | kneel | jump | crossed_arms | angry | smile | cry | grin | n_lie_0 | n_lie_1 | n_stand_0 | n_stand_1 | n_stand_2 | n_sex_0 | n_sex_1 | |-------:|--------:|:----------|:-------------|:--------------|:----------|:--------------------------------------------------------------------------------------------------|:----------------------------------------------|:----------------------------------------------|:------------------------------------------|:----------------------------------------------|:----------------------------------------------|:----------------------------------------------|:--------------------------------------------|:--------------------------------------------|:--------------------------------------------|:----------------------------------------------|:----------------------------------------------|:------------------------------------------|:------------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:--------------------------------|:------------------------------------|:--------------------------------|:----------------------------------|:----------------------------------------|:----------------------------------------|:----------------------------------------|:------------------------------|:----------------------------------|:----------------------------------|:--------------------------------|:------------------------------------------------|:----------------------------------|:----------------------------------|:------------------------------|:--------------------------------|:--------------------------------------|:--------------------------------------|:------------------------------------------|:------------------------------------------|:------------------------------------------|:--------------------------------------|:--------------------------------------| | 2150 | 26 | **0.985** | 0.944 | 0.844 | **0.696** | [Download](https://huggingface.co/CyberHarem/unzen_azurlane/resolve/main/2150/unzen_azurlane.zip) | ![pattern_0_0](2150/previews/pattern_0_0.png) | ![pattern_0_1](2150/previews/pattern_0_1.png) | ![pattern_1](2150/previews/pattern_1.png) | ![pattern_2_0](2150/previews/pattern_2_0.png) | ![pattern_2_1](2150/previews/pattern_2_1.png) | ![pattern_2_2](2150/previews/pattern_2_2.png) | ![portrait_0](2150/previews/portrait_0.png) | ![portrait_1](2150/previews/portrait_1.png) | ![portrait_2](2150/previews/portrait_2.png) | ![full_body_0](2150/previews/full_body_0.png) | ![full_body_1](2150/previews/full_body_1.png) | ![profile_0](2150/previews/profile_0.png) | ![profile_1](2150/previews/profile_1.png) | ![free_0](2150/previews/free_0.png) | ![free_1](2150/previews/free_1.png) | ![shorts](2150/previews/shorts.png) | ![maid_0](2150/previews/maid_0.png) | ![maid_1](2150/previews/maid_1.png) | ![miko](2150/previews/miko.png) | ![yukata](2150/previews/yukata.png) | ![suit](2150/previews/suit.png) | ![china](2150/previews/china.png) | ![bikini_0](2150/previews/bikini_0.png) | ![bikini_1](2150/previews/bikini_1.png) | ![bikini_2](2150/previews/bikini_2.png) | ![sit](2150/previews/sit.png) | ![squat](2150/previews/squat.png) | ![kneel](2150/previews/kneel.png) | ![jump](2150/previews/jump.png) | ![crossed_arms](2150/previews/crossed_arms.png) | ![angry](2150/previews/angry.png) | ![smile](2150/previews/smile.png) | ![cry](2150/previews/cry.png) | ![grin](2150/previews/grin.png) | ![n_lie_0](2150/previews/n_lie_0.png) | ![n_lie_1](2150/previews/n_lie_1.png) | ![n_stand_0](2150/previews/n_stand_0.png) | ![n_stand_1](2150/previews/n_stand_1.png) | ![n_stand_2](2150/previews/n_stand_2.png) | ![n_sex_0](2150/previews/n_sex_0.png) | ![n_sex_1](2150/previews/n_sex_1.png) | | 516 | 7 | 0.971 | **0.981** | **0.859** | 0.693 | [Download](https://huggingface.co/CyberHarem/unzen_azurlane/resolve/main/516/unzen_azurlane.zip) | ![pattern_0_0](516/previews/pattern_0_0.png) | ![pattern_0_1](516/previews/pattern_0_1.png) | ![pattern_1](516/previews/pattern_1.png) | ![pattern_2_0](516/previews/pattern_2_0.png) | ![pattern_2_1](516/previews/pattern_2_1.png) | ![pattern_2_2](516/previews/pattern_2_2.png) | ![portrait_0](516/previews/portrait_0.png) | ![portrait_1](516/previews/portrait_1.png) | ![portrait_2](516/previews/portrait_2.png) | ![full_body_0](516/previews/full_body_0.png) | ![full_body_1](516/previews/full_body_1.png) | ![profile_0](516/previews/profile_0.png) | ![profile_1](516/previews/profile_1.png) | ![free_0](516/previews/free_0.png) | ![free_1](516/previews/free_1.png) | ![shorts](516/previews/shorts.png) | ![maid_0](516/previews/maid_0.png) | ![maid_1](516/previews/maid_1.png) | ![miko](516/previews/miko.png) | ![yukata](516/previews/yukata.png) | ![suit](516/previews/suit.png) | ![china](516/previews/china.png) | ![bikini_0](516/previews/bikini_0.png) | ![bikini_1](516/previews/bikini_1.png) | ![bikini_2](516/previews/bikini_2.png) | ![sit](516/previews/sit.png) | ![squat](516/previews/squat.png) | ![kneel](516/previews/kneel.png) | ![jump](516/previews/jump.png) | ![crossed_arms](516/previews/crossed_arms.png) | ![angry](516/previews/angry.png) | ![smile](516/previews/smile.png) | ![cry](516/previews/cry.png) | ![grin](516/previews/grin.png) | ![n_lie_0](516/previews/n_lie_0.png) | ![n_lie_1](516/previews/n_lie_1.png) | ![n_stand_0](516/previews/n_stand_0.png) | ![n_stand_1](516/previews/n_stand_1.png) | ![n_stand_2](516/previews/n_stand_2.png) | ![n_sex_0](516/previews/n_sex_0.png) | ![n_sex_1](516/previews/n_sex_1.png) | | 3268 | 39 | 0.983 | 0.912 | 0.836 | 0.683 | [Download](https://huggingface.co/CyberHarem/unzen_azurlane/resolve/main/3268/unzen_azurlane.zip) | ![pattern_0_0](3268/previews/pattern_0_0.png) | ![pattern_0_1](3268/previews/pattern_0_1.png) | ![pattern_1](3268/previews/pattern_1.png) | ![pattern_2_0](3268/previews/pattern_2_0.png) | ![pattern_2_1](3268/previews/pattern_2_1.png) | ![pattern_2_2](3268/previews/pattern_2_2.png) | ![portrait_0](3268/previews/portrait_0.png) | ![portrait_1](3268/previews/portrait_1.png) | ![portrait_2](3268/previews/portrait_2.png) | ![full_body_0](3268/previews/full_body_0.png) | ![full_body_1](3268/previews/full_body_1.png) | ![profile_0](3268/previews/profile_0.png) | ![profile_1](3268/previews/profile_1.png) | ![free_0](3268/previews/free_0.png) | ![free_1](3268/previews/free_1.png) | ![shorts](3268/previews/shorts.png) | ![maid_0](3268/previews/maid_0.png) | ![maid_1](3268/previews/maid_1.png) | ![miko](3268/previews/miko.png) | ![yukata](3268/previews/yukata.png) | ![suit](3268/previews/suit.png) | ![china](3268/previews/china.png) | ![bikini_0](3268/previews/bikini_0.png) | ![bikini_1](3268/previews/bikini_1.png) | ![bikini_2](3268/previews/bikini_2.png) | ![sit](3268/previews/sit.png) | ![squat](3268/previews/squat.png) | ![kneel](3268/previews/kneel.png) | ![jump](3268/previews/jump.png) | ![crossed_arms](3268/previews/crossed_arms.png) | ![angry](3268/previews/angry.png) | ![smile](3268/previews/smile.png) | ![cry](3268/previews/cry.png) | ![grin](3268/previews/grin.png) | ![n_lie_0](3268/previews/n_lie_0.png) | ![n_lie_1](3268/previews/n_lie_1.png) | ![n_stand_0](3268/previews/n_stand_0.png) | ![n_stand_1](3268/previews/n_stand_1.png) | ![n_stand_2](3268/previews/n_stand_2.png) | ![n_sex_0](3268/previews/n_sex_0.png) | ![n_sex_1](3268/previews/n_sex_1.png) | | 1204 | 15 | 0.978 | 0.906 | 0.840 | 0.683 | [Download](https://huggingface.co/CyberHarem/unzen_azurlane/resolve/main/1204/unzen_azurlane.zip) | ![pattern_0_0](1204/previews/pattern_0_0.png) | ![pattern_0_1](1204/previews/pattern_0_1.png) | ![pattern_1](1204/previews/pattern_1.png) | ![pattern_2_0](1204/previews/pattern_2_0.png) | ![pattern_2_1](1204/previews/pattern_2_1.png) | ![pattern_2_2](1204/previews/pattern_2_2.png) | ![portrait_0](1204/previews/portrait_0.png) | ![portrait_1](1204/previews/portrait_1.png) | ![portrait_2](1204/previews/portrait_2.png) | ![full_body_0](1204/previews/full_body_0.png) | ![full_body_1](1204/previews/full_body_1.png) | ![profile_0](1204/previews/profile_0.png) | ![profile_1](1204/previews/profile_1.png) | ![free_0](1204/previews/free_0.png) | ![free_1](1204/previews/free_1.png) | ![shorts](1204/previews/shorts.png) | ![maid_0](1204/previews/maid_0.png) | ![maid_1](1204/previews/maid_1.png) | ![miko](1204/previews/miko.png) | ![yukata](1204/previews/yukata.png) | ![suit](1204/previews/suit.png) | ![china](1204/previews/china.png) | ![bikini_0](1204/previews/bikini_0.png) | ![bikini_1](1204/previews/bikini_1.png) | ![bikini_2](1204/previews/bikini_2.png) | ![sit](1204/previews/sit.png) | ![squat](1204/previews/squat.png) | ![kneel](1204/previews/kneel.png) | ![jump](1204/previews/jump.png) | ![crossed_arms](1204/previews/crossed_arms.png) | ![angry](1204/previews/angry.png) | ![smile](1204/previews/smile.png) | ![cry](1204/previews/cry.png) | ![grin](1204/previews/grin.png) | ![n_lie_0](1204/previews/n_lie_0.png) | ![n_lie_1](1204/previews/n_lie_1.png) | ![n_stand_0](1204/previews/n_stand_0.png) | ![n_stand_1](1204/previews/n_stand_1.png) | ![n_stand_2](1204/previews/n_stand_2.png) | ![n_sex_0](1204/previews/n_sex_0.png) | ![n_sex_1](1204/previews/n_sex_1.png) | | 1376 | 17 | 0.977 | 0.942 | 0.839 | 0.680 | [Download](https://huggingface.co/CyberHarem/unzen_azurlane/resolve/main/1376/unzen_azurlane.zip) | ![pattern_0_0](1376/previews/pattern_0_0.png) | ![pattern_0_1](1376/previews/pattern_0_1.png) | ![pattern_1](1376/previews/pattern_1.png) | ![pattern_2_0](1376/previews/pattern_2_0.png) | ![pattern_2_1](1376/previews/pattern_2_1.png) | ![pattern_2_2](1376/previews/pattern_2_2.png) | ![portrait_0](1376/previews/portrait_0.png) | ![portrait_1](1376/previews/portrait_1.png) | ![portrait_2](1376/previews/portrait_2.png) | ![full_body_0](1376/previews/full_body_0.png) | ![full_body_1](1376/previews/full_body_1.png) | ![profile_0](1376/previews/profile_0.png) | ![profile_1](1376/previews/profile_1.png) | ![free_0](1376/previews/free_0.png) | ![free_1](1376/previews/free_1.png) | ![shorts](1376/previews/shorts.png) | ![maid_0](1376/previews/maid_0.png) | ![maid_1](1376/previews/maid_1.png) | ![miko](1376/previews/miko.png) | ![yukata](1376/previews/yukata.png) | ![suit](1376/previews/suit.png) | ![china](1376/previews/china.png) | ![bikini_0](1376/previews/bikini_0.png) | ![bikini_1](1376/previews/bikini_1.png) | ![bikini_2](1376/previews/bikini_2.png) | ![sit](1376/previews/sit.png) | ![squat](1376/previews/squat.png) | ![kneel](1376/previews/kneel.png) | ![jump](1376/previews/jump.png) | ![crossed_arms](1376/previews/crossed_arms.png) | ![angry](1376/previews/angry.png) | ![smile](1376/previews/smile.png) | ![cry](1376/previews/cry.png) | ![grin](1376/previews/grin.png) | ![n_lie_0](1376/previews/n_lie_0.png) | ![n_lie_1](1376/previews/n_lie_1.png) | ![n_stand_0](1376/previews/n_stand_0.png) | ![n_stand_1](1376/previews/n_stand_1.png) | ![n_stand_2](1376/previews/n_stand_2.png) | ![n_sex_0](1376/previews/n_sex_0.png) | ![n_sex_1](1376/previews/n_sex_1.png) | ## Anything Else? Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. ## All Steps We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * [Steps From 2666 to 3440](all/0.md) * [Steps From 1806 to 2580](all/1.md) * [Steps From 946 to 1720](all/2.md) * [Steps From 86 to 860](all/3.md)
{"license": "mit", "tags": ["art", "not-for-all-audiences"], "datasets": ["CyberHarem/unzen_azurlane"], "pipeline_tag": "text-to-image"}
text-to-image
CyberHarem/unzen_azurlane
[ "art", "not-for-all-audiences", "text-to-image", "dataset:CyberHarem/unzen_azurlane", "license:mit", "region:us" ]
2024-02-15T01:57:20+00:00
[]
[]
TAGS #art #not-for-all-audiences #text-to-image #dataset-CyberHarem/unzen_azurlane #license-mit #region-us
Lora of unzen/雲仙/云仙 (Azur Lane) =============================== What Is This? ------------- This is the LoRA model of waifu unzen/雲仙/云仙 (Azur Lane). How Is It Trained? ------------------ * This model is trained with HCP-Diffusion. * The auto-training framework is maintained by DeepGHS Team. * The base model used for training is deepghs/animefull-latest. * Dataset used for training is the 'stage3-p480-800' in CyberHarem/unzen\_azurlane, which contains 341 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 11, resolution is 720x720, clustering into 20 buckets. * Trained for 3440 steps, 40 checkpoints were saved and evaluated. * Trigger word is 'unzen\_azurlane'. * Pruned core tags for this waifu are 'breasts, long\_hair, large\_breasts, white\_hair, purple\_eyes, hair\_over\_one\_eye, multicolored\_hair, bangs, streaked\_hair, very\_long\_hair, ponytail'. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. How to Use It? -------------- ### If You Are Using A1111 WebUI v1.7+ Just use it like the classic LoRA. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 2150, you need to download '2150/unzen\_azurlane.pt' as the embedding and '2150/unzen\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters. Which Step Should I Use? ------------------------ We selected 5 good steps for you to choose. The best one is step 2150. 1640 images (1.77 GiB) were generated for auto-testing. !Metrics Plot The base model used for generating preview images is Meina/MeinaMix\_V11. Here are the preview of the recommended steps: Anything Else? -------------- Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. All Steps --------- We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * Steps From 2666 to 3440 * Steps From 1806 to 2580 * Steps From 946 to 1720 * Steps From 86 to 860
[ "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 2150, you need to download '2150/unzen\\_azurlane.pt' as the embedding and '2150/unzen\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 2150.\n\n\n1640 images (1.77 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 2666 to 3440\n* Steps From 1806 to 2580\n* Steps From 946 to 1720\n* Steps From 86 to 860" ]
[ "TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/unzen_azurlane #license-mit #region-us \n", "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 2150, you need to download '2150/unzen\\_azurlane.pt' as the embedding and '2150/unzen\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 2150.\n\n\n1640 images (1.77 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 2666 to 3440\n* Steps From 1806 to 2580\n* Steps From 946 to 1720\n* Steps From 86 to 860" ]
[ 44, 38, 475 ]
[ "passage: TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/unzen_azurlane #license-mit #region-us \n### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file." ]
[ 0.009051081724464893, -0.013769813813269138, -0.004030069801956415, 0.07651089876890182, 0.07829166948795319, 0.08536243438720703, 0.23320235311985016, 0.07793933153152466, 0.12274285405874252, -0.07519222050905228, 0.0884777083992958, 0.05553790554404259, -0.004647260997444391, 0.025720948353409767, -0.03003322333097458, -0.14360301196575165, -0.06035345792770386, -0.037941232323646545, 0.003925763536244631, 0.014811134897172451, 0.07946082949638367, 0.008592338301241398, 0.0989389717578888, -0.053920596837997437, -0.034549590200185776, 0.048206593841314316, -0.032607633620500565, -0.04028473421931267, 0.027341024950146675, 0.08113672584295273, 0.12426642328500748, 0.009499973617494106, 0.06648167222738266, -0.16113890707492828, 0.06598561257123947, -0.009781315922737122, -0.10825416445732117, -0.00899583101272583, 0.013405641540884972, -0.02753964625298977, 0.12520408630371094, 0.019200261682271957, -0.11424997448921204, 0.041576314717531204, -0.1280352771282196, -0.023574385792016983, -0.05348493903875351, 0.04864545166492462, 0.1367432177066803, 0.056032102555036545, 0.02117118611931801, 0.0539063960313797, -0.051338888704776764, 0.08346729725599289, 0.10583893209695816, -0.13884881138801575, -0.06716588139533997, 0.10098820179700851, 0.008061667904257774, 0.13636523485183716, -0.09135691821575165, 0.09502295404672623, 0.06919339299201965, -0.05079777166247368, -0.15172423422336578, -0.09538327902555466, -0.227415531873703, -0.009074782952666283, 0.01132581103593111, 0.023951133713126183, 0.4042302072048187, 0.06065336987376213, 0.040587592869997025, 0.06797719746828079, -0.06740862131118774, 0.026981733739376068, -0.09951221942901611, 0.1410333216190338, 0.04070034623146057, 0.09339314699172974, -0.0450877770781517, -0.10530833154916763, -0.10806071758270264, -0.06564050167798996, -0.07893633842468262, -0.004705186001956463, 0.027297507971525192, 0.11738668382167816, -0.19295714795589447, 0.001544299186207354, -0.04791193827986717, -0.12883441150188446, 0.021026669070124626, -0.09919800609350204, 0.16530530154705048, 0.06881021708250046, -0.01632530987262726, 0.003567897016182542, 0.25087660551071167, 0.13047344982624054, 0.19022901356220245, 0.05210646986961365, -0.10279348492622375, 0.1299925297498703, 0.0346289686858654, -0.08642441034317017, -0.0108327092602849, -0.10180144011974335, 0.1383725255727768, -0.04707479476928711, 0.11273461580276489, -0.05730290710926056, -0.1109122559428215, 0.02422557957470417, -0.09941017627716064, 0.06845081597566605, 0.03602106124162674, 0.010663382709026337, -0.04515710473060608, 0.046576615422964096, 0.035100966691970825, -0.038903865963220596, -0.004279705695807934, -0.006266044918447733, -0.05231894552707672, 0.051750026643276215, 0.10759922862052917, 0.030918922275304794, 0.05550840124487877, 0.0006224259850569069, -0.018229058012366295, 0.0008649705559946597, -0.04589184373617172, -0.00046622753143310547, 0.04307509958744049, 0.04841083288192749, 0.08898212760686874, -0.15684320032596588, -0.07894528657197952, -0.012241411954164505, 0.06272382289171219, 0.010941749438643456, 0.0972387045621872, -0.011779814027249813, 0.05421774461865425, 0.00952292699366808, -0.023547278717160225, 0.0301270242780447, -0.09821558743715286, 0.08293663710355759, -0.01843639649450779, 0.08651518076658249, -0.1962188482284546, -0.002511328784748912, -0.044056978076696396, 0.011401196010410786, 0.06800658255815506, -0.007577955257147551, -0.11577794700860977, 0.12528707087039948, -0.007475119549781084, 0.07063999027013779, -0.09655427187681198, 0.043631210923194885, 0.02881775051355362, 0.07577390223741531, -0.09048143029212952, 0.009205114096403122, 0.11378499865531921, -0.13796021044254303, -0.16197781264781952, 0.09419431537389755, -0.024934299290180206, 0.026052890345454216, 0.03774550184607506, 0.15419644117355347, 0.1759786754846573, -0.18487754464149475, -0.021868275478482246, 0.06485199183225632, -0.019191324710845947, -0.06893941760063171, -0.013356192037463188, 0.10877415537834167, 0.01172556821256876, 0.03277556598186493, -0.029603956267237663, 0.12800702452659607, -0.031671565026044846, -0.09231150895357132, -0.03268851339817047, -0.08207445591688156, -0.06485965102910995, 0.049018494784832, -0.007418185938149691, -0.0501861572265625, 0.016374943777918816, -0.1511426568031311, 0.16824959218502045, 0.013847636990249157, 0.022200962528586388, -0.07397002726793289, 0.11120285838842392, 0.006596401799470186, 0.009322021156549454, 0.011013954877853394, -0.06556646525859833, -0.1033940389752388, 0.22552090883255005, 0.08449230343103409, 0.08758459985256195, 0.05866019055247307, -0.04553890600800514, -0.07011693716049194, 0.018830686807632446, 0.009109347127377987, -0.03968724608421326, 0.0237893033772707, -0.1066252812743187, 0.05005870386958122, -0.012247967533767223, 0.027040492743253708, -0.003593896282836795, -0.025677431374788284, 0.06008579954504967, 0.012265765108168125, -0.024247583001852036, 0.09210186451673508, 0.049028314650058746, -0.01748424395918846, -0.06692013889551163, 0.003179841209203005, 0.07519450783729553, -0.010885783471167088, -0.0826268196105957, 0.014300057664513588, -0.0013082769000902772, 0.034870073199272156, 0.2010059356689453, -0.22606626152992249, 0.04247531667351723, 0.024595454335212708, 0.0514104925096035, 0.03573035076260567, 0.0028112465515732765, -0.02878340147435665, 0.037528809159994125, -0.022923026233911514, 0.0686744675040245, -0.014002567157149315, 0.06957467645406723, -0.028863338753581047, -0.13540713489055634, -0.007740207947790623, -0.030799780040979385, 0.16569754481315613, -0.1702529937028885, 0.06443364173173904, 0.18929390609264374, -0.127278134226799, 0.13140560686588287, -0.003569653257727623, -0.011099886149168015, 0.012736392207443714, 0.038338255137205124, 0.000055581633205292746, 0.1030297577381134, -0.0746593102812767, -0.03169042244553566, 0.02393999509513378, -0.0862320140004158, 0.032722197473049164, -0.11988750845193863, -0.11095412820577621, -0.07348192483186722, -0.03473226726055145, -0.03856544569134712, 0.028185158967971802, -0.05325959622859955, 0.07419442385435104, -0.09282999485731125, -0.08247637003660202, -0.02261742390692234, -0.08555806428194046, 0.022564753890037537, 0.010740888305008411, -0.05912154167890549, -0.1403903216123581, -0.1179245114326477, -0.08435950428247452, -0.15436291694641113, -0.010585583746433258, 0.0678953230381012, -0.11294028908014297, -0.0416659340262413, 0.01945776678621769, -0.05430544167757034, 0.10142936557531357, -0.07548916339874268, 0.0252106711268425, 0.04877350106835365, -0.03055819496512413, -0.16973689198493958, 0.001019359682686627, -0.06513843685388565, -0.06676920503377914, 0.1515105962753296, -0.14619991183280945, 0.18324601650238037, -0.030856525525450706, 0.05203507840633392, 0.060171738266944885, 0.036921340972185135, 0.12388842552900314, -0.11158057302236557, 0.0742458701133728, 0.17965127527713776, 0.03772910311818123, 0.08115017414093018, 0.11908208578824997, 0.08627743273973465, -0.10590525716543198, 0.0370810404419899, 0.07628314197063446, -0.09592080861330032, -0.07889098674058914, -0.05422945320606232, -0.11957541108131409, -0.06528066098690033, 0.05373134836554527, 0.05979841947555542, 0.049773093312978745, 0.12311960756778717, -0.05427895486354828, -0.005046121310442686, 0.09673302620649338, 0.049539364874362946, 0.0804583802819252, 0.018023835495114326, 0.05626511946320534, -0.14840179681777954, -0.04441061615943909, 0.15947261452674866, 0.22133132815361023, 0.23700174689292908, 0.027093619108200073, 0.07000044733285904, 0.12443451583385468, 0.08602748811244965, 0.0988495722413063, 0.05183251574635506, 0.0008541532442905009, 0.01574838161468506, -0.06876392662525177, -0.05071289464831352, 0.01448812335729599, 0.0039816866628825665, -0.05281006917357445, -0.14070048928260803, 0.10612146556377411, 0.004193281289190054, 0.0942210704088211, 0.14011330902576447, 0.044563546776771545, -0.09902551770210266, 0.153592050075531, 0.0938563421368599, 0.08623547106981277, -0.05651276186108589, 0.12972599267959595, 0.0508759468793869, -0.002721302444115281, 0.16620087623596191, 0.02666495367884636, 0.15001331269741058, -0.032648030668497086, -0.07422088831663132, -0.0684850737452507, -0.047062113881111145, 0.008926990441977978, 0.03435354307293892, -0.22401084005832672, 0.09837290644645691, 0.055104129016399384, 0.016309361904859543, -0.0036485930904746056, -0.057175759226083755, 0.19359730184078217, 0.16339905560016632, 0.0735902488231659, 0.030965888872742653, -0.04498941823840141, -0.010283795185387135, -0.08153855800628662, 0.05684398114681244, 0.01720157079398632, 0.07864569872617722, -0.042286623269319534, -0.10138611495494843, -0.021299943327903748, -0.0033673872239887714, 0.030138913542032242, -0.08618444949388504, -0.11090546101331711, -0.04778210073709488, 0.25210991501808167, -0.06286277621984482, 0.04987110197544098, 0.05951806530356407, 0.01178919430822134, -0.02831118553876877, 0.026320917531847954, -0.034630026668310165, -0.015652067959308624, -0.03985747694969177, -0.00005162559318705462, 0.00635130237787962, -0.05351952090859413, -0.056742336601018906, -0.03334497660398483, -0.10133863985538483, -0.1086115762591362, -0.00017881947860587388, -0.0496131032705307, 0.012082889676094055, -0.02341926284134388, 0.006303705275058746, -0.09745196253061295, -0.03283001109957695, 0.020448273047804832, 0.031712856143713, -0.07492649555206299, -0.1377701312303543, -0.011417361907660961, -0.012293225154280663, -0.058314938098192215, 0.03402222692966461, -0.11896748840808868, -0.09052261710166931, -0.05969247221946716, -0.029560308903455734, 0.12400981038808823, 0.22209416329860687, -0.026320813223719597, 0.00010144849511561915, 0.16263747215270996, -0.09500155597925186, -0.32313501834869385, -0.17386069893836975, -0.16236449778079987, -0.10054721683263779, 0.02830573357641697, -0.08112594485282898, 0.023799538612365723, 0.07722396403551102, -0.036234255880117416, 0.19759750366210938, -0.19585031270980835, -0.09705756604671478, 0.09090310335159302, 0.08848001807928085, 0.3091769814491272, -0.24720996618270874, 0.016320345923304558, -0.11095962673425674, -0.024994902312755585, 0.009704467840492725, -0.07977664470672607, 0.1198471188545227, 0.04006368666887283, 0.08164352923631668, -0.008350485004484653, -0.005402110517024994, 0.14678046107292175, -0.06914971768856049, 0.1375904083251953, -0.11880306899547577, -0.09380847215652466, 0.20366613566875458, -0.03358302637934685, 0.008630525320768356, -0.21244783699512482, -0.0302894189953804, -0.0292899738997221, 0.0404374860227108, -0.006670600268989801, 0.05469777435064316, -0.007854795083403587, -0.013498473912477493, -0.1304553747177124, -0.020655058324337006, -0.021219387650489807, 0.06564416736364365, 0.23646794259548187, -0.05671137571334839, -0.06338734179735184, 0.037156615406274796, -0.01056975219398737, 0.10212421417236328, 0.01956065557897091, -0.059308525174856186, -0.04536515474319458, 0.0959838479757309, -0.2001415491104126, 0.05589061975479126, 0.006935501005500555, -0.003939471207559109, 0.017091970890760422, 0.01218560989946127, 0.02707228623330593, 0.1173722892999649, 0.17621615529060364, -0.0006616413011215627, -0.03079609014093876, -0.018501419574022293, 0.024370163679122925, 0.12237142026424408, -0.01681392639875412, 0.10883966833353043, 0.018161624670028687, 0.04141157120466232, 0.008951064199209213, 0.057095013558864594, -0.08499231189489365, -0.09114338457584381, 0.09796761721372604, -0.043763164430856705, -0.08411835879087448, 0.0917547345161438, 0.05312146246433258, 0.07369878888130188, 0.004069997929036617, 0.04660937935113907, 0.012868506833910942, -0.12751278281211853, 0.013649185188114643, 0.2133145034313202, -0.07374255359172821, -0.06967543810606003, -0.0637020692229271, 0.014227595180273056, -0.11807165294885635, 0.07378625124692917, 0.037580572068691254, -0.03020407445728779, 0.11535385251045227, -0.04314490780234337, -0.029589835554361343, 0.009077458642423153, -0.0636904239654541, 0.03479797765612602, -0.14541089534759521, -0.20322676002979279, 0.05319445580244064, -0.00221730163320899, -0.0662674754858017, -0.08980116993188858, -0.0884285718202591, 0.06303201615810394, -0.16956767439842224, 0.13872765004634857, -0.06728778779506683, 0.0632406696677208, -0.03534045070409775, -0.05654488503932953, -0.11103196442127228, -0.01705099083483219, -0.0513334721326828, -0.026280108839273453, 0.06221969053149223, 0.014558789320290089, -0.12338781356811523, -0.11380045861005783, 0.059360016137361526, -0.0031541287899017334, -0.006883625872433186, 0.015748295933008194, -0.06889615207910538, 0.021964669227600098, -0.22898820042610168, -0.06957542896270752, 0.07775277644395828, 0.04471171274781227, -0.0941888689994812, 0.12392786890268326, 0.043849363923072815, -0.025139348581433296, 0.039906758815050125, 0.006336895748972893, 0.18185019493103027, -0.07410722225904465, 0.03292333707213402, -0.12637411057949066, -0.15872034430503845, -0.027862168848514557, 0.030686862766742706, 0.22431592643260956, 0.08607737720012665, 0.12330509722232819, -0.0561673678457737, 0.01827516034245491, -0.01249533612281084, 0.0723988339304924, 0.01472957618534565, -0.10514522343873978, -0.05141226574778557, -0.17491501569747925, -0.06608253717422485, -0.06343408674001694, 0.16295884549617767, 0.030181534588336945, -0.1420915275812149, -0.0020223865285515785, 0.11290683597326279, -0.19348107278347015, -0.016213394701480865, 0.1684330552816391, -0.043625857681035995, 0.025790195912122726, -0.16168005764484406, 0.02248198911547661, 0.07638459652662277, -0.04138574376702309, 0.0021681212820112705, 0.12271929532289505, -0.0034334566444158554, -0.002932596253231168, 0.03145763278007507, -0.0373862162232399, 0.08413734287023544, -0.061803705990314484, 0.05290756747126579, 0.0027920911088585854, -0.0458851233124733, -0.11167335510253906, 0.19065865874290466, -0.013050098903477192, 0.014779459685087204, -0.058366380631923676, -0.0016104582464322448, -0.09725025296211243, -0.0982813760638237, -0.06700623035430908, -0.13395912945270538, 0.07289042323827744, -0.05599640682339668, 0.010922395624220371, 0.0034242654219269753, 0.01991122029721737, -0.07342718541622162, 0.017344320192933083, -0.19011561572551727, -0.041299089789390564, 0.013832799158990383, -0.017606152221560478, -0.025451598688960075, -0.044770944863557816, -0.04470445215702057, 0.024312706664204597, -0.055440161377191544, -0.06233828514814377, 0.059830907732248306, 0.08822806924581528, 0.06428615748882294, -0.1562616378068924, -0.10440576821565628, -0.0739845484495163, 0.0341251939535141, 0.07723052054643631, 0.1772412806749344, 0.03665601834654808, -0.005460115149617195, 0.0442059300839901, 0.13290272653102875, 0.011363182216882706, -0.09162779897451401, -0.0712311714887619, -0.1223519816994667, -0.1367352455854416, -0.017801618203520775, -0.0647062212228775, -0.024951070547103882, 0.013491439633071423, 0.23482854664325714, 0.19399024546146393, -0.15535368025302887, 0.03944678604602814, -0.07574691623449326, 0.04021822288632393, -0.03410802781581879, 0.15896648168563843, 0.05210955813527107, 0.14654022455215454, -0.029316985979676247, -0.037061478942632675, -0.06332151591777802, 0.019872330129146576, -0.10562392324209213, 0.03170350566506386, -0.013178028166294098, -0.06854262948036194, -0.06424710154533386, 0.09906546771526337, -0.11733246594667435, 0.06616465747356415, 0.1822943240404129, -0.1422264128923416, -0.016360152512788773, -0.038761284202337265, 0.05596889555454254, 0.11282722651958466, 0.00966635998338461, -0.07639224827289581, -0.01918763481080532, 0.007720941677689552, 0.023644311353564262, -0.16103577613830566, -0.1089705303311348, -0.0003122957714367658, -0.1254398375749588, 0.1321529746055603, -0.003089979523792863, -0.0010510693537071347, 0.03567538037896156, -0.07136029750108719, 0.0006731287576258183, 0.1717679351568222, 0.021893374621868134, -0.03143070638179779, -0.025080015882849693, -0.05971817672252655, -0.10382658988237381, 0.07531052082777023, 0.09018315374851227, 0.051399774849414825, -0.0029697646386921406, 0.15699253976345062, -0.021047117188572884, -0.033303506672382355, 0.13478034734725952, -0.16822503507137299, 0.09625657647848129, 0.004777473863214254, -0.016702748835086823, -0.06862062960863113, -0.04317755624651909, 0.04024219885468483, 0.08220448344945908, -0.16708873212337494, -0.04375625401735306, 0.06125885248184204, -0.09209781140089035, 0.055131398141384125, 0.04696342721581459, -0.0857926607131958, 0.016861263662576675, -0.11365381628274918, 0.0011050661560148, -0.0998060330748558, 0.04820295423269272, 0.19907107949256897, -0.03289218991994858, 0.010038448497653008, -0.1460357904434204, 0.05437568202614784, -0.037007976323366165, -0.03611326590180397, -0.07533301413059235 ]
null
null
transformers
## DAVinCI-42dot_LLM-PLM-1.3B-v1.1 This model is a fine-tuned version of [42dot/42dot_LLM-PLM-1.3B](https://huggingface.co/42dot/42dot_LLM-PLM-1.3B) on a custom dataset. ### Model description More information needed ### Intended uses & limitations More information needed ### Training and evaluation data More information needed ### Training procedure ### Training hyperparameters The following hyperparameters were used during training: * learning_rate: 2e-05 * train_batch_size: 24 * eval_batch_size: 8 * seed: 42 * gradient_accumulation_steps: 4 * total_train_batch_size: 96 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr_scheduler_type: linear * num_epochs: 1.0 * mixed_precision_training: Native AMP ### Training results ### Framework versions * Transformers 4.36.2 * Pytorch 2.1.2+cu121 * Datasets 2.0.0 * Tokenizers 0.15.0
{"license": "cc-by-nc-4.0"}
text-generation
jungyuko/DAVinCI-42dot_LLM-PLM-1.3B-v1.1
[ "transformers", "safetensors", "llama", "text-generation", "license:cc-by-nc-4.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T01:58:16+00:00
[]
[]
TAGS #transformers #safetensors #llama #text-generation #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
## DAVinCI-42dot_LLM-PLM-1.3B-v1.1 This model is a fine-tuned version of 42dot/42dot_LLM-PLM-1.3B on a custom dataset. ### Model description More information needed ### Intended uses & limitations More information needed ### Training and evaluation data More information needed ### Training procedure ### Training hyperparameters The following hyperparameters were used during training: * learning_rate: 2e-05 * train_batch_size: 24 * eval_batch_size: 8 * seed: 42 * gradient_accumulation_steps: 4 * total_train_batch_size: 96 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr_scheduler_type: linear * num_epochs: 1.0 * mixed_precision_training: Native AMP ### Training results ### Framework versions * Transformers 4.36.2 * Pytorch 2.1.2+cu121 * Datasets 2.0.0 * Tokenizers 0.15.0
[ "## DAVinCI-42dot_LLM-PLM-1.3B-v1.1\n\nThis model is a fine-tuned version of 42dot/42dot_LLM-PLM-1.3B on a custom dataset.", "### Model description\nMore information needed", "### Intended uses & limitations\nMore information needed", "### Training and evaluation data\nMore information needed", "### Training procedure", "### Training hyperparameters\nThe following hyperparameters were used during training:\n* learning_rate: 2e-05\n* train_batch_size: 24\n* eval_batch_size: 8\n* seed: 42\n* gradient_accumulation_steps: 4\n* total_train_batch_size: 96\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr_scheduler_type: linear\n* num_epochs: 1.0\n* mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n* Transformers 4.36.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.0.0\n* Tokenizers 0.15.0" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## DAVinCI-42dot_LLM-PLM-1.3B-v1.1\n\nThis model is a fine-tuned version of 42dot/42dot_LLM-PLM-1.3B on a custom dataset.", "### Model description\nMore information needed", "### Intended uses & limitations\nMore information needed", "### Training and evaluation data\nMore information needed", "### Training procedure", "### Training hyperparameters\nThe following hyperparameters were used during training:\n* learning_rate: 2e-05\n* train_batch_size: 24\n* eval_batch_size: 8\n* seed: 42\n* gradient_accumulation_steps: 4\n* total_train_batch_size: 96\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr_scheduler_type: linear\n* num_epochs: 1.0\n* mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n* Transformers 4.36.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.0.0\n* Tokenizers 0.15.0" ]
[ 58, 49, 7, 13, 9, 4, 126, 4, 33 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## DAVinCI-42dot_LLM-PLM-1.3B-v1.1\n\nThis model is a fine-tuned version of 42dot/42dot_LLM-PLM-1.3B on a custom dataset.### Model description\nMore information needed### Intended uses & limitations\nMore information needed### Training and evaluation data\nMore information needed### Training procedure### Training hyperparameters\nThe following hyperparameters were used during training:\n* learning_rate: 2e-05\n* train_batch_size: 24\n* eval_batch_size: 8\n* seed: 42\n* gradient_accumulation_steps: 4\n* total_train_batch_size: 96\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr_scheduler_type: linear\n* num_epochs: 1.0\n* mixed_precision_training: Native AMP### Training results### Framework versions\n* Transformers 4.36.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.0.0\n* Tokenizers 0.15.0" ]
[ -0.08858167380094528, 0.08803413063287735, -0.003813729388639331, 0.08055642992258072, 0.13998404145240784, 0.030931711196899414, 0.11715347319841385, 0.1280614137649536, -0.1201416403055191, 0.06240842118859291, 0.042381852865219116, 0.06709027290344238, 0.0638263076543808, 0.12035293877124786, -0.044051505625247955, -0.26000478863716125, 0.007098687347024679, -0.016774078831076622, -0.09147386252880096, 0.11453507840633392, 0.10583250224590302, -0.10764951258897781, 0.05884683132171631, 0.011858166195452213, -0.1257096529006958, 0.018875859677791595, -0.009573042392730713, -0.03804760053753853, 0.10989966988563538, 0.0014662387548014522, 0.12416771054267883, 0.026939086616039276, 0.1505536437034607, -0.21314989030361176, -0.0008558668778277934, 0.09333062171936035, 0.031091172248125076, 0.07413846254348755, 0.08662702143192291, 0.010127536021173, 0.09089017659425735, -0.13601654767990112, 0.08902563899755478, 0.020026320591568947, -0.10148856788873672, -0.21088054776191711, -0.1102643832564354, 0.04918772354722023, 0.12210451811552048, 0.07432954758405685, 0.012785892933607101, 0.10032862424850464, -0.04950401186943054, 0.07986772805452347, 0.21152086555957794, -0.27444806694984436, -0.06555204838514328, 0.07077275961637497, 0.05620536953210831, 0.05801345780491829, -0.10054879635572433, -0.008878080174326897, 0.02706865221261978, 0.02229374833405018, 0.1302197426557541, -0.012882155366241932, -0.04082031920552254, -0.023211129009723663, -0.1299469918012619, -0.04197869449853897, 0.09557542204856873, 0.02992362342774868, -0.05177021771669388, -0.06634321808815002, -0.07498277723789215, -0.13558827340602875, -0.03571533039212227, -0.06416452676057816, 0.0543924979865551, -0.0246741846203804, -0.03859893977642059, -0.05464572831988335, -0.07940550148487091, -0.08317545056343079, -0.0014239330776035786, 0.1278870850801468, 0.04304907098412514, 0.006542198825627565, -0.03163987025618553, 0.11207062005996704, -0.007803789805620909, -0.13704347610473633, -0.0027479776181280613, 0.017182325944304466, -0.08038701117038727, -0.06554614752531052, -0.03886006772518158, -0.05668320879340172, -0.03001980483531952, 0.1407257467508316, -0.08679939806461334, 0.06314648687839508, -0.02775549329817295, 0.00603556539863348, -0.07508797943592072, 0.14728151261806488, -0.05745713412761688, -0.04266947880387306, 0.005406794138252735, 0.09589606523513794, 0.02276480942964554, 0.0034826810006052256, -0.07414949685335159, -0.0276810135692358, 0.10151287913322449, 0.06520402431488037, -0.052836399525403976, 0.03708619624376297, -0.03707776218652725, -0.032707586884498596, 0.06012839078903198, -0.12493399530649185, 0.050420429557561874, -0.003187501337379217, -0.08615098148584366, -0.03903648257255554, 0.029267998412251472, 0.019546959549188614, -0.00425588246434927, 0.14869928359985352, -0.061466727405786514, 0.021629543974995613, -0.09878546744585037, -0.085026316344738, 0.011785108596086502, -0.027754656970500946, -0.023412266746163368, -0.05963016673922539, -0.24085555970668793, -0.04094073921442032, 0.053987666964530945, -0.06120559573173523, -0.012250721454620361, -0.024233411997556686, -0.07561192661523819, 0.01382183376699686, -0.014679751358926296, 0.14274470508098602, -0.04448216035962105, 0.08345868438482285, 0.009510824456810951, 0.054612599313259125, 0.035982705652713776, 0.03326738625764847, -0.07746089994907379, 0.034659650176763535, -0.19429084658622742, 0.057644423097372055, -0.05388892441987991, 0.0051742298528552055, -0.11931836605072021, -0.10487499833106995, -0.029984837397933006, -0.028383173048496246, 0.08159175515174866, 0.10807082056999207, -0.21231994032859802, -0.016306094825267792, 0.18578004837036133, -0.10472824424505234, -0.07272656261920929, 0.10101676732301712, -0.04520820081233978, 0.03561550751328468, 0.05140961334109306, 0.12890870869159698, 0.09759747982025146, -0.1451362818479538, -0.014505078084766865, -0.04616619274020195, 0.05444781482219696, 0.03886345028877258, 0.0641525536775589, -0.008190993219614029, 0.07129983603954315, -0.011340919882059097, -0.05992741137742996, -0.0202550757676363, -0.07806151360273361, -0.08554352074861526, -0.0539260171353817, -0.0793338343501091, 0.04524174705147743, 0.043305691331624985, 0.03277101367712021, -0.07920140027999878, -0.11278839409351349, 0.08731644600629807, 0.11778250336647034, -0.05488191545009613, 0.0356423556804657, -0.07166378945112228, 0.012945287860929966, -0.03426065668463707, -0.025636842474341393, -0.19697806239128113, -0.08979887515306473, 0.03438170254230499, -0.05567265301942825, 0.013750328682363033, 0.009120824746787548, 0.0738813653588295, 0.07786006480455399, -0.08003360778093338, -0.033169738948345184, -0.08412697166204453, -0.0031507997773587704, -0.09445967525243759, -0.18630951642990112, -0.06093123182654381, -0.025866732001304626, 0.11991303414106369, -0.20132328569889069, 0.020914766937494278, 0.004894716665148735, 0.1469895988702774, 0.0303646307438612, -0.046565424650907516, -0.02001451514661312, 0.043120987713336945, -0.009244025684893131, -0.09222467988729477, 0.039118461310863495, -0.007019840646535158, -0.04568857327103615, -0.0437023788690567, -0.1468350887298584, 0.06405457854270935, 0.08747149258852005, 0.017514776438474655, -0.09420355409383774, -0.002591045806184411, -0.048946622759103775, -0.0518859438598156, -0.0817529484629631, -0.015908941626548767, 0.18263699114322662, 0.017286090180277824, 0.13200703263282776, -0.08103588968515396, -0.05323883146047592, 0.000646560569293797, 0.003886188380420208, 0.008676189929246902, 0.05692758411169052, 0.05055084079504013, -0.11375114321708679, 0.08805470913648605, 0.12151453644037247, -0.049556512385606766, 0.11420177668333054, -0.04480105638504028, -0.07901531457901001, -0.035791825503110886, 0.018718251958489418, 0.009890537708997726, 0.10421418398618698, -0.051747072488069534, 0.021340133622288704, 0.014829080551862717, 0.04169574752449989, 0.018876470625400543, -0.19450825452804565, 0.002605293644592166, 0.03982851281762123, -0.07640001177787781, 0.008152613416314125, -0.023386148735880852, 0.0304733794182539, 0.09813352674245834, 0.015804212540388107, -0.04564109817147255, 0.010594247840344906, -0.02074994333088398, -0.07583130896091461, 0.1744602471590042, -0.09925612807273865, -0.14384162425994873, -0.11176616698503494, 0.021635539829730988, -0.03094814345240593, -0.01645825430750847, 0.012435796670615673, -0.0738474428653717, -0.05210704728960991, -0.0858810693025589, -0.006451253313571215, -0.021939055994153023, -0.011980924755334854, 0.05845875293016434, 0.01291718054562807, 0.08113112300634384, -0.13024066388607025, 0.0036875526420772076, -0.014592332765460014, -0.09384340792894363, -0.00004864295260631479, 0.07467958331108093, 0.08005160093307495, 0.135328009724617, -0.03817780688405037, 0.01998709701001644, -0.021069293841719627, 0.1794436275959015, -0.07603751868009567, 0.007454450707882643, 0.16054175794124603, -0.0077966246753931046, 0.07195832580327988, 0.14380259811878204, 0.045050594955682755, -0.0955008938908577, 0.02473384328186512, 0.0613403394818306, -0.019841628149151802, -0.23566296696662903, -0.040124207735061646, -0.01786121167242527, -0.05834067612886429, 0.07693463563919067, 0.03174838423728943, 0.01864277385175228, 0.024365030229091644, -0.01667330227792263, 0.03366846218705177, 0.006365569774061441, 0.0904354602098465, 0.09273830056190491, 0.047753557562828064, 0.10560072958469391, -0.01890498958528042, -0.030002892017364502, 0.043015703558921814, 0.008759299293160439, 0.23842628300189972, -0.012014694511890411, 0.09353072196245193, 0.03157961741089821, 0.11977148801088333, -0.028446950018405914, 0.059733446687459946, 0.002501277020201087, -0.009474234655499458, -0.0023485671263188124, -0.06496452540159225, -0.019939832389354706, 0.03385340794920921, -0.06021364778280258, 0.046374496072530746, -0.084567129611969, 0.02160545438528061, 0.027930596843361855, 0.265285849571228, 0.04712243750691414, -0.2883572578430176, -0.0628005787730217, 0.018361324444413185, -0.03909801319241524, -0.04420386999845505, -0.010192980989813805, 0.10693582147359848, -0.11305564641952515, 0.06560172140598297, -0.09157579392194748, 0.08259282261133194, -0.04533660411834717, 0.008070319890975952, 0.07719289511442184, 0.11166340112686157, 0.007408986799418926, 0.05769697204232216, -0.21573825180530548, 0.2305997908115387, 0.02210063301026821, 0.12594255805015564, -0.04975642263889313, 0.036492716521024704, 0.029097840189933777, 0.07135778665542603, 0.08499559760093689, -0.01566329225897789, -0.08152943104505539, -0.16470803320407867, -0.08167993277311325, 0.007874843664467335, 0.14694999158382416, -0.037718843668699265, 0.10443012416362762, -0.04367249459028244, -0.013410917483270168, 0.05477265641093254, -0.055543892085552216, -0.16685862839221954, -0.07657234370708466, 0.03286154568195343, -0.00381856644526124, 0.008044525980949402, -0.08612925559282303, -0.09034489840269089, -0.049630746245384216, 0.18742239475250244, 0.016357041895389557, -0.042358703911304474, -0.13610605895519257, 0.05551834776997566, 0.14078888297080994, -0.061940740793943405, 0.04243623465299606, 0.03297296538949013, 0.11530639976263046, 0.023792942985892296, -0.06598193198442459, 0.06348898261785507, -0.08394467085599899, -0.19798238575458527, -0.06996539235115051, 0.13665468990802765, 0.028043607249855995, 0.0475303940474987, 0.018390802666544914, 0.03545742481946945, 0.03085094504058361, -0.08419453352689743, 0.005909975152462721, 0.04157024249434471, 0.11196764558553696, 0.05167733132839203, -0.06749099493026733, -0.0035668143536895514, -0.03591620922088623, -0.028797123581171036, 0.1034477949142456, 0.23582717776298523, -0.08622831851243973, 0.05901552364230156, 0.03919904679059982, -0.08119681477546692, -0.17084167897701263, 0.04701916128396988, 0.10853256285190582, 0.0003461376763880253, 0.07719629257917404, -0.16332636773586273, 0.06615158170461655, 0.10335440188646317, -0.04432567581534386, 0.06445743143558502, -0.3317919373512268, -0.14316071569919586, 0.03431381657719612, 0.1203572228550911, 0.018070125952363014, -0.1563561111688614, -0.04876631498336792, -0.02384795993566513, -0.1305411458015442, 0.120026133954525, -0.07625516504049301, 0.10215935111045837, -0.008764225989580154, 0.06978616863489151, 0.02285083755850792, -0.0459352545440197, 0.13624714314937592, 0.010832163505256176, 0.09097721427679062, -0.048795752227306366, 0.04154004529118538, 0.08825729042291641, -0.06440091878175735, 0.04302988201379776, -0.016109049320220947, 0.05565707013010979, -0.11006174981594086, -0.013910142704844475, -0.07043670862913132, 0.03823138028383255, -0.05984337627887726, -0.05787124112248421, -0.04829041659832001, 0.057134807109832764, 0.06218415126204491, -0.03645087033510208, 0.06601014733314514, 0.025172358378767967, 0.11672032624483109, 0.14120717346668243, 0.10470711439847946, 0.020940016955137253, -0.06996061652898788, 0.012740910984575748, -0.011487183161079884, 0.044137004762887955, -0.13060790300369263, 0.04777601733803749, 0.12443619966506958, 0.031636793166399, 0.13326960802078247, 0.0384206660091877, -0.05739079788327217, 0.002709029009565711, 0.053429801017045975, -0.11638353019952774, -0.1311386525630951, 0.012381254695355892, -0.018744071945548058, -0.11889249086380005, 0.02481520175933838, 0.11411461234092712, -0.06721725314855576, -0.007945591583848, 0.0006214787717908621, 0.03727481886744499, -0.013995960354804993, 0.20620086789131165, 0.03273986652493477, 0.06376578658819199, -0.07220803201198578, 0.11330568045377731, 0.06484425812959671, -0.07774648070335388, 0.02201097086071968, 0.07733576744794846, -0.07799984514713287, -0.01642698422074318, 0.059994425624608994, 0.13974922895431519, -0.011252542957663536, -0.043391693383455276, -0.10521510243415833, -0.09234460443258286, 0.03841206058859825, 0.12673358619213104, 0.043660346418619156, 0.0004364110645838082, -0.025255374610424042, 0.038856521248817444, -0.1370563805103302, 0.10760647058486938, 0.07153968513011932, 0.08480316400527954, -0.14072000980377197, 0.16159825026988983, -0.004774218425154686, 0.02089822106063366, -0.012645791284739971, 0.048799000680446625, -0.061633821576833725, -0.016022799536585808, -0.13024014234542847, -0.012851503677666187, -0.022083329036831856, -0.00945600401610136, -0.014028479345142841, -0.053856488317251205, -0.03831791132688522, 0.02752879075706005, -0.07738982886075974, -0.059875793755054474, -0.010680961422622204, 0.02768636681139469, -0.14634528756141663, -0.03940358757972717, 0.024570392444729805, -0.08687896281480789, 0.07344075292348862, 0.03826656565070152, 0.042308393865823746, 0.03537144884467125, -0.10465862601995468, 0.008922005072236061, 0.04269768297672272, 0.015001312829554081, 0.04212133586406708, -0.11892273277044296, -0.006070659961551428, -0.02708832174539566, 0.039920151233673096, 0.015004348941147327, 0.06611224263906479, -0.1321827918291092, 0.012295239605009556, -0.06621848791837692, -0.035431332886219025, -0.038861650973558426, 0.03635477274656296, 0.09878775477409363, 0.028500264510512352, 0.16881722211837769, -0.0775122344493866, 0.05398513749241829, -0.2329895794391632, -0.02176547981798649, -0.007336677052080631, -0.04676941782236099, -0.06712089478969574, -0.03582366183400154, 0.07415930926799774, -0.06597551703453064, 0.08971650153398514, 0.01677750237286091, 0.09641703963279724, 0.040267348289489746, -0.06308580935001373, -0.014329480938613415, 0.01898219808936119, 0.1442192643880844, 0.04903111606836319, -0.025207966566085815, 0.07494037598371506, -0.006870932411402464, 0.059824734926223755, 0.01304536685347557, 0.2002153843641281, 0.15679992735385895, -0.029212331399321556, 0.07389138638973236, 0.059325993061065674, -0.1171724796295166, -0.17525231838226318, 0.05289386957883835, -0.035383936017751694, 0.08742093294858932, -0.04684535786509514, 0.16847467422485352, 0.09618782997131348, -0.19707490503787994, 0.06245655193924904, -0.04185283184051514, -0.09374100714921951, -0.11766517907381058, -0.022258922457695007, -0.08571962267160416, -0.11563637852668762, 0.010009396821260452, -0.12031031399965286, 0.07688774168491364, 0.08318465948104858, -0.0023513392079621553, 0.007832909934222698, 0.15503814816474915, -0.02515990659594536, 0.009569382295012474, 0.06173375993967056, 0.030669288709759712, 0.014554736204445362, -0.01912567764520645, -0.07008829712867737, 0.014650176279246807, 0.00913121085613966, 0.069877989590168, -0.0318024717271328, -0.02619866281747818, 0.011793933808803558, -0.0027031460776925087, -0.08349438011646271, 0.02090701460838318, 0.017833709716796875, 0.07031994313001633, 0.055798593908548355, 0.04208732023835182, 0.012256002053618431, -0.035477593541145325, 0.2639877498149872, -0.08918865025043488, -0.08003868907690048, -0.10014209151268005, 0.22320325672626495, 0.04018739238381386, -0.0011427365243434906, 0.05654432252049446, -0.11105996370315552, 0.004207163117825985, 0.1516944169998169, 0.12210948765277863, -0.06405654549598694, -0.011016534641385078, -0.0030067849438637495, -0.01585354469716549, -0.03348856791853905, 0.11688841134309769, 0.09676901251077652, 0.04416608437895775, -0.06316117942333221, -0.006369336508214474, -0.014567487873136997, -0.02902671881020069, -0.071058489382267, 0.06700153648853302, -0.011943303979933262, 0.02115575224161148, -0.01519464049488306, 0.08351334929466248, 0.021036552265286446, -0.2053770124912262, 0.06448632478713989, -0.18355943262577057, -0.18239432573318481, 0.0031272328924387693, 0.08591899275779724, -0.01734345033764839, 0.05882229655981064, -0.009652910754084587, -0.012621561996638775, 0.11427713930606842, -0.016747789457440376, -0.026724280789494514, -0.092164546251297, 0.06882834434509277, -0.11125855892896652, 0.22797805070877075, -0.009037640877068043, 0.07197516411542892, 0.10173071175813675, 0.01180777233093977, -0.12381511926651001, 0.043687161058187485, 0.062307242304086685, -0.08473492413759232, 0.03451911360025406, 0.18344666063785553, -0.051314905285835266, 0.06020456925034523, 0.04431688413023949, -0.1343661993741989, 0.0014360020868480206, -0.0012485058978199959, -0.04626612365245819, -0.06553877145051956, -0.003525272710248828, -0.05532208830118179, 0.15139837563037872, 0.2079465389251709, -0.026755575090646744, 0.04742530733346939, -0.066594697535038, 0.02632119134068489, 0.0526796355843544, 0.11344296485185623, -0.03612031787633896, -0.23388002812862396, 0.021656567230820656, 0.072688989341259, 0.031481944024562836, -0.2057090401649475, -0.10291953384876251, 0.0690169632434845, -0.06122999265789986, -0.0854777842760086, 0.111505888402462, 0.05471235141158104, 0.03136532008647919, -0.02258353680372238, -0.11829345673322678, -0.0533570796251297, 0.16296851634979248, -0.18218867480754852, -0.05025436356663704 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
feature-extraction
furrutiav/bert_qa_extractor_cockatiel_2022_ulra_two_signal_z_value_mixtral_v2_it_107
[ "transformers", "safetensors", "bert", "feature-extraction", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
2024-02-15T02:01:43+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 39, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.052746038883924484, 0.20255789160728455, -0.0045078229159116745, 0.0248473659157753, 0.10497838258743286, 0.00675728265196085, 0.06521498411893845, 0.11486967653036118, -0.0023755673319101334, 0.12028469145298004, 0.027631845325231552, 0.08119397610425949, 0.12110675126314163, 0.15393014252185822, 0.005160121712833643, -0.24253977835178375, 0.05344875901937485, -0.09366832673549652, 0.004077504388988018, 0.11452110856771469, 0.1343945860862732, -0.10780399292707443, 0.08976872265338898, -0.00683097867295146, -0.01712046191096306, -0.015751034021377563, -0.07134060561656952, -0.06668227165937424, 0.05541034787893295, 0.07649129629135132, 0.0725555345416069, 0.010986946523189545, 0.07830587029457092, -0.2806258797645569, 0.014425364322960377, 0.08005264401435852, 0.0010765197221189737, 0.06795802712440491, 0.08151742070913315, -0.06789936870336533, 0.1251654475927353, -0.0605485662817955, 0.14059753715991974, 0.07639917731285095, -0.08928128331899643, -0.19590547680854797, -0.06669555604457855, 0.07481247186660767, 0.129872128367424, 0.05026249960064888, -0.02990107797086239, 0.1371748298406601, -0.09688840061426163, 0.00786701962351799, 0.12302009761333466, -0.07360870391130447, -0.05524582043290138, 0.031063849106431007, 0.10805318504571915, 0.09297362715005875, -0.11762315034866333, -0.008467874489724636, 0.029582185670733452, 0.022175652906298637, 0.08627551048994064, 0.015828849747776985, 0.1525639444589615, 0.041341137140989304, -0.14141254127025604, -0.0526716373860836, 0.09056255221366882, 0.03701045364141464, -0.050960201770067215, -0.23367193341255188, -0.026245610788464546, -0.012442239560186863, -0.03079850971698761, -0.04234880208969116, 0.053594592958688736, -0.03630254790186882, 0.07596245408058167, -0.007196845952421427, -0.07732249796390533, -0.031211229041218758, 0.05230424553155899, 0.06785056740045547, 0.018615471199154854, -0.006994647905230522, 0.019442738965153694, 0.11387838423252106, 0.07708574831485748, -0.13029205799102783, -0.07214002311229706, -0.0739525631070137, -0.09558356553316116, -0.04332297295331955, 0.03707554563879967, 0.07106684148311615, 0.04390906170010567, 0.20283061265945435, -0.017690327018499374, 0.046562306582927704, 0.0476159006357193, 0.005842953454703093, 0.07147589325904846, 0.10925443470478058, -0.06689215451478958, -0.14432233572006226, -0.06022803485393524, 0.08875485509634018, -0.009834992699325085, -0.03670760244131088, -0.049119677394628525, 0.04676154628396034, 0.03209913894534111, 0.11318106204271317, 0.08643888682126999, -0.003593706525862217, -0.0628826767206192, -0.042073074728250504, 0.22331053018569946, -0.14625342190265656, 0.043256524950265884, 0.007445589639246464, -0.0429743155837059, -0.0076383077539503574, 0.005870272871106863, 0.014089803211390972, -0.03238216042518616, 0.10351061820983887, -0.0778173878788948, -0.035906463861465454, -0.1116463914513588, -0.06868703663349152, 0.024910317733883858, 0.0025890374090522528, -0.018393149599432945, -0.04424213990569115, -0.11253650486469269, -0.051282741129398346, 0.0724339634180069, -0.07579848170280457, -0.05524555593729019, 0.009976830333471298, -0.04834962263703346, 0.0031978494953364134, 0.00010397454752819613, 0.11258035898208618, -0.03314845636487007, 0.025259260088205338, -0.04850656911730766, 0.06803499162197113, 0.10959596186876297, 0.038730688393116, -0.0804535374045372, 0.07286878675222397, -0.22788093984127045, 0.10223092138767242, -0.09346398711204529, 0.025767935439944267, -0.14578653872013092, -0.04199126362800598, 0.02854149229824543, 0.02887420728802681, -0.010361229069530964, 0.1268649846315384, -0.1982942521572113, -0.035082314163446426, 0.15190726518630981, -0.11336656659841537, -0.09347330778837204, 0.065653957426548, -0.05610617995262146, 0.11296144872903824, 0.04835578054189682, -0.019556574523448944, 0.06953749805688858, -0.1281629204750061, -0.04506009817123413, -0.021473335102200508, -0.008493004366755486, 0.14857245981693268, 0.06750676780939102, -0.05737153813242912, 0.07104712724685669, 0.02051553688943386, -0.037109848111867905, -0.03301886469125748, -0.03470754995942116, -0.09331934154033661, 0.009520708583295345, -0.07244295626878738, 0.03737799823284149, -0.02224314957857132, -0.08870045095682144, -0.030656753107905388, -0.17619828879833221, 0.043274905532598495, 0.08050142228603363, 0.008233942091464996, -0.021131468936800957, -0.09287237375974655, 0.02556683123111725, -0.009385489858686924, -0.021018607541918755, -0.1641797423362732, -0.044834475964307785, 0.04416196420788765, -0.1971662938594818, 0.023802341893315315, -0.03283040598034859, 0.05093098804354668, 0.03247829154133797, -0.04019762575626373, -0.005096070934087038, 0.0028117431793361902, 0.01809627003967762, -0.026984719559550285, -0.200385183095932, -0.031109308823943138, -0.029154371470212936, 0.1362139731645584, -0.22226740419864655, 0.028292208909988403, 0.07483648508787155, 0.13521188497543335, 0.0009690870065242052, -0.04426588490605354, 0.010693409480154514, -0.05366935580968857, -0.053671274334192276, -0.06512755900621414, -0.007102466654032469, -0.03287021815776825, -0.04422381520271301, 0.06460095942020416, -0.19425635039806366, -0.03641216829419136, 0.10608077049255371, 0.10164625942707062, -0.14719000458717346, -0.028969714418053627, -0.04096706584095955, -0.06081128865480423, -0.09094393998384476, -0.0630471333861351, 0.14371246099472046, 0.04861542955040932, 0.048413511365652084, -0.08624191582202911, -0.0630124881863594, 0.00895135197788477, 0.0006565740332007408, -0.03649118170142174, 0.08907787501811981, 0.08782777935266495, -0.10737399011850357, 0.08881597965955734, 0.08605224639177322, 0.06605713814496994, 0.10539878904819489, 0.001256609451957047, -0.10750970244407654, -0.029154706746339798, 0.005644100718200207, 0.01547710970044136, 0.14092515408992767, -0.044270921498537064, 0.04743899777531624, 0.05656488984823227, -0.027443327009677887, 0.01715722121298313, -0.10313762724399567, 0.02984124980866909, 0.046840768307447433, -0.010507673025131226, 0.012429861351847649, -0.03895113617181778, 0.025837475433945656, 0.08796556293964386, 0.03584056720137596, 0.027896199375391006, 0.0029043578542768955, -0.03437814116477966, -0.10392027348279953, 0.17429527640342712, -0.0878753736615181, -0.28357240557670593, -0.1356295943260193, -0.00747122336179018, 0.05167245492339134, -0.022715993225574493, 0.013256389647722244, -0.04903135821223259, -0.11467588692903519, -0.10348290205001831, 0.008818334899842739, 0.0437844917178154, -0.07700283080339432, -0.07256268709897995, 0.046553414314985275, 0.033613573759794235, -0.14174877107143402, 0.022300107404589653, 0.048012908548116684, -0.03855963796377182, -0.015413837507367134, 0.07170835882425308, 0.10258439928293228, 0.17387451231479645, -0.004228805657476187, -0.01945391111075878, 0.023280048742890358, 0.24459126591682434, -0.14296141266822815, 0.10647262632846832, 0.15432609617710114, -0.06630013138055801, 0.1025824174284935, 0.19176462292671204, 0.02610800787806511, -0.07571171224117279, 0.03370760753750801, 0.03715203329920769, -0.053104497492313385, -0.23274335265159607, -0.060641512274742126, 0.0011178229469805956, -0.06850682199001312, 0.09104112535715103, 0.08915619552135468, 0.11183936148881912, 0.0454646460711956, -0.08415863662958145, -0.06847929954528809, 0.019614145159721375, 0.10642454773187637, -0.03275766968727112, 0.007264797575771809, 0.09054313600063324, -0.04184457287192345, -0.005177726969122887, 0.10835286974906921, 0.007426192983984947, 0.1962665617465973, 0.031048519536852837, 0.15333782136440277, 0.07211130857467651, 0.0342402458190918, 0.026680786162614822, 0.025636766105890274, 0.023090654984116554, 0.009547512046992779, -0.01598707027733326, -0.08795502036809921, 0.027014199644327164, 0.13500221073627472, 0.07871367782354355, 0.029795078560709953, 0.020392734557390213, -0.0429922379553318, 0.062152985483407974, 0.15964233875274658, 0.006258485373109579, -0.2136749029159546, -0.03950631618499756, 0.08867984265089035, -0.0793125256896019, -0.1237078458070755, -0.02518491819500923, 0.03823186457157135, -0.1809074580669403, 0.04127289727330208, -0.01795332506299019, 0.11453432589769363, -0.11700457334518433, -0.028958700597286224, 0.039744846522808075, 0.08327627927064896, -0.03253408893942833, 0.07922478020191193, -0.1647184044122696, 0.1165376752614975, 0.012328862212598324, 0.05802180990576744, -0.11617794632911682, 0.09878876805305481, 0.012594180181622505, -0.009003117680549622, 0.16720694303512573, -0.0008162438753060997, -0.07339610159397125, -0.06517832726240158, -0.07867198437452316, -0.022016214206814766, 0.09116258472204208, -0.11647430807352066, 0.08271238952875137, -0.012302344664931297, -0.03819865360856056, 0.002976413816213608, -0.1073245257139206, -0.12343364208936691, -0.191313698887825, 0.05862122401595116, -0.11746024340391159, 0.00024363139527849853, -0.10003595799207687, -0.05551697313785553, -0.04721582680940628, 0.19990667700767517, -0.14306047558784485, -0.09675363451242447, -0.1526252180337906, -0.09468596428632736, 0.1679719239473343, -0.04768168181180954, 0.08716544508934021, -0.00014324963558465242, 0.22273695468902588, 0.00589721417054534, -0.010143720544874668, 0.07824880629777908, -0.08608578145503998, -0.17828822135925293, -0.07740302383899689, 0.12055730819702148, 0.12802201509475708, 0.05279289186000824, -0.012038013897836208, 0.020934196189045906, -0.036648161709308624, -0.11678951978683472, 0.003050430677831173, 0.1217387318611145, 0.05949230119585991, 0.039503831416368484, -0.002558275358751416, -0.10200468450784683, -0.07551230490207672, -0.0352395698428154, 0.02261841483414173, 0.18903005123138428, -0.08441178500652313, 0.15781226754188538, 0.13112787902355194, -0.05333179607987404, -0.21253353357315063, 0.030583804473280907, 0.043237145990133286, 0.004318034742027521, 0.0612679123878479, -0.17720702290534973, 0.08167627453804016, 0.025727098807692528, -0.05116020143032074, 0.15224720537662506, -0.16569727659225464, -0.15514664351940155, 0.0824643224477768, 0.05010354146361351, -0.22108957171440125, -0.12386278063058853, -0.0879128947854042, -0.06589758396148682, -0.1396872103214264, 0.08584427833557129, 0.014041651971638203, -0.0018043812597170472, 0.05013851076364517, 0.033740755170583725, 0.018914686515927315, -0.048698488622903824, 0.21615906059741974, -0.0022440196480602026, 0.03326340764760971, -0.07553089410066605, -0.10180798172950745, 0.06950566172599792, -0.05141735449433327, 0.08518881350755692, -0.03099823370575905, 0.005753061734139919, -0.08320630341768265, -0.057475052773952484, -0.05255331099033356, 0.03318103775382042, -0.08139406144618988, -0.10520965605974197, -0.06759276986122131, 0.09429939836263657, 0.09139011800289154, -0.03298058733344078, -0.04032526910305023, -0.08896728605031967, 0.039150089025497437, 0.20617929100990295, 0.17360219359397888, 0.05333937704563141, -0.10111589729785919, 0.002542630536481738, -0.01915728859603405, 0.040264517068862915, -0.21200114488601685, 0.04798245429992676, 0.04617756977677345, 0.024147402495145798, 0.12109645456075668, -0.0176423080265522, -0.1646004468202591, -0.047221194952726364, 0.0562983863055706, -0.03494611009955406, -0.20504815876483917, -0.01314060389995575, 0.04864202439785004, -0.18736153841018677, -0.06957933306694031, 0.016700902953743935, -0.014444489032030106, -0.027432914823293686, 0.013032985851168633, 0.06286440044641495, 0.025481918826699257, 0.10238313674926758, 0.05989401787519455, 0.1000840812921524, -0.112981878221035, 0.0795830711722374, 0.09043775498867035, -0.08344172686338425, 0.009394102729856968, 0.06964189559221268, -0.05280066654086113, -0.02294989861547947, 0.022772129625082016, 0.06757686287164688, -0.003049787599593401, -0.057536181062459946, -0.02079189568758011, -0.10809285193681717, 0.06586270034313202, 0.1269281655550003, 0.0400845967233181, -0.006831571459770203, 0.04905473813414574, 0.02419281378388405, -0.07880669087171555, 0.11321208626031876, 0.03362756222486496, 0.03722309693694115, -0.05989459529519081, -0.01674187369644642, 0.04316421225667, 0.005734616424888372, -0.02047782577574253, -0.025104478001594543, -0.05658029392361641, -0.013948953710496426, -0.18932224810123444, 0.014544147998094559, -0.07588981091976166, 0.005138450767844915, 0.014814606867730618, -0.040141742676496506, -0.018671197816729546, 0.012856033630669117, -0.08163223415613174, -0.05027473345398903, -0.0038707295898348093, 0.09766460955142975, -0.1400173306465149, 0.008230311796069145, 0.09175591170787811, -0.11852382868528366, 0.06848865002393723, -0.019968708977103233, -0.014717686921358109, 0.0038272906094789505, -0.1270400881767273, 0.04572216048836708, -0.004586559720337391, 0.02062096633017063, 0.04444560408592224, -0.17065683007240295, 0.004877567756921053, -0.0423397533595562, -0.0478336401283741, -0.015323328785598278, -0.08405033499002457, -0.11406292766332626, 0.10921793431043625, 0.002206311793997884, -0.08430022746324539, -0.010287429206073284, 0.04696008190512657, 0.10919637978076935, -0.03898061811923981, 0.124757781624794, 0.0047785635106265545, 0.06639395654201508, -0.18268363177776337, -0.024298490956425667, -0.014514438807964325, 0.007352736312896013, 0.027192458510398865, -0.016180848702788353, 0.04238643869757652, -0.01372526679188013, 0.2601816952228546, -0.021822240203619003, 0.07231466472148895, 0.0637383759021759, 0.042024899274110794, 0.016651110723614693, 0.08318763226270676, 0.06755662709474564, 0.016758481040596962, 0.004258559085428715, 0.02265608124434948, -0.03241465613245964, -0.016654497012495995, -0.15768693387508392, 0.07677853107452393, 0.14623822271823883, 0.08591317385435104, 0.007676990237087011, 0.06586159020662308, -0.10330242663621902, -0.10554943233728409, 0.08015866577625275, -0.03888537734746933, -0.0009790018666535616, -0.058588381856679916, 0.15355949103832245, 0.14971502125263214, -0.17422176897525787, 0.08231138437986374, -0.03791337087750435, -0.04883022606372833, -0.11436772346496582, -0.15839459002017975, -0.06608819216489792, -0.029153592884540558, -0.0041826991364359856, -0.05528274551033974, 0.06748054921627045, 0.10802645981311798, -0.0021057529374957085, -0.00038325722562149167, 0.09545762091875076, -0.026331622153520584, -0.01757199876010418, 0.03465426340699196, 0.04817976430058479, 0.033562518656253815, -0.04831063002347946, 0.020485511049628258, 0.004976877011358738, 0.03976510092616081, 0.05864322930574417, 0.023703020066022873, -0.03892989084124565, 0.014479226432740688, -0.01092575490474701, -0.1049860492348671, 0.022427968680858612, -0.029776830226182938, -0.07360642403364182, 0.13104131817817688, 0.029177764430642128, 0.019099419936537743, -0.03228067234158516, 0.20109383761882782, -0.07107947021722794, -0.06925153732299805, -0.14109766483306885, 0.10889512300491333, -0.03372858464717865, 0.06323269009590149, 0.058447178453207016, -0.1133023053407669, -0.002398417331278324, 0.1314154714345932, 0.133079394698143, -0.033533163368701935, 0.005780258681625128, 0.03008044883608818, 0.00756559893488884, -0.0482633113861084, 0.045497048646211624, 0.031092669814825058, 0.15440985560417175, -0.06949599832296371, 0.07780899107456207, 0.00008295764564536512, -0.08774317800998688, -0.036128852516412735, 0.1405542492866516, 0.006535779219120741, 0.03079606406390667, -0.06559351831674576, 0.10371401906013489, -0.07252706587314606, -0.23936228454113007, 0.045033879578113556, -0.07753164321184158, -0.15683837234973907, -0.013978141359984875, 0.02726292423903942, -0.009009851142764091, 0.02702206000685692, 0.0654432401061058, -0.06469112634658813, 0.161378413438797, 0.03472336754202843, -0.08781957626342773, -0.05673113837838173, 0.07957270741462708, -0.09192227572202682, 0.2958409786224365, 0.013188840821385384, 0.029593972489237785, 0.10327941924333572, -0.019989576190710068, -0.13285429775714874, 0.030561091378331184, 0.10066051781177521, -0.09982595592737198, 0.06684590131044388, 0.18159176409244537, -0.009470577351748943, 0.10021016746759415, 0.07437440752983093, -0.061603669077157974, 0.05807222053408623, -0.0826035663485527, -0.06770919263362885, -0.09389114379882812, 0.05970105528831482, -0.06468918174505234, 0.14543601870536804, 0.1228262409567833, -0.04243761673569679, -0.004415105562657118, -0.02816380001604557, 0.043726447969675064, 0.012194468639791012, 0.12871193885803223, 0.008576037362217903, -0.1618158370256424, 0.026840461418032646, 0.0030557403806596994, 0.10387714207172394, -0.21997274458408356, -0.08367477357387543, 0.04838619381189346, -0.029553698375821114, -0.05334814265370369, 0.10579082369804382, 0.06295353919267654, 0.0504634715616703, -0.04548325017094612, -0.05543007701635361, -0.008723298087716103, 0.14979462325572968, -0.1187625601887703, -0.006005466915667057 ]
null
null
stable-baselines3
# **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
{"library_name": "stable-baselines3", "tags": ["LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "PPO", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "LunarLander-v2", "type": "LunarLander-v2"}, "metrics": [{"type": "mean_reward", "value": "256.99 +/- 16.71", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
kaljr/ppo-LunarLander-v2
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
2024-02-15T02:06:29+00:00
[]
[]
TAGS #stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
# PPO Agent playing LunarLander-v2 This is a trained model of a PPO agent playing LunarLander-v2 using the stable-baselines3 library. ## Usage (with Stable-baselines3) TODO: Add your code
[ "# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.", "## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ "TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n", "# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.", "## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ 39, 41, 17 ]
[ "passage: TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ 0.03942384943366051, 0.04900386184453964, -0.005304091144353151, 0.026427261531352997, 0.107408307492733, -0.026511888951063156, 0.11188238859176636, 0.0814051404595375, 0.10722193866968155, 0.04762078449130058, 0.08338645845651627, 0.06030960753560066, 0.05080918222665787, 0.2571701407432556, 0.04754156619310379, -0.22987541556358337, 0.036159250885248184, -0.04869936779141426, 0.12395193427801132, 0.07178173214197159, -0.0038484656251966953, -0.06485428661108017, 0.020415637642145157, -0.013290755450725555, 0.05367108806967735, 0.04282612353563309, -0.01716216839849949, -0.08207534998655319, 0.07169748842716217, -0.06345846503973007, 0.06986866891384125, 0.07677983492612839, 0.13218913972377777, -0.17832116782665253, 0.029566360637545586, 0.02571309357881546, -0.07189024239778519, 0.01342033501714468, 0.008019951172173023, 0.05120139941573143, 0.17303818464279175, 0.019879888743162155, 0.07844575494527817, -0.0025605305563658476, -0.15412317216396332, -0.018950799480080605, 0.0436202734708786, 0.12546207010746002, 0.08808347582817078, 0.04605821147561073, 0.01970590092241764, 0.17503218352794647, -0.054352790117263794, -0.028833400458097458, 0.21759237349033356, -0.2881564497947693, -0.031460098922252655, 0.321048766374588, 0.06997483223676682, 0.09725230932235718, -0.07540661096572876, -0.03619609400629997, 0.007783263456076384, -0.013137873262166977, -0.028666524216532707, -0.07447073608636856, 0.17313385009765625, 0.05152064561843872, -0.05057951435446739, -0.09541505575180054, 0.16948209702968597, 0.006921638268977404, 0.0018855923553928733, -0.019282981753349304, 0.009060598909854889, 0.07402525842189789, -0.016097044572234154, -0.07255112379789352, 0.057438433170318604, 0.05330665782094002, 0.019649166613817215, -0.1435653269290924, -0.10762494057416916, -0.022740179672837257, -0.008012006990611553, 0.17786912620067596, -0.009255532175302505, 0.042902372777462006, 0.003065188182517886, 0.10384012013673782, -0.12480384111404419, -0.03354184702038765, -0.0454259067773819, -0.07565800100564957, -0.0223417766392231, -0.02058211714029312, -0.03580251708626747, 0.07184842973947525, 0.11971849203109741, 0.027368178591132164, 0.09350208193063736, 0.047715865075588226, -0.03206788748502731, 0.06343851238489151, 0.05555703118443489, 0.14222665131092072, 0.05807621404528618, 0.012854371219873428, 0.13179877400398254, 0.055213116109371185, 0.033023182302713394, -0.0613492950797081, -0.18252409994602203, 0.07489913702011108, -0.07031869143247604, 0.007941240444779396, 0.12051256000995636, -0.04480670019984245, -0.1183447614312172, -0.037500523030757904, -0.017392054200172424, -0.06224250793457031, -0.025395862758159637, 0.0547584593296051, -0.02883218228816986, -0.03973718360066414, 0.0011496668448671699, 0.09384800493717194, 0.00953749567270279, -0.1752052903175354, 0.03303423151373863, -0.025042934343218803, -0.10782608389854431, 0.009975161403417587, 0.0022444494534283876, 0.03394931182265282, 0.04408763721585274, -0.11822668462991714, -0.30899152159690857, -0.07652641832828522, 0.05490870401263237, -0.06516939401626587, -0.18425025045871735, -0.13193942606449127, 0.02454492449760437, -0.09037084132432938, -0.044885024428367615, -0.12759265303611755, -0.028549788519740105, 0.01743689924478531, 0.011519349180161953, 0.10758619755506516, -0.0106219332665205, -0.012188062071800232, -0.1571401208639145, 0.008273907005786896, -0.20951123535633087, 0.0890483483672142, -0.019150104373693466, 0.037884220480918884, -0.032381169497966766, -0.07404014468193054, 0.030707746744155884, 0.052499737590551376, -0.01474119070917368, 0.13510210812091827, -0.15592676401138306, -0.03691192343831062, -0.007996266707777977, -0.13611900806427002, -0.04786273464560509, -0.10358831286430359, -0.04357128217816353, 0.13354332745075226, 0.018664736300706863, 0.15356586873531342, -0.08709818124771118, -0.0722038671374321, 0.20489206910133362, -0.010411538183689117, -0.12820468842983246, -0.076752208173275, 0.10165707021951675, 0.021510310471057892, -0.056606587022542953, -0.02523270808160305, -0.1839766949415207, -0.0152357779443264, -0.04550420492887497, -0.047039128839969635, 0.01796751655638218, -0.010888241231441498, 0.13837894797325134, 0.08494598418474197, 0.05018039792776108, -0.06086122244596481, -0.006730288732796907, 0.10779471695423126, 0.08823856711387634, 0.008680110797286034, 0.023406028747558594, -0.05774238705635071, 0.09552932530641556, -0.04003755748271942, -0.0142367510125041, -0.08283266425132751, -0.036246106028556824, -0.026256313547492027, 0.17507147789001465, 0.09440762549638748, 0.2257927656173706, 0.09567736834287643, 0.039160262793302536, 0.031270865350961685, -0.13181598484516144, -0.1425403207540512, -0.0017254541162401438, 0.09020978957414627, -0.14270411431789398, -0.04119925573468208, -0.08974775671958923, -0.17768175899982452, -0.12202505767345428, 0.0006432619411498308, -0.17960017919540405, 0.06390921026468277, 0.05408334732055664, -0.035177867859601974, 0.03272094577550888, 0.13032332062721252, -0.011533179320394993, -0.03967514634132385, 0.0831870287656784, 0.0379033200442791, -0.041234664618968964, -0.021742934361100197, 0.11885567009449005, 0.15673065185546875, 0.13124459981918335, -0.03511447086930275, 0.004914294462651014, 0.07076404243707657, -0.02309088408946991, 0.06539414077997208, 0.0558244064450264, 0.20973342657089233, 0.188301220536232, 0.038996949791908264, 0.008822928182780743, -0.07048165798187256, 0.0855446457862854, -0.0742373839020729, -0.14302679896354675, -0.05579735338687897, 0.08729292452335358, 0.016605578362941742, 0.023469142615795135, 0.08711627870798111, 0.024545932188630104, 0.09132762253284454, 0.15968108177185059, 0.01990218088030815, -0.09659269452095032, -0.050218869000673294, 0.01175848301500082, 0.027713103219866753, 0.04794301092624664, -0.04514073207974434, -0.00937939714640379, 0.017020760104060173, -0.10303554683923721, 0.031789086759090424, -0.1413339376449585, -0.1358717679977417, 0.044326696544885635, 0.003906996920704842, 0.010907664895057678, 0.02786896750330925, -0.0038291432429105043, 0.019039705395698547, 0.04351753741502762, -0.06975466758012772, 0.047416772693395615, -0.024745507165789604, -0.020031947642564774, 0.03340689837932587, -0.057257164269685745, -0.205775648355484, -0.17696654796600342, 0.00013708483311347663, -0.09910997003316879, 0.10194740444421768, 0.018308809027075768, -0.12373185902833939, 0.047737859189510345, -0.05822649225592613, 0.027574289590120316, -0.01875593699514866, -0.049130141735076904, 0.10507171601057053, 0.1525275856256485, -0.016146350651979446, 0.018018173053860664, -0.04865182936191559, -0.10157987475395203, -0.19632206857204437, 0.0691583976149559, 0.04680244252085686, 0.014610917307436466, 0.10669491440057755, 0.018072687089443207, 0.02367905154824257, -0.007674071006476879, -0.016521066427230835, -0.011659215204417706, -0.08781040459871292, 0.31909599900245667, 0.04510033503174782, -0.025173069909214973, 0.02041010931134224, -0.0043001663871109486, -0.028083480894565582, 0.03263787180185318, -0.0985708013176918, -0.07548979669809341, -0.08774089068174362, -0.04367410019040108, -0.09784720093011856, 0.053299110382795334, 0.05916472524404526, 0.003188040340319276, -0.07727594673633575, 0.04221395403146744, 0.11369874328374863, -0.0923808291554451, -0.07137343287467957, 0.07477962225675583, 0.0972946360707283, -0.07331304252147675, 0.00012658814375754446, 0.00874367356300354, 0.023951783776283264, 0.037102166563272476, 0.06778035312891006, -0.03966575115919113, 0.08589404821395874, -0.19917890429496765, 0.0372927263379097, 0.106058269739151, 0.023754918947815895, 0.0638108178973198, 0.07643651217222214, -0.1058402881026268, -0.008500572293996811, -0.032518330961465836, -0.21341575682163239, 0.1668180525302887, 0.1355515867471695, 0.06788124144077301, -0.025637222453951836, -0.00461410591378808, -0.0649740919470787, 0.05773647129535675, 0.02723747305572033, -0.14758841693401337, 0.004883295856416225, 0.06064270809292793, 0.026899009943008423, 0.01614922471344471, 0.07971042394638062, 0.014697225764393806, -0.1801026314496994, -0.014406266622245312, 0.10730406641960144, 0.002390873385593295, 0.0053148469887673855, -0.03175045922398567, -0.1755964607000351, 0.0751047357916832, 0.004285442177206278, 0.07233936339616776, -0.1676585078239441, 0.14297930896282196, -0.10089799761772156, 0.07726949453353882, -0.004285062663257122, -0.021311495453119278, 0.02507244050502777, -0.0541163794696331, 0.15163759887218475, 0.01058570109307766, -0.021810131147503853, -0.1200498715043068, -0.1717042326927185, -0.019227758049964905, -0.11788936704397202, -0.11679866164922714, 0.050424277782440186, 0.062185097485780716, 0.04923136904835701, -0.061147067695856094, 0.1518532931804657, -0.047422297298908234, 0.060713399201631546, -0.06893875449895859, -0.06755045056343079, 0.03764858841896057, -0.12588608264923096, -0.08176055550575256, 0.05573027580976486, 0.19166934490203857, 0.15833087265491486, -0.02816431224346161, -0.03472423925995827, -0.047419581562280655, -0.006212298292666674, -0.007802055217325687, 0.0275666993111372, 0.023223137483000755, 0.07315318286418915, -0.07681374251842499, -0.11649256944656372, 0.033787861466407776, -0.06713802367448807, -0.055589709430933, -0.015439179725944996, 0.1513158082962036, 0.04671623185276985, 0.07720734924077988, -0.018946662545204163, 0.03887668624520302, -0.001724981120787561, -0.056474871933460236, 0.16197094321250916, 0.03885216265916824, -0.05193585529923439, 0.06837689876556396, 0.053174007683992386, 0.043745119124650955, 0.03011113777756691, -0.026783017441630363, 0.206032395362854, 0.1980147808790207, 0.014206883497536182, 0.2175983190536499, 0.03177616000175476, -0.03772832080721855, -0.1300560086965561, -0.065880686044693, -0.006372632458806038, 0.03559038043022156, 0.08070417493581772, -0.18207235634326935, -0.015011128038167953, -0.05689644813537598, -0.034518610686063766, -0.15059494972229004, -0.28553900122642517, -0.05957856774330139, 0.20075850188732147, 0.14706264436244965, 0.27519428730010986, -0.10432573407888412, 0.035197313874959946, 0.02663275972008705, -0.04912831634283066, -0.006501141935586929, 0.00018665487004909664, 0.10268618166446686, -0.15421873331069946, 0.1176437959074974, 0.08486983180046082, -0.019002694636583328, 0.01058861706405878, -0.1619086116552353, 0.00936629343777895, -0.12191236019134521, 0.05354422330856323, 0.1400289237499237, -0.048128653317689896, -0.054873593151569366, 0.14033560454845428, -0.024562934413552284, -0.22685599327087402, -0.04648222774267197, -0.043600670993328094, -0.010640020482242107, 0.026607351377606392, -0.1013401448726654, 0.04101909324526787, 0.1330099105834961, 0.009380043484270573, 0.1147187277674675, 0.11749245226383209, -0.052566803991794586, 0.10792597383260727, 0.2257719188928604, -0.018785694614052773, 0.04689010605216026, -0.12743118405342102, -0.0012336712097749114, -0.028270328417420387, 0.013657891191542149, -0.09504974633455276, -0.09938385337591171, 0.02366873063147068, 0.02872389927506447, 0.009118586778640747, 0.0921793207526207, -0.029922157526016235, 0.0759170651435852, 0.06817561388015747, -0.13014446198940277, -0.16288450360298157, 0.015828335657715797, -0.007344507612287998, 0.08354310691356659, 0.00027861111448146403, 0.08878035843372345, -0.11932205408811569, -0.018093237653374672, -0.03153328225016594, -0.03319635987281799, -0.130486860871315, -0.07138993591070175, 0.06156524643301964, 0.028095467016100883, -0.06602972000837326, 0.1398407518863678, 0.026440169662237167, 0.15942534804344177, 0.049197953194379807, 0.012499804608523846, 0.07227300107479095, -0.05345509201288223, 0.1283530443906784, 0.13818155229091644, -0.00868943240493536, -0.05460423603653908, -0.1013643890619278, -0.10236792266368866, 0.08925779908895493, -0.05773641914129257, 0.07476430386304855, -0.14885357022285461, -0.06675903499126434, 0.015772046521306038, 0.016141414642333984, -0.09562095999717712, 0.02571965754032135, -0.01625603251159191, -0.18119946122169495, 0.056570518761873245, -0.048285093158483505, 0.0440407395362854, -0.06347788125276566, -0.1110161691904068, -0.17226378619670868, 0.06091433763504028, 0.08593481779098511, -0.053876690566539764, -0.12229149043560028, 0.011023230850696564, -0.00012518465518951416, -0.06341652572154999, -0.05023367330431938, 0.09722746908664703, -0.11020902544260025, 0.031452205032110214, -0.012567701749503613, 0.08853451162576675, -0.03510405123233795, -0.011538895778357983, 0.044220831245183945, -0.08039166033267975, -0.009481523185968399, 0.03534642979502678, -0.026372017338871956, -0.04127239063382149, -0.2689029574394226, 0.0036654395516961813, 0.0341104120016098, 0.02497158572077751, 0.07856601476669312, 0.011906822212040424, 0.021174922585487366, 0.03993808850646019, -0.15396519005298615, -0.013395369984209538, 0.14574195444583893, -0.07689505815505981, -0.022186370566487312, 0.05703273415565491, -0.09054436534643173, 0.013882770203053951, -0.030287226662039757, 0.1345842480659485, 0.023923413828015327, 0.06404478847980499, -0.0851147472858429, 0.10106813907623291, -0.1451139897108078, -0.04998219385743141, -0.01244612317532301, 0.09761348366737366, 0.07019034773111343, -0.10272270441055298, 0.014697125181555748, 0.04210108891129494, 0.19416837394237518, 0.016384804621338844, -0.0356343574821949, -0.03396720811724663, 0.004015897400677204, 0.22076453268527985, 0.03044266067445278, 0.10457023978233337, 0.07281364500522614, -0.026583973318338394, 0.12624378502368927, 0.09929762035608292, 0.11280370503664017, -0.055645186454057693, 0.13904185593128204, 0.04667386785149574, 0.038641396909952164, 0.0614289753139019, 0.06836545467376709, 0.09098632633686066, -0.0008288522367365658, 0.1138714924454689, 0.013811973854899406, -0.02422109805047512, -0.021335409954190254, 0.17759373784065247, 0.10501719266176224, -0.14769648015499115, 0.029047364369034767, -0.01258957851678133, 0.039933037012815475, -0.014194529503583908, -0.15634691715240479, -0.07240267097949982, -0.3315149247646332, 0.1226184144616127, -0.07119352370500565, 0.019930170848965645, 0.007913772016763687, -0.037425633519887924, -0.03296699747443199, -0.04477746784687042, 0.13151589035987854, -0.013641550205647945, -0.006079165264964104, -0.04815853759646416, -0.015360191464424133, -0.11607866734266281, -0.11200575530529022, -0.013207737356424332, -0.13671602308750153, -0.010119039565324783, 0.05595948174595833, 0.003977729007601738, 0.01821410097181797, -0.03142618387937546, 0.0024383175186812878, 0.06541839241981506, -0.05751744285225868, 0.056182678788900375, 0.12097269296646118, 0.08766137808561325, -0.1058853268623352, 0.031048951670527458, 0.2011747509241104, 0.04359564557671547, -0.12483977526426315, 0.01449228823184967, 0.1819491684436798, 0.004885740112513304, 0.017068125307559967, -0.006097703706473112, -0.0540788508951664, -0.07554277032613754, 0.1251034289598465, 0.08296554535627365, -0.09985227137804031, 0.015833314508199692, -0.0726347416639328, -0.01594804972410202, -0.06374675035476685, 0.10130585730075836, 0.09538925439119339, 0.04440245032310486, -0.10621760785579681, -0.08487539738416672, -0.10891728103160858, 0.040588874369859695, -0.08629853278398514, -0.07311757653951645, 0.09629398584365845, -0.07057105004787445, -0.07029950618743896, 0.025521177798509598, -0.17978744208812714, -0.009467960335314274, 0.1711762249469757, -0.24654000997543335, -0.0916430801153183, -0.10857923328876495, 0.14477859437465668, 0.016497576609253883, 0.1013975441455841, -0.006207061931490898, -0.007889035157859325, -0.20577777922153473, 0.024890204891562462, -0.05293011665344238, -0.02073732763528824, 0.07814782857894897, -0.09476397186517715, 0.22629831731319427, -0.08276885002851486, 0.020940175279974937, 0.012659613974392414, 0.0870661810040474, -0.030675338581204414, 0.09283176809549332, -0.03660329803824425, -0.12576518952846527, -0.03620953485369682, 0.03001813031733036, 0.013904244638979435, 0.10071761906147003, 0.09772487729787827, -0.03414725139737129, 0.03389119729399681, 0.09747414290904999, 0.04172342270612717, -0.023843804374337196, 0.0360250361263752, -0.17077107727527618, 0.02182629331946373, -0.018498148769140244, -0.06935930997133255, 0.03687669709324837, -0.06603235751390457, 0.1639697551727295, 0.04022442549467087, 0.0670473501086235, -0.036152735352516174, 0.0073931049555540085, -0.014454689808189869, -0.013775371946394444, -0.026180334389209747, -0.17259705066680908, -0.10422050207853317, -0.1347656100988388, -0.012701659463346004, -0.034971047192811966, 0.04591470584273338, 0.023234914988279343, -0.0003200018545612693, -0.014577031135559082, -0.12090865522623062, 0.04360328987240791, 0.11146783083677292, -0.04631396010518074, -0.026193076744675636 ]
null
null
null
# Lora of atlanta/アトランタ/亚特兰大 (Azur Lane) ## What Is This? This is the LoRA model of waifu atlanta/アトランタ/亚特兰大 (Azur Lane). ## How Is It Trained? * This model is trained with [HCP-Diffusion](https://github.com/7eu7d7/HCP-Diffusion). * The [auto-training framework](https://github.com/deepghs/cyberharem) is maintained by [DeepGHS Team](https://huggingface.co/deepghs). * The base model used for training is [deepghs/animefull-latest](https://huggingface.co/deepghs/animefull-latest). * Dataset used for training is the `stage3-p480-800` in [CyberHarem/atlanta_azurlane](https://huggingface.co/datasets/CyberHarem/atlanta_azurlane), which contains 37 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 800 steps, 40 checkpoints were saved and evaluated. * **Trigger word is `atlanta_azurlane`.** * Pruned core tags for this waifu are `pink_hair, blue_eyes, braid, long_hair, ahoge, bangs, crown_braid, black_ribbon, hair_ribbon, ribbon, breasts, hair_ornament, hair_between_eyes`. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. ## How to Use It? ### If You Are Using A1111 WebUI v1.7+ **Just use it like the classic LoRA**. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 680, you need to download [`680/atlanta_azurlane.pt`](https://huggingface.co/CyberHarem/atlanta_azurlane/resolve/main/680/atlanta_azurlane.pt) as the embedding and [`680/atlanta_azurlane.safetensors`](https://huggingface.co/CyberHarem/atlanta_azurlane/resolve/main/680/atlanta_azurlane.safetensors) for loading Lora. By using both files together, you can generate images for the desired characters. ## Which Step Should I Use? We selected 5 good steps for you to choose. The best one is step 680. 1480 images (1.61 GiB) were generated for auto-testing. ![Metrics Plot](metrics_plot.png) The base model used for generating preview images is [Meina/MeinaMix_V11](https://huggingface.co/Meina/MeinaMix_V11). Here are the preview of the recommended steps: | Step | Epoch | CCIP | AI Corrupt | Bikini Plus | Score | Download | pattern_0_0 | pattern_0_1 | portrait_0 | portrait_1 | portrait_2 | full_body_0 | full_body_1 | profile_0 | profile_1 | free_0 | free_1 | shorts | maid_0 | maid_1 | miko | yukata | suit | china | bikini_0 | bikini_1 | bikini_2 | sit | squat | kneel | jump | crossed_arms | angry | smile | cry | grin | n_lie_0 | n_lie_1 | n_stand_0 | n_stand_1 | n_stand_2 | n_sex_0 | n_sex_1 | |-------:|--------:|:----------|:-------------|:--------------|:----------|:-----------------------------------------------------------------------------------------------------|:---------------------------------------------|:---------------------------------------------|:-------------------------------------------|:-------------------------------------------|:-------------------------------------------|:---------------------------------------------|:---------------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-------------------------------|:-----------------------------------|:-------------------------------|:---------------------------------|:---------------------------------------|:---------------------------------------|:---------------------------------------|:-----------------------------|:---------------------------------|:---------------------------------|:-------------------------------|:-----------------------------------------------|:---------------------------------|:---------------------------------|:-----------------------------|:-------------------------------|:-------------------------------------|:-------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------------|:-------------------------------------|:-------------------------------------| | 680 | 74 | 0.931 | **0.976** | 0.841 | **0.723** | [Download](https://huggingface.co/CyberHarem/atlanta_azurlane/resolve/main/680/atlanta_azurlane.zip) | ![pattern_0_0](680/previews/pattern_0_0.png) | ![pattern_0_1](680/previews/pattern_0_1.png) | ![portrait_0](680/previews/portrait_0.png) | ![portrait_1](680/previews/portrait_1.png) | ![portrait_2](680/previews/portrait_2.png) | ![full_body_0](680/previews/full_body_0.png) | ![full_body_1](680/previews/full_body_1.png) | ![profile_0](680/previews/profile_0.png) | ![profile_1](680/previews/profile_1.png) | ![free_0](680/previews/free_0.png) | ![free_1](680/previews/free_1.png) | ![shorts](680/previews/shorts.png) | ![maid_0](680/previews/maid_0.png) | ![maid_1](680/previews/maid_1.png) | ![miko](680/previews/miko.png) | ![yukata](680/previews/yukata.png) | ![suit](680/previews/suit.png) | ![china](680/previews/china.png) | ![bikini_0](680/previews/bikini_0.png) | ![bikini_1](680/previews/bikini_1.png) | ![bikini_2](680/previews/bikini_2.png) | ![sit](680/previews/sit.png) | ![squat](680/previews/squat.png) | ![kneel](680/previews/kneel.png) | ![jump](680/previews/jump.png) | ![crossed_arms](680/previews/crossed_arms.png) | ![angry](680/previews/angry.png) | ![smile](680/previews/smile.png) | ![cry](680/previews/cry.png) | ![grin](680/previews/grin.png) | ![n_lie_0](680/previews/n_lie_0.png) | ![n_lie_1](680/previews/n_lie_1.png) | ![n_stand_0](680/previews/n_stand_0.png) | ![n_stand_1](680/previews/n_stand_1.png) | ![n_stand_2](680/previews/n_stand_2.png) | ![n_sex_0](680/previews/n_sex_0.png) | ![n_sex_1](680/previews/n_sex_1.png) | | 560 | 61 | 0.912 | 0.975 | 0.844 | 0.718 | [Download](https://huggingface.co/CyberHarem/atlanta_azurlane/resolve/main/560/atlanta_azurlane.zip) | ![pattern_0_0](560/previews/pattern_0_0.png) | ![pattern_0_1](560/previews/pattern_0_1.png) | ![portrait_0](560/previews/portrait_0.png) | ![portrait_1](560/previews/portrait_1.png) | ![portrait_2](560/previews/portrait_2.png) | ![full_body_0](560/previews/full_body_0.png) | ![full_body_1](560/previews/full_body_1.png) | ![profile_0](560/previews/profile_0.png) | ![profile_1](560/previews/profile_1.png) | ![free_0](560/previews/free_0.png) | ![free_1](560/previews/free_1.png) | ![shorts](560/previews/shorts.png) | ![maid_0](560/previews/maid_0.png) | ![maid_1](560/previews/maid_1.png) | ![miko](560/previews/miko.png) | ![yukata](560/previews/yukata.png) | ![suit](560/previews/suit.png) | ![china](560/previews/china.png) | ![bikini_0](560/previews/bikini_0.png) | ![bikini_1](560/previews/bikini_1.png) | ![bikini_2](560/previews/bikini_2.png) | ![sit](560/previews/sit.png) | ![squat](560/previews/squat.png) | ![kneel](560/previews/kneel.png) | ![jump](560/previews/jump.png) | ![crossed_arms](560/previews/crossed_arms.png) | ![angry](560/previews/angry.png) | ![smile](560/previews/smile.png) | ![cry](560/previews/cry.png) | ![grin](560/previews/grin.png) | ![n_lie_0](560/previews/n_lie_0.png) | ![n_lie_1](560/previews/n_lie_1.png) | ![n_stand_0](560/previews/n_stand_0.png) | ![n_stand_1](560/previews/n_stand_1.png) | ![n_stand_2](560/previews/n_stand_2.png) | ![n_sex_0](560/previews/n_sex_0.png) | ![n_sex_1](560/previews/n_sex_1.png) | | 580 | 63 | 0.907 | 0.967 | **0.845** | 0.716 | [Download](https://huggingface.co/CyberHarem/atlanta_azurlane/resolve/main/580/atlanta_azurlane.zip) | ![pattern_0_0](580/previews/pattern_0_0.png) | ![pattern_0_1](580/previews/pattern_0_1.png) | ![portrait_0](580/previews/portrait_0.png) | ![portrait_1](580/previews/portrait_1.png) | ![portrait_2](580/previews/portrait_2.png) | ![full_body_0](580/previews/full_body_0.png) | ![full_body_1](580/previews/full_body_1.png) | ![profile_0](580/previews/profile_0.png) | ![profile_1](580/previews/profile_1.png) | ![free_0](580/previews/free_0.png) | ![free_1](580/previews/free_1.png) | ![shorts](580/previews/shorts.png) | ![maid_0](580/previews/maid_0.png) | ![maid_1](580/previews/maid_1.png) | ![miko](580/previews/miko.png) | ![yukata](580/previews/yukata.png) | ![suit](580/previews/suit.png) | ![china](580/previews/china.png) | ![bikini_0](580/previews/bikini_0.png) | ![bikini_1](580/previews/bikini_1.png) | ![bikini_2](580/previews/bikini_2.png) | ![sit](580/previews/sit.png) | ![squat](580/previews/squat.png) | ![kneel](580/previews/kneel.png) | ![jump](580/previews/jump.png) | ![crossed_arms](580/previews/crossed_arms.png) | ![angry](580/previews/angry.png) | ![smile](580/previews/smile.png) | ![cry](580/previews/cry.png) | ![grin](580/previews/grin.png) | ![n_lie_0](580/previews/n_lie_0.png) | ![n_lie_1](580/previews/n_lie_1.png) | ![n_stand_0](580/previews/n_stand_0.png) | ![n_stand_1](580/previews/n_stand_1.png) | ![n_stand_2](580/previews/n_stand_2.png) | ![n_sex_0](580/previews/n_sex_0.png) | ![n_sex_1](580/previews/n_sex_1.png) | | 700 | 76 | **0.934** | 0.975 | 0.836 | 0.716 | [Download](https://huggingface.co/CyberHarem/atlanta_azurlane/resolve/main/700/atlanta_azurlane.zip) | ![pattern_0_0](700/previews/pattern_0_0.png) | ![pattern_0_1](700/previews/pattern_0_1.png) | ![portrait_0](700/previews/portrait_0.png) | ![portrait_1](700/previews/portrait_1.png) | ![portrait_2](700/previews/portrait_2.png) | ![full_body_0](700/previews/full_body_0.png) | ![full_body_1](700/previews/full_body_1.png) | ![profile_0](700/previews/profile_0.png) | ![profile_1](700/previews/profile_1.png) | ![free_0](700/previews/free_0.png) | ![free_1](700/previews/free_1.png) | ![shorts](700/previews/shorts.png) | ![maid_0](700/previews/maid_0.png) | ![maid_1](700/previews/maid_1.png) | ![miko](700/previews/miko.png) | ![yukata](700/previews/yukata.png) | ![suit](700/previews/suit.png) | ![china](700/previews/china.png) | ![bikini_0](700/previews/bikini_0.png) | ![bikini_1](700/previews/bikini_1.png) | ![bikini_2](700/previews/bikini_2.png) | ![sit](700/previews/sit.png) | ![squat](700/previews/squat.png) | ![kneel](700/previews/kneel.png) | ![jump](700/previews/jump.png) | ![crossed_arms](700/previews/crossed_arms.png) | ![angry](700/previews/angry.png) | ![smile](700/previews/smile.png) | ![cry](700/previews/cry.png) | ![grin](700/previews/grin.png) | ![n_lie_0](700/previews/n_lie_0.png) | ![n_lie_1](700/previews/n_lie_1.png) | ![n_stand_0](700/previews/n_stand_0.png) | ![n_stand_1](700/previews/n_stand_1.png) | ![n_stand_2](700/previews/n_stand_2.png) | ![n_sex_0](700/previews/n_sex_0.png) | ![n_sex_1](700/previews/n_sex_1.png) | | 500 | 55 | 0.912 | 0.968 | 0.841 | 0.713 | [Download](https://huggingface.co/CyberHarem/atlanta_azurlane/resolve/main/500/atlanta_azurlane.zip) | ![pattern_0_0](500/previews/pattern_0_0.png) | ![pattern_0_1](500/previews/pattern_0_1.png) | ![portrait_0](500/previews/portrait_0.png) | ![portrait_1](500/previews/portrait_1.png) | ![portrait_2](500/previews/portrait_2.png) | ![full_body_0](500/previews/full_body_0.png) | ![full_body_1](500/previews/full_body_1.png) | ![profile_0](500/previews/profile_0.png) | ![profile_1](500/previews/profile_1.png) | ![free_0](500/previews/free_0.png) | ![free_1](500/previews/free_1.png) | ![shorts](500/previews/shorts.png) | ![maid_0](500/previews/maid_0.png) | ![maid_1](500/previews/maid_1.png) | ![miko](500/previews/miko.png) | ![yukata](500/previews/yukata.png) | ![suit](500/previews/suit.png) | ![china](500/previews/china.png) | ![bikini_0](500/previews/bikini_0.png) | ![bikini_1](500/previews/bikini_1.png) | ![bikini_2](500/previews/bikini_2.png) | ![sit](500/previews/sit.png) | ![squat](500/previews/squat.png) | ![kneel](500/previews/kneel.png) | ![jump](500/previews/jump.png) | ![crossed_arms](500/previews/crossed_arms.png) | ![angry](500/previews/angry.png) | ![smile](500/previews/smile.png) | ![cry](500/previews/cry.png) | ![grin](500/previews/grin.png) | ![n_lie_0](500/previews/n_lie_0.png) | ![n_lie_1](500/previews/n_lie_1.png) | ![n_stand_0](500/previews/n_stand_0.png) | ![n_stand_1](500/previews/n_stand_1.png) | ![n_stand_2](500/previews/n_stand_2.png) | ![n_sex_0](500/previews/n_sex_0.png) | ![n_sex_1](500/previews/n_sex_1.png) | ## Anything Else? Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. ## All Steps We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * [Steps From 620 to 800](all/0.md) * [Steps From 420 to 600](all/1.md) * [Steps From 220 to 400](all/2.md) * [Steps From 20 to 200](all/3.md)
{"license": "mit", "tags": ["art", "not-for-all-audiences"], "datasets": ["CyberHarem/atlanta_azurlane"], "pipeline_tag": "text-to-image"}
text-to-image
CyberHarem/atlanta_azurlane
[ "art", "not-for-all-audiences", "text-to-image", "dataset:CyberHarem/atlanta_azurlane", "license:mit", "region:us" ]
2024-02-15T02:10:01+00:00
[]
[]
TAGS #art #not-for-all-audiences #text-to-image #dataset-CyberHarem/atlanta_azurlane #license-mit #region-us
Lora of atlanta/アトランタ/亚特兰大 (Azur Lane) ====================================== What Is This? ------------- This is the LoRA model of waifu atlanta/アトランタ/亚特兰大 (Azur Lane). How Is It Trained? ------------------ * This model is trained with HCP-Diffusion. * The auto-training framework is maintained by DeepGHS Team. * The base model used for training is deepghs/animefull-latest. * Dataset used for training is the 'stage3-p480-800' in CyberHarem/atlanta\_azurlane, which contains 37 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 800 steps, 40 checkpoints were saved and evaluated. * Trigger word is 'atlanta\_azurlane'. * Pruned core tags for this waifu are 'pink\_hair, blue\_eyes, braid, long\_hair, ahoge, bangs, crown\_braid, black\_ribbon, hair\_ribbon, ribbon, breasts, hair\_ornament, hair\_between\_eyes'. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. How to Use It? -------------- ### If You Are Using A1111 WebUI v1.7+ Just use it like the classic LoRA. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 680, you need to download '680/atlanta\_azurlane.pt' as the embedding and '680/atlanta\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters. Which Step Should I Use? ------------------------ We selected 5 good steps for you to choose. The best one is step 680. 1480 images (1.61 GiB) were generated for auto-testing. !Metrics Plot The base model used for generating preview images is Meina/MeinaMix\_V11. Here are the preview of the recommended steps: Anything Else? -------------- Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. All Steps --------- We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * Steps From 620 to 800 * Steps From 420 to 600 * Steps From 220 to 400 * Steps From 20 to 200
[ "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 680, you need to download '680/atlanta\\_azurlane.pt' as the embedding and '680/atlanta\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 680.\n\n\n1480 images (1.61 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 620 to 800\n* Steps From 420 to 600\n* Steps From 220 to 400\n* Steps From 20 to 200" ]
[ "TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/atlanta_azurlane #license-mit #region-us \n", "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 680, you need to download '680/atlanta\\_azurlane.pt' as the embedding and '680/atlanta\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 680.\n\n\n1480 images (1.61 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 620 to 800\n* Steps From 420 to 600\n* Steps From 220 to 400\n* Steps From 20 to 200" ]
[ 44, 38, 467 ]
[ "passage: TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/atlanta_azurlane #license-mit #region-us \n### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file." ]
[ 0.006354877259582281, 0.003995741251856089, -0.004206218756735325, 0.07873811572790146, 0.08142488449811935, 0.0847165659070015, 0.22895285487174988, 0.07880040258169174, 0.12053811550140381, -0.07502641528844833, 0.0858553871512413, 0.056191276758909225, -0.001999315805733204, 0.04246993735432625, -0.0330052487552166, -0.1422153264284134, -0.06599566340446472, -0.030083615332841873, 0.002715218812227249, 0.0188597459346056, 0.07632157951593399, 0.006282452028244734, 0.10324709862470627, -0.053063392639160156, -0.036338936537504196, 0.05314311385154724, -0.030940230935811996, -0.05160362273454666, 0.030965283513069153, 0.08407652378082275, 0.12830284237861633, 0.009348808787763119, 0.06671887636184692, -0.16227126121520996, 0.06585223972797394, -0.00616150489076972, -0.10599351674318314, -0.0024578983429819345, 0.0207810178399086, -0.03554464131593704, 0.1271742284297943, 0.037871357053518295, -0.1060500219464302, 0.04120652750134468, -0.132785826921463, -0.03716302290558815, -0.053124379366636276, 0.03586964309215546, 0.14429831504821777, 0.05647099390625954, 0.02411326952278614, 0.057400841265916824, -0.04468931257724762, 0.07953384518623352, 0.10734935104846954, -0.13838835060596466, -0.06808687001466751, 0.09581488370895386, 0.01466995757073164, 0.1368812471628189, -0.0819154754281044, 0.10009113699197769, 0.07094163447618484, -0.04174274206161499, -0.13976134359836578, -0.09146466851234436, -0.2191578894853592, -0.011763855814933777, 0.01173042319715023, 0.019433490931987762, 0.4132442772388458, 0.06036898121237755, 0.03538493067026138, 0.06495200842618942, -0.06750521808862686, 0.02481253817677498, -0.10146047919988632, 0.14025893807411194, 0.046165838837623596, 0.09696799516677856, -0.03558235988020897, -0.10648062080144882, -0.1165977492928505, -0.06894004344940186, -0.0857519805431366, -0.01169637031853199, 0.023031264543533325, 0.12115196883678436, -0.19814103841781616, 0.006660803221166134, -0.04553896561264992, -0.12784861028194427, 0.027684086933732033, -0.10407289117574692, 0.1679764688014984, 0.06923293322324753, -0.011759335175156593, 0.0004671829519793391, 0.24456091225147247, 0.12284773588180542, 0.19541336596012115, 0.0490565150976181, -0.10992145538330078, 0.13207414746284485, 0.029904933646321297, -0.08510930836200714, -0.015489336103200912, -0.10363224893808365, 0.1439725011587143, -0.050965581089258194, 0.10963321477174759, -0.06200959533452988, -0.11342137306928635, 0.021753715351223946, -0.1070941686630249, 0.06829576194286346, 0.04360280558466911, 0.0016696446109563112, -0.05275196209549904, 0.044889386743307114, 0.03025812841951847, -0.03646817430853844, -0.0037644796539098024, -0.010596987791359425, -0.055492233484983444, 0.04851696640253067, 0.11188484728336334, 0.034177761524915695, 0.06141972914338112, -0.0011110714403912425, -0.02027316950261593, -0.005283141974359751, -0.05026568844914436, -0.001575105357915163, 0.04585864022374153, 0.04676692560315132, 0.08604476600885391, -0.16006413102149963, -0.08833105117082596, -0.013542995788156986, 0.06302321702241898, 0.0010835621505975723, 0.09787807613611221, -0.003658408299088478, 0.05761163681745529, 0.007580887991935015, -0.025943882763385773, 0.03730200603604317, -0.1056632325053215, 0.08346109837293625, -0.013736555352807045, 0.09361587464809418, -0.19744442403316498, -0.0029049578588455915, -0.04228387400507927, 0.012553166598081589, 0.06118309870362282, -0.004179337527602911, -0.11123556643724442, 0.12910789251327515, -0.01606251671910286, 0.07313503324985504, -0.1066836416721344, 0.04906894266605377, 0.0273633673787117, 0.08228322118520737, -0.10311875492334366, 0.012215474620461464, 0.12039852142333984, -0.13756713271141052, -0.16538269817829132, 0.09103872627019882, -0.02250581793487072, 0.030333293601870537, 0.04495343938469887, 0.1546146720647812, 0.16556316614151, -0.180265873670578, -0.00930291973054409, 0.05221758782863617, -0.021020663902163506, -0.08891098201274872, -0.01292109489440918, 0.1057891920208931, 0.0168145764619112, 0.03187683969736099, -0.028581732884049416, 0.1313847154378891, -0.030927401036024094, -0.08889970928430557, -0.028885195031762123, -0.08183230459690094, -0.07984843105077744, 0.050393491983413696, -0.009250164963304996, -0.05134520307183266, 0.012954103760421276, -0.16251416504383087, 0.15767832100391388, 0.014794912189245224, 0.018344514071941376, -0.07509041577577591, 0.12021616101264954, 0.0033765542320907116, 0.0024866617750376463, 0.008349314332008362, -0.06351260095834732, -0.10775038599967957, 0.24265547096729279, 0.08560376614332199, 0.08065822720527649, 0.06225094571709633, -0.05340494588017464, -0.06584040820598602, 0.021535417065024376, 0.01774432137608528, -0.03803015872836113, 0.02539956010878086, -0.10710659623146057, 0.05226588621735573, -0.013716241344809532, 0.03381839394569397, -0.014872905798256397, -0.026333725079894066, 0.0688953697681427, 0.013546275906264782, -0.015581605024635792, 0.09031526744365692, 0.05201506242156029, -0.020573001354932785, -0.0684347152709961, 0.0009751284960657358, 0.07387346774339676, -0.0127939498052001, -0.08221402019262314, 0.02676883339881897, -0.003177432809025049, 0.05885813385248184, 0.20032887160778046, -0.21771211922168732, 0.04038141295313835, 0.011629649437963963, 0.051988065242767334, 0.039294131100177765, -0.00026113487547263503, -0.029595280066132545, 0.03425005450844765, -0.01778477057814598, 0.07245967537164688, -0.017769938334822655, 0.06979488581418991, -0.02689266949892044, -0.13547608256340027, -0.013214428909122944, -0.03458942845463753, 0.16552938520908356, -0.1669943779706955, 0.06276435405015945, 0.18619419634342194, -0.12337959557771683, 0.13377805054187775, 0.003965218551456928, -0.011014726012945175, 0.017954343929886818, 0.032272931188344955, 0.0019292458891868591, 0.10278218239545822, -0.07010530680418015, -0.02863544411957264, 0.027718327939510345, -0.08626091480255127, 0.03147149458527565, -0.12373417615890503, -0.11427096277475357, -0.06586006283760071, -0.03620303422212601, -0.04084307700395584, 0.019914347678422928, -0.04954862967133522, 0.07788750529289246, -0.09015224874019623, -0.08747702091932297, -0.02708526886999607, -0.08638893067836761, 0.023775123059749603, 0.012731977738440037, -0.0561441145837307, -0.14259786903858185, -0.13206413388252258, -0.08904208987951279, -0.13802751898765564, -0.004351657349616289, 0.065678671002388, -0.10915286093950272, -0.04614662006497383, 0.015962183475494385, -0.05427165329456329, 0.09829510003328323, -0.07756490260362625, 0.021739516407251358, 0.054686304181814194, -0.0315229669213295, -0.17029252648353577, 0.0011916906805709004, -0.06285255402326584, -0.05209307745099068, 0.15400438010692596, -0.15658308565616608, 0.18504342436790466, -0.024407420307397842, 0.04972591996192932, 0.06218442693352699, 0.03298475965857506, 0.12359803169965744, -0.11818947643041611, 0.07684913277626038, 0.1898682862520218, 0.04097078740596771, 0.07646732777357101, 0.12447214126586914, 0.08646681904792786, -0.11308474838733673, 0.03370555490255356, 0.07378403097391129, -0.10203421115875244, -0.0940205454826355, -0.05585866793990135, -0.10889479517936707, -0.06477786600589752, 0.057211488485336304, 0.057253915816545486, 0.046732958406209946, 0.12584829330444336, -0.055353086441755295, -0.0028434402775019407, 0.10134363174438477, 0.0458873026072979, 0.08836153894662857, 0.021137339994311333, 0.06379872560501099, -0.1476389765739441, -0.045412998646497726, 0.16238203644752502, 0.2280067801475525, 0.23131321370601654, 0.02433529868721962, 0.06594720482826233, 0.1189882680773735, 0.07086823880672455, 0.09771890193223953, 0.05839226394891739, 0.004938661586493254, 0.018292995169758797, -0.07615789026021957, -0.04984358325600624, 0.013409222476184368, 0.004272762686014175, -0.05025186017155647, -0.15025290846824646, 0.1039692610502243, -0.004145262762904167, 0.08095388859510422, 0.13498209416866302, 0.036380719393491745, -0.11643514782190323, 0.16421136260032654, 0.10142432898283005, 0.08932501822710037, -0.06743302196264267, 0.12898574769496918, 0.051891960203647614, -0.0036217470187693834, 0.16913044452667236, 0.025359081104397774, 0.1536102145910263, -0.03649193421006203, -0.07810040563344955, -0.06770937144756317, -0.055719152092933655, 0.007857263088226318, 0.035325706005096436, -0.22079995274543762, 0.10693325847387314, 0.05860443413257599, 0.010706058703362942, -0.004529467783868313, -0.05245189368724823, 0.1824832558631897, 0.15799225866794586, 0.07976070791482925, 0.022368932142853737, -0.03889048844575882, -0.011522512882947922, -0.08624919503927231, 0.05644764006137848, 0.005813904572278261, 0.07639506459236145, -0.037223927676677704, -0.09452703595161438, -0.019847270101308823, -0.002488667843863368, 0.02451854757964611, -0.07290264219045639, -0.11466087400913239, -0.0491781122982502, 0.259985089302063, -0.06412286311388016, 0.04814529046416283, 0.053708892315626144, 0.014024254865944386, -0.032454073429107666, 0.03070037066936493, -0.03558439761400223, -0.01734800450503826, -0.040310122072696686, 0.010617672465741634, 0.010273756459355354, -0.04346946254372597, -0.053783778101205826, -0.02687421254813671, -0.09713973850011826, -0.10182182490825653, 0.009164644405245781, -0.04848511144518852, 0.017381703481078148, -0.026876982301473618, 0.011374169960618019, -0.10491088032722473, -0.03269730135798454, 0.026907745748758316, 0.032418981194496155, -0.08162243664264679, -0.13099810481071472, -0.0022683655843138695, -0.006805770564824343, -0.05458177998661995, 0.03310918062925339, -0.11370937526226044, -0.09886881709098816, -0.05481396242976189, -0.029547810554504395, 0.12672068178653717, 0.2275119125843048, -0.026249591261148453, 0.00042516461689956486, 0.1476716846227646, -0.10271594673395157, -0.31253135204315186, -0.16368071734905243, -0.16146545112133026, -0.0975174829363823, 0.03592243790626526, -0.07955380529165268, 0.025161392986774445, 0.08313681930303574, -0.04099629446864128, 0.20747312903404236, -0.19651181995868683, -0.094664067029953, 0.08918624371290207, 0.091084323823452, 0.30884018540382385, -0.2485802173614502, 0.01332135684788227, -0.11828599870204926, -0.0385877899825573, 0.012098490260541439, -0.08470172435045242, 0.118568055331707, 0.03802494332194328, 0.07848452031612396, -0.005011971108615398, -0.006470260210335255, 0.14649735391139984, -0.07932528108358383, 0.13698311150074005, -0.11528044939041138, -0.085053451359272, 0.20768606662750244, -0.0326455719769001, 0.011950348503887653, -0.20888546109199524, -0.03749595582485199, -0.03868759050965309, 0.03557460010051727, -0.005603952333331108, 0.05630890652537346, -0.0032850352581590414, -0.008870086632668972, -0.12294647097587585, -0.026984142139554024, -0.02953351102769375, 0.061770759522914886, 0.22336050868034363, -0.06361693888902664, -0.053956203162670135, 0.042348455637693405, 0.0048746950924396515, 0.09782638400793076, 0.01461568009108305, -0.05452127382159233, -0.043183136731386185, 0.09119392931461334, -0.20706507563591003, 0.05810222029685974, 0.008790946565568447, -0.011330516077578068, 0.010957867838442326, 0.013614486902952194, 0.017768053337931633, 0.11692705750465393, 0.18758024275302887, 0.0031835115514695644, -0.023431342095136642, -0.019302284345030785, 0.038267720490694046, 0.1258774846792221, -0.02254636585712433, 0.11093904823064804, 0.02401871420443058, 0.035391565412282944, 0.012939776293933392, 0.0541953481733799, -0.07984089106321335, -0.0812007412314415, 0.10109696537256241, -0.044511958956718445, -0.08532576262950897, 0.09387362003326416, 0.05385315790772438, 0.07004215568304062, 0.0072892652824521065, 0.04548553749918938, 0.020305095240473747, -0.12741099298000336, 0.024387087672948837, 0.20900152623653412, -0.07488615065813065, -0.07015854865312576, -0.06794153898954391, 0.012228219769895077, -0.12081050127744675, 0.07437899708747864, 0.033347293734550476, -0.027130790054798126, 0.11482229828834534, -0.048257194459438324, -0.026156295090913773, 0.009116428904235363, -0.061328135430812836, 0.036640290170907974, -0.15455135703086853, -0.19469453394412994, 0.04804074019193649, -0.011719428934156895, -0.06718297302722931, -0.08655566722154617, -0.08391546458005905, 0.06934162974357605, -0.1637679636478424, 0.14051318168640137, -0.06636252254247665, 0.06116911396384239, -0.03972947970032692, -0.047542717307806015, -0.11200995743274689, -0.020102698355913162, -0.05529216676950455, -0.020546283572912216, 0.056618303060531616, 0.010395022109150887, -0.12491060048341751, -0.11761678755283356, 0.057315438985824585, -0.0035324397031217813, -0.0058771828189492226, 0.009639794006943703, -0.06949839740991592, 0.012972814962267876, -0.22348298132419586, -0.06399168074131012, 0.09173918515443802, 0.03963495418429375, -0.09271973371505737, 0.12064400315284729, 0.05153008550405502, -0.021652109920978546, 0.03592654690146446, 0.006209885701537132, 0.17193681001663208, -0.07599720358848572, 0.020641960203647614, -0.11493764072656631, -0.1673872321844101, -0.02074608951807022, 0.030678866431117058, 0.23468708992004395, 0.08438798040151596, 0.12156958878040314, -0.04890431836247444, 0.01824609562754631, -0.014373932033777237, 0.07375440746545792, 0.0072530112229287624, -0.10383978486061096, -0.04970138147473335, -0.16746622323989868, -0.06542231887578964, -0.0643547996878624, 0.17180778086185455, 0.048114750534296036, -0.13823528587818146, -0.0034954596776515245, 0.11204719543457031, -0.17675848305225372, -0.013039842247962952, 0.1686563938856125, -0.04484618082642555, 0.026511868461966515, -0.1499384641647339, 0.024417458102107048, 0.07883473485708237, -0.023302199319005013, -0.006283066235482693, 0.11923518031835556, 0.0034618780482560396, 0.003979546949267387, 0.040413111448287964, -0.03438098356127739, 0.07322906702756882, -0.07700824737548828, 0.06227515637874603, 0.0035400250926613808, -0.044789303094148636, -0.12002997100353241, 0.18834766745567322, -0.01521190907806158, 0.012462385930120945, -0.06239258497953415, 0.00009109565144171938, -0.10720618069171906, -0.10136537998914719, -0.06995366513729095, -0.13243526220321655, 0.07063306868076324, -0.06123168393969536, 0.010430257767438889, 0.002593370620161295, 0.013320268131792545, -0.07490994781255722, 0.01265634223818779, -0.18951416015625, -0.05124467611312866, 0.01666831783950329, -0.010678131133317947, -0.019497670233249664, -0.04393346980214119, -0.032283417880535126, 0.017871923744678497, -0.058271411806344986, -0.06855501979589462, 0.059811782091856, 0.0865146741271019, 0.062422387301921844, -0.16135774552822113, -0.1059357151389122, -0.07616955786943436, 0.03287797421216965, 0.07784398645162582, 0.17360876500606537, 0.03142518550157547, -0.006962625775486231, 0.0478939563035965, 0.14762845635414124, 0.010761373676359653, -0.07734479755163193, -0.0697600394487381, -0.12234454602003098, -0.1389404535293579, -0.014403332024812698, -0.07197567820549011, -0.02586394175887108, 0.024566566571593285, 0.24118182063102722, 0.1939423382282257, -0.15402528643608093, 0.03620629385113716, -0.07278579473495483, 0.03826688975095749, -0.03492707014083862, 0.16203059256076813, 0.05245298147201538, 0.15117482841014862, -0.03140464052557945, -0.03739574924111366, -0.05976620316505432, 0.0217597559094429, -0.10457313060760498, 0.040359023958444595, -0.015616735443472862, -0.06788057833909988, -0.06309632211923599, 0.11136411130428314, -0.10922951251268387, 0.07631681859493256, 0.17335058748722076, -0.15093214809894562, -0.022009406238794327, -0.041773706674575806, 0.05184045061469078, 0.10914444178342819, 0.01198683399707079, -0.07795841991901398, -0.028555624186992645, 0.006303791422396898, 0.02788837067782879, -0.17676326632499695, -0.11328213661909103, -0.002142511773854494, -0.12667745351791382, 0.13750125467777252, -0.005653502885252237, -0.003790313843637705, 0.03536362200975418, -0.06692896038293839, -0.007112143561244011, 0.1676127165555954, 0.017704883590340614, -0.02421579323709011, -0.0206284336745739, -0.06316613405942917, -0.10536332428455353, 0.07173781096935272, 0.09143739193677902, 0.0493776835501194, -0.005608194507658482, 0.15765321254730225, -0.02043057046830654, -0.03300094231963158, 0.13175955414772034, -0.17464184761047363, 0.09151238203048706, -0.0011979861883446574, -0.01917205937206745, -0.06947088986635208, -0.043530356138944626, 0.038738396018743515, 0.08461818844079971, -0.15968450903892517, -0.04593125730752945, 0.06146622821688652, -0.10392109304666519, 0.059440694749355316, 0.04780346900224686, -0.09997923672199249, 0.021122608333826065, -0.12339962273836136, -0.006205285899341106, -0.10282055288553238, 0.04703715816140175, 0.20772966742515564, -0.030860736966133118, 0.010621289722621441, -0.13394342362880707, 0.05831749364733696, -0.03293439745903015, -0.04559849202632904, -0.07387298345565796 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-imdb This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 2.3276 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 128 - eval_batch_size: 128 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | No log | 1.0 | 479 | 2.3738 | | 2.5112 | 2.0 | 958 | 2.3320 | | 2.5112 | 3.0 | 1437 | 2.3276 | ### Framework versions - Transformers 4.38.0.dev0 - Pytorch 2.1.2+cu118 - Datasets 2.17.0 - Tokenizers 0.15.1
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "distilbert-base-uncased", "model-index": [{"name": "distilbert-base-uncased-finetuned-imdb", "results": []}]}
fill-mask
sjunique/distilbert-base-uncased-finetuned-imdb
[ "transformers", "tensorboard", "safetensors", "distilbert", "fill-mask", "generated_from_trainer", "base_model:distilbert-base-uncased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-15T02:10:34+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #distilbert #fill-mask #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-imdb ====================================== This model is a fine-tuned version of distilbert-base-uncased on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 2.3276 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 128 * eval\_batch\_size: 128 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.38.0.dev0 * Pytorch 2.1.2+cu118 * Datasets 2.17.0 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #safetensors #distilbert #fill-mask #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ 72, 113, 4, 38 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #distilbert #fill-mask #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ -0.1208956390619278, 0.11539584398269653, -0.0038115112110972404, 0.09731951355934143, 0.10213534533977509, 0.0073190899565815926, 0.15777279436588287, 0.1397729516029358, -0.08092784136533737, 0.07735689729452133, 0.1252320110797882, 0.09881896525621414, 0.04527072235941887, 0.1892971694469452, -0.061125680804252625, -0.21515151858329773, 0.04148176684975624, 0.02331903949379921, -0.07644440233707428, 0.1024843379855156, 0.0924086943268776, -0.11586707085371017, 0.07490500062704086, 0.020648302510380745, -0.14146146178245544, 0.000993217807263136, 0.020457185804843903, -0.07652311027050018, 0.09893934428691864, 0.02588745951652527, 0.12511640787124634, 0.03192257508635521, 0.07714544236660004, -0.16183696687221527, 0.007948052138090134, 0.08124437183141708, -0.0003099030873272568, 0.08483034372329712, 0.04163403809070587, 0.0026635013055056334, 0.066740021109581, -0.10470593720674515, 0.07603269815444946, 0.022487251088023186, -0.12209092080593109, -0.2844066321849823, -0.11586525291204453, 0.0471954308450222, 0.10064774006605148, 0.06325697898864746, 0.007044341415166855, 0.1548590213060379, -0.016869617626070976, 0.09039875864982605, 0.23581676185131073, -0.296050488948822, -0.05747811496257782, 0.01020011119544506, 0.03635686635971069, 0.040609151124954224, -0.08025746047496796, -0.02403699792921543, 0.03660190477967262, 0.027268070727586746, 0.1481526494026184, -0.008180063217878342, -0.028086481615900993, -0.02441759593784809, -0.1353854537010193, -0.03665845841169357, 0.12432065606117249, 0.04927847161889076, -0.059160053730010986, -0.06276723742485046, -0.08262836933135986, -0.14259102940559387, -0.05326300114393234, -0.002887365175411105, 0.040446240454912186, -0.03257954865694046, -0.0741296038031578, -0.012225515209138393, -0.07171377539634705, -0.06056590750813484, -0.04395829886198044, 0.14730937778949738, 0.05419549718499184, 0.028938017785549164, -0.025329846888780594, 0.0709066167473793, -0.057906147092580795, -0.17368654906749725, -0.011097891256213188, 0.013446004129946232, -0.008594702929258347, -0.036767151206731796, -0.035434525460004807, -0.10088331252336502, 0.027113577350974083, 0.19078850746154785, -0.08742310106754303, 0.06995872408151627, -0.02109142392873764, 0.02983919531106949, -0.10756225883960724, 0.1536405235528946, -0.03288155049085617, -0.004286871757358313, 0.033079955726861954, 0.08538173139095306, 0.08074627071619034, -0.019707992672920227, -0.09139999002218246, 0.03458046540617943, 0.09421593695878983, 0.027203166857361794, -0.029126346111297607, 0.04310573637485504, -0.054257627576589584, -0.005714209284633398, 0.09369123727083206, -0.11111757904291153, 0.03877978026866913, -0.004791829735040665, -0.053487155586481094, -0.035580288618803024, 0.04101033881306648, 0.00323297968134284, -0.0008588783093728125, 0.09922090172767639, -0.07501426339149475, -0.0005622185999527574, -0.09484324604272842, -0.13701856136322021, 0.03724420443177223, -0.0855637714266777, 0.0019151248270645738, -0.09518292546272278, -0.17409372329711914, -0.0006489121005870402, 0.07557359337806702, -0.03965376690030098, -0.01674373261630535, -0.01585981994867325, -0.08364392817020416, 0.04435794800519943, -0.013843224383890629, 0.06349499523639679, -0.06398890167474747, 0.0852624922990799, 0.04687805846333504, 0.08755135536193848, -0.04498834162950516, 0.03138107433915138, -0.09000139683485031, 0.056337956339120865, -0.22666409611701965, 0.019891399890184402, -0.07344017177820206, 0.057667870074510574, -0.10075961798429489, -0.08564631640911102, 0.006856138817965984, -0.019836949184536934, 0.10610214620828629, 0.09310869127511978, -0.16892313957214355, -0.05300101265311241, 0.19456356763839722, -0.11216393858194351, -0.14063699543476105, 0.13213422894477844, -0.03899132087826729, -0.009657005779445171, 0.04484988749027252, 0.15288685262203217, 0.07662786543369293, -0.13809999823570251, -0.012353264726698399, -0.029442984610795975, 0.07172556221485138, -0.04459038004279137, 0.0819927230477333, -0.013969644904136658, 0.03316060081124306, 0.007218322716653347, -0.019190525636076927, 0.0364278219640255, -0.07170403748750687, -0.09192811697721481, -0.05873401463031769, -0.08384357392787933, 0.047574181109666824, 0.042035914957523346, 0.04425046220421791, -0.12199266254901886, -0.1253659725189209, 0.03924345597624779, 0.08488266915082932, -0.07622706890106201, 0.03720016032457352, -0.09269807487726212, 0.10433689504861832, -0.07351803034543991, -0.015721436589956284, -0.16112563014030457, -0.06289607286453247, 0.03191651403903961, -0.06624975055456161, -0.0036862557753920555, -0.0639505535364151, 0.08218038082122803, 0.09353770315647125, -0.06226465851068497, -0.032702479511499405, -0.03990063816308975, 0.012513833120465279, -0.09976179897785187, -0.20794729888439178, -0.043563056737184525, -0.04461755231022835, 0.08387354761362076, -0.161765456199646, 0.043105561286211014, 0.006066164001822472, 0.12969735264778137, 0.052012450993061066, -0.029332507401704788, -0.02307792566716671, 0.06391001492738724, -0.033968620002269745, -0.08848007023334503, 0.040004435926675797, 0.01879025250673294, -0.08858431130647659, -0.023277226835489273, -0.15528760850429535, 0.1558682918548584, 0.12512964010238647, 0.01155321579426527, -0.07136639952659607, 0.020558971911668777, -0.04991814121603966, -0.030908579006791115, -0.043403249233961105, 0.02006775513291359, 0.08022726327180862, 0.008725885301828384, 0.13526925444602966, -0.08952344208955765, -0.026587730273604393, 0.047108445316553116, -0.03706182539463043, -0.011391147039830685, 0.07298372685909271, 0.06767670065164566, -0.08857118338346481, 0.1341610550880432, 0.18033190071582794, -0.08879277110099792, 0.11654479801654816, -0.06826049089431763, -0.06842244416475296, -0.036119211465120316, 0.005975665058940649, 0.042693499475717545, 0.13377240300178528, -0.0784563198685646, 0.012098372913897038, 0.02870345674455166, 0.013278950937092304, -0.005158646497875452, -0.19337019324302673, 0.0013616645010188222, 0.03999575972557068, -0.06076956167817116, -0.029560642316937447, 0.0029140624683350325, 0.0025194480549544096, 0.08492375165224075, -0.00016858860908541828, -0.0678916797041893, 0.03878088667988777, 0.004833349026739597, -0.06409548223018646, 0.18224970996379852, -0.11085784435272217, -0.14442561566829681, -0.1340080201625824, -0.06594471633434296, -0.051294442266225815, 0.010339602828025818, 0.06484991312026978, -0.06331939250230789, -0.06896264851093292, -0.10502290725708008, -0.0366094671189785, 0.036296818405389786, 0.041422098875045776, 0.04977753013372421, -0.015261847525835037, 0.10884672403335571, -0.10486211627721786, -0.015421411022543907, -0.010579759255051613, -0.014043631963431835, 0.03571309894323349, 0.04359607771039009, 0.10271164774894714, 0.11807036399841309, -0.02931448072195053, 0.0159242395311594, -0.024078883230686188, 0.225978285074234, -0.058361977338790894, 0.005709238350391388, 0.14507511258125305, -0.010360023938119411, 0.08071634918451309, 0.1420227438211441, 0.05753881111741066, -0.08805244415998459, 0.007062096614390612, 0.025120683014392853, -0.04073366895318031, -0.21084710955619812, -0.024596745148301125, -0.04715443029999733, -0.02279721200466156, 0.10281349718570709, 0.03557628393173218, 0.03418036177754402, 0.0568835586309433, -0.004647726658731699, 0.06042419746518135, -0.017623869702219963, 0.09596967697143555, 0.09168791770935059, 0.06965405493974686, 0.13386864960193634, -0.052459992468357086, -0.021654648706316948, 0.04146095737814903, -0.00729419244453311, 0.22154410183429718, 0.00689743272960186, 0.14881698787212372, 0.05364309623837471, 0.1768541932106018, 0.008581530302762985, 0.07047366350889206, -0.011497956700623035, -0.019891956821084023, -0.004046982154250145, -0.06492570042610168, -0.010363460518419743, 0.030889233574271202, -0.07755097001791, 0.05629406124353409, -0.10396280884742737, 0.03627970814704895, 0.0398440957069397, 0.27346697449684143, 0.06572456657886505, -0.34992629289627075, -0.09991957992315292, 0.020620130002498627, -0.01621319353580475, -0.03716983273625374, 0.0028826461639255285, 0.09520097821950912, -0.04161101207137108, 0.08902467042207718, -0.0857991874217987, 0.08630793541669846, 0.0025206911377608776, 0.019905921071767807, 0.06147347018122673, 0.12158365547657013, -0.0028374292887747288, 0.04189431294798851, -0.26963990926742554, 0.27392542362213135, 0.028203163295984268, 0.09881187230348587, -0.05443074554204941, 0.033695608377456665, 0.032952629029750824, 0.052992165088653564, 0.07259811460971832, -0.01836385391652584, -0.13937023282051086, -0.16123656928539276, -0.09262029826641083, 0.012718923389911652, 0.08576668798923492, 0.016424942761659622, 0.1069975197315216, -0.017572948709130287, -0.01665743812918663, 0.06999285519123077, -0.01849018782377243, -0.1171744093298912, -0.09152756631374359, 0.015097672119736671, 0.05526968836784363, -0.047419097274541855, -0.09442682564258575, -0.10353874415159225, -0.08767320215702057, 0.16176974773406982, -0.02394878678023815, -0.05248326435685158, -0.1251591593027115, 0.047821927815675735, 0.10447535663843155, -0.08778475970029831, 0.060278598219156265, -0.01645345613360405, 0.12023623287677765, 0.010460845194756985, -0.07539622485637665, 0.11775378882884979, -0.0814664214849472, -0.18040284514427185, -0.07083038240671158, 0.103786401450634, 0.014117957092821598, 0.047817669808864594, -0.009677080437541008, 0.03923993185162544, 0.002417723648250103, -0.06778563559055328, 0.038554802536964417, 0.00799634214490652, 0.0664684846997261, 0.005248297471553087, -0.03537459298968315, -0.009445303119719028, -0.038032930344343185, -0.018614396452903748, 0.10258279740810394, 0.2858559191226959, -0.09174595773220062, 0.043435703963041306, 0.06287512928247452, -0.05134572088718414, -0.18663649260997772, 0.014912489801645279, 0.046128153800964355, 0.012244705110788345, 0.027633991092443466, -0.15300917625427246, 0.06398411840200424, 0.08751890808343887, -0.03932946175336838, 0.09447719901800156, -0.28028127551078796, -0.13679492473602295, 0.11844620108604431, 0.12983833253383636, 0.08792094141244888, -0.16382531821727753, -0.047787584364414215, -0.03486673906445503, -0.11903063207864761, 0.08798649907112122, -0.12250135093927383, 0.0944283977150917, -0.022650016471743584, 0.043348267674446106, 0.0012730059679597616, -0.07059711217880249, 0.14546525478363037, -0.026982290670275688, 0.10106116533279419, -0.05398477986454964, 0.05826641991734505, 0.10176613926887512, -0.08417128026485443, 0.04378439858555794, -0.09797298908233643, 0.04975639656186104, -0.043307360261678696, -0.014738280326128006, -0.05067186430096626, 0.0411379411816597, -0.03175129368901253, -0.03624480590224266, -0.049566760659217834, 0.02381240203976631, 0.0601099468767643, -0.029525136575102806, 0.18412794172763824, 0.04490215703845024, 0.1616552770137787, 0.14288583397865295, 0.04926128312945366, -0.07655595242977142, -0.05355421453714371, 0.009288115426898003, -0.043236635625362396, 0.07229422777891159, -0.15639398992061615, 0.047403816133737564, 0.11368545889854431, 0.010317266918718815, 0.12312614172697067, 0.05759540572762489, -0.06620526313781738, 0.01928558386862278, 0.06994428485631943, -0.14683598279953003, -0.09526818245649338, 0.018299881368875504, 0.02451479248702526, -0.13082453608512878, 0.04962817206978798, 0.11752869933843613, -0.06532765924930573, 0.0018087526550516486, -0.00023059823433868587, 0.03102825954556465, -0.022211289033293724, 0.19370624423027039, 0.04785937815904617, 0.057808127254247665, -0.0892355665564537, 0.0865539014339447, 0.052678775042295456, -0.10799000412225723, 0.0393902063369751, 0.059997621923685074, -0.08282417058944702, -0.02082247845828533, 0.06989431381225586, 0.1571425199508667, -0.013914833776652813, -0.06830068677663803, -0.15158089995384216, -0.10554589331150055, 0.06552886217832565, 0.180522620677948, 0.06690268218517303, 0.0014143859734758735, -0.007165413815528154, 0.009674154222011566, -0.10738212615251541, 0.11309722065925598, 0.06435650587081909, 0.08952701836824417, -0.13694007694721222, 0.10108441859483719, -0.0021844746079295874, 0.012739493511617184, -0.014728045091032982, 0.04987829551100731, -0.10191471874713898, -0.016882428899407387, -0.15259550511837006, 0.008324503898620605, -0.050296053290367126, -0.008255255408585072, -0.006849725265055895, -0.0545625239610672, -0.06210707500576973, 0.035100117325782776, -0.0956798791885376, -0.04529031738638878, 0.018412161618471146, 0.024944061413407326, -0.1356160193681717, -0.04178032651543617, 0.02352851629257202, -0.0879603698849678, 0.0517205111682415, 0.025408975780010223, 0.01577322743833065, 0.030724624171853065, -0.12014003098011017, -0.0005793813616037369, 0.05054594203829765, 0.011951831169426441, 0.055667247623205185, -0.12021429091691971, -0.01930922456085682, 0.006084160413593054, 0.02458237297832966, 0.004717313684523106, 0.0918571949005127, -0.12409979104995728, 0.002772481879219413, -0.02528727799654007, -0.05138152837753296, -0.04341578111052513, 0.0384109690785408, 0.1144997626543045, 0.01068863831460476, 0.1953538954257965, -0.10335207730531693, 0.021899618208408356, -0.19999021291732788, 0.004757632967084646, -0.00924597680568695, -0.09985886514186859, -0.08506413549184799, -0.009663772769272327, 0.0640118420124054, -0.06539292633533478, 0.11809621751308441, -0.008905558846890926, 0.0359329991042614, 0.049590956419706345, -0.06126805767416954, 0.001733968616463244, 0.02207482047379017, 0.19430483877658844, 0.02635541372001171, -0.03717627748847008, 0.0569160096347332, 0.00881314929574728, 0.09610021859407425, 0.08553659915924072, 0.1849120408296585, 0.15546059608459473, 0.024266615509986877, 0.10373647511005402, 0.0434119738638401, -0.05644964799284935, -0.15392062067985535, 0.04901226982474327, -0.03169780597090721, 0.13040824234485626, 0.005825640633702278, 0.1670171618461609, 0.11394451558589935, -0.15800145268440247, 0.037929873913526535, -0.02798941731452942, -0.08302612602710724, -0.12528777122497559, -0.079337939620018, -0.09820064157247543, -0.1389428824186325, 0.0005781484069302678, -0.09768421947956085, 0.03814869001507759, 0.07416105270385742, 0.007288954686373472, 0.008760210126638412, 0.16157862544059753, 0.00874693039804697, 0.04663119837641716, 0.042165372520685196, 0.01368609257042408, -0.0435195155441761, -0.002899890299886465, -0.0970175489783287, 0.01821286976337433, -0.0027463024016469717, 0.04007018357515335, -0.023479441180825233, -0.01635018177330494, 0.06704368442296982, 0.00470003392547369, -0.11859958618879318, 0.010266878642141819, 0.022855916991829872, 0.052282821387052536, 0.06411406397819519, 0.021748967468738556, 0.023412570357322693, -0.01386073138564825, 0.17622603476047516, -0.07799076288938522, -0.06961801648139954, -0.1244911476969719, 0.20642900466918945, 0.01882910542190075, -0.027404239401221275, 0.026043137535452843, -0.08486543595790863, -0.001503948587924242, 0.16362667083740234, 0.16495612263679504, -0.020722387358546257, -0.003615807509049773, -0.02542387694120407, -0.02053884044289589, -0.05625138431787491, 0.10192037373781204, 0.10782063752412796, 0.026898043230175972, -0.05532519146800041, -0.051367346197366714, -0.05169776454567909, -0.016959121450781822, -0.08912970125675201, 0.035254254937171936, -0.0007759317522868514, 0.0005033444613218307, -0.02881760150194168, 0.054481297731399536, -0.003194700926542282, -0.10600126534700394, 0.06664863228797913, -0.17649529874324799, -0.16212904453277588, -0.002728618448600173, 0.09534201771020889, 0.0007279158453457057, 0.03953177109360695, -0.019249537959694862, 0.005554592702537775, 0.10760483145713806, -0.02477426454424858, -0.05016166716814041, -0.09279587864875793, 0.0753636509180069, -0.11122320592403412, 0.23952442407608032, -0.011236261576414108, 0.05652843043208122, 0.1320514678955078, 0.040049564093351364, -0.10619137436151505, 0.06340714544057846, 0.05588085949420929, -0.07518181949853897, 0.022369548678398132, 0.12421630322933197, -0.04549291729927063, 0.09230838716030121, 0.04732559993863106, -0.11155465245246887, -0.020714731886982918, -0.061260659247636795, -0.07135137915611267, -0.04557742923498154, -0.026436805725097656, -0.05927091836929321, 0.12585482001304626, 0.18182623386383057, -0.04161764681339264, 0.0014059336390346289, -0.04396931082010269, 0.03491686284542084, 0.07208522409200668, 0.017739124596118927, -0.040379252284765244, -0.2386053204536438, 0.04439062252640724, 0.05500084161758423, 0.005819625686854124, -0.2587345838546753, -0.09532502293586731, -0.005179466679692268, -0.052698686718940735, -0.09506890177726746, 0.08695334196090698, 0.0908050686120987, 0.051425494253635406, -0.06356833130121231, -0.05953836068511009, -0.07336734980344772, 0.14384245872497559, -0.14825864136219025, -0.07860464602708817 ]
null
null
transformers
# merged This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details Made as a test model, not sure about quality, probably will not make any quants unless someone finds out it's good and asks. ### Merge Method This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [TheBloke/Llama-2-13B-fp16](https://huggingface.co/TheBloke/Llama-2-13B-fp16) as a base. ### Models Merged The following models were included in the merge: * [Masterjp123/SnowyRP-FinalV1-L2-13B](https://huggingface.co/Masterjp123/SnowyRP-FinalV1-L2-13B) * [posicube/Llama2-chat-AYB-13B](https://huggingface.co/posicube/Llama2-chat-AYB-13B) * [Sao10K/Stheno-1.8-L2-13B](https://huggingface.co/Sao10K/Stheno-1.8-L2-13B) * [ValiantLabs/ShiningValiantXS](https://huggingface.co/ValiantLabs/ShiningValiantXS) * [sauce1337/BerrySauce-L2-13b](https://huggingface.co/sauce1337/BerrySauce-L2-13b) ### Configuration The following YAML configuration was used to produce this model: ```yaml base_model: model: path: TheBloke/Llama-2-13B-fp16 dtype: bfloat16 merge_method: ties parameters: int8_mask: 1.0 normalize: 1.0 slices: - sources: - layer_range: [0, 40] model: model: path: Masterjp123/Snowyrp-V2B-P1 parameters: density: [1.0, 0.7, 0.1] weight: 1.0 - layer_range: [0, 40] model: model: path: Masterjp123/SnowyRP-FinalV1-L2-13B parameters: density: 0.5 weight: [0.0, 0.3, 0.7, 1.0] - layer_range: [0, 40] model: model: path: sauce1337/BerrySauce-L2-13b parameters: density: 0.33 weight: - filter: mlp value: 0.5 - value: 0.0 - layer_range: [0, 40] model: model: path: TheBloke/Llama-2-13B-fp16 ``` for Masterjp123/Snowyrp-V2B-P1 ```yaml base_model: model: path: TheBloke/Llama-2-13B-fp16 dtype: bfloat16 merge_method: ties parameters: int8_mask: 1.0 normalize: 1.0 slices: - sources: - layer_range: [0, 40] model: model: path: Sao10K/Stheno-1.8-L2-13B parameters: density: [1.0, 0.7, 0.1] weight: 1.0 - layer_range: [0, 40] model: model: path: ValiantLabs/ShiningValiantXS parameters: density: 0.5 weight: [0.0, 0.3, 0.7, 1.0] - layer_range: [0, 40] model: model: path: posicube/Llama2-chat-AYB-13B parameters: density: 0.33 weight: - filter: mlp value: 0.5 - value: 0.0 - layer_range: [0, 40] model: model: path: TheBloke/Llama-2-13B-fp16 ```
{"library_name": "transformers", "tags": ["mergekit", "merge"], "base_model": ["TheBloke/Llama-2-13B-fp16", "Masterjp123/SnowyRP-FinalV1-L2-13B", "Masterjp123/Snowyrp-V2B-P1", "sauce1337/BerrySauce-L2-13b"]}
text-generation
Masterjp123/SnowyRP-V2-13B-L2_BetaTest
[ "transformers", "safetensors", "llama", "text-generation", "mergekit", "merge", "arxiv:2306.01708", "base_model:TheBloke/Llama-2-13B-fp16", "base_model:Masterjp123/SnowyRP-FinalV1-L2-13B", "base_model:Masterjp123/Snowyrp-V2B-P1", "base_model:sauce1337/BerrySauce-L2-13b", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T02:12:00+00:00
[ "2306.01708" ]
[]
TAGS #transformers #safetensors #llama #text-generation #mergekit #merge #arxiv-2306.01708 #base_model-TheBloke/Llama-2-13B-fp16 #base_model-Masterjp123/SnowyRP-FinalV1-L2-13B #base_model-Masterjp123/Snowyrp-V2B-P1 #base_model-sauce1337/BerrySauce-L2-13b #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# merged This is a merge of pre-trained language models created using mergekit. ## Merge Details Made as a test model, not sure about quality, probably will not make any quants unless someone finds out it's good and asks. ### Merge Method This model was merged using the TIES merge method using TheBloke/Llama-2-13B-fp16 as a base. ### Models Merged The following models were included in the merge: * Masterjp123/SnowyRP-FinalV1-L2-13B * posicube/Llama2-chat-AYB-13B * Sao10K/Stheno-1.8-L2-13B * ValiantLabs/ShiningValiantXS * sauce1337/BerrySauce-L2-13b ### Configuration The following YAML configuration was used to produce this model: for Masterjp123/Snowyrp-V2B-P1
[ "# merged\n\nThis is a merge of pre-trained language models created using mergekit.", "## Merge Details\n\nMade as a test model, not sure about quality, probably will not make any quants unless someone finds out it's good and asks.", "### Merge Method\n\nThis model was merged using the TIES merge method using TheBloke/Llama-2-13B-fp16 as a base.", "### Models Merged\n\nThe following models were included in the merge:\n* Masterjp123/SnowyRP-FinalV1-L2-13B\n* posicube/Llama2-chat-AYB-13B\n* Sao10K/Stheno-1.8-L2-13B\n* ValiantLabs/ShiningValiantXS\n* sauce1337/BerrySauce-L2-13b", "### Configuration\n\nThe following YAML configuration was used to produce this model:\n\n\nfor Masterjp123/Snowyrp-V2B-P1" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #arxiv-2306.01708 #base_model-TheBloke/Llama-2-13B-fp16 #base_model-Masterjp123/SnowyRP-FinalV1-L2-13B #base_model-Masterjp123/Snowyrp-V2B-P1 #base_model-sauce1337/BerrySauce-L2-13b #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# merged\n\nThis is a merge of pre-trained language models created using mergekit.", "## Merge Details\n\nMade as a test model, not sure about quality, probably will not make any quants unless someone finds out it's good and asks.", "### Merge Method\n\nThis model was merged using the TIES merge method using TheBloke/Llama-2-13B-fp16 as a base.", "### Models Merged\n\nThe following models were included in the merge:\n* Masterjp123/SnowyRP-FinalV1-L2-13B\n* posicube/Llama2-chat-AYB-13B\n* Sao10K/Stheno-1.8-L2-13B\n* ValiantLabs/ShiningValiantXS\n* sauce1337/BerrySauce-L2-13b", "### Configuration\n\nThe following YAML configuration was used to produce this model:\n\n\nfor Masterjp123/Snowyrp-V2B-P1" ]
[ 142, 19, 35, 35, 88, 33 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #arxiv-2306.01708 #base_model-TheBloke/Llama-2-13B-fp16 #base_model-Masterjp123/SnowyRP-FinalV1-L2-13B #base_model-Masterjp123/Snowyrp-V2B-P1 #base_model-sauce1337/BerrySauce-L2-13b #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# merged\n\nThis is a merge of pre-trained language models created using mergekit.## Merge Details\n\nMade as a test model, not sure about quality, probably will not make any quants unless someone finds out it's good and asks.### Merge Method\n\nThis model was merged using the TIES merge method using TheBloke/Llama-2-13B-fp16 as a base.### Models Merged\n\nThe following models were included in the merge:\n* Masterjp123/SnowyRP-FinalV1-L2-13B\n* posicube/Llama2-chat-AYB-13B\n* Sao10K/Stheno-1.8-L2-13B\n* ValiantLabs/ShiningValiantXS\n* sauce1337/BerrySauce-L2-13b### Configuration\n\nThe following YAML configuration was used to produce this model:\n\n\nfor Masterjp123/Snowyrp-V2B-P1" ]
[ -0.06799020618200302, 0.07541470974683762, -0.005155785474926233, 0.009567887522280216, 0.050647083669900894, 0.029259109869599342, 0.1316298097372055, 0.10773096978664398, 0.0024154328275471926, 0.08616536110639572, 0.02339641936123371, 0.03101377934217453, 0.09737295657396317, 0.20185697078704834, 0.00857671070843935, -0.14554528892040253, 0.06759124249219894, -0.03733186051249504, -0.10327411442995071, 0.1008431613445282, 0.11907374113798141, -0.03977331891655922, 0.08403754234313965, 0.0136281568557024, -0.10131736844778061, 0.03886411711573601, -0.003323118668049574, -0.004318751394748688, 0.06368225812911987, 0.08301066607236862, 0.047606319189071655, 0.04100482165813446, -0.00030084452009759843, -0.17324496805667877, 0.019041379913687706, 0.01645820215344429, -0.011157783679664135, 0.0477256253361702, 0.08270556479692459, -0.029070045799016953, 0.08433569967746735, -0.07232329994440079, 0.020011039450764656, 0.0966881811618805, -0.13520890474319458, -0.12654393911361694, -0.15500739216804504, 0.17256984114646912, 0.0790364071726799, 0.04218218848109245, -0.030718961730599403, 0.07052802294492722, 0.03488422930240631, 0.06509310752153397, 0.06351271271705627, -0.21929602324962616, -0.04847535118460655, 0.055669691413640976, 0.03642153739929199, -0.06165097653865814, 0.027202535420656204, 0.017916876822710037, 0.021030044183135033, 0.003826157422736287, -0.011226614005863667, -0.04198579862713814, 0.13589046895503998, -0.0722346305847168, -0.1048608049750328, -0.0518641322851181, 0.08969399333000183, 0.04592069983482361, -0.043249573558568954, -0.07324213534593582, -0.07202795147895813, -0.06664713472127914, -0.0015724549302831292, -0.05114365741610527, -0.0010123047977685928, -0.01664881780743599, 0.11125344783067703, -0.06753063201904297, -0.03987054526805878, -0.024524973705410957, -0.016588861122727394, 0.1394096463918686, 0.005497843027114868, 0.033343326300382614, -0.009831869974732399, 0.046536557376384735, -0.16861435770988464, -0.1367582082748413, 0.010489609092473984, -0.050670064985752106, -0.12139295786619186, -0.0072048120200634, -0.0248549971729517, -0.08068329095840454, 0.05066447705030441, 0.204121932387352, -0.12157688289880753, 0.040262091904878616, 0.03269822895526886, 0.03267166018486023, 0.015495711006224155, 0.01019047200679779, -0.0866478905081749, -0.09441550076007843, 0.006411131005734205, 0.04572445526719093, 0.0665554627776146, 0.00904410146176815, -0.054283928126096725, -0.03206872567534447, -0.005789242219179869, 0.0033891794737428427, 0.07338374108076096, 0.024018680676817894, -0.06956548988819122, -0.03429080545902252, 0.11277884244918823, -0.11376851052045822, 0.027081569656729698, 0.014370121993124485, -0.030617060139775276, -0.0022071513812988997, 0.08200032263994217, 0.01815260387957096, -0.01447385735809803, 0.07533856481313705, -0.024670137092471123, -0.03935593366622925, -0.08352413028478622, -0.056427258998155594, 0.03304228186607361, -0.06002317741513252, -0.05656671151518822, -0.09533163160085678, -0.14202982187271118, -0.08992260694503784, 0.01989104598760605, -0.06945096701383591, -0.049817342311143875, -0.004420683719217777, -0.024898312985897064, 0.02676742896437645, 0.011679514311254025, -0.011055640876293182, 0.017687493935227394, 0.0006982923368923366, 0.041582345962524414, 0.02983701229095459, -0.08503027260303497, 0.006161809433251619, -0.07404104620218277, 0.11371918767690659, -0.21999968588352203, 0.08519721031188965, -0.08226945251226425, 0.04006972536444664, -0.13551239669322968, -0.00659277755767107, -0.02255813404917717, -0.0036699865013360977, 0.06592918932437897, 0.15520671010017395, -0.16969570517539978, -0.03416571021080017, 0.14570307731628418, -0.14390908181667328, -0.08984269946813583, 0.09956970065832138, 0.027243033051490784, 0.01385418325662613, 0.03316070884466171, 0.16878297924995422, 0.1506810337305069, -0.00217943312600255, -0.04757373034954071, -0.048363685607910156, 0.05149213224649429, 0.07773452997207642, 0.08817458897829056, -0.08159137517213821, -0.10377207398414612, 0.030372094362974167, -0.04751182720065117, 0.03788309171795845, -0.02820557728409767, -0.06832782179117203, -0.049780528992414474, -0.0417233444750309, 0.07040446251630783, -0.019823184236884117, 0.02302587404847145, -0.0454193651676178, -0.09645017981529236, -0.00963133480399847, 0.08556094765663147, -0.007670531049370766, 0.009810999035835266, -0.07191382348537445, 0.16612191498279572, -0.016962306573987007, 0.010625721886754036, -0.1257731169462204, -0.037556737661361694, 0.01509099081158638, -0.10378661751747131, 0.03172331303358078, -0.0646488294005394, 0.06268081068992615, 0.03152208775281906, -0.028196970000863075, -0.05900149047374725, 0.016222892329096794, 0.015948040410876274, -0.009882657788693905, -0.16028200089931488, -0.11767082661390305, -0.0380685031414032, 0.22652599215507507, -0.05541418492794037, 0.05437471717596054, 0.00477732066065073, 0.20721474289894104, -0.030480697751045227, -0.05098685994744301, 0.03497319668531418, 0.039241522550582886, -0.0003239985671825707, 0.003440545406192541, 0.052016813308000565, 0.01802295818924904, -0.1427353322505951, 0.07293421030044556, -0.11259189993143082, -0.08027241379022598, 0.06025664508342743, 0.06341123580932617, -0.10078176856040955, -0.02676643803715706, -0.022993789985775948, -0.02806224301457405, 0.039080265909433365, -0.059409406036138535, 0.06495349109172821, 0.05395224317908287, 0.08759862184524536, -0.03140562027692795, -0.017555367201566696, 0.03043377585709095, -0.03544872626662254, -0.041858039796352386, 0.10374906659126282, 0.008227691985666752, -0.15404057502746582, 0.1050863191485405, 0.08500435203313828, 0.060250718146562576, 0.05150517076253891, 0.0031461361795663834, -0.02145126461982727, -0.09935756027698517, 0.043678056448698044, 0.01643167994916439, 0.0419515036046505, -0.060158345848321915, 0.05379008129239082, 0.03878246992826462, -0.01080599520355463, 0.022965388372540474, -0.07670509070158005, 0.036153458058834076, 0.04615242779254913, -0.024946684017777443, 0.06876666098833084, 0.05203241854906082, 0.00754490727558732, 0.04897463694214821, 0.04930780827999115, 0.05019048973917961, -0.0001611918123671785, -0.04647724702954292, -0.10843255370855331, 0.16143399477005005, -0.13869057595729828, -0.18355974555015564, -0.19378791749477386, -0.011564764194190502, -0.0666041150689125, 0.0009802988497540355, 0.03902396932244301, -0.045746099203825, -0.040712740272283554, -0.07432965934276581, 0.0652264729142189, 0.025489583611488342, -0.03896188363432884, 0.0035277598071843386, -0.003306394210085273, 0.09480686485767365, -0.05732959881424904, -0.027527617290616035, 0.041355330497026443, -0.015045268461108208, 0.03537276014685631, 0.014276571571826935, 0.08866717666387558, 0.10564074665307999, -0.0035331714898347855, 0.008460222743451595, -0.0071428739465773106, 0.2126501500606537, -0.05348632112145424, 0.07748160511255264, 0.15851125121116638, -0.04835741966962814, 0.07691625505685806, 0.19035585224628448, 0.01885947585105896, -0.02363070845603943, -0.007888342253863811, 0.0019403229234740138, -0.022607767954468727, -0.20645710825920105, -0.11276907473802567, -0.04761222377419472, -0.007487457245588303, 0.033619459718465805, 0.04123018682003021, -0.03971433266997337, 0.06916075199842453, -0.08179724961519241, -0.02685314044356346, -0.0205826535820961, 0.04126216471195221, 0.1415814906358719, -0.005109524819999933, 0.048176608979701996, -0.02701379545032978, 0.03476063907146454, 0.06513148546218872, 0.030535832047462463, 0.00903500709682703, 0.04757993295788765, 0.10247103124856949, 0.11061682552099228, 0.1179545670747757, 0.01781894452869892, -0.007942069321870804, 0.006858705077320337, 0.020810546353459358, -0.004363254643976688, -0.11431537568569183, -0.03263086825609207, 0.06048649549484253, -0.05352920666337013, 0.11348427087068558, -0.08853058516979218, 0.05431511998176575, 0.0330343097448349, 0.21956457197666168, 0.08700764924287796, -0.21623755991458893, -0.08702398091554642, 0.05892825871706009, 0.05276312679052353, -0.06399279087781906, -0.02818729169666767, -0.010919923894107342, -0.11348014324903488, 0.1422882080078125, -0.016261830925941467, 0.06790794432163239, 0.04103311896324158, -0.00713821267709136, 0.060953982174396515, 0.05589866265654564, 0.022839359939098358, 0.04198018088936806, -0.13120995461940765, 0.17263537645339966, 0.016562750563025475, -0.0011964282020926476, 0.04541494697332382, 0.07168442755937576, -0.010918961837887764, 0.10652266442775726, 0.14532117545604706, 0.04048655554652214, -0.04546637833118439, -0.09030425548553467, -0.1056453064084053, -0.03373704105615616, 0.04192781448364258, -0.06399234384298325, 0.09275020658969879, -0.00901807565242052, -0.03476161137223244, -0.024778710678219795, 0.10695018619298935, -0.16034972667694092, -0.1094956025481224, 0.08704308420419693, -0.027533061802387238, 0.0139243733137846, -0.09780935198068619, -0.029233448207378387, -0.09593352675437927, 0.22407571971416473, -0.07693023234605789, -0.0775609165430069, -0.10942380130290985, 0.03303808718919754, 0.15161851048469543, -0.08545487374067307, 0.056678999215364456, -0.05849427729845047, 0.10773178935050964, -0.07866308838129044, -0.10082345455884933, 0.0631251260638237, -0.07997152954339981, -0.15137970447540283, -0.023388845846056938, 0.1682211011648178, -0.007034895941615105, 0.05298668518662453, 0.012496309354901314, 0.05329097434878349, 0.03437920659780502, -0.05267566815018654, 0.05104726552963257, 0.14356310665607452, -0.01630067452788353, 0.10931456834077835, -0.023983152583241463, -0.1315286010503769, -0.06969086825847626, -0.02172447182238102, 0.1165536642074585, 0.2940150797367096, -0.0864664614200592, 0.09588085114955902, 0.08373280614614487, -0.09787137806415558, -0.15799450874328613, -0.01701611652970314, 0.07070469111204147, 0.06659837812185287, -0.022008704021573067, -0.07302950322628021, 0.04555734992027283, 0.078931525349617, -0.023679478093981743, -0.0302265677601099, -0.2941291928291321, -0.1289454996585846, 0.03189842775464058, 0.0094910254701972, 0.02522403560578823, -0.16250286996364594, -0.1352681666612625, -0.06858526170253754, -0.18523910641670227, 0.025934362784028053, -0.032505691051483154, 0.02030600979924202, -0.03290838748216629, 0.015892133116722107, 0.048509370535612106, -0.04116872698068619, 0.138995960354805, -0.02902906946837902, -0.010229389183223248, -0.06955771893262863, -0.03235283121466637, 0.052692461758852005, -0.05454796552658081, 0.050593458116054535, 0.05936812236905098, 0.03566037118434906, -0.15628774464130402, 0.017120905220508575, -0.09991058707237244, 0.0772738829255104, -0.08743716776371002, -0.013707410544157028, -0.022927913814783096, 0.0732002779841423, 0.045489538460969925, -0.004376649856567383, 0.037785544991493225, -0.07403503358364105, 0.146783709526062, 0.2262619584798813, 0.06121448054909706, 0.029321886599063873, -0.07224742323160172, 0.047109950333833694, -0.04655548185110092, 0.024718379601836205, -0.0667988508939743, -0.002560568507760763, 0.08746913820505142, 0.014762714505195618, 0.10700392723083496, -0.0019241963746026158, -0.11420626193284988, -0.014355342835187912, 0.08597546070814133, -0.16292056441307068, -0.19209647178649902, -0.03950348123908043, 0.0590161569416523, -0.08194934576749802, -0.028668906539678574, 0.1635604351758957, -0.03942961245775223, -0.034057628363370895, 0.01654307171702385, 0.045256856828927994, -0.09890937060117722, 0.10938327759504318, 0.03974071517586708, 0.06317665427923203, -0.06620830297470093, 0.07756142318248749, 0.08317792415618896, -0.1130189597606659, 0.017672410234808922, 0.14707832038402557, -0.08467567712068558, -0.06133782118558884, -0.050174396485090256, 0.09031544625759125, -0.03484529256820679, -0.023890864104032516, -0.08038569241762161, -0.13760873675346375, 0.014944677241146564, 0.11712634563446045, 0.038942985236644745, -0.004627944435924292, 0.05185754969716072, -0.05183955281972885, -0.012470759451389313, 0.06726255267858505, 0.06698545813560486, 0.0734928771853447, -0.08035343140363693, 0.1250813752412796, -0.03796021267771721, 0.056705836206674576, -0.002625886583700776, 0.010563792660832405, -0.1195620745420456, -0.015135755762457848, -0.1121009811758995, 0.005102819763123989, -0.10446961224079132, -0.05216260254383087, -0.021808303892612457, -0.009606094099581242, 0.0006696734926663339, -0.002323688706383109, -0.04882727935910225, -0.06791208684444427, -0.03910176083445549, 0.03682912513613701, -0.08308294415473938, -0.011797056533396244, 0.04055619612336159, -0.059554751962423325, 0.052703600376844406, 0.0172024667263031, 0.02241312339901924, -0.051204778254032135, -0.09582964330911636, -0.05189263075590134, 0.006723927333950996, 0.0006128610111773014, 0.02466222271323204, -0.17374743521213531, -0.008137949742376804, -0.022717278450727463, -0.06729425489902496, -0.012992171570658684, 0.08298096060752869, -0.10684843361377716, -0.0076409438624978065, 0.027929050847887993, -0.017695186659693718, -0.0865316167473793, 0.019881708547472954, 0.028754446655511856, 0.0582294836640358, 0.07868731021881104, -0.06335022300481796, 0.08821024000644684, -0.1648785024881363, -0.022289711982011795, -0.02666422165930271, -0.01122727245092392, -0.05670635402202606, -0.02346956916153431, 0.04628957435488701, 0.0007334351539611816, 0.04431794583797455, -0.0201362706720829, 0.0011494802311062813, 0.052616219967603683, -0.14393377304077148, 0.011515077203512192, 0.046123865991830826, 0.12126592546701431, 0.036583803594112396, -0.002615647856146097, 0.027066688984632492, 0.0038645665626972914, -0.05854150652885437, -0.02684791013598442, 0.08877205848693848, 0.14331252872943878, 0.08672889322042465, 0.06500270962715149, 0.13844716548919678, -0.008421315811574459, 0.003212918294593692, -0.04989755526185036, -0.0502169169485569, 0.07941809296607971, -0.04774775356054306, 0.13014495372772217, 0.07747799158096313, -0.20709463953971863, 0.10123134404420853, -0.035995833575725555, -0.05542192608118057, -0.1004541888833046, -0.13587459921836853, -0.06970774382352829, -0.06810607761144638, -0.008792187087237835, -0.08913400024175644, 0.017624374479055405, -0.011097400449216366, -0.013991636224091053, 0.0060655586421489716, 0.1201833188533783, -0.05289554223418236, -0.02908177860081196, 0.04104814678430557, 0.0370674654841423, 0.02225431054830551, -0.017276108264923096, -0.01039296854287386, 0.05653354525566101, 0.05161658301949501, 0.0072947614826262, 0.0453699566423893, 0.026210477575659752, 0.03065836988389492, -0.0056998333893716335, -0.108124740421772, 0.01027858629822731, 0.01776532642543316, 0.08422192931175232, 0.037742361426353455, 0.028352225199341774, 0.001601855969056487, -0.04394166171550751, 0.10974209755659103, -0.05229000002145767, -0.08403422683477402, -0.10063815861940384, 0.15927471220493317, -0.027362443506717682, 0.020688271149992943, -0.011627043597400188, -0.1035882979631424, 0.03368694335222244, 0.14058329164981842, 0.26281094551086426, 0.01492584589868784, -0.0006803294527344406, 0.02472733147442341, 0.01905571110546589, 0.016189631074666977, 0.0822034701704979, 0.018473487347364426, 0.11382436007261276, -0.056623540818691254, 0.11965162307024002, -0.018025172874331474, -0.06216008961200714, -0.06671547144651413, 0.05570089444518089, -0.014530796557664871, 0.008782911114394665, 0.028270872309803963, 0.11451251059770584, -0.03544265776872635, -0.11626987904310226, 0.03319542855024338, -0.1314055323600769, -0.13152606785297394, -0.0644579753279686, 0.04249519482254982, 0.03600650280714035, 0.09091976284980774, -0.02884848602116108, -0.03571785241365433, 0.19379062950611115, 0.00552706653252244, -0.05436750128865242, -0.10978763550519943, 0.03961498290300369, -0.09685828536748886, 0.12381584942340851, 0.014387146569788456, 0.04810890927910805, 0.11479704082012177, -0.013380306772887707, -0.1383296698331833, -0.015457669273018837, 0.05666455253958702, -0.016714006662368774, 0.02474713698029518, 0.11473926156759262, -0.0010776014532893896, 0.09892130643129349, 0.026007678359746933, -0.1093532145023346, 0.05532291159033775, 0.025110816583037376, -0.009051084518432617, -0.07828828692436218, 0.09301130473613739, -0.07643944025039673, 0.1540406495332718, 0.22858625650405884, -0.0775531530380249, -0.0006683453684672713, -0.05220068246126175, 0.02582068182528019, 0.053201816976070404, 0.11292269825935364, -0.0008597323321737349, -0.16507484018802643, 0.047637950628995895, 0.03448212519288063, 0.057443972676992416, -0.2620425522327423, -0.08636151254177094, -0.05527028813958168, -0.003487078472971916, -0.012575479224324226, 0.12424211204051971, 0.06356550008058548, 0.051032647490501404, -0.024324491620063782, -0.13876157999038696, -0.009983903728425503, 0.12382519245147705, -0.09405086189508438, -0.07970596849918365 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Whisper Large Korean/English This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.8019 - Wer: 198.2263 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 773 - training_steps: 7728 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.5546 | 1.0 | 773 | 0.5308 | 240.1463 | | 0.3284 | 2.0 | 1546 | 0.5160 | 133.6395 | | 0.176 | 3.0 | 2319 | 0.5582 | 264.5033 | | 0.0977 | 4.0 | 3092 | 0.6110 | 155.6417 | | 0.065 | 5.0 | 3865 | 0.6577 | 194.4118 | | 0.0298 | 6.0 | 4638 | 0.7021 | 235.0691 | | 0.0109 | 7.0 | 5411 | 0.7408 | 158.8282 | | 0.0069 | 8.0 | 6184 | 0.7550 | 201.9574 | | 0.0057 | 9.0 | 6957 | 0.8019 | 198.2263 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.0.1 - Datasets 2.17.0 - Tokenizers 0.15.2
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "openai/whisper-large-v3", "model-index": [{"name": "Whisper Large Korean/English", "results": []}]}
automatic-speech-recognition
gcasey2/whisper-large-v3-ko-en
[ "transformers", "tensorboard", "safetensors", "whisper", "automatic-speech-recognition", "generated_from_trainer", "base_model:openai/whisper-large-v3", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2024-02-15T02:12:14+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #base_model-openai/whisper-large-v3 #license-apache-2.0 #endpoints_compatible #region-us
Whisper Large Korean/English ============================ This model is a fine-tuned version of openai/whisper-large-v3 on the None dataset. It achieves the following results on the evaluation set: * Loss: 0.8019 * Wer: 198.2263 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 1e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 773 * training\_steps: 7728 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.37.2 * Pytorch 2.0.1 * Datasets 2.17.0 * Tokenizers 0.15.2
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 773\n* training\\_steps: 7728\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.0.1\n* Datasets 2.17.0\n* Tokenizers 0.15.2" ]
[ "TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #base_model-openai/whisper-large-v3 #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 773\n* training\\_steps: 7728\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.0.1\n* Datasets 2.17.0\n* Tokenizers 0.15.2" ]
[ 72, 132, 4, 30 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #base_model-openai/whisper-large-v3 #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 773\n* training\\_steps: 7728\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.0.1\n* Datasets 2.17.0\n* Tokenizers 0.15.2" ]
[ -0.1295376867055893, 0.11741089820861816, -0.003631713567301631, 0.07757934927940369, 0.1003364771604538, -0.000902308791410178, 0.12844978272914886, 0.15018734335899353, -0.06001100316643715, 0.060835100710392, 0.08525029569864273, 0.09183002263307571, 0.0593733936548233, 0.16503985226154327, -0.04780369997024536, -0.27810677886009216, 0.027347689494490623, 0.010087779723107815, -0.06842724978923798, 0.12133549898862839, 0.09258700907230377, -0.11179850995540619, 0.029420332983136177, 0.009699585847556591, -0.08774469047784805, -0.018542451784014702, 0.0054466878063976765, -0.09899623692035675, 0.10841196030378342, -0.0022666577715426683, 0.08785979449748993, 0.058723773807287216, 0.08022847026586533, -0.19199925661087036, 0.011398185975849628, 0.06399961560964584, 0.03676942363381386, 0.07868444174528122, 0.0672757476568222, -0.0069983708672225475, 0.0723102018237114, -0.07301387190818787, 0.08242794871330261, 0.0392688512802124, -0.0955989882349968, -0.33976367115974426, -0.07750299572944641, 0.06329608708620071, 0.10946352779865265, 0.0780940055847168, -0.00861443392932415, 0.11433383077383041, -0.03729397431015968, 0.09621527045965195, 0.2629376947879791, -0.27637559175491333, -0.04982168972492218, -0.04957956820726395, 0.0349808894097805, 0.0686127170920372, -0.08561491966247559, -0.01975201442837715, 0.020364707335829735, 0.0403381772339344, 0.11819937825202942, 0.0048712934367358685, -0.045941490679979324, -0.030979951843619347, -0.1403874009847641, -0.063113272190094, 0.14156010746955872, 0.025659391656517982, -0.04856090247631073, -0.1085454449057579, -0.06716056913137436, -0.14240285754203796, -0.05237918347120285, -0.004836532287299633, 0.03236956521868706, -0.04850444942712784, -0.07661174982786179, -0.0244075246155262, -0.08645012229681015, -0.09813253581523895, 0.0064142849296331406, 0.18346697092056274, 0.06317444145679474, 0.015182691626250744, -0.01370447501540184, 0.09782707691192627, 0.032324619591236115, -0.1680227369070053, -0.016757193952798843, 0.025629857555031776, -0.02354195900261402, 0.010446268133819103, -0.04277804121375084, -0.041764624416828156, 0.02920745499432087, 0.14055579900741577, -0.0840296596288681, 0.06901728361845016, -0.010254095308482647, 0.039654552936553955, -0.11615262925624847, 0.1735021471977234, -0.03198488801717758, -0.012571987695991993, 0.008623173460364342, 0.13252481818199158, 0.06715176999568939, -0.0224315095692873, -0.07185297459363937, 0.008341549895703793, 0.11982186883687973, 0.05582543462514877, -0.008401439525187016, 0.04030503332614899, -0.05512285977602005, -0.02927078679203987, 0.02143065072596073, -0.12764465808868408, 0.008537188172340393, 0.027027282863855362, -0.07901739329099655, -0.04741726443171501, 0.038224875926971436, 0.004490263294428587, -0.04026707261800766, 0.06670793890953064, -0.0436137430369854, -0.0045809014700353146, -0.06773706525564194, -0.08766388893127441, 0.025483930483460426, -0.059682223945856094, -0.0042731547728180885, -0.0948227122426033, -0.1526612639427185, -0.01860959269106388, 0.04905349388718605, -0.03451230376958847, -0.06186462566256523, -0.042794227600097656, -0.09766015410423279, 0.03324928507208824, -0.03359679505228996, 0.09181483834981918, -0.05245296284556389, 0.114975206553936, 0.04470442235469818, 0.08004461973905563, 0.024411289021372795, 0.04949774965643883, -0.07959596812725067, 0.03973541408777237, -0.17396125197410583, 0.07244792580604553, -0.09208908677101135, 0.03623628616333008, -0.10925016552209854, -0.11238567531108856, 0.022999830543994904, -0.018263783305883408, 0.09986866265535355, 0.1091001108288765, -0.16052651405334473, -0.0845320001244545, 0.19710379838943481, -0.1384957730770111, -0.12056463211774826, 0.13929963111877441, 0.0014292397536337376, -0.02193624898791313, 0.05543458089232445, 0.2221170961856842, 0.10444021224975586, -0.11903348565101624, 0.008615195751190186, -0.009574724361300468, 0.06378203630447388, -0.028601359575986862, 0.10584191232919693, -0.006252259947359562, 0.01901855133473873, 0.02638566866517067, -0.04496212676167488, 0.0498916432261467, -0.08637188374996185, -0.09783336520195007, -0.04190460965037346, -0.09331301599740982, 0.05284550040960312, 0.06212455406785011, 0.04347636550664902, -0.11320207267999649, -0.10734739154577255, 0.017340246587991714, 0.12825754284858704, -0.09018964320421219, 0.030642541125416756, -0.10354779660701752, 0.09014322608709335, -0.02581951767206192, -0.012636351399123669, -0.1675838828086853, -0.003434273647144437, 0.0358998142182827, -0.07774024456739426, 0.0009255380718968809, -0.05865625664591789, 0.08355925977230072, 0.071803517639637, -0.058204732835292816, -0.07728476822376251, -0.050450168550014496, -0.005119116976857185, -0.08259840309619904, -0.20787879824638367, -0.07046445459127426, -0.03696460649371147, 0.12537576258182526, -0.16251727938652039, 0.03895308077335358, -0.000016543021047255024, 0.11652868241071701, 0.04263843595981598, -0.025393638759851456, 0.0034317844547331333, 0.0632501021027565, 0.0037558285985141993, -0.05910687893629074, 0.060120098292827606, 0.025860818102955818, -0.09686972945928574, 0.03263038024306297, -0.1489192396402359, 0.13991330564022064, 0.11339583247900009, -0.002190919825807214, -0.04477117955684662, 0.00654958700761199, -0.07442169636487961, -0.043580152094364166, -0.020780164748430252, 0.005618008319288492, 0.15891017019748688, 0.0026536965742707253, 0.14338858425617218, -0.10883071273565292, -0.04255419597029686, 0.029443569481372833, -0.03773042932152748, -0.008131303824484348, 0.12310922890901566, -0.014441213570535183, -0.05222802236676216, 0.11272752285003662, 0.12327470630407333, -0.08973528444766998, 0.14741820096969604, -0.08246553689241409, -0.0794539600610733, -0.017139876261353493, 0.019246403127908707, 0.026855703443288803, 0.10988268256187439, -0.12337756156921387, -0.0013559000799432397, 0.03564409166574478, 0.003908856306225061, 0.027811532840132713, -0.21197381615638733, -0.0043667638674378395, 0.021814744919538498, -0.08316446840763092, -0.021639561280608177, 0.0025632574688643217, 0.019985415041446686, 0.11111672222614288, -0.007951817475259304, -0.0702979639172554, 0.013345679268240929, -0.027090925723314285, -0.07572223991155624, 0.18132777512073517, -0.09618331491947174, -0.1904270201921463, -0.1251685470342636, -0.0021730647422373295, -0.029573457315564156, 0.0011732999701052904, 0.06514128297567368, -0.0838649645447731, -0.024097252637147903, -0.0958881676197052, 0.0059926011599600315, 0.012422507628798485, 0.03416655957698822, 0.05176093801856041, 0.025682607665657997, 0.1141284853219986, -0.10618133842945099, -0.0006767248269170523, -0.048432961106300354, -0.043054692447185516, 0.02631080336868763, 0.05203419178724289, 0.06956472992897034, 0.15166538953781128, 0.0021679389756172895, 0.022396011278033257, -0.03136737644672394, 0.16891001164913177, -0.07752560824155807, -0.0327468104660511, 0.18282033503055573, -0.010862195864319801, 0.06416083872318268, 0.13740862905979156, 0.04214762523770332, -0.07323720306158066, -0.008533542975783348, 0.0000945399806369096, -0.04202360287308693, -0.22730353474617004, -0.059860553592443466, -0.038063038140535355, 0.028260596096515656, 0.09374701976776123, 0.03484398126602173, 0.018271619454026222, 0.030826568603515625, -0.024424055591225624, -0.027870813384652138, 0.024480607360601425, 0.067755788564682, 0.10059414803981781, 0.02142246812582016, 0.12822775542736053, -0.032666102051734924, -0.03422614559531212, 0.02258029580116272, 0.0033368177246302366, 0.21455198526382446, -0.04488339647650719, 0.13788311183452606, 0.05073174461722374, 0.1592360883951187, 0.03197319060564041, 0.06011076644062996, -0.006296138744801283, -0.006128295324742794, 0.012146497145295143, -0.06005809083580971, -0.05593135207891464, 0.02020133286714554, -0.012114817276597023, 0.06323084980249405, -0.12719197571277618, 0.010025654919445515, 0.03513320907950401, 0.3237801194190979, 0.0668720155954361, -0.3085552453994751, -0.11104198545217514, 0.011282281018793583, -0.07007638365030289, -0.026741506531834602, 0.03514235094189644, 0.13803724944591522, -0.0726446732878685, 0.05751853436231613, -0.07083471864461899, 0.07592439651489258, -0.05642043054103851, 0.016567235812544823, 0.054873935878276825, 0.09352347254753113, 0.022462977096438408, 0.056004371494054794, -0.2415640950202942, 0.29590633511543274, -0.009865314699709415, 0.08785133063793182, -0.05465567111968994, 0.019485119730234146, 0.029216308146715164, -0.004108890425413847, 0.08906979113817215, -0.0162347499281168, -0.08104784041643143, -0.1869075447320938, -0.11716748774051666, 0.019922874867916107, 0.11427041888237, -0.011453714221715927, 0.09762925654649734, -0.03249908983707428, -0.021192526444792747, 0.045248404145240784, -0.08710476011037827, -0.06952030956745148, -0.07099325954914093, 0.02549964375793934, 0.06107570230960846, 0.008679060265421867, -0.11125288903713226, -0.11586213856935501, -0.05687464401125908, 0.11861179769039154, -0.06548480689525604, -0.07425156235694885, -0.11510331183671951, 0.05070875585079193, 0.13576802611351013, -0.08362466841936111, 0.04246682673692703, 0.01525137759745121, 0.10955918580293655, 0.024526601657271385, -0.06574895232915878, 0.08262906968593597, -0.0930682048201561, -0.23373979330062866, -0.02877742052078247, 0.20023445785045624, 0.013981355354189873, 0.06736984848976135, -0.011107026599347591, 0.02171938866376877, -0.001098015927709639, -0.06425228714942932, 0.05225289240479469, 0.043096765875816345, 0.028474371880292892, 0.03850250318646431, -0.023338258266448975, -0.039864495396614075, -0.07107683271169662, -0.02000253275036812, 0.14832083880901337, 0.2570730447769165, -0.08132815361022949, 0.09649623930454254, 0.10152356326580048, -0.045497555285692215, -0.19953161478042603, -0.03755451366305351, 0.11356647312641144, 0.003965773154050112, -0.002013077028095722, -0.17711789906024933, 0.0542142316699028, 0.07264672964811325, -0.05672788619995117, 0.06382421404123306, -0.33302217721939087, -0.13959172368049622, 0.11559044569730759, 0.1138707771897316, 0.07340739667415619, -0.14979414641857147, -0.05941898003220558, 0.0005702715134248137, -0.06326175481081009, 0.07358645647764206, -0.08577920496463776, 0.1267787665128708, -0.024992715567350388, 0.024319473654031754, 0.028650810942053795, -0.06402819603681564, 0.1308530867099762, -0.04253155738115311, 0.057943955063819885, -0.04319275543093681, 0.0455964058637619, 0.05718027427792549, -0.06544672697782516, 0.04307260364294052, -0.0735667273402214, 0.06509947031736374, -0.071373850107193, -0.029910219833254814, -0.103265181183815, 0.025097724050283432, -0.016217930242419243, -0.03813912719488144, 0.004361816681921482, 0.03522182255983353, 0.0591445155441761, -0.007586267311125994, 0.12808586657047272, -0.009426436387002468, 0.16444113850593567, 0.15810446441173553, 0.10823903977870941, -0.05029354989528656, -0.06564269214868546, -0.020307397469878197, -0.029434852302074432, 0.06525144726037979, -0.1163020133972168, 0.04671434313058853, 0.1188800036907196, 0.0378413200378418, 0.11516709625720978, 0.0630783662199974, -0.06274350732564926, 0.026715191081166267, 0.06816946715116501, -0.13002917170524597, -0.1419307440519333, -0.025728439912199974, 0.04545116052031517, -0.1296422928571701, 0.07413507252931595, 0.11660455912351608, -0.06719593703746796, -0.008283109404146671, -0.008689125999808311, 0.03537677973508835, -0.02790389023721218, 0.22109851241111755, 0.04947251081466675, 0.0771256610751152, -0.11101031303405762, 0.1072918176651001, 0.03306872025132179, -0.11947005242109299, 0.048814792186021805, 0.10366750508546829, -0.07674574851989746, -0.033336520195007324, 0.00982037466019392, 0.0670870691537857, 0.03750229999423027, -0.0767676904797554, -0.1408427208662033, -0.13093909621238708, 0.07476331293582916, 0.1444651335477829, 0.05993588641285896, 0.022248847410082817, -0.025764083489775658, 0.027901945635676384, -0.10485334694385529, 0.1244000717997551, 0.06422256678342819, 0.05181548744440079, -0.15510375797748566, 0.15043073892593384, 0.023371579125523567, 0.04812505096197128, -0.02499687485396862, 0.002687479369342327, -0.09127797931432724, 0.024061821401119232, -0.12818686664104462, 0.009880306199193, -0.03516513109207153, -0.0007560601807199419, -0.007791831158101559, -0.06273721903562546, -0.05326224863529205, 0.05222021043300629, -0.11114752292633057, -0.027080990374088287, -0.007089673541486263, 0.034645501524209976, -0.11566855013370514, -0.03902869299054146, 0.027546556666493416, -0.10072146356105804, 0.09966807067394257, 0.07013256102800369, -0.003748176386579871, 0.04754894599318504, -0.09581753611564636, -0.019912278279662132, 0.05563433840870857, 0.012854337692260742, 0.019729800522327423, -0.12943795323371887, -0.014601939357817173, 0.005510503426194191, 0.001499062986113131, -0.00014324286894407123, 0.0956559106707573, -0.12864063680171967, 0.01381958182901144, -0.004623635672032833, -0.040622737258672714, -0.06930378824472427, 0.022316886112093925, 0.0719330683350563, 0.03907530754804611, 0.157719686627388, -0.10353817045688629, 0.046602241694927216, -0.225841224193573, 0.006479719653725624, -0.02173488773405552, -0.10160959511995316, -0.08689092099666595, 0.002087686210870743, 0.08850419521331787, -0.06020421162247658, 0.09317520260810852, -0.02491656132042408, 0.05036969855427742, 0.031146595254540443, -0.06435544043779373, 0.027604928240180016, 0.0452435202896595, 0.20936298370361328, 0.030521810054779053, -0.03058745339512825, 0.07920455932617188, -0.010545062832534313, 0.061220280826091766, 0.09120488166809082, 0.14460046589374542, 0.17911213636398315, 0.06018030643463135, 0.08104312419891357, 0.07579537481069565, -0.05657525733113289, -0.19977852702140808, 0.03818487748503685, -0.04579295217990875, 0.1181262880563736, -0.011944567784667015, 0.2102441042661667, 0.12110897898674011, -0.14795631170272827, 0.06155778840184212, -0.026773013174533844, -0.06904701143503189, -0.11967561393976212, -0.09158332645893097, -0.08416877686977386, -0.15875789523124695, 0.001382212620228529, -0.11129115521907806, 0.03521048277616501, 0.08348395675420761, 0.020215097814798355, 0.018099967390298843, 0.13351774215698242, 0.03366643190383911, 0.02600175514817238, 0.08634956181049347, 0.005627279169857502, -0.02471901848912239, -0.016395805403590202, -0.10284676402807236, 0.052600931376218796, 0.01670895703136921, 0.04399598389863968, -0.036048371344804764, -0.09328877925872803, 0.04280582442879677, -0.011099832132458687, -0.1184200868010521, 0.023819586262106895, -0.0010364390909671783, 0.0852431207895279, 0.05075893551111221, 0.03148103505373001, -0.009894416667521, -0.01434494461864233, 0.26852455735206604, -0.09785773605108261, -0.1112682893872261, -0.12223250418901443, 0.26526549458503723, 0.007115943357348442, -0.037980854511260986, 0.039274733513593674, -0.08048705011606216, -0.03902106359601021, 0.17135615646839142, 0.1857309490442276, -0.016823071986436844, 0.0033437141682952642, -0.015371071174740791, -0.012074182741343975, -0.06009382754564285, 0.10065431147813797, 0.13655589520931244, 0.09922461211681366, -0.06877995282411575, -0.028050260618329048, -0.03103320673108101, -0.022768745198845863, -0.06386948376893997, 0.08764638751745224, -0.00853663682937622, -0.013140757568180561, -0.04234858229756355, 0.07006625086069107, -0.050348952412605286, -0.11139939725399017, 0.0104127312079072, -0.20881971716880798, -0.17631056904792786, -0.021940866485238075, 0.10056948661804199, 0.0323251448571682, 0.03566206991672516, 0.007715598214417696, 0.010363065637648106, 0.05935106799006462, -0.014315719716250896, -0.0497235506772995, -0.08135942369699478, 0.08829643577337265, -0.11750157922506332, 0.21359123289585114, -0.029678918421268463, 0.05673439800739288, 0.10837800055742264, 0.045560162514448166, -0.10216714441776276, 0.07210330665111542, 0.05670321360230446, -0.13465949892997742, 0.03216983750462532, 0.2034401297569275, -0.03662235662341118, 0.0989534854888916, 0.03738606721162796, -0.1310274600982666, -0.025446467101573944, -0.05350695550441742, -0.06118924543261528, -0.04589572176337242, -0.010642653331160545, -0.0439271405339241, 0.12102529406547546, 0.1878296434879303, -0.06996141374111176, -0.017791740596294403, -0.057886701077222824, 0.00656154053285718, 0.05276842042803764, 0.02863970585167408, -0.015312797389924526, -0.2662360668182373, 0.011045859195291996, 0.0065613617189228535, 0.007553659845143557, -0.23493944108486176, -0.08396759629249573, 0.015286856330931187, -0.06077595055103302, -0.08937156200408936, 0.07306680083274841, 0.08395315706729889, 0.04072483256459236, -0.0380094088613987, -0.03602434694766998, -0.0376506932079792, 0.175574392080307, -0.18675613403320312, -0.06944449990987778 ]
null
null
peft
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.7.2.dev0
{"library_name": "peft", "base_model": "mistralai/Mixtral-8x7B-v0.1"}
null
Krisbiantoro/mixtral-id-llama-1500
[ "peft", "tensorboard", "safetensors", "arxiv:1910.09700", "base_model:mistralai/Mixtral-8x7B-v0.1", "region:us" ]
2024-02-15T02:13:03+00:00
[ "1910.09700" ]
[]
TAGS #peft #tensorboard #safetensors #arxiv-1910.09700 #base_model-mistralai/Mixtral-8x7B-v0.1 #region-us
# Model Card for Model ID ## Model Details ### Model Description - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact ### Framework versions - PEFT 0.7.2.dev0
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.7.2.dev0" ]
[ "TAGS\n#peft #tensorboard #safetensors #arxiv-1910.09700 #base_model-mistralai/Mixtral-8x7B-v0.1 #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.7.2.dev0" ]
[ 46, 6, 3, 54, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4, 14 ]
[ "passage: TAGS\n#peft #tensorboard #safetensors #arxiv-1910.09700 #base_model-mistralai/Mixtral-8x7B-v0.1 #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.7.2.dev0" ]
[ -0.11452840268611908, 0.20770102739334106, -0.003536192001774907, 0.02621976099908352, 0.07501813769340515, 0.015439551323652267, 0.07265464961528778, 0.13011525571346283, 0.02282065525650978, 0.13219135999679565, 0.06033185124397278, 0.11527051031589508, 0.11358703672885895, 0.20430400967597961, -0.004787145182490349, -0.17786675691604614, 0.020795417949557304, -0.058679983019828796, 0.01895478181540966, 0.1268366277217865, 0.1342477798461914, -0.09384733438491821, 0.07286559045314789, -0.020514531061053276, -0.01174569595605135, -0.032934386283159256, -0.06397457420825958, -0.013232933357357979, 0.046645332127809525, 0.034228354692459106, 0.056560516357421875, 0.000003203717369615333, 0.08704663813114166, -0.2597137689590454, 0.01380851585417986, 0.05536508932709694, -0.013196664862334728, 0.0848989486694336, 0.10902126133441925, -0.047565311193466187, 0.1174967810511589, -0.044861625880002975, 0.13424715399742126, 0.08057372272014618, -0.10789141058921814, -0.21456874907016754, -0.0696173831820488, 0.07879366725683212, 0.17983606457710266, 0.07013759762048721, -0.03803468495607376, 0.12072862684726715, -0.06071881949901581, 0.026014212518930435, 0.08346635103225708, -0.11320272833108902, -0.06688769906759262, 0.08295276761054993, 0.13053524494171143, 0.08263832330703735, -0.12136472016572952, -0.035555168986320496, 0.026441801339387894, 0.045679934322834015, 0.07493790984153748, 0.008041405119001865, 0.17827191948890686, 0.0347827672958374, -0.14244239032268524, -0.0525832362473011, 0.09567738324403763, 0.004614652134478092, -0.037006184458732605, -0.21892938017845154, -0.014539807103574276, -0.09590781480073929, -0.03787452355027199, -0.053406618535518646, 0.033534739166498184, 0.013801265507936478, 0.10555539280176163, -0.050470829010009766, -0.07915300875902176, -0.013118650764226913, 0.11569813638925552, 0.06496059149503708, 0.01083049364387989, -0.019382381811738014, -0.004321618005633354, 0.12142791599035263, 0.06706342846155167, -0.12573719024658203, -0.05647919699549675, -0.05888354405760765, -0.032105788588523865, -0.020021723583340645, 0.05174713954329491, 0.007970472797751427, 0.03480026125907898, 0.26218515634536743, -0.018572648987174034, 0.06716626137495041, 0.04290563613176346, 0.01441981177777052, 0.02546624280512333, 0.10453013330698013, -0.038005709648132324, -0.19566266238689423, -0.011417774483561516, 0.1005411148071289, 0.008120528422296047, -0.028938889503479004, -0.051063645631074905, 0.021086756139993668, 0.041739825159311295, 0.11806681007146835, 0.10752354562282562, -0.0219286922365427, -0.059081487357616425, -0.061145003885030746, 0.20788201689720154, -0.15233662724494934, 0.05721485614776611, 0.02737128548324108, -0.004114609211683273, -0.06418909132480621, 0.019402893260121346, 0.00026014246395789087, -0.044234734028577805, 0.10251492261886597, -0.06164507940411568, -0.04185623303055763, -0.11931819468736649, -0.0486174076795578, 0.03645549342036247, -0.030055470764636993, -0.0546887069940567, -0.02748965099453926, -0.08251608163118362, -0.10694020241498947, 0.1005372703075409, -0.05368606373667717, -0.05146199092268944, -0.027836203575134277, -0.06834883242845535, 0.029708627611398697, 0.027706164866685867, 0.06215353310108185, -0.028687767684459686, 0.044549040496349335, -0.030384736135601997, 0.0687129870057106, 0.08494442701339722, 0.03864484652876854, -0.07210583984851837, 0.06792531162500381, -0.17781159281730652, 0.08191125839948654, -0.0603812113404274, 0.02576405741274357, -0.16198010742664337, 0.004793447908014059, 0.0014810212887823582, 0.026848457753658295, 0.05331256985664368, 0.15025873482227325, -0.19463177025318146, -0.03159056976437569, 0.18135789036750793, -0.10052265971899033, -0.11115431785583496, 0.03346474468708038, -0.045914944261312485, 0.16868172585964203, 0.040353234857320786, 0.01149996742606163, 0.09953942149877548, -0.1552104353904724, -0.01659570261836052, -0.03602045774459839, 0.013108503073453903, 0.049677442759275436, 0.06522632390260696, -0.08453000336885452, 0.009743398055434227, 0.00810884591192007, -0.04818582907319069, -0.018982579931616783, -0.036398082971572876, -0.09056159108877182, 0.005165010690689087, -0.08006231486797333, 0.0020761964842677116, 0.009097333997488022, -0.08539512753486633, -0.011553426273167133, -0.14077363908290863, -0.01294606365263462, 0.07150174677371979, 0.0061983149498701096, -0.0056957281194627285, -0.07900595664978027, 0.03666229918599129, -0.06022055074572563, -0.013052350841462612, -0.15405796468257904, -0.0011209127260372043, 0.021514395251870155, -0.14344380795955658, 0.011375870555639267, -0.1465400755405426, 0.07573793828487396, 0.012197467498481274, -0.05993088707327843, -0.0361594632267952, 0.023308657109737396, -0.008612209931015968, -0.07369867712259293, -0.22385218739509583, -0.03060087189078331, -0.053917452692985535, 0.11325854063034058, -0.21450376510620117, 0.04937378689646721, 0.007769700139760971, 0.1163010373711586, 0.014042928814888, -0.07048200815916061, 0.029664969071745872, -0.05663533881306648, -0.021938154473900795, -0.0733657106757164, -0.005101561080664396, 0.0037095907609909773, -0.021105578169226646, 0.0297686830163002, -0.15196722745895386, -0.06844176352024078, 0.09194651246070862, 0.08414904028177261, -0.14253166317939758, 0.011338298209011555, -0.03578273206949234, -0.061446547508239746, -0.07427825033664703, -0.0663641095161438, 0.06358431279659271, 0.0498393252491951, 0.05237707123160362, -0.08674903213977814, -0.07225654274225235, 0.0023003313690423965, -0.010184342041611671, -0.026508202776312828, 0.12386228889226913, 0.07886422425508499, -0.09605099260807037, 0.09936383366584778, 0.07285447418689728, 0.03449197858572006, 0.09759418666362762, -0.005811323411762714, -0.10186150670051575, -0.03597366064786911, 0.057352423667907715, 0.020977696403861046, 0.15591169893741608, -0.062008656561374664, 0.0387609489262104, 0.042126718908548355, -0.03967028856277466, 0.04254807159304619, -0.09696927666664124, 0.015093177556991577, 0.009825685992836952, -0.01528007909655571, 0.030523475259542465, -0.027259353548288345, 0.009716561064124107, 0.08810757100582123, 0.06728649884462357, 0.03156159073114395, 0.01618625968694687, -0.03871751204133034, -0.13905608654022217, 0.17378836870193481, -0.08862102776765823, -0.21783599257469177, -0.15387338399887085, 0.028309915214776993, 0.05257178097963333, -0.012421171180903912, 0.034166302531957626, -0.04071762412786484, -0.08829396218061447, -0.08770029991865158, 0.025261204689741135, 0.04058610647916794, -0.06022367998957634, -0.07307644933462143, 0.03595113381743431, 0.02365167997777462, -0.13261526823043823, 0.02703234553337097, 0.048834484070539474, 0.0056540523655712605, -0.004964405205100775, 0.03467266634106636, 0.08565102517604828, 0.20501336455345154, -0.0002010849566431716, 0.0046296934597194195, 0.054197851568460464, 0.27317512035369873, -0.15372560918331146, 0.12575364112854004, 0.11664876341819763, -0.05427989736199379, 0.08420481532812119, 0.19652311503887177, 0.03389941155910492, -0.08230147510766983, 0.01336789969354868, 0.0426209457218647, -0.04025545343756676, -0.2642473578453064, -0.044152434915304184, -0.02653048187494278, -0.0684066191315651, 0.08959687501192093, 0.08407518267631531, 0.10131971538066864, 0.03285384550690651, -0.07003041356801987, -0.07872873544692993, 0.05533747375011444, 0.11766940355300903, -0.049764737486839294, 0.020680652931332588, 0.08408129960298538, -0.048763688653707504, 0.008013000711798668, 0.08679667115211487, -0.005688415840268135, 0.14030693471431732, 0.051567792892456055, 0.12348796427249908, 0.06968045979738235, 0.0693393126130104, 0.002459020586684346, 0.05258777737617493, -0.004705403000116348, 0.03311220183968544, 0.017901591956615448, -0.09116077423095703, 0.025585653260350227, 0.1095600351691246, 0.008045485243201256, 0.02616061083972454, 0.02189583145081997, -0.07584165036678314, 0.033071957528591156, 0.21635879576206207, 0.02735283225774765, -0.20684963464736938, -0.07963833957910538, 0.05450638756155968, -0.06666528433561325, -0.1536678820848465, -0.009909657761454582, 0.012975439429283142, -0.15129342675209045, 0.014404869638383389, -0.04879938066005707, 0.11188439279794693, -0.07026029378175735, -0.04830925911664963, 0.09698721021413803, 0.05520874261856079, -0.04804028943181038, 0.033032044768333435, -0.18392425775527954, 0.10946333408355713, 0.03430570289492607, 0.07782432436943054, -0.08425302058458328, 0.08401565998792648, 0.00047603138955309987, -0.017727840691804886, 0.1554093360900879, -0.0009620614582672715, -0.07469870895147324, -0.08039069920778275, -0.07468555867671967, -0.02345626801252365, 0.08858277648687363, -0.13007862865924835, 0.07364977151155472, -0.01781565137207508, -0.03672971948981285, 0.0016959878848865628, -0.11467589437961578, -0.10480616241693497, -0.16475458443164825, 0.05713718757033348, -0.07893310487270355, 0.0029578336980193853, -0.07508930563926697, -0.05098648741841316, 0.032134123146533966, 0.16703282296657562, -0.192268505692482, -0.11536376923322678, -0.1481693685054779, -0.10817303508520126, 0.15488873422145844, -0.05034187063574791, 0.08288701623678207, -0.01257662195712328, 0.15090392529964447, -0.016638340428471565, -0.03024102747440338, 0.0885392278432846, -0.08645318448543549, -0.18723361194133759, -0.052763309329748154, 0.18649865686893463, 0.13614371418952942, 0.02547452226281166, -0.020595064386725426, 0.03136712685227394, -0.047516148537397385, -0.1017124354839325, 0.02700764685869217, 0.13657541573047638, 0.06314714252948761, -0.01185546349734068, -0.03150719404220581, -0.11491791903972626, -0.054597191512584686, -0.03572273999452591, -0.014585461467504501, 0.2017887830734253, -0.07335415482521057, 0.1650594174861908, 0.1335376352071762, -0.06580524146556854, -0.20846593379974365, 0.04284961149096489, 0.025463977828621864, 0.016937198117375374, 0.01926443539559841, -0.19204504787921906, 0.07315020263195038, -0.020290689542889595, -0.07393594086170197, 0.17565391957759857, -0.206150621175766, -0.13305725157260895, 0.09299658238887787, 0.02024700492620468, -0.20486868917942047, -0.14726123213768005, -0.11147042363882065, -0.010884441435337067, -0.13067300617694855, 0.06284159421920776, 0.020502617582678795, 0.01307047437876463, 0.010912534780800343, 0.010171090252697468, 0.04284116253256798, -0.05117832496762276, 0.18942798674106598, -0.02902309037744999, 0.008497115224599838, -0.05767281353473663, -0.0980190634727478, 0.013812048360705376, -0.06849948316812515, 0.11382661014795303, -0.019054099917411804, 0.02455807849764824, -0.15543852746486664, -0.0463736429810524, -0.0720645934343338, 0.01588205061852932, -0.0955205038189888, -0.08415448665618896, -0.0489051453769207, 0.07662037760019302, 0.10406865179538727, -0.026342172175645828, 0.03740102797746658, -0.0829169824719429, 0.09051726013422012, 0.21100100874900818, 0.16494737565517426, 0.04659244045615196, -0.0586048848927021, 0.02460010163486004, -0.03560258448123932, 0.04211777448654175, -0.21980687975883484, 0.036559101194143295, 0.06083698570728302, 0.03783959522843361, 0.08109793812036514, -0.0011528779286891222, -0.16170573234558105, -0.080106221139431, 0.07717148959636688, -0.058753661811351776, -0.16197118163108826, -0.021787647157907486, 0.027164185419678688, -0.1952822357416153, -0.0428120456635952, 0.041221097111701965, -0.017557458952069283, -0.03762128949165344, 0.02382640540599823, 0.08402124792337418, -0.013940392062067986, 0.1014292761683464, 0.0815148800611496, 0.08829054981470108, -0.09581964462995529, 0.06928030401468277, 0.08638858795166016, -0.03547874093055725, 0.015707770362496376, 0.14948128163814545, -0.04551633819937706, -0.03074737824499607, 0.08592642843723297, 0.11152059584856033, 0.0014519047690555453, -0.04181824252009392, 0.011927596293389797, -0.0540725402534008, 0.07731564342975616, 0.13972114026546478, 0.01624494232237339, -0.010308886878192425, 0.06636767089366913, 0.03130646422505379, -0.09943098574876785, 0.12240882217884064, 0.06688526272773743, 0.025326311588287354, -0.01841406524181366, -0.022282330319285393, -0.016317106783390045, -0.010105324909090996, -0.014414812438189983, -0.005314622074365616, -0.09370731562376022, -0.0014889362500980496, -0.11718988418579102, 0.02436661906540394, -0.08920998126268387, 0.00047505018301308155, 0.010445394553244114, -0.039952050894498825, -0.001386559451930225, -0.006219012197107077, -0.0757206380367279, -0.05979738011956215, -0.03306205943226814, 0.07057064771652222, -0.14004233479499817, 0.025948751717805862, 0.06630770862102509, -0.10957234352827072, 0.059108175337314606, -0.006454480811953545, 0.012765670195221901, -0.004378998186439276, -0.14525574445724487, 0.05351549759507179, -0.027448533102869987, -0.021082593128085136, 0.011305967345833778, -0.16909156739711761, -0.0032004897948354483, -0.049763113260269165, -0.07658127695322037, 0.009984966367483139, -0.010866035707294941, -0.1260809749364853, 0.12572389841079712, -0.006664713844656944, -0.05785495042800903, -0.018860645592212677, 0.059379857033491135, 0.07700713723897934, -0.01577453687787056, 0.08684093505144119, -0.028773248195648193, 0.08328206092119217, -0.1838465929031372, -0.008156179450452328, -0.012411403469741344, 0.03456897661089897, -0.021582840010523796, -0.029359323903918266, 0.05510886758565903, -0.011163104325532913, 0.14745619893074036, -0.005421650130301714, 0.06213729456067085, 0.047948576509952545, 0.005773589946329594, 0.028613541275262833, 0.06913561373949051, 0.052690211683511734, -0.028578251600265503, -0.016660526394844055, 0.040090836584568024, 0.002109561813995242, -0.045490264892578125, -0.12496952712535858, 0.059310752898454666, 0.1875685155391693, 0.0829269215464592, 0.03755662590265274, -0.0003121798799838871, -0.12900538742542267, -0.08976776152849197, 0.09315481781959534, -0.013385774567723274, -0.02705681324005127, -0.06750230491161346, 0.21955528855323792, 0.13430620729923248, -0.19219763576984406, 0.08348270505666733, -0.04099658876657486, -0.03317305073142052, -0.12926112115383148, -0.16190366446971893, -0.054124895483255386, -0.027780264616012573, -0.03483838587999344, -0.06283318251371384, 0.058772314339876175, 0.03741983696818352, 0.0061728814616799355, -0.002139490330591798, 0.10254166275262833, 0.011964769102633, -0.03390606492757797, 0.05222131311893463, 0.07035569101572037, 0.049639709293842316, -0.08671849220991135, 0.010194049216806889, 0.0000975059883785434, 0.005476527847349644, 0.06317199021577835, 0.0283930916339159, -0.049071360379457474, 0.02817521058022976, -0.009419801644980907, -0.1188378557562828, 0.04655039682984352, -0.003185717388987541, -0.013362590223550797, 0.15737038850784302, 0.032149799168109894, 0.0019228443270549178, -0.016063153743743896, 0.2271793782711029, -0.0715511217713356, -0.07855004072189331, -0.12559549510478973, 0.07396107912063599, -0.04819689318537712, 0.024803325533866882, 0.016908083111047745, -0.12807413935661316, 0.013680718839168549, 0.16455142199993134, 0.13542717695236206, 0.0036380626261234283, 0.008076501078903675, 0.04257560521364212, 0.00828926544636488, -0.011894640512764454, 0.014841936528682709, 0.03697746992111206, 0.2131565362215042, -0.07267003506422043, 0.0744846910238266, -0.012813287787139416, -0.07088397443294525, -0.022845201194286346, 0.129685640335083, -0.012596954591572285, -0.00846902746707201, -0.05795341357588768, 0.13774684071540833, -0.06214897707104683, -0.21679344773292542, 0.06393483281135559, -0.09342589229345322, -0.13173362612724304, -0.03810613974928856, 0.007220853120088577, -0.030618321150541306, 0.0142881590873003, 0.07212823629379272, -0.0502297580242157, 0.16402657330036163, 0.02918355166912079, -0.06188825145363808, -0.10028684139251709, 0.04905257746577263, -0.13988856971263885, 0.28659459948539734, 0.02218535728752613, 0.024889541789889336, 0.10945302993059158, -0.01846194639801979, -0.14584749937057495, 0.00889171939343214, 0.1033577173948288, -0.060198165476322174, 0.04902920499444008, 0.16175156831741333, -0.008938096463680267, 0.12212442606687546, 0.05099421367049217, -0.057479314506053925, 0.031144099310040474, -0.0711895003914833, -0.057231154292821884, -0.12264922261238098, 0.06726675480604172, -0.07378930598497391, 0.1490534543991089, 0.12405599653720856, -0.06682117283344269, -0.004887878429144621, -0.017162831500172615, 0.0735495463013649, 0.015170223079621792, 0.12069540470838547, 0.014757768251001835, -0.17845170199871063, 0.04950382187962532, 0.0060271224938333035, 0.11256583034992218, -0.2193717509508133, -0.06195097044110298, 0.04887834191322327, -0.02858596295118332, -0.08820034563541412, 0.1206444799900055, 0.0398920439183712, 0.022829413414001465, -0.031116722151637077, -0.08921553939580917, 0.01626606658101082, 0.15651091933250427, -0.10311020165681839, -0.016579244285821915 ]
null
null
transformers
This model is [sparsetral-16x7B-v2](https://huggingface.co/serpdotai/sparsetral-16x7B-v2) further tuned utilizing [SPIN](https://arxiv.org/abs/2401.01335) on [OpenHermes-2.5](https://huggingface.co/datasets/teknium/OpenHermes-2.5) mixed with traditional DPO samples. This is iteration_0, plan to keep making iterations until improvements stop. Kuru~ Kuru~ ![Kuru~ Kuru~](https://github.com/duiqt/herta_kuru/raw/main/static/img/hertaa_github.gif) ## Training - 8x A6000s - Base model is [sparsetral-16x7B-v2](https://huggingface.co/serpdotai/sparsetral-16x7B-v2) - [Forked version of unsloth](https://github.com/serp-ai/unsloth) for efficient training - Sequence Length: 4096 - Effective batch size: 64 - Learning Rate: 5e-7 with linear decay (0.1 warmup ratio) - Epochs: 2 - 50k samples (~15k traditional dpo samples, rest SPIN) - QLoRA: - 256 r and 256 alpha - ```python target_modules=[ "q_proj", "k_proj", "v_proj", "o_proj", "gate_proj", "up_proj", "down_proj", "adapter_down", "adapter_up", ] ``` ## Prompt Format ``` <|im_start|>system\n{message}<|im_end|>\n<|im_start|>user\n{message}<|im_end|>\n<|im_start|>assistant\n ``` ## Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("serpdotai/sparsetral-16x7B-v2-SPIN_iter0", trust_remote_code=True) model = AutoModelForCausalLM.from_pretrained("serpdotai/sparsetral-16x7B-v2-SPIN_iter0", device_map="auto", trust_remote_code=True).eval() system_str = "<|im_start|>system\n{message}<|im_end|>\n" user_str = "<|im_start|>user\n{message}<|im_end|>\n" assistant_str = "<|im_start|>assistant\n{message}<|im_end|>\n" def construct_prompt(messages): prompt = "" for message in messages: if message["from"] in ["human", "user"]: prompt += user_str.format( message=message["value"] ) elif message["from"] in ["gpt", "assistant"]: prompt += assistant_str.format( message=message["value"] ) elif message["from"] in ["system", "instruction"]: prompt += system_str.format( message=message["value"] ) else: raise ValueError( f"Unknown message type: {message['from']}" ) return prompt + "<|im_start|>assistant\n" system = "You are a helpful assistant who will help the user to the best of their ability. If you don't know something, say \"I don't know\"" user = "Are you sentient?" messages = [ {"from": "system", "value": system}, {"from": "user", "value": user}, ] prompt = construct_prompt(messages) inputs = tokenizer(prompt, return_tensors="pt") inputs = inputs.to(model.device) pred = model.generate(**inputs, max_length=4096, do_sample=True, top_k=50, top_p=0.99, temperature=0.9, num_return_sequences=1) print(tokenizer.decode(pred.cpu()[0], skip_special_tokens=True)) ``` ## Other Information Paper reference: [Parameter-Efficient Sparsity Crafting from Dense to Mixture-of-Experts for Instruction Tuning on General Tasks](https://arxiv.org/abs/2401.02731) [Original Paper repo](https://github.com/wuhy68/Parameter-Efficient-MoE) [Forked repo with mistral support (sparsetral)](https://github.com/serp-ai/Parameter-Efficient-MoE) If you are interested in faster inferencing, check out our [fork of vLLM](https://github.com/serp-ai/vllm) that adds sparsetral support
{"language": ["en"], "license": "apache-2.0", "datasets": ["teknium/OpenHermes-2.5", "jondurbin/truthy-dpo-v0.1", "jondurbin/gutenberg-dpo-v0.1", "argilla/dpo-mix-7k"]}
text-generation
serpdotai/sparsetral-16x7B-v2-SPIN_iter0
[ "transformers", "safetensors", "sparsetral", "text-generation", "conversational", "custom_code", "en", "dataset:teknium/OpenHermes-2.5", "dataset:jondurbin/truthy-dpo-v0.1", "dataset:jondurbin/gutenberg-dpo-v0.1", "dataset:argilla/dpo-mix-7k", "arxiv:2401.01335", "arxiv:2401.02731", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-15T02:14:42+00:00
[ "2401.01335", "2401.02731" ]
[ "en" ]
TAGS #transformers #safetensors #sparsetral #text-generation #conversational #custom_code #en #dataset-teknium/OpenHermes-2.5 #dataset-jondurbin/truthy-dpo-v0.1 #dataset-jondurbin/gutenberg-dpo-v0.1 #dataset-argilla/dpo-mix-7k #arxiv-2401.01335 #arxiv-2401.02731 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
This model is sparsetral-16x7B-v2 further tuned utilizing SPIN on OpenHermes-2.5 mixed with traditional DPO samples. This is iteration_0, plan to keep making iterations until improvements stop. Kuru~ Kuru~ !Kuru~ Kuru~ ## Training - 8x A6000s - Base model is sparsetral-16x7B-v2 - Forked version of unsloth for efficient training - Sequence Length: 4096 - Effective batch size: 64 - Learning Rate: 5e-7 with linear decay (0.1 warmup ratio) - Epochs: 2 - 50k samples (~15k traditional dpo samples, rest SPIN) - QLoRA: - 256 r and 256 alpha - ## Prompt Format ## Usage ## Other Information Paper reference: Parameter-Efficient Sparsity Crafting from Dense to Mixture-of-Experts for Instruction Tuning on General Tasks Original Paper repo Forked repo with mistral support (sparsetral) If you are interested in faster inferencing, check out our fork of vLLM that adds sparsetral support
[ "## Training\n- 8x A6000s\n- Base model is sparsetral-16x7B-v2\n- Forked version of unsloth for efficient training\n- Sequence Length: 4096\n- Effective batch size: 64\n- Learning Rate: 5e-7 with linear decay (0.1 warmup ratio)\n- Epochs: 2\n- 50k samples (~15k traditional dpo samples, rest SPIN)\n- QLoRA:\n - 256 r and 256 alpha\n -", "## Prompt Format", "## Usage", "## Other Information\nPaper reference: Parameter-Efficient Sparsity Crafting from Dense to Mixture-of-Experts for Instruction Tuning on General Tasks\n\nOriginal Paper repo\n\nForked repo with mistral support (sparsetral)\n\nIf you are interested in faster inferencing, check out our fork of vLLM that adds sparsetral support" ]
[ "TAGS\n#transformers #safetensors #sparsetral #text-generation #conversational #custom_code #en #dataset-teknium/OpenHermes-2.5 #dataset-jondurbin/truthy-dpo-v0.1 #dataset-jondurbin/gutenberg-dpo-v0.1 #dataset-argilla/dpo-mix-7k #arxiv-2401.01335 #arxiv-2401.02731 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "## Training\n- 8x A6000s\n- Base model is sparsetral-16x7B-v2\n- Forked version of unsloth for efficient training\n- Sequence Length: 4096\n- Effective batch size: 64\n- Learning Rate: 5e-7 with linear decay (0.1 warmup ratio)\n- Epochs: 2\n- 50k samples (~15k traditional dpo samples, rest SPIN)\n- QLoRA:\n - 256 r and 256 alpha\n -", "## Prompt Format", "## Usage", "## Other Information\nPaper reference: Parameter-Efficient Sparsity Crafting from Dense to Mixture-of-Experts for Instruction Tuning on General Tasks\n\nOriginal Paper repo\n\nForked repo with mistral support (sparsetral)\n\nIf you are interested in faster inferencing, check out our fork of vLLM that adds sparsetral support" ]
[ 133, 105, 5, 3, 80 ]
[ "passage: TAGS\n#transformers #safetensors #sparsetral #text-generation #conversational #custom_code #en #dataset-teknium/OpenHermes-2.5 #dataset-jondurbin/truthy-dpo-v0.1 #dataset-jondurbin/gutenberg-dpo-v0.1 #dataset-argilla/dpo-mix-7k #arxiv-2401.01335 #arxiv-2401.02731 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n## Training\n- 8x A6000s\n- Base model is sparsetral-16x7B-v2\n- Forked version of unsloth for efficient training\n- Sequence Length: 4096\n- Effective batch size: 64\n- Learning Rate: 5e-7 with linear decay (0.1 warmup ratio)\n- Epochs: 2\n- 50k samples (~15k traditional dpo samples, rest SPIN)\n- QLoRA:\n - 256 r and 256 alpha\n -## Prompt Format## Usage## Other Information\nPaper reference: Parameter-Efficient Sparsity Crafting from Dense to Mixture-of-Experts for Instruction Tuning on General Tasks\n\nOriginal Paper repo\n\nForked repo with mistral support (sparsetral)\n\nIf you are interested in faster inferencing, check out our fork of vLLM that adds sparsetral support" ]
[ -0.07244923710823059, 0.08543276786804199, -0.004937367979437113, -0.0025482636410743, 0.012124363332986832, 0.01683899201452732, 0.01461674552410841, 0.15956082940101624, -0.05757681652903557, 0.08095552027225494, 0.004080668557435274, 0.034174125641584396, 0.09151662886142731, 0.09585287421941757, -0.08471260219812393, -0.2727574408054352, 0.024670876562595367, 0.004450117703527212, -0.018525034189224243, 0.07890427857637405, 0.11593236029148102, -0.06807009130716324, 0.022978996858000755, -0.0035219192504882812, -0.03682256117463112, -0.03363179787993431, -0.018853815272450447, -0.057829998433589935, 0.09921582788228989, 0.0714387372136116, 0.13791488111019135, 0.07687123864889145, 0.08450543880462646, -0.1874934285879135, 0.03580307215452194, 0.03622370585799217, 0.029171548783779144, 0.10613678395748138, 0.05217844992876053, 0.02211879752576351, 0.0121723972260952, -0.09569445997476578, 0.07443957030773163, 0.03814539313316345, -0.07630456238985062, -0.16193027794361115, -0.15283286571502686, 0.05808033421635628, 0.0846150740981102, 0.038329947739839554, -0.025324111804366112, 0.13027577102184296, -0.06365843117237091, 0.034523479640483856, 0.18493346869945526, -0.2055545449256897, -0.0797090157866478, 0.12442732602357864, 0.03242022916674614, 0.029291482642292976, -0.05017886683344841, -0.02034100890159607, 0.03772873058915138, 0.01842992566525936, -0.0033366980496793985, 0.005336406175047159, 0.11812932044267654, 0.01057536993175745, -0.13593721389770508, -0.0017673776019364595, 0.06987825781106949, 0.022241221740841866, -0.053548675030469894, -0.14069317281246185, -0.0789577066898346, 0.03616296872496605, -0.067644402384758, -0.06752870976924896, -0.0041302465833723545, 0.01913398876786232, 0.032375484704971313, -0.015128682367503643, -0.05201555788516998, -0.06821972876787186, -0.04218621551990509, 0.06351288408041, 0.04451749846339226, -0.02122262492775917, 0.04287007823586464, 0.15129616856575012, -0.02018461376428604, -0.10490762442350388, -0.044833771884441376, -0.07683141529560089, -0.0934387668967247, -0.06884701550006866, -0.018026011064648628, -0.1037280485033989, 0.06280102580785751, 0.2484712302684784, 0.002953087678179145, 0.028470337390899658, -0.04223346710205078, -0.02459554374217987, -0.00687621533870697, 0.031863193958997726, -0.032917577773332596, -0.019097082316875458, 0.023800255730748177, 0.10254064202308655, 0.025281378999352455, 0.02895769104361534, 0.01235431618988514, -0.07800360769033432, 0.024062709882855415, 0.07334895431995392, 0.06774197518825531, 0.013088478706777096, -0.07202276587486267, -0.05382196605205536, 0.20879927277565002, -0.14755216240882874, -0.04494939744472504, -0.00394615950062871, -0.046607233583927155, 0.11800801753997803, 0.017708061262965202, 0.056838978081941605, -0.041290126740932465, 0.09120161831378937, -0.029757505282759666, -0.05692771077156067, -0.059870652854442596, -0.044888392090797424, 0.025605810806155205, 0.025105765089392662, -0.029186712577939034, -0.09761309623718262, -0.16261769831180573, -0.04480573907494545, 0.027651099488139153, -0.07466243952512741, -0.06821230053901672, 0.003325818805024028, 0.012610145844519138, 0.0024668222758919, -0.0032768624369055033, 0.13644346594810486, -0.05710914358496666, 0.06169036775827408, -0.03566235303878784, 0.021201571449637413, 0.015369598753750324, 0.037661582231521606, -0.03857605531811714, 0.05155601724982262, -0.17565473914146423, 0.0903373435139656, -0.1358610838651657, 0.036151573061943054, -0.17816133797168732, -0.06452671438455582, 0.031040724366903305, -0.021189123392105103, 0.023052861914038658, 0.12021717429161072, -0.23667889833450317, -0.01113087683916092, 0.1490216702222824, -0.13432660698890686, -0.055998433381319046, 0.1321878433227539, -0.019074322655797005, -0.024366889148950577, 0.06888964027166367, 0.007000823505222797, 0.13643595576286316, -0.10259141772985458, -0.04137725755572319, -0.030991846695542336, 0.05218890681862831, 0.11270638555288315, 0.08337534964084625, -0.04727545008063316, 0.014350245706737041, 0.016219552606344223, -0.05139860138297081, -0.00937793217599392, -0.0255876574665308, -0.0894898846745491, -0.04688752815127373, -0.007091837469488382, 0.05371443182229996, -0.020434478297829628, 0.004559403285384178, -0.031708501279354095, -0.1339443325996399, -0.04198211058974266, 0.15828107297420502, -0.012645854614675045, 0.006764765828847885, -0.08875378221273422, 0.07920801639556885, 0.013217215426266193, -0.017245221883058548, -0.15287044644355774, -0.12233985960483551, 0.024106180295348167, -0.07947637885808945, -0.04397011548280716, 0.06902779638767242, 0.04933813586831093, 0.027276059612631798, -0.07863826304674149, -0.021585961803793907, -0.04087494686245918, -0.017231550067663193, -0.02364192344248295, -0.0866079106926918, -0.037880074232816696, -0.07242164015769958, 0.057982299476861954, -0.1562146246433258, 0.019682999700307846, 0.0716627836227417, 0.19671988487243652, 0.056489817798137665, -0.020675798878073692, 0.02952421084046364, 0.00901686493307352, 0.04419355466961861, -0.05552108585834503, 0.04511277750134468, -0.0067422413267195225, -0.07672127336263657, -0.042113758623600006, -0.026875481009483337, -0.0429195798933506, 0.07083558291196823, 0.08902380615472794, -0.08545688539743423, -0.17298129200935364, -0.06551914662122726, -0.04947366937994957, -0.12246549129486084, -0.027545498684048653, 0.1415010690689087, 0.05604856088757515, 0.04210641607642174, -0.06759436428546906, -0.04375578463077545, -0.0677725300192833, 0.017542008310556412, -0.0009116759756579995, 0.0913432389497757, -0.03246762603521347, -0.0761200338602066, 0.011941100470721722, 0.050949737429618835, -0.018813837319612503, 0.05809905752539635, -0.096051886677742, -0.09744241833686829, 0.002191175241023302, 0.036621276289224625, 0.033208511769771576, 0.025456136092543602, 0.0056116716004908085, 0.055070534348487854, 0.07858431339263916, -0.021376056596636772, 0.01850508525967598, -0.09979686886072159, 0.0032039228826761246, -0.01570826955139637, -0.07672122865915298, 0.006494227331131697, 0.05138426274061203, 0.029298147186636925, 0.08332151919603348, 0.026407970115542412, 0.0011877540964633226, 0.010379386134445667, -0.041951246559619904, -0.03980811685323715, 0.11226543039083481, -0.07101790606975555, -0.08435045182704926, -0.10076220333576202, 0.008576136082410812, 0.008374445140361786, -0.04114334285259247, 0.0018435657257214189, -0.07464883476495743, -0.12147622555494308, -0.049532946199178696, 0.0900968462228775, 0.08227771520614624, -0.014377433806657791, 0.004181600641459227, 0.035947561264038086, 0.03921511396765709, -0.12381582707166672, 0.023129070177674294, 0.006717845797538757, -0.03296259418129921, 0.069020576775074, 0.0017882324755191803, 0.02596127986907959, 0.08770246058702469, 0.07390004396438599, -0.024469928815960884, -0.0064107864163815975, 0.1601838618516922, -0.06813972443342209, 0.07794686406850815, 0.17027290165424347, -0.018440453335642815, 0.06081383675336838, 0.17738285660743713, 0.04896300286054611, -0.06412968039512634, 0.050575245171785355, 0.063782699406147, 0.026465505361557007, -0.252399742603302, -0.08285121619701385, -0.009997445158660412, -0.027100984007120132, 0.07298936694860458, 0.02451666072010994, -0.05200335010886192, 0.07481838017702103, -0.09130024909973145, 0.008357958868145943, 0.03301393985748291, 0.03617597743868828, 0.07981445640325546, 0.02763538248836994, 0.10593489557504654, -0.019824916496872902, -0.026998763903975487, 0.09239555895328522, 0.027906246483325958, 0.2501537799835205, -0.045008864253759384, 0.17840886116027832, 0.0009576923330314457, 0.13104844093322754, 0.09707775712013245, 0.05756402760744095, 0.012166926637291908, -0.003241556929424405, -0.0014328428078442812, -0.06043092906475067, -0.021719256415963173, 0.044958602637052536, 0.04675563424825668, -0.01701800897717476, -0.0008882724214345217, -0.008355872705578804, 0.01966797560453415, 0.29298079013824463, 0.03275628015398979, -0.1519959419965744, -0.07027478516101837, 0.07939807325601578, -0.08440153300762177, -0.034342169761657715, 0.01483816560357809, 0.003261311911046505, -0.1230776458978653, 0.06115630269050598, -0.10405617952346802, 0.08793124556541443, -0.10537052899599075, -0.018063245341181755, -0.011042822152376175, 0.1160527691245079, -0.03848749399185181, 0.05549834668636322, -0.15835928916931152, 0.18305054306983948, -0.011377014219760895, 0.05232438072562218, -0.04418680816888809, 0.07098057866096497, 0.06177853047847748, -0.0052229491993784904, 0.12736448645591736, -0.02021101303398609, -0.07319328188896179, -0.028347456827759743, -0.1445198804140091, 0.04260619357228279, 0.03199586644768715, -0.07783748209476471, 0.0998011976480484, -0.030292047187685966, -0.001444786204956472, 0.017324164509773254, -0.04216725006699562, -0.1206270381808281, -0.09231950342655182, 0.08391103893518448, -0.13950203359127045, 0.10309921950101852, -0.1263623684644699, -0.09499271959066391, -0.10556618124246597, 0.11587628722190857, -0.011624359525740147, -0.1179514229297638, -0.1383948177099228, 0.06681779026985168, 0.14986641705036163, -0.0641530454158783, 0.04388565942645073, 0.034547217190265656, 0.1347881406545639, -0.016085337847471237, -0.04765480384230614, 0.051495011895895004, -0.038631170988082886, -0.18955574929714203, -0.013075032271444798, 0.12594915926456451, 0.05512142926454544, 0.05705057829618454, 0.010653100907802582, 0.026021867990493774, 0.00539093604311347, -0.07246008515357971, 0.011108842678368092, 0.09762673825025558, 0.08970865607261658, 0.10715116560459137, -0.04182293266057968, -0.13661889731884003, -0.06811265647411346, -0.028670141473412514, 0.10346226394176483, 0.23878134787082672, -0.013703633099794388, 0.060759514570236206, 0.014146790839731693, -0.0227077417075634, -0.1465473473072052, -0.028918374329805374, 0.07087868452072144, 0.012032541446387768, -0.0294268149882555, -0.16781851649284363, 0.09633836150169373, 0.14994965493679047, -0.010561615228652954, -0.035917263478040695, -0.24899235367774963, -0.11068394780158997, 0.11577533930540085, 0.04392153397202492, -0.004565035458654165, -0.0878220647573471, -0.09654896706342697, -0.012323657050728798, -0.17503705620765686, 0.014185739681124687, -0.07126227021217346, 0.04032628610730171, -0.03673110902309418, 0.02760038897395134, 0.04836432263255119, -0.03920436650514603, 0.18102727830410004, -0.006131551694124937, 0.026791764423251152, -0.07576937228441238, 0.0034624170511960983, -0.09882429987192154, -0.026795519515872, 0.043771784752607346, 0.026385681703686714, 0.046687427908182144, -0.08588621020317078, -0.030829695984721184, -0.03753252699971199, 0.03964344784617424, -0.04817892238497734, -0.09712040424346924, -0.06455601006746292, 0.10330507159233093, 0.04869437962770462, -0.02471337653696537, 0.12021006643772125, -0.03136739134788513, 0.08698160201311111, 0.05404219776391983, 0.11305908113718033, -0.0066960593685507774, -0.05686252936720848, -0.012077364139258862, -0.006916734389960766, 0.04091404378414154, -0.08019062876701355, 0.03625492379069328, 0.1316683143377304, 0.039348017424345016, 0.11597687751054764, 0.004003411158919334, -0.09749165922403336, 0.0444473959505558, 0.04432172700762749, -0.08822055906057358, -0.17652617394924164, -0.031344447284936905, 0.018037302419543266, -0.1717800498008728, -0.06257615983486176, 0.16042831540107727, -0.013981569558382034, -0.03574785590171814, 0.05185472592711449, 0.07588153332471848, -0.004597424063831568, 0.15146711468696594, 0.024257084354758263, 0.06053842604160309, -0.08558212965726852, 0.07257441431283951, 0.08603501319885254, -0.09525096416473389, -0.07733248174190521, 0.07868180423974991, -0.09806359559297562, -0.061565060168504715, 0.011956294067203999, 0.058066338300704956, -0.0640566349029541, 0.03343746066093445, -0.0649615228176117, -0.1718447506427765, 0.06332065910100937, 0.049625203013420105, 0.06445196270942688, 0.01533765159547329, 0.013810043223202229, -0.003925904631614685, -0.002455807290971279, 0.11119715124368668, 0.13730394840240479, 0.07828263193368912, -0.0868266373872757, 0.10905147343873978, -0.0743502527475357, -0.019772935658693314, -0.02091754786670208, 0.035324785858392715, -0.08315728604793549, -0.040727559477090836, -0.2858753800392151, 0.024778563529253006, -0.08774743974208832, 0.026694415137171745, -0.05131940916180611, 0.0007489959243685007, 0.024385951459407806, 0.0008817545021884143, -0.0777854323387146, -0.028253750875592232, -0.023669729009270668, 0.06948000937700272, -0.08815447241067886, 0.02344697341322899, 0.08351299166679382, -0.0784587636590004, 0.09189923107624054, 0.017663082107901573, 0.02651427686214447, 0.04590154066681862, -0.053813304752111435, -0.025728311389684677, 0.030009429901838303, 0.07206457108259201, 0.01956743560731411, -0.15280814468860626, -0.0378568135201931, -0.024636810645461082, -0.06492214649915695, 0.023493818938732147, 0.021248364821076393, -0.10835099965333939, -0.00658015301451087, -0.022617418318986893, -0.015068397857248783, -0.07055795937776566, -0.006140691693872213, 0.12681342661380768, 0.07657922804355621, 0.12564362585544586, -0.06909723579883575, 0.05987268313765526, -0.18094110488891602, 0.0012417702237144113, 0.004540547728538513, -0.03509039059281349, 0.0611843578517437, 0.004017602186650038, 0.04388469085097313, -0.0314229391515255, 0.04473524168133736, -0.11401394754648209, -0.02003537490963936, 0.06551159918308258, -0.07129957526922226, 0.007038915995508432, 0.05252329632639885, 0.21039918065071106, 0.10338157415390015, 0.007113612722605467, 0.005707064643502235, -0.04477102309465408, -0.001026787213049829, -0.040403224527835846, 0.09775456041097641, 0.1797337830066681, 0.07741642743349075, -0.00992303341627121, 0.03785895183682442, -0.08210825175046921, -0.01977677270770073, 0.1807582676410675, -0.004722311161458492, 0.03964413329958916, -0.017830727621912956, 0.08390513807535172, 0.09864009916782379, -0.1665782332420349, 0.04798540472984314, -0.015418116003274918, -0.06735262274742126, -0.0727221816778183, -0.11683598905801773, -0.07631246745586395, -0.09769682586193085, 0.00355802895501256, -0.11775704473257065, 0.011744861491024494, 0.10443545132875443, 0.023547042161226273, 0.02942078374326229, 0.08690714836120605, -0.08153901994228363, -0.06503451615571976, 0.10574102401733398, 0.02565120719373226, -0.04013819247484207, -0.0034482358023524284, -0.049266986548900604, 0.04245946183800697, 0.0732431635260582, -0.027339110150933266, -0.012288357131183147, -0.010747465305030346, 0.0011814888566732407, 0.0014664017362520099, -0.06267522275447845, 0.012532202526926994, 0.0014545547310262918, -0.010671117343008518, 0.12978599965572357, 0.07951773703098297, 0.04760316014289856, -0.0064769284799695015, 0.245932474732399, -0.0441843643784523, -0.13036294281482697, -0.20843730866909027, 0.021589307114481926, -0.027247225865721703, 0.016287805512547493, 0.11402634531259537, -0.08765079826116562, 0.05383570119738579, 0.03792193904519081, 0.09416452795267105, -0.023159209638834, 0.016122139990329742, 0.025904299691319466, -0.012625880539417267, -0.04705192521214485, 0.03960251435637474, 0.13144908845424652, 0.012170765548944473, -0.06879131495952606, -0.03598315641283989, 0.015779733657836914, -0.05424921214580536, -0.06581563502550125, 0.034106459468603134, 0.02862139232456684, -0.012810807675123215, -0.02945614792406559, 0.1555483639240265, -0.05548910051584244, -0.17201101779937744, -0.05015086382627487, -0.058947619050741196, -0.1688307374715805, -0.05372760444879532, -0.037140969187021255, 0.010235856287181377, 0.027772461995482445, -0.005171776283532381, -0.0075214519165456295, 0.23804070055484772, -0.0030243557412177324, -0.03685412555932999, -0.08008291572332382, 0.12185340374708176, 0.03501022607088089, 0.07508213073015213, 0.051075056195259094, 0.028723252937197685, 0.09583064913749695, 0.005118783563375473, -0.10002940148115158, 0.06426884979009628, 0.07831180095672607, -0.12239306420087814, 0.06491810828447342, 0.13615258038043976, 0.004969845525920391, 0.1184936985373497, 0.09173726290464401, -0.08871939033269882, 0.005898450035601854, 0.1293312907218933, -0.056583136320114136, -0.08123327046632767, 0.11145108938217163, -0.11408093571662903, 0.12178578227758408, 0.21478621661663055, -0.03520141541957855, 0.04616634175181389, -0.026239395141601562, 0.05974166467785835, 0.020434413105249405, 0.1079343855381012, -0.019289372488856316, -0.14388002455234528, -0.007786649279296398, -0.01869136281311512, 0.04102464020252228, -0.1383800059556961, -0.11680300533771515, -0.03991535305976868, -0.02146242931485176, -0.06146000325679779, 0.12346313893795013, 0.08494894206523895, -0.013695497997105122, 0.004765808116644621, -0.01597486063838005, -0.02620724029839039, 0.07665523141622543, -0.13842879235744476, -0.02602984756231308 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # whisper-small-en-nonnative-maritime This model was trained from scratch on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.0000 - Wer: 34.0415 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - training_steps: 5000 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:------:|:----:|:---------------:|:-------:| | 0.0001 | 142.86 | 1000 | 0.0001 | 39.4007 | | 0.0001 | 285.71 | 2000 | 0.0001 | 39.2263 | | 0.0 | 428.57 | 3000 | 0.0000 | 36.6735 | | 0.0 | 571.43 | 4000 | 0.0000 | 35.1039 | | 0.0 | 714.29 | 5000 | 0.0000 | 34.0415 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
{"tags": ["generated_from_trainer"], "metrics": ["wer"], "model-index": [{"name": "whisper-small-en-nonnative-maritime", "results": []}]}
automatic-speech-recognition
vishakha-lall/whisper-small-en-nonnative-maritime
[ "transformers", "tensorboard", "safetensors", "whisper", "automatic-speech-recognition", "generated_from_trainer", "endpoints_compatible", "region:us" ]
2024-02-15T02:14:58+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #endpoints_compatible #region-us
whisper-small-en-nonnative-maritime =================================== This model was trained from scratch on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 0.0000 * Wer: 34.0415 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 1e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 500 * training\_steps: 5000 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.37.2 * Pytorch 2.2.0+cu121 * Datasets 2.16.1 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 5000\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 5000\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1" ]
[ 47, 130, 4, 33 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 5000\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1" ]
[ -0.1048581600189209, 0.07307299226522446, -0.0018195786979049444, 0.07808683812618256, 0.14389918744564056, -0.004991959314793348, 0.11343612521886826, 0.1106838509440422, -0.05023760721087456, 0.07940381020307541, 0.10198436677455902, 0.11674610525369644, 0.031624674797058105, 0.13005566596984863, -0.0496794655919075, -0.28210318088531494, -0.00032563976128585637, -0.0018182150088250637, -0.04642023891210556, 0.12060297280550003, 0.07311111688613892, -0.1506311446428299, 0.039322394877672195, -0.020853566005825996, -0.14235638082027435, -0.002610378433018923, 0.022935625165700912, -0.06347044557332993, 0.13689270615577698, 0.002633107593283057, 0.10905245691537857, 0.041223619133234024, 0.10336943715810776, -0.2463860809803009, 0.01727740280330181, 0.045632150024175644, 0.047900017350912094, 0.0589175708591938, 0.04682871699333191, -0.031106075271964073, 0.059901852160692215, -0.09708193689584732, 0.08412047475576401, 0.023832622915506363, -0.12268611788749695, -0.2364518940448761, -0.0656953975558281, 0.0029898430220782757, 0.09599176794290543, 0.09632838517427444, -0.02985071763396263, 0.11320045590400696, -0.08541574329137802, 0.09964171797037125, 0.2381812483072281, -0.28672873973846436, -0.06474663317203522, -0.0167684368789196, 0.038568541407585144, 0.08647767454385757, -0.1275436133146286, 0.00018891284707933664, 0.042505841702222824, 0.04406427592039108, 0.11520329862833023, -0.01868661493062973, -0.0853920429944992, 0.00309875444509089, -0.1493646800518036, -0.012121272273361683, 0.06134755164384842, 0.033712778240442276, -0.02670322358608246, -0.10293988138437271, -0.04240962862968445, -0.17430651187896729, -0.06327367573976517, -0.0226630587130785, 0.04250174015760422, -0.05716889351606369, -0.12178869545459747, -0.007029336411505938, -0.09202057123184204, -0.1008402407169342, -0.019908783957362175, 0.17206333577632904, 0.030996832996606827, -0.018871435895562172, -0.012523162178695202, 0.08569251000881195, -0.001923596253618598, -0.1287207007408142, 0.013795153237879276, 0.045189663767814636, -0.06935635209083557, -0.03712098300457001, -0.05408402159810066, -0.06731313467025757, 0.019106967374682426, 0.11410695314407349, -0.0594634935259819, 0.08431988209486008, -0.006841633468866348, 0.03459457680583, -0.10933398455381393, 0.19344036281108856, -0.05584549903869629, -0.031641386449337006, -0.025638483464717865, 0.09302642941474915, 0.01603045128285885, -0.02040485106408596, -0.08303114771842957, 0.02623891830444336, 0.09799446165561676, 0.04104287177324295, -0.06537363678216934, 0.05282103270292282, -0.0460900142788887, 0.009549200534820557, -0.06838129460811615, -0.10906478017568588, 0.03418755531311035, 0.026303423568606377, -0.0568900890648365, -0.04733336716890335, 0.013540350832045078, 0.032599739730358124, -0.013355336152017117, 0.11002396047115326, -0.06670071184635162, 0.031508203595876694, -0.07094711810350418, -0.11829949170351028, 0.005339286755770445, -0.07086601108312607, 0.009800996631383896, -0.09462922811508179, -0.0997922271490097, -0.032515376806259155, 0.03455638512969017, -0.039813414216041565, -0.006697401870042086, -0.1035163477063179, -0.07464655488729477, 0.02410806529223919, -0.027121732011437416, 0.10562144964933395, -0.058029819279909134, 0.11278831213712692, 0.032570645213127136, 0.08880362659692764, 0.012121369130909443, 0.04559170827269554, -0.07117611914873123, 0.0247303806245327, -0.19936473667621613, 0.08542171865701675, -0.09351105242967606, 0.053987741470336914, -0.1109260618686676, -0.10294541716575623, 0.008502907119691372, 0.007691430859267712, 0.10326546430587769, 0.11939162760972977, -0.19909168779850006, -0.08506632596254349, 0.21187838912010193, -0.08307556062936783, -0.08080461621284485, 0.13952042162418365, -0.04055459424853325, 0.023414961993694305, 0.07412896305322647, 0.27205613255500793, 0.047147829085588455, -0.11856161803007126, 0.01707966811954975, -0.037431471049785614, 0.06743810325860977, 0.009187557734549046, 0.04939442500472069, -0.02454739809036255, 0.03044307231903076, 0.011060143820941448, -0.01520138792693615, 0.0737513080239296, -0.1033046767115593, -0.07852619141340256, -0.01771245151758194, -0.11436790227890015, 0.06244204565882683, 0.05357144773006439, 0.055770087987184525, -0.10799118131399155, -0.07981473952531815, 0.03961385786533356, 0.07571721822023392, -0.10046588629484177, 0.04250069335103035, -0.09480106085538864, 0.04094754904508591, -0.016093512997031212, -0.02587987668812275, -0.16813506186008453, 0.01898280531167984, 0.02017149142920971, 0.009589186869561672, 0.03790285438299179, -0.053932659327983856, 0.09616301953792572, 0.05545671284198761, -0.06264640390872955, -0.04131398722529411, -0.00855543278157711, 0.015186665579676628, -0.11604350805282593, -0.22720764577388763, -0.019519805908203125, -0.03567175194621086, 0.13391296565532684, -0.22376631200313568, 0.02518327161669731, 0.007346609141677618, 0.09181968122720718, 0.04054682329297066, -0.03804168850183487, -0.00176376907620579, 0.08516131341457367, 0.0011093111243098974, -0.058300066739320755, 0.05583368241786957, -0.0006592270801775157, -0.11487961560487747, 0.009885020554065704, -0.16839398443698883, 0.10308261215686798, 0.12606750428676605, -0.05453374981880188, -0.0939253717660904, -0.010962742380797863, -0.05028143152594566, -0.03287770226597786, -0.034747835248708725, 0.026243610307574272, 0.21993963420391083, 0.0031788412015885115, 0.13402487337589264, -0.07053102552890778, -0.03479177877306938, 0.02870715968310833, -0.02920224890112877, 0.014069710858166218, 0.1449730098247528, 0.057555485516786575, -0.05676952004432678, 0.10717304795980453, 0.09626326709985733, -0.0827041044831276, 0.1522451788187027, -0.04074348881840706, -0.07885677367448807, -0.016382863745093346, 0.013905275613069534, 0.01930883154273033, 0.12012934684753418, -0.125905379652977, -0.02314680814743042, -0.007431042846292257, 0.0102100083604455, 0.02277868613600731, -0.22922609746456146, -0.0247156023979187, 0.030850522220134735, -0.06180848926305771, -0.020664032548666, -0.020142996683716774, 0.01068307738751173, 0.10756565630435944, 0.0014357486506924033, -0.07063530385494232, -0.004265355411916971, -0.023500990122556686, -0.08405701071023941, 0.20113638043403625, -0.09347731620073318, -0.15448148548603058, -0.0798855721950531, -0.04767508804798126, 0.01407550647854805, 0.012289140373468399, 0.048160120844841, -0.1199248656630516, -0.00767033314332366, -0.08462510257959366, 0.027534864842891693, -0.0035771073307842016, 0.04452541843056679, 0.0024948017671704292, 0.004069158807396889, 0.06692726165056229, -0.09064099192619324, 0.004024548456072807, -0.06856090575456619, -0.039822012186050415, 0.0386064313352108, 0.05760795623064041, 0.10992264002561569, 0.18426406383514404, -0.005781239364296198, 0.015435170382261276, -0.052448567003011703, 0.18800774216651917, -0.09882375597953796, -0.040508512407541275, 0.0965050533413887, -0.028727862983942032, 0.0366758294403553, 0.14617033302783966, 0.04910938814282417, -0.10988039523363113, 0.018921099603176117, 0.0310229305177927, -0.03839929401874542, -0.2088906317949295, -0.042962346225976944, -0.03290407732129097, 0.010884338989853859, 0.08564796298742294, 0.012721828185021877, -0.0002385590341873467, 0.021590309217572212, 0.041124291718006134, -0.02280971221625805, 0.006636311765760183, 0.051110077649354935, 0.06577084213495255, 0.01188751868903637, 0.1170642301440239, -0.020320050418376923, -0.072444848716259, 0.0032598257530480623, -0.007138949353247881, 0.22813856601715088, 0.009471847675740719, 0.11965391039848328, 0.04440677911043167, 0.1478598266839981, 0.012485216371715069, 0.05457115173339844, 0.013759125024080276, -0.05015401542186737, 0.017421355471014977, -0.048917289823293686, -0.04175584763288498, 0.03448681905865669, 0.0339629091322422, 0.0417129248380661, -0.12677262723445892, -0.004668511915951967, 0.05159781873226166, 0.29923215508461, 0.07066386938095093, -0.2767058312892914, -0.09525426477193832, 0.009205920621752739, -0.07725021243095398, -0.006189435720443726, 0.048233263194561005, 0.1737382858991623, -0.07111796736717224, 0.027362041175365448, -0.06339370459318161, 0.0712849572300911, -0.055246226489543915, 0.03999245911836624, 0.015226895920932293, 0.07082754373550415, -0.0162209402769804, 0.03231828659772873, -0.28429821133613586, 0.31029221415519714, 0.0011986163444817066, 0.09169712662696838, -0.05146582052111626, -0.02652939409017563, 0.020893670618534088, -0.005258568096905947, 0.10958470404148102, -0.021673599258065224, -0.10243343561887741, -0.22216831147670746, -0.06967849284410477, 0.023182230070233345, 0.16049613058567047, -0.012258266098797321, 0.11737929284572601, -0.016077086329460144, -0.013802350498735905, 0.07305503636598587, -0.10481885820627213, -0.08235990256071091, -0.07990175485610962, -0.021208448335528374, 0.0645807757973671, 0.0009277124772779644, -0.07880162447690964, -0.09981528669595718, -0.09590829908847809, 0.11154493689537048, -0.0892668217420578, -0.004782318137586117, -0.11225160211324692, 0.07300475239753723, 0.10914581269025803, -0.07051336020231247, 0.044404152780771255, 0.04699547588825226, 0.08172248303890228, 0.013151895254850388, -0.029184114187955856, 0.12757404148578644, -0.07533539831638336, -0.18545302748680115, -0.0597412995994091, 0.16368436813354492, 0.058965280652046204, 0.07864910364151001, -0.008656562305986881, 0.027485068887472153, 0.007956210523843765, -0.07133349031209946, 0.06358645856380463, 0.013619082048535347, 0.006226865574717522, 0.045703593641519547, -0.036039769649505615, -0.026671377941966057, -0.09528416395187378, -0.0347273014485836, 0.20346902310848236, 0.2644710838794708, -0.0908910408616066, 0.05905623733997345, 0.07318457961082458, -0.0502828024327755, -0.21212263405323029, 0.04655294120311737, 0.07950638234615326, 0.02477050945162773, 0.029918737709522247, -0.1688326597213745, 0.07986441999673843, 0.05753438174724579, -0.01969606988132, 0.0894784927368164, -0.31904786825180054, -0.14978548884391785, 0.11563628166913986, 0.13650676608085632, 0.06211729347705841, -0.1342824250459671, -0.03851233422756195, -0.0075760786421597, -0.04617507383227348, 0.06718514859676361, -0.10319062322378159, 0.1430014818906784, 0.013562790118157864, 0.08128097653388977, 0.031248709186911583, -0.06404634565114975, 0.10027755796909332, -0.016649095341563225, 0.08164740353822708, -0.03808736056089401, -0.032688625156879425, -0.012784741818904877, -0.036618951708078384, -0.0027989638037979603, -0.0622953437268734, 0.030567392706871033, -0.03577891364693642, -0.04194752871990204, -0.08997179567813873, 0.012616426683962345, -0.01594812422990799, -0.062144551426172256, -0.012527474202215672, 0.04847012087702751, 0.06580136716365814, 0.006633874028921127, 0.09117473661899567, -0.07207933068275452, 0.1561063826084137, 0.09666918218135834, 0.12554511427879333, -0.031689759343862534, -0.02023046463727951, 0.013223651796579361, -0.03295358642935753, 0.04495260491967201, -0.11780395358800888, 0.035578932613134384, 0.14455120265483856, 0.049516137689352036, 0.13682803511619568, 0.0735841616988182, -0.05499548465013504, 0.03328321501612663, 0.05524631217122078, -0.12552984058856964, -0.17107287049293518, 0.005011483561247587, -0.005295912269502878, -0.10259857028722763, 0.051755718886852264, 0.11026841402053833, -0.062025588005781174, -0.001793412957340479, -0.02018183283507824, 0.016843479126691818, -0.05819571390748024, 0.22374746203422546, 0.05162270739674568, 0.055952150374650955, -0.0959213450551033, 0.07349611818790436, 0.034072745591402054, -0.1328122466802597, 0.053264569491147995, 0.09296504408121109, -0.06931019574403763, -0.0327560119330883, 0.032683197408914566, 0.1313253492116928, 0.030571231618523598, -0.040415678173303604, -0.13320687413215637, -0.1338537484407425, 0.06949934363365173, 0.21964141726493835, 0.04942631721496582, 0.008563860319554806, -0.04897460341453552, 0.03774720057845116, -0.13920629024505615, 0.0908212959766388, 0.0349992960691452, 0.04622767120599747, -0.12860284745693207, 0.20151029527187347, 0.003881271695718169, 0.021595217287540436, -0.028669578954577446, 0.01921936124563217, -0.1343410611152649, 0.046683453023433685, -0.09093082696199417, -0.01579323597252369, -0.02816624753177166, -0.007766608614474535, 0.006319056730717421, -0.06512380391359329, -0.06335505843162537, 0.01656957156956196, -0.11302769929170609, -0.01601850800216198, 0.01229291781783104, 0.032803576439619064, -0.1074068695306778, -0.046468667685985565, 0.030757669359445572, -0.07511882483959198, 0.08996619284152985, 0.0581984780728817, -0.006446146406233311, 0.075163334608078, -0.18264757096767426, -0.01169278658926487, 0.05783824995160103, -0.013937068171799183, 0.033312711864709854, -0.1002986803650856, -0.0350450724363327, -0.02677314355969429, 0.06530051678419113, 0.024314995855093002, 0.07280121743679047, -0.11571335792541504, 0.005307708401232958, -0.02698533423244953, -0.06084452196955681, -0.0679541751742363, 0.02553417533636093, 0.05410841107368469, 0.02228759601712227, 0.14017564058303833, -0.11588584631681442, 0.05893852934241295, -0.20197343826293945, 0.0015868673799559474, -0.02691839449107647, -0.09125935286283493, -0.0851990357041359, -0.03456255793571472, 0.09504547715187073, -0.06988661736249924, 0.10769051313400269, -0.0271491389721632, 0.06059962883591652, 0.030723635107278824, -0.09523545950651169, -0.0017024778062477708, 0.06349709630012512, 0.2230585813522339, 0.0403946153819561, -0.05847551301121712, 0.05815497040748596, 0.03967234864830971, 0.08684889227151871, 0.1534259170293808, 0.18020136654376984, 0.1981029212474823, 0.0365082323551178, 0.09386779367923737, 0.05703950673341751, -0.07342797517776489, -0.13453207910060883, 0.07986865192651749, -0.06894306093454361, 0.08350721746683121, -0.032439760863780975, 0.25355857610702515, 0.08700919896364212, -0.15285103023052216, 0.056919995695352554, -0.06048684939742088, -0.09919805824756622, -0.1120237410068512, -0.02856682986021042, -0.0895775631070137, -0.14749391376972198, 0.011026449501514435, -0.10716061294078827, 0.03742712363600731, 0.07354573160409927, 0.03550061210989952, 0.01331980712711811, 0.186380073428154, 0.02604096196591854, 0.03934590891003609, 0.1087435781955719, 0.010536760091781616, -0.017882978543639183, -0.06687624752521515, -0.08983606100082397, 0.049914050847291946, -0.023152891546487808, 0.036610573530197144, -0.05488888546824455, -0.11129742860794067, 0.042695824056863785, -0.013935339637100697, -0.10976725816726685, 0.025321325287222862, 0.015528528951108456, 0.08466783910989761, 0.036871537566185, 0.03574708476662636, -0.02810516767203808, -0.01994580216705799, 0.25668179988861084, -0.10720452666282654, -0.0963725671172142, -0.10983160883188248, 0.2755523920059204, 0.023905884474515915, 0.022039363160729408, -0.009461223147809505, -0.08987654000520706, -0.003953215666115284, 0.2230842560529709, 0.17585450410842896, -0.07823452353477478, 0.01177610270678997, -0.00716535747051239, -0.0024505213368684053, -0.054933976382017136, 0.08036120980978012, 0.13488329946994781, 0.0561646893620491, -0.08306947350502014, -0.051436059176921844, -0.028934631496667862, -0.041562292724847794, -0.031237972900271416, 0.07010716944932938, 0.035673316568136215, 0.016515903174877167, -0.061122797429561615, 0.0933796763420105, -0.08855647593736649, -0.12291647493839264, 0.026394106447696686, -0.24042600393295288, -0.1575072854757309, -0.01798500306904316, 0.08155764639377594, 0.021729985252022743, 0.03978720307350159, -0.006744956132024527, 0.0009909938089549541, 0.03515535220503807, -0.008144501596689224, -0.04766973853111267, -0.05855657905340195, 0.05493473261594772, -0.1019381657242775, 0.18754638731479645, -0.05211586505174637, 0.03871863707900047, 0.12407790124416351, 0.07391487061977386, -0.046682026237249374, 0.10273103415966034, 0.05478890612721443, -0.1403772383928299, 0.03320016711950302, 0.19564533233642578, -0.054514314979314804, 0.13030584156513214, 0.04590681567788124, -0.159378319978714, 0.04270622506737709, -0.09802395850419998, -0.07787465304136276, -0.05198068171739578, -0.022688766941428185, -0.0459018312394619, 0.12449964135885239, 0.18768784403800964, -0.052733954042196274, 0.02194327488541603, -0.047943927347660065, -0.00437404727563262, 0.04074440896511078, 0.09046586602926254, -0.06336335092782974, -0.2760789096355438, 0.013951332308351994, 0.018539635464549065, -0.007872647605836391, -0.2859798073768616, -0.08007141947746277, 0.01931476593017578, -0.049466922879219055, -0.07429458200931549, 0.10141762346029282, 0.10665765404701233, 0.0507291778922081, -0.049186937510967255, -0.1413087248802185, -0.01596112735569477, 0.19789209961891174, -0.1591741293668747, -0.07555712014436722 ]
null
null
transformers
![Reyna aloobun qwen0.5B](https://i.imgur.com/QfbOY6c.jpeg) - Finetuned [Qwen/Qwen1.5-1.8B-Chat](https://huggingface.co/Qwen/Qwen1.5-1.8B-Chat), with SFT on teknium's OpenHermes-2.5 dataset. - This marks the inception of my Qwen1.5 LLM series, with this model laying the foundation for what lies ahead. - Format: ChatML - ``` <|im_start|>system {system}<|im_end|> <|im_start|>user {prompt}<|im_end|> <|im_start|>assistant ``` - Next step would be to do a DPO train on top. ## Benchamrks: |Avg. | Arc | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K | |--|--|--|--|--|--|--| |41.46 | 35.24 |60.42 | 45.37 | 41.4 | 60.85 | 5.46 | ## Example: ``` from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer, StoppingCriteria import torch class MyStoppingCriteria(StoppingCriteria): def __init__(self, target_sequence, prompt): self.target_sequence = target_sequence self.prompt=prompt def __call__(self, input_ids, scores, **kwargs): generated_text = tokenizer.decode(input_ids[0]) generated_text = generated_text.replace(self.prompt,'') if self.target_sequence in generated_text: return True return False def __len__(self): return 1 def __iter__(self): yield self modelpath="aloobun/Reyna-Mini-1.8B-v0.1" model = AutoModelForCausalLM.from_pretrained( modelpath, torch_dtype=torch.bfloat16, device_map="cuda", trust_remote_code=True, ) tokenizer = AutoTokenizer.from_pretrained( modelpath, trust_remote_code=True, use_fast=False, ) prompt = "<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n<|im_start|>user\nReflect on a real-world scenario where understanding probability theory could make a significant difference in decision-making.\n<|im_start|>assistant\n" encoded_input = tokenizer(prompt, return_tensors='pt') input_ids=encoded_input['input_ids'].cuda() streamer = TextStreamer(tokenizer=tokenizer, skip_prompt=True) op = model.generate( input_ids, streamer=streamer, pad_token_id=tokenizer.eos_token_id, do_sample=True, temperature=0.6, top_p=0.8, max_new_tokens=512, stopping_criteria=MyStoppingCriteria("<|im_end|>", prompt) ) ``` ## Output: > One real-world scenario where understanding probability theory can make a significant difference in decision-making is in the field of finance. Financial institutions, such as banks and investment firms, must make decisions about lending money to individuals or businesses, and how much risk they should take on. > In this case, understanding probability theory would help financial analysts and investors make more informed decisions by providing them with information about the likelihood of different outcomes. For example, if an investor wants to invest in a particular stock, they might want to understand the probability that it will perform well over time, based on historical data and market trends. They might also be interested in understanding the probability of defaulting on a loan, which would help them evaluate whether it's worth taking on that risk. > Probability theory provides valuable insights into how events are likely to occur and what factors contribute to those probabilities. By using statistical models and simulations, financial professionals can estimate the likelihood of different scenarios and make better-informed decisions about how to allocate their resources. This can lead to increased profits for financial institutions and improved customer satisfaction for individual investors.<|im_end|>
{"license": "other", "library_name": "transformers", "tags": ["chatml", "finetune", "gpt4", "synthetic data", "custom_code", "qwen2"], "datasets": ["teknium/OpenHermes-2.5"], "license_name": "tongyi-qianwen-research", "license_link": "https://huggingface.co/Qwen/Qwen1.5-1.8B-Chat/raw/main/LICENSE"}
text-generation
aloobun/Reyna-Mini-1.8B-v0.1
[ "transformers", "safetensors", "qwen2", "text-generation", "chatml", "finetune", "gpt4", "synthetic data", "custom_code", "conversational", "dataset:teknium/OpenHermes-2.5", "license:other", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-15T02:15:13+00:00
[]
[]
TAGS #transformers #safetensors #qwen2 #text-generation #chatml #finetune #gpt4 #synthetic data #custom_code #conversational #dataset-teknium/OpenHermes-2.5 #license-other #autotrain_compatible #endpoints_compatible #region-us
!Reyna aloobun qwen0.5B * Finetuned Qwen/Qwen1.5-1.8B-Chat, with SFT on teknium's OpenHermes-2.5 dataset. * This marks the inception of my Qwen1.5 LLM series, with this model laying the foundation for what lies ahead. * Format: ChatML -------------- * Next step would be to do a DPO train on top. Benchamrks: ----------- Example: -------- Output: ------- > > One real-world scenario where understanding probability theory can make a significant difference in decision-making is in the field of finance. Financial institutions, such as banks and investment firms, must make decisions about lending money to individuals or businesses, and how much risk they should take on. > In this case, understanding probability theory would help financial analysts and investors make more informed decisions by providing them with information about the likelihood of different outcomes. For example, if an investor wants to invest in a particular stock, they might want to understand the probability that it will perform well over time, based on historical data and market trends. They might also be interested in understanding the probability of defaulting on a loan, which would help them evaluate whether it's worth taking on that risk. > Probability theory provides valuable insights into how events are likely to occur and what factors contribute to those probabilities. By using statistical models and simulations, financial professionals can estimate the likelihood of different scenarios and make better-informed decisions about how to allocate their resources. This can lead to increased profits for financial institutions and improved customer satisfaction for individual investors.<|im\_end|> > > >
[]
[ "TAGS\n#transformers #safetensors #qwen2 #text-generation #chatml #finetune #gpt4 #synthetic data #custom_code #conversational #dataset-teknium/OpenHermes-2.5 #license-other #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 80 ]
[ "passage: TAGS\n#transformers #safetensors #qwen2 #text-generation #chatml #finetune #gpt4 #synthetic data #custom_code #conversational #dataset-teknium/OpenHermes-2.5 #license-other #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.1020805612206459, 0.14314869046211243, -0.003159529762342572, 0.05144061893224716, 0.08320766687393188, -0.027747532352805138, 0.13506582379341125, 0.11245603859424591, 0.05934979394078255, -0.03346659615635872, 0.12416880577802658, 0.14627505838871002, 0.009358714334666729, 0.20879901945590973, -0.08970733731985092, -0.1555982381105423, 0.09588802605867386, -0.031120598316192627, -0.017420465126633644, 0.05005338042974472, 0.1266998052597046, -0.05000864341855049, 0.10906898230314255, -0.03806750848889351, -0.07555948942899704, -0.00927825178951025, 0.04280354082584381, -0.11843553185462952, 0.1066095381975174, 0.07987260818481445, 0.02828785963356495, 0.11194486916065216, -0.014488372020423412, -0.16965606808662415, 0.04180312529206276, -0.012610231526196003, -0.09523407369852066, 0.023116059601306915, 0.045640937983989716, -0.01601906307041645, 0.004468053113669157, 0.04737173765897751, -0.015901973471045494, 0.08734666556119919, -0.10977866500616074, -0.06776069104671478, -0.059802036732435226, -0.03616827726364136, 0.012706602923572063, 0.06720056384801865, 0.003310688305646181, 0.18885885179042816, -0.04750737175345421, 0.11428650468587875, 0.0776360034942627, -0.38468611240386963, 0.0044416626915335655, 0.10474710166454315, 0.0767679214477539, 0.050024811178445816, -0.04270270839333534, 0.025584854185581207, 0.07044783234596252, -0.013536282815039158, 0.00813332013785839, -0.044090636074543, -0.16108199954032898, 0.018896352499723434, -0.07066293805837631, -0.00931092631071806, 0.2544039487838745, 0.0083262138068676, 0.01180760096758604, -0.08249486237764359, -0.0996125116944313, -0.008545443415641785, 0.0030592726543545723, 0.018670465797185898, -0.022805241867899895, 0.07502110302448273, -0.02316856011748314, -0.02857314608991146, -0.11272531747817993, 0.019159190356731415, -0.18461325764656067, 0.08041258156299591, 0.0041859569028019905, 0.06048641726374626, -0.1425340473651886, 0.02806589938700199, -0.051355261355638504, -0.13127796351909637, -0.06296934932470322, -0.12245018035173416, 0.03631968051195145, -0.015167799778282642, -0.027887258678674698, 0.006490355357527733, 0.19895099103450775, 0.17594008147716522, 0.005755869206041098, 0.05302419513463974, -0.0316813588142395, 0.03287625312805176, 0.039571911096572876, 0.008879171684384346, -0.000796561420429498, -0.08568896353244781, 0.10386843979358673, -0.07534844428300858, 0.05604400485754013, -0.039791397750377655, -0.1287965327501297, 0.007606343366205692, 0.0074279336258769035, 0.09646456688642502, 0.05910347029566765, 0.09682152420282364, -0.03981968015432358, 0.01297066081315279, 0.1890043318271637, -0.050952985882759094, -0.01608831062912941, 0.027313778176903725, -0.0033464464358985424, -0.07642825692892075, -0.0075002396479249, 0.08013888448476791, -0.00023467412393074483, -0.06646699458360672, -0.043933186680078506, -0.07585872709751129, -0.004032785072922707, -0.07703892141580582, 0.06199844181537628, -0.0508863627910614, 0.06493912637233734, -0.18021658062934875, -0.18325094878673553, 0.03152550011873245, 0.03567282482981682, -0.011081579141318798, -0.03881525248289108, -0.042071159929037094, -0.05945175513625145, 0.006358879618346691, -0.06452862918376923, -0.08754146844148636, -0.10960718244314194, 0.04712087661027908, -0.014536364004015923, 0.04065908491611481, -0.22100821137428284, 0.0015662105288356543, -0.12698517739772797, -0.007633306551724672, 0.06683460623025894, 0.034468233585357666, -0.09313153475522995, 0.12568718194961548, -0.0059172604233026505, -0.00006964501517359167, -0.0015725638950243592, -0.0038201729767024517, -0.0034543697256594896, 0.15842409431934357, -0.1442597508430481, -0.01332920789718628, 0.2918665111064911, -0.1120595782995224, -0.25778433680534363, 0.1243276596069336, -0.009997133165597916, 0.02602851390838623, 0.07241539657115936, 0.1946936398744583, 0.11206503212451935, -0.09921043366193771, -0.010463499464094639, 0.09261608123779297, -0.06019801273941994, -0.0852128341794014, 0.03814404830336571, 0.006763429846614599, -0.03761499002575874, 0.059547536075115204, 0.07130371034145355, 0.07265592366456985, -0.051110606640577316, -0.08372686803340912, -0.05110543593764305, -0.0940328910946846, 0.04029996693134308, -0.02294997312128544, 0.026115933433175087, -0.11168710142374039, 0.021057050675153732, -0.07483430206775665, 0.03936592489480972, -0.0010287826880812645, -0.027271803468465805, -0.14684835076332092, 0.10053783655166626, 0.01406969502568245, 0.026374464854598045, -0.07951720803976059, -0.16193436086177826, 0.03968866169452667, 0.028994407504796982, 0.0028879994060844183, 0.05162572115659714, 0.07621802389621735, 0.03274904564023018, 0.004871605895459652, -0.03184729069471359, 0.1394425928592682, 0.06509721279144287, -0.04600037261843681, -0.12494953721761703, 0.06337818503379822, -0.08048871159553528, 0.10384304076433182, -0.08558466285467148, 0.021343745291233063, 0.11436264961957932, 0.09799040853977203, 0.0448731891810894, 0.052577726542949677, -0.0004749245708808303, 0.0017009590519592166, -0.07623755186796188, -0.011246972717344761, 0.06843379884958267, 0.0622752383351326, -0.09451405704021454, 0.1388670951128006, -0.15255577862262726, 0.31403979659080505, 0.1654004007577896, -0.11813484132289886, 0.05139970779418945, -0.0664990171790123, -0.03094002977013588, 0.001288771047256887, 0.020357653498649597, 0.007224222179502249, -0.021007779985666275, -0.036078423261642456, 0.12299459427595139, -0.08029855787754059, -0.013463448733091354, 0.034946732223033905, -0.053806617856025696, -0.04407469183206558, 0.05719981715083122, 0.055679332464933395, -0.1802336424589157, 0.14312097430229187, 0.2094603031873703, 0.06484197825193405, 0.15074455738067627, -0.03504631668329239, 0.04110899195075035, 0.009406094439327717, 0.04211572930216789, -0.02077665738761425, 0.06976466625928879, -0.15334230661392212, 0.006344854831695557, 0.07376714795827866, 0.035784024745225906, 0.053277093917131424, -0.10451622307300568, -0.04741073399782181, -0.009152709506452084, -0.036400336772203445, -0.025406505912542343, 0.08000656962394714, -0.012224568985402584, 0.11637090146541595, -0.021346958354115486, -0.07012387365102768, 0.10442614555358887, -0.027631500735878944, -0.10455384850502014, 0.17139048874378204, -0.16634692251682281, -0.19122180342674255, -0.06446872651576996, -0.14031897485256195, -0.08334170281887054, 0.031661003828048706, 0.12265823781490326, -0.05638733506202698, -0.06170115992426872, -0.06887020170688629, -0.08786465227603912, 0.04596484825015068, 0.02892693504691124, -0.016749009490013123, 0.05141698569059372, -0.024594230577349663, -0.12869679927825928, -0.033759746700525284, 0.029647989198565483, -0.06968409568071365, 0.13238561153411865, -0.09166919440031052, 0.12034158408641815, 0.08926623314619064, -0.002807264681905508, 0.02921443060040474, -0.06429407745599747, 0.264190137386322, -0.02553645148873329, 0.005585165228694677, 0.1767122596502304, -0.005874451249837875, 0.04797758907079697, 0.18727490305900574, 0.014077384024858475, -0.11200138181447983, 0.03988025337457657, -0.08528990298509598, -0.06160251796245575, -0.20698656141757965, -0.14355704188346863, -0.08689714968204498, 0.16130401194095612, -0.01990273781120777, 0.06070951744914055, 0.07317736744880676, 0.08881019800901413, -0.013558999635279179, -0.03303080052137375, 0.04419117793440819, 0.07439731806516647, 0.148185133934021, 0.004401296377182007, 0.10745839774608612, -0.08094556629657745, -0.021181127056479454, 0.13318359851837158, 0.10643484443426132, 0.11189905554056168, 0.05335231125354767, 0.13488347828388214, 0.056520331650972366, 0.22159737348556519, 0.0863616093993187, 0.04070865362882614, 0.04257960617542267, -0.026513870805501938, -0.0070035746321082115, -0.039339520037174225, -0.054363857954740524, 0.042512886226177216, -0.050027936697006226, -0.11212797462940216, -0.04486696049571037, 0.033573321998119354, 0.10407749563455582, 0.06775103509426117, 0.06422022730112076, -0.18428659439086914, -0.0609605647623539, 0.08411349356174469, 0.03483312949538231, -0.0469772070646286, 0.12040131539106369, 0.05714414268732071, -0.09997204691171646, 0.10327248275279999, -0.028613420203328133, 0.08732849359512329, 0.02849552407860756, 0.0363072007894516, -0.038840025663375854, -0.07411137968301773, 0.018140796571969986, 0.11823959648609161, -0.30178171396255493, 0.19997842609882355, -0.0007617404335178435, 0.012954478152096272, -0.0985754206776619, -0.017585154622793198, 0.023387577384710312, 0.10509620606899261, 0.123532235622406, -0.009558826684951782, -0.07183555513620377, -0.045245081186294556, -0.04759374260902405, 0.07965465635061264, 0.07486269623041153, 0.05722365900874138, 0.017417283728718758, -0.032858408987522125, -0.014285271055996418, 0.01474649179726839, -0.01717853918671608, -0.09981183707714081, -0.1645687371492386, 0.037088725715875626, 0.10852208733558655, 0.0862061008810997, -0.05153542011976242, -0.003382578492164612, -0.1787499487400055, 0.19011612236499786, -0.14106933772563934, -0.07785582542419434, -0.08494766056537628, -0.09200435131788254, 0.024039002135396004, -0.07136853039264679, 0.08314177393913269, -0.05146736651659012, 0.061612170189619064, -0.043542202562093735, -0.18501447141170502, 0.1316343992948532, -0.1575099676847458, -0.10818906128406525, -0.0477212592959404, 0.10210717469453812, -0.05755748227238655, 0.0013161411043256521, 0.05418099835515022, 0.0004321351007092744, -0.09760642051696777, -0.11932245641946793, -0.011306979693472385, 0.013033043593168259, -0.006020809058099985, 0.016729554161429405, -0.05372362583875656, -0.1650906801223755, 0.0232982374727726, -0.0793612152338028, 0.22033296525478363, 0.291384756565094, -0.08113020658493042, 0.14291732013225555, 0.177536278963089, -0.04131917655467987, -0.31132572889328003, -0.07500234246253967, -0.1383989006280899, -0.052893538028001785, 0.003618067130446434, -0.09291938692331314, 0.07394716888666153, 0.014131344854831696, -0.07096229493618011, 0.06442634761333466, -0.1938401460647583, -0.08354710042476654, 0.14374825358390808, 0.036138348281383514, 0.2571566104888916, -0.15736201405525208, -0.08583997935056686, -0.06086764484643936, -0.24625131487846375, 0.1797744482755661, -0.20187963545322418, 0.04444572702050209, 0.04373137652873993, 0.009123380295932293, 0.0005529160844162107, -0.053193025290966034, 0.08240006864070892, -0.042107049375772476, 0.04239051416516304, -0.09956089407205582, 0.00848050881177187, 0.10002263635396957, -0.015131199732422829, 0.041080325841903687, -0.0622706301510334, 0.07448209822177887, -0.024908343330025673, -0.013434765860438347, -0.06353329122066498, 0.07274310290813446, 0.00730171101167798, -0.09806615114212036, -0.03293302655220032, -0.02359955571591854, 0.05980898067355156, 0.011127335019409657, 0.18808621168136597, -0.005301924422383308, 0.1056923195719719, 0.1883728802204132, 0.14994585514068604, -0.1936195194721222, 0.1428365260362625, -0.02889144979417324, -0.09126178175210953, 0.03303803130984306, -0.06906846910715103, 0.051915209740400314, 0.10992466658353806, -0.0418354831635952, 0.06018614023923874, 0.06854532659053802, 0.03162148594856262, 0.05272641032934189, 0.12526562809944153, -0.200704887509346, -0.1022830381989479, -0.03019694983959198, 0.14476601779460907, 0.037902310490608215, 0.13931886851787567, 0.1446988433599472, -0.0009965378558263183, -0.027334749698638916, -0.04356298968195915, 0.04489975422620773, -0.015285049565136433, 0.05215643718838692, 0.05008796229958534, -0.0034019805025309324, -0.14141124486923218, 0.08920832723379135, -0.006133119110018015, -0.1623765379190445, 0.022706924006342888, 0.0681999996304512, -0.17828094959259033, -0.11978396028280258, -0.031349170953035355, 0.09834422171115875, -0.14769132435321808, -0.09270618110895157, -0.05493616312742233, -0.11079980432987213, 0.008124972693622112, 0.1781974881887436, 0.07816047221422195, 0.08746252208948135, 0.06954742968082428, -0.017940703779459, -0.014218516647815704, 0.0652504488825798, -0.017860576510429382, 0.03966589644551277, -0.1563456952571869, 0.012112384662032127, -0.07212233543395996, 0.10464312136173248, -0.06886526942253113, -0.029511835426092148, -0.13066008687019348, -0.012385436333715916, -0.1382473260164261, 0.016450073570013046, -0.10814490914344788, 0.006861214060336351, -0.022740351036190987, -0.06194600462913513, -0.03824526444077492, 0.0148268798366189, -0.08122306317090988, 0.022880638018250465, -0.008134828880429268, 0.05206877365708351, -0.12101607024669647, -0.021472837775945663, 0.0421842522919178, -0.01482921838760376, 0.1473200023174286, 0.06219702586531639, -0.08894925564527512, 0.06946316361427307, -0.14965398609638214, -0.07557389885187149, 0.07681377977132797, 0.043336935341358185, 0.056269265711307526, 0.04489721357822418, -0.019493451341986656, 0.11788734048604965, -0.01760534942150116, 0.026377959176898003, 0.11397051811218262, -0.0938258022069931, 0.004123962949961424, -0.028881952166557312, -0.0514882393181324, -0.012016748078167439, -0.059711702167987823, 0.08551625162363052, -0.05360950902104378, 0.1607985496520996, -0.08529689908027649, 0.0074599809013307095, -0.08980484306812286, -0.00658091064542532, 0.014469284564256668, -0.13562816381454468, -0.1931791603565216, -0.0037960580084472895, 0.02846178226172924, -0.036428965628147125, 0.27831190824508667, -0.05297786369919777, -0.03417111560702324, 0.060906682163476944, -0.0402519553899765, 0.030562005937099457, 0.02760455757379532, 0.2722121775150299, 0.05987468361854553, -0.015897054225206375, -0.11248679459095001, 0.038914136588573456, 0.011246131733059883, -0.057731565088033676, 0.05535949394106865, 0.05588328093290329, 0.013346615247428417, 0.05228879302740097, 0.043896954506635666, -0.03681280091404915, -0.05315956473350525, -0.11794751137495041, -0.05488763004541397, 0.07409800589084625, -0.030111638829112053, 0.025775020942091942, 0.13998916745185852, -0.02435719035565853, -0.01632862724363804, -0.07071703672409058, -0.04808781296014786, -0.13910771906375885, -0.08204732835292816, -0.1307935118675232, -0.1206749975681305, -0.004030519165098667, -0.12616895139217377, 0.011889709159731865, -0.007414825726300478, 0.052587009966373444, -0.05574702098965645, 0.10927651822566986, -0.06871969252824783, -0.030081115663051605, 0.004954962059855461, -0.024965811520814896, 0.011618896387517452, 0.029884062707424164, -0.058306749910116196, -0.0574486181139946, 0.0023754280991852283, -0.04226023703813553, 0.07178106904029846, -0.01737259514629841, 0.05322503298521042, -0.09880273789167404, -0.06208030506968498, -0.041486479341983795, 0.06260959059000015, -0.040562935173511505, 0.09769868850708008, 0.04415864497423172, 0.002933267969638109, 0.11748813092708588, 0.21537022292613983, -0.061454709619283676, -0.1390165239572525, -0.11077506095170975, 0.10947099328041077, 0.012943245470523834, 0.10113610327243805, 0.005350533872842789, -0.030378738418221474, -0.020463688299059868, 0.2635117471218109, 0.3057822585105896, -0.10060960054397583, 0.03829672187566757, -0.01392883900552988, 0.019595693796873093, 0.02134125865995884, 0.10845991224050522, 0.05065964534878731, 0.26804807782173157, -0.0644209235906601, -0.037830617278814316, 0.0018479136051610112, 0.0194599200040102, -0.12385851889848709, 0.058825403451919556, -0.019085686653852463, -0.04858803749084473, -0.040209025144577026, 0.09453917294740677, -0.1662966012954712, 0.0962991714477539, -0.12491649389266968, -0.16858269274234772, -0.05930950492620468, -0.02267267554998398, 0.09338971972465515, 0.05276630446314812, 0.03684554621577263, -0.03094136342406273, -0.00805115420371294, 0.07594740390777588, -0.003254888579249382, -0.17368541657924652, 0.030620239675045013, 0.03816632926464081, -0.041651591658592224, 0.13015864789485931, -0.0031646944116801023, 0.09423784911632538, 0.105653777718544, 0.01731123775243759, -0.10098257660865784, 0.09663964807987213, 0.06994391977787018, -0.0463828518986702, 0.007267913781106472, -0.08986575156450272, 0.0016174940392374992, 0.041492778807878494, 0.07476619631052017, -0.10256277769804001, 0.04859617352485657, 0.08703044801950455, -0.02997884526848793, -0.057999156415462494, 0.07405979186296463, -0.04601183533668518, 0.09544362127780914, -0.007226546760648489, -0.057801503688097, -0.0018425834132358432, -0.0105383126065135, 0.036315061151981354, 0.023586535826325417, -0.07910054177045822, -0.04886365681886673, -0.1554580181837082, -0.04057062417268753, 0.0791434794664383, 0.04436424747109413, -0.16869713366031647, -0.007046863436698914, -0.12797348201274872, 0.06438635289669037, -0.11894512176513672, 0.09077990055084229, 0.16532714664936066, -0.0051690018735826015, -0.029168639332056046, 0.06108022481203079, -0.0017929964233189821, 0.06476528942584991, -0.0657183974981308, -0.13912333548069 ]
null
null
stable-baselines3
# **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
{"library_name": "stable-baselines3", "tags": ["LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "PPO", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "LunarLander-v2", "type": "LunarLander-v2"}, "metrics": [{"type": "mean_reward", "value": "248.05 +/- 26.13", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
Noname08/ppo-LunarLander-v2
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
2024-02-15T02:22:44+00:00
[]
[]
TAGS #stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
# PPO Agent playing LunarLander-v2 This is a trained model of a PPO agent playing LunarLander-v2 using the stable-baselines3 library. ## Usage (with Stable-baselines3) TODO: Add your code
[ "# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.", "## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ "TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n", "# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.", "## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ 39, 41, 17 ]
[ "passage: TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ 0.03942384943366051, 0.04900386184453964, -0.005304091144353151, 0.026427261531352997, 0.107408307492733, -0.026511888951063156, 0.11188238859176636, 0.0814051404595375, 0.10722193866968155, 0.04762078449130058, 0.08338645845651627, 0.06030960753560066, 0.05080918222665787, 0.2571701407432556, 0.04754156619310379, -0.22987541556358337, 0.036159250885248184, -0.04869936779141426, 0.12395193427801132, 0.07178173214197159, -0.0038484656251966953, -0.06485428661108017, 0.020415637642145157, -0.013290755450725555, 0.05367108806967735, 0.04282612353563309, -0.01716216839849949, -0.08207534998655319, 0.07169748842716217, -0.06345846503973007, 0.06986866891384125, 0.07677983492612839, 0.13218913972377777, -0.17832116782665253, 0.029566360637545586, 0.02571309357881546, -0.07189024239778519, 0.01342033501714468, 0.008019951172173023, 0.05120139941573143, 0.17303818464279175, 0.019879888743162155, 0.07844575494527817, -0.0025605305563658476, -0.15412317216396332, -0.018950799480080605, 0.0436202734708786, 0.12546207010746002, 0.08808347582817078, 0.04605821147561073, 0.01970590092241764, 0.17503218352794647, -0.054352790117263794, -0.028833400458097458, 0.21759237349033356, -0.2881564497947693, -0.031460098922252655, 0.321048766374588, 0.06997483223676682, 0.09725230932235718, -0.07540661096572876, -0.03619609400629997, 0.007783263456076384, -0.013137873262166977, -0.028666524216532707, -0.07447073608636856, 0.17313385009765625, 0.05152064561843872, -0.05057951435446739, -0.09541505575180054, 0.16948209702968597, 0.006921638268977404, 0.0018855923553928733, -0.019282981753349304, 0.009060598909854889, 0.07402525842189789, -0.016097044572234154, -0.07255112379789352, 0.057438433170318604, 0.05330665782094002, 0.019649166613817215, -0.1435653269290924, -0.10762494057416916, -0.022740179672837257, -0.008012006990611553, 0.17786912620067596, -0.009255532175302505, 0.042902372777462006, 0.003065188182517886, 0.10384012013673782, -0.12480384111404419, -0.03354184702038765, -0.0454259067773819, -0.07565800100564957, -0.0223417766392231, -0.02058211714029312, -0.03580251708626747, 0.07184842973947525, 0.11971849203109741, 0.027368178591132164, 0.09350208193063736, 0.047715865075588226, -0.03206788748502731, 0.06343851238489151, 0.05555703118443489, 0.14222665131092072, 0.05807621404528618, 0.012854371219873428, 0.13179877400398254, 0.055213116109371185, 0.033023182302713394, -0.0613492950797081, -0.18252409994602203, 0.07489913702011108, -0.07031869143247604, 0.007941240444779396, 0.12051256000995636, -0.04480670019984245, -0.1183447614312172, -0.037500523030757904, -0.017392054200172424, -0.06224250793457031, -0.025395862758159637, 0.0547584593296051, -0.02883218228816986, -0.03973718360066414, 0.0011496668448671699, 0.09384800493717194, 0.00953749567270279, -0.1752052903175354, 0.03303423151373863, -0.025042934343218803, -0.10782608389854431, 0.009975161403417587, 0.0022444494534283876, 0.03394931182265282, 0.04408763721585274, -0.11822668462991714, -0.30899152159690857, -0.07652641832828522, 0.05490870401263237, -0.06516939401626587, -0.18425025045871735, -0.13193942606449127, 0.02454492449760437, -0.09037084132432938, -0.044885024428367615, -0.12759265303611755, -0.028549788519740105, 0.01743689924478531, 0.011519349180161953, 0.10758619755506516, -0.0106219332665205, -0.012188062071800232, -0.1571401208639145, 0.008273907005786896, -0.20951123535633087, 0.0890483483672142, -0.019150104373693466, 0.037884220480918884, -0.032381169497966766, -0.07404014468193054, 0.030707746744155884, 0.052499737590551376, -0.01474119070917368, 0.13510210812091827, -0.15592676401138306, -0.03691192343831062, -0.007996266707777977, -0.13611900806427002, -0.04786273464560509, -0.10358831286430359, -0.04357128217816353, 0.13354332745075226, 0.018664736300706863, 0.15356586873531342, -0.08709818124771118, -0.0722038671374321, 0.20489206910133362, -0.010411538183689117, -0.12820468842983246, -0.076752208173275, 0.10165707021951675, 0.021510310471057892, -0.056606587022542953, -0.02523270808160305, -0.1839766949415207, -0.0152357779443264, -0.04550420492887497, -0.047039128839969635, 0.01796751655638218, -0.010888241231441498, 0.13837894797325134, 0.08494598418474197, 0.05018039792776108, -0.06086122244596481, -0.006730288732796907, 0.10779471695423126, 0.08823856711387634, 0.008680110797286034, 0.023406028747558594, -0.05774238705635071, 0.09552932530641556, -0.04003755748271942, -0.0142367510125041, -0.08283266425132751, -0.036246106028556824, -0.026256313547492027, 0.17507147789001465, 0.09440762549638748, 0.2257927656173706, 0.09567736834287643, 0.039160262793302536, 0.031270865350961685, -0.13181598484516144, -0.1425403207540512, -0.0017254541162401438, 0.09020978957414627, -0.14270411431789398, -0.04119925573468208, -0.08974775671958923, -0.17768175899982452, -0.12202505767345428, 0.0006432619411498308, -0.17960017919540405, 0.06390921026468277, 0.05408334732055664, -0.035177867859601974, 0.03272094577550888, 0.13032332062721252, -0.011533179320394993, -0.03967514634132385, 0.0831870287656784, 0.0379033200442791, -0.041234664618968964, -0.021742934361100197, 0.11885567009449005, 0.15673065185546875, 0.13124459981918335, -0.03511447086930275, 0.004914294462651014, 0.07076404243707657, -0.02309088408946991, 0.06539414077997208, 0.0558244064450264, 0.20973342657089233, 0.188301220536232, 0.038996949791908264, 0.008822928182780743, -0.07048165798187256, 0.0855446457862854, -0.0742373839020729, -0.14302679896354675, -0.05579735338687897, 0.08729292452335358, 0.016605578362941742, 0.023469142615795135, 0.08711627870798111, 0.024545932188630104, 0.09132762253284454, 0.15968108177185059, 0.01990218088030815, -0.09659269452095032, -0.050218869000673294, 0.01175848301500082, 0.027713103219866753, 0.04794301092624664, -0.04514073207974434, -0.00937939714640379, 0.017020760104060173, -0.10303554683923721, 0.031789086759090424, -0.1413339376449585, -0.1358717679977417, 0.044326696544885635, 0.003906996920704842, 0.010907664895057678, 0.02786896750330925, -0.0038291432429105043, 0.019039705395698547, 0.04351753741502762, -0.06975466758012772, 0.047416772693395615, -0.024745507165789604, -0.020031947642564774, 0.03340689837932587, -0.057257164269685745, -0.205775648355484, -0.17696654796600342, 0.00013708483311347663, -0.09910997003316879, 0.10194740444421768, 0.018308809027075768, -0.12373185902833939, 0.047737859189510345, -0.05822649225592613, 0.027574289590120316, -0.01875593699514866, -0.049130141735076904, 0.10507171601057053, 0.1525275856256485, -0.016146350651979446, 0.018018173053860664, -0.04865182936191559, -0.10157987475395203, -0.19632206857204437, 0.0691583976149559, 0.04680244252085686, 0.014610917307436466, 0.10669491440057755, 0.018072687089443207, 0.02367905154824257, -0.007674071006476879, -0.016521066427230835, -0.011659215204417706, -0.08781040459871292, 0.31909599900245667, 0.04510033503174782, -0.025173069909214973, 0.02041010931134224, -0.0043001663871109486, -0.028083480894565582, 0.03263787180185318, -0.0985708013176918, -0.07548979669809341, -0.08774089068174362, -0.04367410019040108, -0.09784720093011856, 0.053299110382795334, 0.05916472524404526, 0.003188040340319276, -0.07727594673633575, 0.04221395403146744, 0.11369874328374863, -0.0923808291554451, -0.07137343287467957, 0.07477962225675583, 0.0972946360707283, -0.07331304252147675, 0.00012658814375754446, 0.00874367356300354, 0.023951783776283264, 0.037102166563272476, 0.06778035312891006, -0.03966575115919113, 0.08589404821395874, -0.19917890429496765, 0.0372927263379097, 0.106058269739151, 0.023754918947815895, 0.0638108178973198, 0.07643651217222214, -0.1058402881026268, -0.008500572293996811, -0.032518330961465836, -0.21341575682163239, 0.1668180525302887, 0.1355515867471695, 0.06788124144077301, -0.025637222453951836, -0.00461410591378808, -0.0649740919470787, 0.05773647129535675, 0.02723747305572033, -0.14758841693401337, 0.004883295856416225, 0.06064270809292793, 0.026899009943008423, 0.01614922471344471, 0.07971042394638062, 0.014697225764393806, -0.1801026314496994, -0.014406266622245312, 0.10730406641960144, 0.002390873385593295, 0.0053148469887673855, -0.03175045922398567, -0.1755964607000351, 0.0751047357916832, 0.004285442177206278, 0.07233936339616776, -0.1676585078239441, 0.14297930896282196, -0.10089799761772156, 0.07726949453353882, -0.004285062663257122, -0.021311495453119278, 0.02507244050502777, -0.0541163794696331, 0.15163759887218475, 0.01058570109307766, -0.021810131147503853, -0.1200498715043068, -0.1717042326927185, -0.019227758049964905, -0.11788936704397202, -0.11679866164922714, 0.050424277782440186, 0.062185097485780716, 0.04923136904835701, -0.061147067695856094, 0.1518532931804657, -0.047422297298908234, 0.060713399201631546, -0.06893875449895859, -0.06755045056343079, 0.03764858841896057, -0.12588608264923096, -0.08176055550575256, 0.05573027580976486, 0.19166934490203857, 0.15833087265491486, -0.02816431224346161, -0.03472423925995827, -0.047419581562280655, -0.006212298292666674, -0.007802055217325687, 0.0275666993111372, 0.023223137483000755, 0.07315318286418915, -0.07681374251842499, -0.11649256944656372, 0.033787861466407776, -0.06713802367448807, -0.055589709430933, -0.015439179725944996, 0.1513158082962036, 0.04671623185276985, 0.07720734924077988, -0.018946662545204163, 0.03887668624520302, -0.001724981120787561, -0.056474871933460236, 0.16197094321250916, 0.03885216265916824, -0.05193585529923439, 0.06837689876556396, 0.053174007683992386, 0.043745119124650955, 0.03011113777756691, -0.026783017441630363, 0.206032395362854, 0.1980147808790207, 0.014206883497536182, 0.2175983190536499, 0.03177616000175476, -0.03772832080721855, -0.1300560086965561, -0.065880686044693, -0.006372632458806038, 0.03559038043022156, 0.08070417493581772, -0.18207235634326935, -0.015011128038167953, -0.05689644813537598, -0.034518610686063766, -0.15059494972229004, -0.28553900122642517, -0.05957856774330139, 0.20075850188732147, 0.14706264436244965, 0.27519428730010986, -0.10432573407888412, 0.035197313874959946, 0.02663275972008705, -0.04912831634283066, -0.006501141935586929, 0.00018665487004909664, 0.10268618166446686, -0.15421873331069946, 0.1176437959074974, 0.08486983180046082, -0.019002694636583328, 0.01058861706405878, -0.1619086116552353, 0.00936629343777895, -0.12191236019134521, 0.05354422330856323, 0.1400289237499237, -0.048128653317689896, -0.054873593151569366, 0.14033560454845428, -0.024562934413552284, -0.22685599327087402, -0.04648222774267197, -0.043600670993328094, -0.010640020482242107, 0.026607351377606392, -0.1013401448726654, 0.04101909324526787, 0.1330099105834961, 0.009380043484270573, 0.1147187277674675, 0.11749245226383209, -0.052566803991794586, 0.10792597383260727, 0.2257719188928604, -0.018785694614052773, 0.04689010605216026, -0.12743118405342102, -0.0012336712097749114, -0.028270328417420387, 0.013657891191542149, -0.09504974633455276, -0.09938385337591171, 0.02366873063147068, 0.02872389927506447, 0.009118586778640747, 0.0921793207526207, -0.029922157526016235, 0.0759170651435852, 0.06817561388015747, -0.13014446198940277, -0.16288450360298157, 0.015828335657715797, -0.007344507612287998, 0.08354310691356659, 0.00027861111448146403, 0.08878035843372345, -0.11932205408811569, -0.018093237653374672, -0.03153328225016594, -0.03319635987281799, -0.130486860871315, -0.07138993591070175, 0.06156524643301964, 0.028095467016100883, -0.06602972000837326, 0.1398407518863678, 0.026440169662237167, 0.15942534804344177, 0.049197953194379807, 0.012499804608523846, 0.07227300107479095, -0.05345509201288223, 0.1283530443906784, 0.13818155229091644, -0.00868943240493536, -0.05460423603653908, -0.1013643890619278, -0.10236792266368866, 0.08925779908895493, -0.05773641914129257, 0.07476430386304855, -0.14885357022285461, -0.06675903499126434, 0.015772046521306038, 0.016141414642333984, -0.09562095999717712, 0.02571965754032135, -0.01625603251159191, -0.18119946122169495, 0.056570518761873245, -0.048285093158483505, 0.0440407395362854, -0.06347788125276566, -0.1110161691904068, -0.17226378619670868, 0.06091433763504028, 0.08593481779098511, -0.053876690566539764, -0.12229149043560028, 0.011023230850696564, -0.00012518465518951416, -0.06341652572154999, -0.05023367330431938, 0.09722746908664703, -0.11020902544260025, 0.031452205032110214, -0.012567701749503613, 0.08853451162576675, -0.03510405123233795, -0.011538895778357983, 0.044220831245183945, -0.08039166033267975, -0.009481523185968399, 0.03534642979502678, -0.026372017338871956, -0.04127239063382149, -0.2689029574394226, 0.0036654395516961813, 0.0341104120016098, 0.02497158572077751, 0.07856601476669312, 0.011906822212040424, 0.021174922585487366, 0.03993808850646019, -0.15396519005298615, -0.013395369984209538, 0.14574195444583893, -0.07689505815505981, -0.022186370566487312, 0.05703273415565491, -0.09054436534643173, 0.013882770203053951, -0.030287226662039757, 0.1345842480659485, 0.023923413828015327, 0.06404478847980499, -0.0851147472858429, 0.10106813907623291, -0.1451139897108078, -0.04998219385743141, -0.01244612317532301, 0.09761348366737366, 0.07019034773111343, -0.10272270441055298, 0.014697125181555748, 0.04210108891129494, 0.19416837394237518, 0.016384804621338844, -0.0356343574821949, -0.03396720811724663, 0.004015897400677204, 0.22076453268527985, 0.03044266067445278, 0.10457023978233337, 0.07281364500522614, -0.026583973318338394, 0.12624378502368927, 0.09929762035608292, 0.11280370503664017, -0.055645186454057693, 0.13904185593128204, 0.04667386785149574, 0.038641396909952164, 0.0614289753139019, 0.06836545467376709, 0.09098632633686066, -0.0008288522367365658, 0.1138714924454689, 0.013811973854899406, -0.02422109805047512, -0.021335409954190254, 0.17759373784065247, 0.10501719266176224, -0.14769648015499115, 0.029047364369034767, -0.01258957851678133, 0.039933037012815475, -0.014194529503583908, -0.15634691715240479, -0.07240267097949982, -0.3315149247646332, 0.1226184144616127, -0.07119352370500565, 0.019930170848965645, 0.007913772016763687, -0.037425633519887924, -0.03296699747443199, -0.04477746784687042, 0.13151589035987854, -0.013641550205647945, -0.006079165264964104, -0.04815853759646416, -0.015360191464424133, -0.11607866734266281, -0.11200575530529022, -0.013207737356424332, -0.13671602308750153, -0.010119039565324783, 0.05595948174595833, 0.003977729007601738, 0.01821410097181797, -0.03142618387937546, 0.0024383175186812878, 0.06541839241981506, -0.05751744285225868, 0.056182678788900375, 0.12097269296646118, 0.08766137808561325, -0.1058853268623352, 0.031048951670527458, 0.2011747509241104, 0.04359564557671547, -0.12483977526426315, 0.01449228823184967, 0.1819491684436798, 0.004885740112513304, 0.017068125307559967, -0.006097703706473112, -0.0540788508951664, -0.07554277032613754, 0.1251034289598465, 0.08296554535627365, -0.09985227137804031, 0.015833314508199692, -0.0726347416639328, -0.01594804972410202, -0.06374675035476685, 0.10130585730075836, 0.09538925439119339, 0.04440245032310486, -0.10621760785579681, -0.08487539738416672, -0.10891728103160858, 0.040588874369859695, -0.08629853278398514, -0.07311757653951645, 0.09629398584365845, -0.07057105004787445, -0.07029950618743896, 0.025521177798509598, -0.17978744208812714, -0.009467960335314274, 0.1711762249469757, -0.24654000997543335, -0.0916430801153183, -0.10857923328876495, 0.14477859437465668, 0.016497576609253883, 0.1013975441455841, -0.006207061931490898, -0.007889035157859325, -0.20577777922153473, 0.024890204891562462, -0.05293011665344238, -0.02073732763528824, 0.07814782857894897, -0.09476397186517715, 0.22629831731319427, -0.08276885002851486, 0.020940175279974937, 0.012659613974392414, 0.0870661810040474, -0.030675338581204414, 0.09283176809549332, -0.03660329803824425, -0.12576518952846527, -0.03620953485369682, 0.03001813031733036, 0.013904244638979435, 0.10071761906147003, 0.09772487729787827, -0.03414725139737129, 0.03389119729399681, 0.09747414290904999, 0.04172342270612717, -0.023843804374337196, 0.0360250361263752, -0.17077107727527618, 0.02182629331946373, -0.018498148769140244, -0.06935930997133255, 0.03687669709324837, -0.06603235751390457, 0.1639697551727295, 0.04022442549467087, 0.0670473501086235, -0.036152735352516174, 0.0073931049555540085, -0.014454689808189869, -0.013775371946394444, -0.026180334389209747, -0.17259705066680908, -0.10422050207853317, -0.1347656100988388, -0.012701659463346004, -0.034971047192811966, 0.04591470584273338, 0.023234914988279343, -0.0003200018545612693, -0.014577031135559082, -0.12090865522623062, 0.04360328987240791, 0.11146783083677292, -0.04631396010518074, -0.026193076744675636 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
sajedjalil/mistral-pregnancy-instruct
[ "transformers", "tensorboard", "safetensors", "mistral", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "4-bit", "region:us" ]
2024-02-15T02:24:33+00:00
[ "1910.09700" ]
[]
TAGS #transformers #tensorboard #safetensors #mistral #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #tensorboard #safetensors #mistral #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 67, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #mistral #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.052834488451480865, 0.20097468793392181, -0.005400892347097397, 0.010345357470214367, 0.09612055867910385, 0.0044621252454817295, 0.06419585645198822, 0.11483705043792725, -0.05335899814963341, 0.1245594173669815, 0.04259124770760536, 0.11028710752725601, 0.11966550350189209, 0.13622677326202393, -0.009136650711297989, -0.21554915606975555, 0.051239609718322754, -0.09125407040119171, -0.019092371687293053, 0.12837672233581543, 0.15019121766090393, -0.10151926428079605, 0.06721150875091553, -0.023830661550164223, -0.031376853585243225, -0.039496615529060364, -0.054336655884981155, -0.04739919677376747, 0.03964413329958916, 0.049113959074020386, 0.07550079375505447, 0.005664164666086435, 0.08115866780281067, -0.2878633141517639, 0.017135092988610268, 0.0675499364733696, -0.004797542467713356, 0.07008560746908188, 0.09030074626207352, -0.06694217026233673, 0.1188211664557457, -0.05924297496676445, 0.14172065258026123, 0.08049197494983673, -0.08682595938444138, -0.16861778497695923, -0.07765386998653412, 0.11839823424816132, 0.1738077849149704, 0.050799738615751266, -0.033571336418390274, 0.11054174602031708, -0.0838150903582573, 0.023086678236722946, 0.04701470583677292, -0.09320878982543945, -0.06606195122003555, 0.0925995260477066, 0.1036270260810852, 0.051852039992809296, -0.1306041032075882, -0.03030187264084816, 0.014773973263800144, 0.018889451399445534, 0.08437798172235489, 0.020138725638389587, 0.1560133695602417, 0.03462229296565056, -0.13147738575935364, -0.06462032347917557, 0.10889264941215515, 0.035270918160676956, -0.036616988480091095, -0.23675133287906647, -0.017707886174321175, -0.02744358964264393, -0.03218868747353554, -0.047032084316015244, 0.03966495394706726, -0.00232007191516459, 0.08815167844295502, -0.010654046200215816, -0.08137903362512589, -0.04403276741504669, 0.0837794691324234, 0.05812852457165718, 0.027523308992385864, -0.020004788413643837, 0.01929345354437828, 0.11309055984020233, 0.09316704422235489, -0.10987255722284317, -0.05073172599077225, -0.05845588445663452, -0.07336867600679398, -0.04278547689318657, 0.03941237926483154, 0.02383175492286682, 0.07387880980968475, 0.25050491094589233, 0.03487619385123253, 0.05740058794617653, 0.03299154341220856, 0.009766368195414543, 0.04354500770568848, 0.09902188181877136, -0.05393790826201439, -0.11967076361179352, -0.025471940636634827, 0.09969540685415268, 0.007152476813644171, -0.030917100608348846, -0.034157779067754745, 0.05977104231715202, 0.05254608765244484, 0.10806621611118317, 0.09098601341247559, 0.025506848469376564, -0.07453421503305435, -0.050475019961595535, 0.19672119617462158, -0.15638388693332672, 0.03549297899007797, 0.03984997794032097, -0.02883821725845337, -0.03997192531824112, 0.008641471154987812, 0.03237761929631233, -0.038104183971881866, 0.09882140904664993, -0.056019753217697144, -0.04624668508768082, -0.11108994483947754, -0.02722025103867054, 0.04398069530725479, 0.004522797651588917, -0.03378827124834061, -0.02674330398440361, -0.08117839694023132, -0.08551941066980362, 0.09190534055233002, -0.06605688482522964, -0.0690089613199234, -0.030248409137129784, -0.08102237433195114, 0.02253611944615841, 0.022220514714717865, 0.09316928684711456, -0.027028946205973625, 0.05474551394581795, -0.05834097042679787, 0.04209844022989273, 0.10234063118696213, 0.03464803099632263, -0.06545072793960571, 0.07023508846759796, -0.2124854028224945, 0.08670587837696075, -0.08269090950489044, 0.047311387956142426, -0.1593446582555771, -0.024496665224432945, 0.031188564375042915, 0.016910942271351814, -0.00731250923126936, 0.12617072463035583, -0.19719818234443665, -0.01822635903954506, 0.17863377928733826, -0.10461100190877914, -0.0735471099615097, 0.06104305014014244, -0.05694979056715965, 0.11276907473802567, 0.032507818192243576, 0.016837026923894882, 0.057141415774822235, -0.10731352120637894, -0.009754782542586327, -0.053310293704271317, -0.01508351694792509, 0.1277005821466446, 0.07986477017402649, -0.08929067105054855, 0.03966565057635307, 0.01970377378165722, -0.040117714554071426, -0.07162562757730484, -0.02594946324825287, -0.09966299682855606, 0.010024922899901867, -0.07607714086771011, 0.015032773837447166, -0.011939003132283688, -0.09229084849357605, -0.03313123434782028, -0.16445350646972656, -0.01792353019118309, 0.08117435872554779, -0.007835064083337784, -0.02236165478825569, -0.10611458122730255, 0.02960303984582424, 0.014810004271566868, -0.00987476110458374, -0.12626680731773376, -0.027383431792259216, 0.029385773465037346, -0.14307525753974915, 0.02692902833223343, -0.06661265343427658, 0.050006452947854996, 0.014281078241765499, -0.03317156806588173, -0.022567102685570717, 0.016117658466100693, 0.013163539581000805, -0.029023366048932076, -0.22410458326339722, -0.019631661474704742, -0.03548116609454155, 0.15508712828159332, -0.23027491569519043, 0.04255750775337219, 0.08180125057697296, 0.1496315449476242, 0.002845378592610359, -0.060339100658893585, 0.027865778654813766, -0.0643223226070404, -0.026854638010263443, -0.05526704713702202, 0.0040493919514119625, -0.014216830022633076, -0.04673163965344429, 0.018559571355581284, -0.1800202876329422, -0.03300604596734047, 0.09991921484470367, 0.051655370742082596, -0.12397440522909164, -0.03804778307676315, -0.031878430396318436, -0.05654427036643028, -0.046401266008615494, -0.06258269399404526, 0.10518386214971542, 0.059099599719047546, 0.03870894014835358, -0.07163000106811523, -0.07554290443658829, -0.001899047987535596, -0.020858222618699074, -0.02447977289557457, 0.09685797989368439, 0.08302325755357742, -0.11507602035999298, 0.09765294939279556, 0.07297831028699875, 0.053551942110061646, 0.09086218476295471, -0.024014752358198166, -0.0713641420006752, -0.02430378459393978, 0.042651381343603134, 0.021311495453119278, 0.13261832296848297, -0.07990746200084686, 0.037976980209350586, 0.04146222025156021, -0.034816015511751175, 0.02526203729212284, -0.0917649194598198, 0.02330339327454567, 0.02158508077263832, -0.02227955125272274, 0.034679848700761795, -0.041841983795166016, 0.017881741747260094, 0.08489806950092316, 0.0443115234375, 0.025322148576378822, 0.02263973467051983, -0.051010217517614365, -0.11479286104440689, 0.16156119108200073, -0.1080920547246933, -0.21137776970863342, -0.13906092941761017, 0.04211707040667534, 0.034783005714416504, -0.016383640468120575, -0.0033858916722238064, -0.05154045298695564, -0.10119808465242386, -0.09479478001594543, 0.004654999822378159, 0.043191589415073395, -0.08905842900276184, -0.050187595188617706, 0.04112578555941582, 0.0466192327439785, -0.141516774892807, 0.01760321855545044, 0.042856208980083466, -0.09013157337903976, -0.006824543233960867, 0.06852949410676956, 0.08653214573860168, 0.19158309698104858, 0.010567589662969112, -0.01327406708151102, 0.02329033985733986, 0.2191559374332428, -0.13998360931873322, 0.10701308399438858, 0.12081018090248108, -0.08242674171924591, 0.08283291757106781, 0.21218040585517883, 0.043551888316869736, -0.09422869980335236, 0.022803621366620064, 0.04062064737081528, -0.027208451181650162, -0.24665962159633636, -0.06791229546070099, -0.0035661973524838686, -0.062158115208148956, 0.08249735832214355, 0.09320490062236786, 0.09711391478776932, 0.03728710860013962, -0.08616939932107925, -0.09025566279888153, 0.06258123368024826, 0.10880918800830841, -0.014983030967414379, 0.006177902687340975, 0.09033512324094772, -0.03321417048573494, 0.01641284115612507, 0.08443867415189743, 0.0188874714076519, 0.14888599514961243, 0.04413631930947304, 0.17554119229316711, 0.0861152708530426, 0.08592135459184647, 0.008067927323281765, 0.021572953090071678, 0.009305214509367943, 0.041279856115579605, 0.0016330410726368427, -0.07800181210041046, -0.022309783846139908, 0.1177462562918663, 0.06055058166384697, 0.015063058584928513, 0.025537792593240738, -0.04258476570248604, 0.07791682332754135, 0.198396697640419, -0.012044759467244148, -0.19261567294597626, -0.05608483403921127, 0.07379072159528732, -0.09469746798276901, -0.1083059012889862, 0.0016497891629114747, 0.028950732201337814, -0.16726277768611908, 0.03891708329319954, -0.037286631762981415, 0.10906007140874863, -0.11611509323120117, -0.023631032556295395, 0.07255537807941437, 0.0634618028998375, -0.01942109316587448, 0.06924676895141602, -0.20475053787231445, 0.10880783200263977, 0.0066001517698168755, 0.07553447782993317, -0.09947476536035538, 0.08740273118019104, -0.0007526907720603049, -0.023901157081127167, 0.1537390649318695, -0.005621455609798431, -0.07554616034030914, -0.06774292141199112, -0.099391408264637, -0.005713432561606169, 0.10416537523269653, -0.12824052572250366, 0.08237942308187485, -0.03154075890779495, -0.0289112888276577, -0.0011822570813819766, -0.10427870601415634, -0.11393633484840393, -0.1736748367547989, 0.04930219426751137, -0.12380670756101608, 0.03436396270990372, -0.11089404672384262, -0.031757570803165436, -0.03204645216464996, 0.18001362681388855, -0.22215892374515533, -0.08196759968996048, -0.13797107338905334, -0.10522904247045517, 0.13072219491004944, -0.038753706961870193, 0.09259927272796631, -0.011674202978610992, 0.16536912322044373, 0.011018362827599049, -0.010310528799891472, 0.07906722277402878, -0.08910427987575531, -0.20677298307418823, -0.06536334753036499, 0.17066635191440582, 0.11525268852710724, 0.03545212000608444, 0.000287556933471933, 0.035471249371767044, -0.021072793751955032, -0.11005019396543503, 0.02238122932612896, 0.1439078003168106, 0.07187364995479584, 0.004129808396100998, -0.011103570461273193, -0.13524897396564484, -0.0827161967754364, -0.04307263344526291, 0.021800002083182335, 0.15324190258979797, -0.07222111523151398, 0.16097372770309448, 0.13479404151439667, -0.06552577018737793, -0.20236515998840332, 0.01051738765090704, 0.030328264459967613, -0.010007081553339958, 0.008130890317261219, -0.17668405175209045, 0.08620604127645493, 0.011234555393457413, -0.056860510259866714, 0.09279665350914001, -0.1702776551246643, -0.1401495635509491, 0.08297594636678696, 0.06001194193959236, -0.2043580710887909, -0.14006857573986053, -0.09796732664108276, -0.032844237983226776, -0.1621166467666626, 0.09755008667707443, 0.007708913646638393, 0.0037532122805714607, 0.035734739154577255, 0.008165989071130753, 0.02857157029211521, -0.058729082345962524, 0.18171481788158417, 0.00027578818844631314, 0.03282225877046585, -0.08759060502052307, -0.10179191082715988, 0.03070172667503357, -0.044745516031980515, 0.07093081623315811, -0.036149706691503525, 0.00850291084498167, -0.10935381054878235, -0.042256783694028854, -0.0577796995639801, 0.016240792348980904, -0.1020742654800415, -0.09217087179422379, -0.05101632699370384, 0.08228278905153275, 0.10086700320243835, -0.021262628957629204, -0.0342043861746788, -0.08324651420116425, 0.06856834888458252, 0.22493121027946472, 0.1966019570827484, 0.06725580245256424, -0.07912718504667282, 0.00033585348865017295, -0.020831260830163956, 0.0481177419424057, -0.1963294893503189, 0.05629334971308708, 0.059084050357341766, 0.01804365962743759, 0.11463156342506409, -0.026349589228630066, -0.14717429876327515, -0.07080743461847305, 0.06234912574291229, -0.061654895544052124, -0.2110869437456131, 0.013404684141278267, 0.05121738091111183, -0.16916172206401825, -0.04016730189323425, 0.03341935947537422, -0.01679670251905918, -0.033391449600458145, 0.012059276923537254, 0.09350721538066864, -0.0015329624293372035, 0.08846935629844666, 0.07468053698539734, 0.09647973626852036, -0.09671876579523087, 0.08520160615444183, 0.09399132430553436, -0.06869548559188843, 0.030684588477015495, 0.08796190470457077, -0.054721105843782425, -0.03408265486359596, 0.05272325128316879, 0.09408900141716003, 0.028605924919247627, -0.057578157633543015, 0.00033376188366673887, -0.09633384644985199, 0.060371965169906616, 0.10807838290929794, 0.026266643777489662, 0.0167081318795681, 0.050699274986982346, 0.039948832243680954, -0.08407669514417648, 0.12055125832557678, 0.05308898165822029, 0.015593260526657104, -0.04697481170296669, -0.013851642608642578, 0.014918781816959381, -0.041015710681676865, -0.0028622217942029238, -0.010037712752819061, -0.07532934099435806, -0.00953451357781887, -0.1338423639535904, 0.021328842267394066, -0.07621663063764572, 0.015507661737501621, 0.02891606278717518, -0.023744206875562668, 0.009069782681763172, 0.001147671602666378, -0.07515859603881836, -0.054316043853759766, -0.007334519177675247, 0.09763575345277786, -0.16248705983161926, 0.02700991928577423, 0.08528634905815125, -0.10886508226394653, 0.09264004975557327, -0.006348124705255032, -0.013947706669569016, 0.009765660390257835, -0.1670515388250351, 0.05764392763376236, -0.024716202169656754, 0.0021372572518885136, 0.0003869409265462309, -0.17661292850971222, 0.00008348759001819417, -0.030598962679505348, -0.06450269371271133, -0.007133767008781433, -0.020635565742850304, -0.11376550793647766, 0.10409335047006607, 0.01058728527277708, -0.08431603759527206, -0.025131171569228172, 0.03712434694170952, 0.08527004718780518, -0.03837678208947182, 0.14522838592529297, -0.0195770226418972, 0.05875604599714279, -0.16669924557209015, -0.008504304103553295, -0.0049210600554943085, 0.015928806737065315, -0.05591136962175369, -0.005450711119920015, 0.04978008568286896, -0.019212210550904274, 0.18645983934402466, -0.026213057339191437, -0.0003956495493184775, 0.0614234060049057, 0.03633897006511688, -0.005665165837854147, 0.0988188162446022, 0.06276490539312363, 0.0065427785739302635, 0.005723139736801386, 0.010450510308146477, -0.05481121316552162, -0.03706621378660202, -0.17841604351997375, 0.06984421610832214, 0.20357003808021545, 0.09210344403982162, -0.02057892642915249, 0.06632699817419052, -0.10828245431184769, -0.10231878608465195, 0.141780287027359, -0.04210126772522926, -0.0012839960400015116, -0.07574470341205597, 0.14354974031448364, 0.1475493609905243, -0.18999400734901428, 0.07106184959411621, -0.07504885643720627, -0.0507393553853035, -0.09940293431282043, -0.21181391179561615, -0.0631086528301239, -0.038815148174762726, -0.013242577202618122, -0.0515957847237587, 0.06344034522771835, 0.09011968970298767, -0.0003513377159833908, -0.011897272430360317, 0.0804000124335289, -0.03300968185067177, 0.0007827926892787218, 0.035289280116558075, 0.06083391606807709, 0.014630911871790886, -0.06082017347216606, 0.009482069872319698, -0.010187007486820221, 0.050416722893714905, 0.0807848796248436, 0.04022103548049927, -0.02279188670217991, 0.013614661991596222, -0.029341207817196846, -0.10319007933139801, 0.051779381930828094, -0.02582433819770813, -0.06043047085404396, 0.15102724730968475, 0.022875646129250526, 0.0070898099802434444, -0.01308579370379448, 0.22499017417430878, -0.06373954564332962, -0.09914014488458633, -0.14789533615112305, 0.07047948986291885, -0.05500692129135132, 0.0462321862578392, 0.052696723490953445, -0.11287195235490799, 0.03247882425785065, 0.15157677233219147, 0.16244743764400482, -0.022908447310328484, 0.01675555482506752, 0.024786820635199547, 0.005874602124094963, -0.023882437497377396, 0.041918110102415085, 0.04422337934374809, 0.13726219534873962, -0.058842673897743225, 0.07276668399572372, 0.0078765619546175, -0.08557147532701492, -0.020382171496748924, 0.12293543666601181, -0.013597259297966957, 0.003893569577485323, -0.06270778924226761, 0.11952853202819824, -0.06941436976194382, -0.2198021560907364, 0.04258532449603081, -0.0739661380648613, -0.1316194385290146, -0.024106398224830627, 0.05017030984163284, -0.005124331451952457, 0.017861103639006615, 0.07797049731016159, -0.07110268622636795, 0.16910569369792938, 0.036577459424734116, -0.07144417613744736, -0.04361674562096596, 0.0775807574391365, -0.0796847939491272, 0.30380067229270935, 0.019320085644721985, 0.04682227224111557, 0.10782097280025482, -0.018694713711738586, -0.12453937530517578, 0.02970682829618454, 0.09843869507312775, -0.0747128576040268, 0.05450698360800743, 0.1679740697145462, 0.0009233925375156105, 0.12669584155082703, 0.06975183635950089, -0.07647688686847687, 0.04657592624425888, -0.0829431489109993, -0.07340765744447708, -0.10605278611183167, 0.10098569095134735, -0.09048886597156525, 0.14528542757034302, 0.1259707808494568, -0.054054636508226395, 0.010301806963980198, -0.02867099642753601, 0.05575840175151825, -0.008431506343185902, 0.12303311377763748, 0.01731177419424057, -0.19737546145915985, 0.03218993544578552, -0.032990291714668274, 0.10002017021179199, -0.18788345158100128, -0.0809892937541008, 0.0369843989610672, 0.001126440824009478, -0.07855185121297836, 0.11744636297225952, 0.07008904218673706, 0.03504768759012222, -0.048617780208587646, -0.019252369180321693, -0.013791046105325222, 0.14716915786266327, -0.09431074559688568, -0.0007206687005236745 ]
null
null
null
<!-- header start --> <!-- 200823 --> <div style="width: auto; margin-left: auto; margin-right: auto"> <img src="https://github.com/janhq/jan/assets/89722390/35daac7d-b895-487c-a6ac-6663daaad78e" alt="Jan banner" style="width: 100%; min-width: 400px; display: block; margin: auto;"> </div> <p align="center"> <a href="https://jan.ai/">Jan</a> - <a href="https://discord.gg/AsJ8krTT3N">Discord</a> </p> <!-- header end --> # Model Description This is a GGUF version of [CultriX/NeuralTrix-7B-dpo](https://huggingface.co/CultriX/NeuralTrix-7B-dpo) - Model creator: [CultriX](https://huggingface.co/CultriX) - Original model: [NeuralTrix-7B-dpo](https://huggingface.co/CultriX/NeuralTrix-7B-dpo) - Model description: [Readme](https://huggingface.co/CultriX/NeuralTrix-7B-dpo/blob/main/README.md) # About Jan Jan believes in the need for an open-source AI ecosystem and is building the infra and tooling to allow open-source AIs to compete on a level playing field with proprietary ones. Jan's long-term vision is to build a cognitive framework for future robots, who are practical, useful assistants for humans and businesses in everyday life. # Jan Model Converter This is a repository for the [open-source converter](https://github.com/janhq/model-converter. We would be grateful if the community could contribute and strengthen this repository. We are aiming to expand the repo that can convert into various types of format
{"license": "apache-2.0", "tags": ["merge", "mergekit", "lazymergekit", "mlabonne/OmniBeagle-7B", "flemmingmiguel/MBX-7B-v3", "AiMavenAi/AiMaven-Prometheus"], "model_name": "NeuralTrix-7B-dpo", "base_model": "CultriX/NeuralTrix-7B-dpo", "model_creator": "CultriX", "quantized_by": "JanHQ"}
null
janhq/neuraltrix-7b-dpo-GGUF
[ "gguf", "merge", "mergekit", "lazymergekit", "mlabonne/OmniBeagle-7B", "flemmingmiguel/MBX-7B-v3", "AiMavenAi/AiMaven-Prometheus", "base_model:CultriX/NeuralTrix-7B-dpo", "license:apache-2.0", "region:us" ]
2024-02-15T02:25:21+00:00
[]
[]
TAGS #gguf #merge #mergekit #lazymergekit #mlabonne/OmniBeagle-7B #flemmingmiguel/MBX-7B-v3 #AiMavenAi/AiMaven-Prometheus #base_model-CultriX/NeuralTrix-7B-dpo #license-apache-2.0 #region-us
<div style="width: auto; margin-left: auto; margin-right: auto"> <img src="URL alt="Jan banner" style="width: 100%; min-width: 400px; display: block; margin: auto;"> </div> <p align="center"> <a href="URL - <a href="URL </p> # Model Description This is a GGUF version of CultriX/NeuralTrix-7B-dpo - Model creator: CultriX - Original model: NeuralTrix-7B-dpo - Model description: Readme # About Jan Jan believes in the need for an open-source AI ecosystem and is building the infra and tooling to allow open-source AIs to compete on a level playing field with proprietary ones. Jan's long-term vision is to build a cognitive framework for future robots, who are practical, useful assistants for humans and businesses in everyday life. # Jan Model Converter This is a repository for the [open-source converter](URL We would be grateful if the community could contribute and strengthen this repository. We are aiming to expand the repo that can convert into various types of format
[ "# Model Description\nThis is a GGUF version of CultriX/NeuralTrix-7B-dpo\n- Model creator: CultriX\n- Original model: NeuralTrix-7B-dpo\n- Model description: Readme", "# About Jan\nJan believes in the need for an open-source AI ecosystem and is building the infra and tooling to allow open-source AIs to compete on a level playing field with proprietary ones.\n\nJan's long-term vision is to build a cognitive framework for future robots, who are practical, useful assistants for humans and businesses in everyday life.", "# Jan Model Converter\nThis is a repository for the [open-source converter](URL We would be grateful if the community could contribute and strengthen this repository. We are aiming to expand the repo that can convert into various types of format" ]
[ "TAGS\n#gguf #merge #mergekit #lazymergekit #mlabonne/OmniBeagle-7B #flemmingmiguel/MBX-7B-v3 #AiMavenAi/AiMaven-Prometheus #base_model-CultriX/NeuralTrix-7B-dpo #license-apache-2.0 #region-us \n", "# Model Description\nThis is a GGUF version of CultriX/NeuralTrix-7B-dpo\n- Model creator: CultriX\n- Original model: NeuralTrix-7B-dpo\n- Model description: Readme", "# About Jan\nJan believes in the need for an open-source AI ecosystem and is building the infra and tooling to allow open-source AIs to compete on a level playing field with proprietary ones.\n\nJan's long-term vision is to build a cognitive framework for future robots, who are practical, useful assistants for humans and businesses in everyday life.", "# Jan Model Converter\nThis is a repository for the [open-source converter](URL We would be grateful if the community could contribute and strengthen this repository. We are aiming to expand the repo that can convert into various types of format" ]
[ 87, 50, 77, 53 ]
[ "passage: TAGS\n#gguf #merge #mergekit #lazymergekit #mlabonne/OmniBeagle-7B #flemmingmiguel/MBX-7B-v3 #AiMavenAi/AiMaven-Prometheus #base_model-CultriX/NeuralTrix-7B-dpo #license-apache-2.0 #region-us \n# Model Description\nThis is a GGUF version of CultriX/NeuralTrix-7B-dpo\n- Model creator: CultriX\n- Original model: NeuralTrix-7B-dpo\n- Model description: Readme# About Jan\nJan believes in the need for an open-source AI ecosystem and is building the infra and tooling to allow open-source AIs to compete on a level playing field with proprietary ones.\n\nJan's long-term vision is to build a cognitive framework for future robots, who are practical, useful assistants for humans and businesses in everyday life.# Jan Model Converter\nThis is a repository for the [open-source converter](URL We would be grateful if the community could contribute and strengthen this repository. We are aiming to expand the repo that can convert into various types of format" ]
[ -0.029163427650928497, 0.08221867680549622, -0.003204237436875701, 0.021092956885695457, 0.06428578495979309, 0.01598728820681572, 0.13171851634979248, 0.0975784882903099, 0.07746290415525436, 0.021357091143727303, 0.04871530458331108, 0.020408257842063904, 0.07273127138614655, 0.13394463062286377, 0.0596727691590786, -0.2495301067829132, 0.0643276795744896, 0.023875372484326363, -0.031745847314596176, 0.020973077043890953, 0.07296772301197052, -0.023210860788822174, 0.07041147351264954, 0.008094726130366325, -0.07297216355800629, -0.05231622979044914, -0.029616806656122208, 0.016460437327623367, 0.0777985230088234, 0.09670163691043854, -0.03287268802523613, -0.009265240281820297, -0.012310019694268703, -0.010880723595619202, 0.01949213817715645, -0.015368866734206676, -0.0002558796259108931, 0.062305886298418045, -0.05155129358172417, 0.04522301256656647, 0.17091961205005646, -0.08559592813253403, -0.04643120616674423, 0.07661273330450058, -0.09989480674266815, -0.19548186659812927, -0.07909461110830307, -0.009014828130602837, -0.006795934401452541, 0.08369859308004379, -0.022821804508566856, 0.1569550484418869, -0.020044205710291862, 0.0029714645352214575, -0.016582971438765526, -0.2103801667690277, -0.04727809876203537, 0.173998162150383, 0.04958564043045044, 0.08620214462280273, 0.0013861668994650245, 0.10272422432899475, 0.03553108125925064, 0.012033411301672459, 0.047366898506879807, -0.04995633289217949, 0.13097147643566132, -0.06933756172657013, -0.09042251110076904, -0.025686604902148247, 0.2623739242553711, 0.026015913113951683, -0.05813731998205185, -0.03225596994161606, 0.04118506610393524, 0.13642984628677368, -0.039141323417425156, 0.0009570528636686504, 0.034684937447309494, 0.02769480273127556, 0.0319620706140995, -0.13876885175704956, -0.06100761517882347, -0.032738152891397476, -0.09546957910060883, 0.24044552445411682, 0.004969735164195299, 0.04812197387218475, -0.009723568335175514, 0.014948264695703983, -0.24416452646255493, -0.057077255100011826, -0.10280919075012207, -0.048743750900030136, 0.0004953425377607346, -0.023561105132102966, -0.05225365608930588, -0.0861871987581253, 0.08772390335798264, 0.06394583731889725, -0.08941658586263657, -0.04114234447479248, -0.03292965516448021, 0.035956624895334244, 0.08741353452205658, 0.00254698283970356, -0.04056815430521965, -0.02012203447520733, 0.10276147723197937, -0.029190432280302048, 0.06845221668481827, 0.0004205355071462691, -0.1037239357829094, -0.04179965704679489, -0.044906049966812134, 0.022620705887675285, 0.10316765308380127, 0.06347138434648514, 0.004923726432025433, -0.06613297015428543, 0.07574761658906937, -0.011654151603579521, 0.015736257657408714, -0.018072472885251045, -0.08119907230138779, 0.04521692171692848, -0.040780726820230484, 0.01042249333113432, -0.02492610737681389, -0.03918301314115524, -0.07442718744277954, -0.016589729115366936, -0.1157519742846489, -0.08001435548067093, 0.05426453799009323, -0.010393415577709675, 0.05250803753733635, -0.13459883630275726, -0.13231000304222107, -0.003555045463144779, 0.02300010621547699, -0.061262574046850204, -0.08825128525495529, 0.02770274318754673, -0.012981544248759747, -0.035153355449438095, -0.000031117804610403255, 0.019261138513684273, -0.01927940361201763, 0.023654034361243248, 0.02300434373319149, 0.038900598883628845, -0.19708262383937836, 0.030617274343967438, -0.07384790480136871, -0.01863260567188263, -0.2677628695964813, 0.06758385896682739, -0.12499500066041946, 0.043235745280981064, -0.07148642838001251, 0.021095173433423042, -0.08851248770952225, 0.01990451291203499, 0.014000602066516876, 0.172868013381958, -0.07508131861686707, 0.004199015907943249, 0.11108431220054626, -0.10158352553844452, -0.07144884765148163, 0.16014130413532257, 0.02584865875542164, 0.04501587152481079, 0.08777061104774475, 0.17587260901927948, 0.043611444532871246, -0.02135074883699417, -0.09553401917219162, -0.0028066113591194153, -0.024694692343473434, -0.07020800560712814, 0.08261161297559738, 0.030526936054229736, -0.026733821257948875, 0.008096725679934025, -0.1364516168832779, 0.08945102244615555, -0.01534515991806984, -0.00030968853388912976, -0.029351485893130302, -0.07169772684574127, 0.04653976485133171, 0.011298133991658688, 0.049846190959215164, -0.029357751831412315, -0.0365481860935688, -0.03492100536823273, 0.093276247382164, 0.011170721612870693, -0.017974920570850372, -0.1343802809715271, 0.07301180809736252, -0.0680900290608406, 0.08613193035125732, -0.04397952929139137, -0.08220550417900085, 0.0825948491692543, -0.061410173773765564, 0.12057319283485413, 0.08434449881315231, 0.04362739995121956, -0.019036222249269485, 0.0023460029624402523, 0.014336683787405491, -0.04393327236175537, -0.026586635038256645, -0.018636731430888176, -0.14609012007713318, 0.028557565063238144, -0.04754255712032318, 0.17410002648830414, -0.06308752298355103, 0.040649186819791794, 0.08323288708925247, 0.07862701267004013, 0.0049677216447889805, -0.02455301769077778, 0.0746421292424202, 0.024320775642991066, 0.010755938477814198, -0.0020866612903773785, 0.04450858384370804, -0.007098148576915264, -0.15662503242492676, 0.19638656079769135, -0.018587488681077957, 0.011089244857430458, 0.10816391557455063, 0.04082009941339493, -0.0036974591203033924, 0.033636920154094696, -0.004444833379238844, -0.006250420585274696, -0.004758334718644619, -0.11482115089893341, 0.22748887538909912, 0.00244233850389719, 0.01288632582873106, -0.05192454531788826, -0.01165478490293026, -0.0016634783241897821, -0.06095895916223526, 0.031828224658966064, 0.024321354925632477, 0.016574865207076073, -0.1245897188782692, 0.11567198485136032, -0.01850135810673237, 0.00830046646296978, 0.11384551227092743, -0.0019796923734247684, -0.027669986709952354, -0.05023773014545441, 0.021136295050382614, 0.02046925202012062, 0.16130134463310242, -0.16291819512844086, -0.02212991565465927, 0.017710072919726372, 0.022191256284713745, 0.018921609967947006, -0.03702164068818092, 0.002895692829042673, -0.022636301815509796, -0.031033720821142197, 0.11200836300849915, 0.034580301493406296, -0.03706103563308716, 0.07248108088970184, 0.06674020737409592, 0.07526027411222458, 0.0317390039563179, 0.020706232637166977, -0.08185295760631561, 0.09787643700838089, -0.04084724187850952, -0.24301451444625854, -0.08995326608419418, 0.03582948073744774, -0.057783883064985275, 0.017878234386444092, 0.059963565319776535, -0.046061307191848755, -0.04154534265398979, -0.04478500783443451, 0.14851054549217224, -0.058269862085580826, -0.04366851598024368, -0.021347157657146454, -0.022370431572198868, -0.023397674784064293, -0.077290840446949, -0.04071200639009476, 0.003806163091212511, -0.06882622838020325, 0.06152324005961418, -0.007859357632696629, 0.09485998004674911, 0.012169872410595417, 0.021089257672429085, -0.02038579061627388, -0.018151843920350075, 0.23878364264965057, -0.12656450271606445, 0.1131516769528389, 0.17269794642925262, 0.07298140972852707, 0.006700593512505293, 0.09179659187793732, 0.04677095636725426, -0.0837850421667099, 0.013486221432685852, 0.005006772931665182, -0.07456840574741364, -0.15406623482704163, -0.109214186668396, -0.07745308429002762, 0.060632333159446716, -0.05503697693347931, 0.03243726119399071, 0.06094653904438019, 0.12326487898826599, -0.06411655247211456, 0.12524232268333435, -0.04572824761271477, 0.06371574103832245, 0.1697211116552353, -0.03874015435576439, 0.06305677443742752, -0.03345619887113571, -0.018935000523924828, 0.1387157291173935, 0.05300477147102356, 0.1587604582309723, 0.010579856112599373, 0.12688963115215302, 0.05141496658325195, 0.030194884166121483, 0.04836535081267357, -0.021033136174082756, -0.014953585341572762, -0.034869927912950516, -0.06279962509870529, -0.07464321702718735, -0.06351862102746964, 0.07863270491361618, 0.01945647783577442, -0.0250351931899786, 0.012926905415952206, -0.04994548112154007, 0.023393038660287857, 0.14580488204956055, -0.008385359309613705, -0.1204938292503357, -0.13619853556156158, 0.05001356080174446, 0.030504342168569565, 0.010521181859076023, -0.046102870255708694, 0.11291340738534927, -0.08120469748973846, -0.020170878618955612, 0.002022623084485531, 0.06542771309614182, -0.10743805766105652, 0.03456541523337364, -0.15547382831573486, 0.08657168596982956, 0.013504219241440296, 0.0627613291144371, -0.1973261684179306, 0.11712753027677536, 0.04451059550046921, 0.09826253354549408, -0.05488607659935951, 0.007868774235248566, 0.07379045337438583, 0.16900332272052765, 0.06843242794275284, 0.05300222709774971, 0.017478708177804947, -0.01253951620310545, -0.12679912149906158, 0.04531947895884514, -0.030043071135878563, -0.06586690247058868, 0.06348109990358353, 0.03422524780035019, 0.047008492052555084, -0.07083632797002792, 0.07293358445167542, -0.16345162689685822, -0.08664606511592865, 0.07859569787979126, 0.08264598995447159, 0.1373981386423111, -0.04517337679862976, 0.009042205289006233, 0.05715765804052353, 0.18124376237392426, 0.043273258954286575, -0.11289328336715698, -0.11247658729553223, 0.11237403750419617, 0.02627282217144966, -0.06231055408716202, -0.033626168966293335, -0.009274297393858433, 0.02515232376754284, -0.028382210060954094, -0.08026999980211258, 0.028246533125638962, -0.09160162508487701, -0.030023416504263878, -0.02026495896279812, 0.01081762369722128, 0.06632234901189804, 0.09368057548999786, 0.02365901507437229, -0.007888156920671463, 0.03648855537176132, -0.11136215925216675, 0.023400025442242622, 0.1288636326789856, -0.060594961047172546, 0.019458288326859474, -0.09983514249324799, -0.030776550993323326, -0.06699857115745544, -0.05447336286306381, 0.07193512469530106, 0.16792650520801544, -0.05657308176159859, -0.008398638106882572, 0.17879141867160797, -0.035823360085487366, -0.22193840146064758, 0.00678320974111557, 0.00771871255710721, 0.02207757532596588, -0.019568728283047676, -0.24440819025039673, 0.09457731246948242, 0.10419303178787231, -0.033526644110679626, 0.1213676780462265, -0.22811956703662872, -0.005973918363451958, -0.03437631204724312, 0.08094125241041183, 0.2781970202922821, -0.13668647408485413, -0.06991682946681976, 0.07496358454227448, -0.13415858149528503, 0.058528851717710495, -0.03527097404003143, 0.1022377610206604, -0.011539137922227383, 0.1350439339876175, 0.040970850735902786, -0.0154030229896307, 0.1089279055595398, -0.05893561989068985, -0.004790716338902712, -0.0950247272849083, 0.07617408782243729, 0.010171981528401375, -0.023615511134266853, 0.16963234543800354, -0.07418803870677948, 0.0053374688141047955, -0.13102424144744873, -0.015925953164696693, -0.06182835251092911, 0.12058566510677338, -0.009810551069676876, -0.11449693888425827, -0.06930676847696304, 0.1064998060464859, -0.020007118582725525, 0.05697318911552429, 0.016813669353723526, -0.04911063238978386, 0.043111588805913925, 0.03797983005642891, 0.13938160240650177, -0.09469148516654968, 0.00729268416762352, 0.01683815009891987, -0.03710547834634781, 0.10295803099870682, -0.09739945828914642, -0.033016424626111984, 0.08920831233263016, -0.008654293604195118, 0.07578685134649277, 0.0006919601000845432, -0.16571743786334991, 0.051048148423433304, 0.03162829950451851, -0.09202437102794647, -0.2843319773674011, -0.0363646000623703, 0.1391020566225052, 0.025342317298054695, 0.05909759923815727, 0.2052503526210785, -0.09170788526535034, -0.02175898104906082, -0.006602554582059383, 0.015167586505413055, -0.027613213285803795, 0.042068324983119965, -0.039414629340171814, 0.020973220467567444, -0.052583884447813034, -0.038211092352867126, 0.03231576830148697, -0.13018755614757538, 0.05395479500293732, 0.13245296478271484, -0.0750972330570221, -0.12298731505870819, -0.13735224306583405, 0.043227896094322205, -0.08168594539165497, -0.01688787341117859, -0.027942119166254997, -0.11413702368736267, 0.005240052007138729, 0.06734040379524231, 0.029892098158597946, -0.00459081307053566, 0.059302136301994324, 0.007078791037201881, -0.0563698336482048, 0.015682034194469452, -0.12658147513866425, 0.028330866247415543, -0.020574763417243958, -0.05129712074995041, 0.02451120689511299, -0.005702553782612085, -0.054245978593826294, -0.00969316903501749, -0.14910921454429626, -0.08123546838760376, -0.030132519081234932, -0.04592368006706238, -0.0848381519317627, -0.04250562563538551, -0.030870208516716957, -0.075681671500206, -0.012901919893920422, 0.07138244062662125, -0.05230648070573807, -0.010690082795917988, -0.02805372327566147, 0.08828768879175186, -0.016770608723163605, -0.015266981907188892, 0.0052620782516896725, -0.0008226390928030014, 0.10857648402452469, 0.03886409476399422, 0.004295019432902336, -0.0669955238699913, -0.036320414394140244, 0.08176543563604355, -0.007594263646751642, 0.01980462670326233, 0.04500332474708557, -0.08923549950122833, -0.026803355664014816, 0.001394671038724482, 0.0354691743850708, 0.028137246146798134, 0.16295382380485535, -0.07237671315670013, 0.02578488551080227, 0.02102871797978878, 0.015055199153721333, -0.057767707854509354, 0.01144854910671711, 0.06789819151163101, 0.05709458142518997, 0.07220982760190964, -0.03897426649928093, 0.03243793919682503, -0.08492343127727509, 0.0271366648375988, 0.0030299120116978884, -0.0011315220035612583, -0.028058474883437157, -0.04073973000049591, 0.019009992480278015, 0.007357517722994089, 0.09800306707620621, 0.013254258781671524, -0.05792265012860298, 0.057040683925151825, -0.021330537274479866, 0.03932077810168266, -0.064652219414711, 0.02227739244699478, 0.042911332100629807, -0.0023091407492756844, 0.004436607472598553, 0.08289839327335358, -0.02041059173643589, -0.11420264840126038, 0.0541258342564106, 0.07230917364358902, 0.11457201093435287, 0.07250529527664185, 0.0737743228673935, 0.038775306195020676, -0.06841171532869339, 0.014142796397209167, 0.03552113100886345, 0.04366651922464371, -0.051889944821596146, 0.20320312678813934, 0.11057914793491364, -0.04078756645321846, 0.09052001684904099, 0.047305624932050705, -0.00838809460401535, -0.01227671280503273, -0.17505162954330444, -0.04523368552327156, -0.14028066396713257, -0.005528101697564125, -0.1330409198999405, -0.015867700800299644, -0.01190076395869255, 0.01460038498044014, -0.05701606720685959, 0.0730692446231842, -0.025158505886793137, -0.04087825119495392, 0.044329337775707245, -0.030922362580895424, -0.03538396954536438, -0.06139906495809555, -0.0169009268283844, -0.02376243658363819, 0.10778980702161789, 0.0458189956843853, 0.06020694598555565, 0.011438463814556599, -0.005518313962966204, -0.014871755614876747, -0.05500595271587372, -0.020068014040589333, 0.0051034498028457165, 0.02039700374007225, 0.15195396542549133, 0.022853214293718338, -0.07993526011705399, 0.0047328718937933445, 0.16490410268306732, -0.017196422442793846, -0.028168775141239166, -0.10941965132951736, 0.16778577864170074, -0.09073814749717712, -0.01855509914457798, -0.05394666641950607, -0.06504456698894501, -0.010174307972192764, 0.2253689169883728, 0.12076430767774582, -0.03819701448082924, -0.008427510038018227, -0.06788268685340881, -0.009253784082829952, 0.004766174126416445, 0.09721749275922775, 0.05042792856693268, 0.1832546442747116, -0.04023219272494316, 0.05132296681404114, -0.06382223963737488, 0.036016903817653656, -0.1115359216928482, -0.03852206841111183, -0.003550577675923705, -0.0023680024314671755, -0.03442559763789177, 0.06780821830034256, -0.08555134385824203, -0.09655071794986725, -0.03505109250545502, -0.11068925261497498, -0.04640396684408188, -0.03794721141457558, 0.022713880985975266, 0.0012505280319601297, 0.12388281524181366, -0.01904415152966976, 0.01600179448723793, 0.04459480568766594, -0.01592359133064747, -0.11347615718841553, -0.044943634420633316, 0.08853347599506378, -0.06266843527555466, 0.15974728763103485, -0.048388492316007614, -0.0018161004409193993, 0.10953556001186371, -0.0592288039624691, -0.17582644522190094, 0.0467645525932312, -0.03987901657819748, -0.08277586102485657, -0.018756192177534103, 0.11362792551517487, 0.002879116218537092, 0.1088353618979454, 0.06028446927666664, -0.12957556545734406, -0.010470244102180004, 0.04492328688502312, 0.03676741197705269, -0.03390321880578995, 0.0691089853644371, -0.1438034623861313, 0.1210380345582962, 0.10774627327919006, -0.027317050844430923, -0.0790075808763504, -0.06654681265354156, 0.003046538680791855, 0.04912533238530159, 0.01700487732887268, -0.08835574239492416, -0.163203626871109, -0.027412671595811844, 0.022451719269156456, -0.0002115375391440466, -0.23467956483364105, -0.03170536831021309, -0.08294603228569031, 0.055931251496076584, 0.03295820951461792, 0.014573521912097931, 0.14426389336585999, -0.013151087798178196, -0.032081808894872665, -0.22583307325839996, -0.000002104527766277897, 0.04222695901989937, -0.06339491903781891, -0.07077782601118088 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
null
kenchenxingyu/flan-large-single-label-stance-human4
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
2024-02-15T02:25:58+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 31, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.06646376848220825, 0.2168014943599701, -0.00225935154594481, 0.023818302899599075, 0.1271018385887146, -0.001635765191167593, 0.04218708351254463, 0.13324736058712006, -0.020175931975245476, 0.11144465953111649, 0.046588581055402756, 0.09377603232860565, 0.09928803145885468, 0.18404334783554077, 0.04859916493296623, -0.2059975117444992, 0.007056170143187046, -0.09090408682823181, 0.014076028019189835, 0.1116579994559288, 0.13719257712364197, -0.10291384905576706, 0.08272874355316162, -0.04045208916068077, -0.02019004337489605, 0.00012576708104461432, -0.09259183704853058, -0.07032395154237747, 0.06885425746440887, 0.06264153122901917, 0.051234472543001175, 0.001456156256608665, 0.09140396863222122, -0.2864592671394348, 0.017265573143959045, 0.08406311273574829, 0.0027674848679453135, 0.06290827691555023, 0.07236549258232117, -0.07389893382787704, 0.11328595131635666, -0.08021481335163116, 0.13019037246704102, 0.08625296503305435, -0.062064990401268005, -0.23071379959583282, -0.07525765895843506, 0.0963398814201355, 0.12251301854848862, 0.06215599179267883, -0.022921854630112648, 0.15455181896686554, -0.06248689442873001, 0.012971068732440472, 0.1294165402650833, -0.11526761949062347, -0.05572471022605896, 0.061741601675748825, 0.11775490641593933, 0.10740239918231964, -0.14110268652439117, -0.0017287094378843904, 0.04900608956813812, 0.029121357947587967, 0.08589313924312592, 0.022661056369543076, 0.12003941088914871, 0.04652795568108559, -0.13695219159126282, -0.04037507623434067, 0.12011898308992386, 0.038862764835357666, -0.06446044892072678, -0.2168138176202774, -0.006778308190405369, -0.0601806715130806, -0.014732478186488152, -0.07019448280334473, 0.039128515869379044, -0.02470310963690281, 0.07317749410867691, -0.04465159401297569, -0.1063927412033081, -0.0421026237308979, 0.0892222449183464, 0.07748593389987946, 0.011527054943144321, -0.02519804798066616, 0.04627908393740654, 0.13455867767333984, 0.05402068421244621, -0.10399353504180908, -0.07017925381660461, -0.06942764669656754, -0.09420394152402878, -0.04035796597599983, 0.056760527193546295, 0.031942449510097504, 0.02665667235851288, 0.22703726589679718, 0.016653569415211678, 0.04155244305729866, 0.0224777739495039, 0.01032855175435543, 0.043662428855895996, 0.0955500528216362, -0.05303520709276199, -0.15660029649734497, -0.04072032496333122, 0.09077946096658707, -0.0027527001220732927, -0.036689214408397675, -0.03966725245118141, 0.03849169611930847, 0.06843466311693192, 0.13122352957725525, 0.07552056759595871, -0.017929591238498688, -0.04813180863857269, -0.030096933245658875, 0.23523783683776855, -0.1493375599384308, 0.04426715523004532, -0.02271856553852558, -0.01804111897945404, -0.03908449783921242, 0.03597262129187584, 0.022118929773569107, -0.000004518366949923802, 0.09706240892410278, -0.058981191366910934, -0.05378659814596176, -0.10168042778968811, -0.03272576630115509, 0.04088849574327469, -0.013975566253066063, -0.010589460842311382, -0.09025166928768158, -0.09490354359149933, -0.04766594246029854, 0.05537205561995506, -0.05123869329690933, -0.03770573064684868, 0.009465423412621021, -0.08151785284280777, -0.005444355774670839, -0.005417742300778627, 0.10699385404586792, -0.03222226724028587, 0.04445803165435791, -0.027600755915045738, 0.05225523188710213, 0.09919606149196625, 0.031576547771692276, -0.0773419588804245, 0.0561848059296608, -0.22559374570846558, 0.07503069192171097, -0.11481974273920059, 0.04335082694888115, -0.1704932004213333, -0.042439818382263184, 0.005444696638733149, 0.0139949731528759, 0.013206101022660732, 0.12720820307731628, -0.19255615770816803, -0.01654396951198578, 0.13260798156261444, -0.09212633967399597, -0.118110790848732, 0.07884611934423447, -0.029701577499508858, 0.1624738723039627, 0.04682036489248276, -0.027025915682315826, 0.09224298596382141, -0.16434773802757263, -0.07092688232660294, -0.00949116237461567, -0.01727987825870514, 0.12109188735485077, 0.07512219995260239, -0.05991523340344429, 0.046571120619773865, 0.02832140028476715, -0.038078423589468, -0.04424772411584854, -0.050857074558734894, -0.10884185880422592, -0.01070026308298111, -0.08987759798765182, 0.04065500199794769, -0.01250192429870367, -0.07916021347045898, -0.029885273426771164, -0.18612512946128845, -0.0030564051121473312, 0.10038342326879501, 0.0035033065360039473, -0.005652366206049919, -0.08666291832923889, 0.026358824223279953, -0.03112892620265484, -0.008404186926782131, -0.16764774918556213, -0.04399421438574791, 0.046902090311050415, -0.16094985604286194, 0.020117372274398804, -0.06413903087377548, 0.06334125250577927, 0.03641495108604431, -0.05590536445379257, -0.0248766727745533, -0.01730942726135254, 0.011945613659918308, -0.05083848536014557, -0.18994836509227753, -0.056277405470609665, -0.037882111966609955, 0.149809330701828, -0.25956398248672485, 0.032966937869787216, 0.051140617579221725, 0.14649195969104767, 0.00406361510977149, -0.05115427449345589, 0.01429014839231968, -0.05360214412212372, -0.054652128368616104, -0.06746816635131836, -0.006135428790003061, -0.027576493099331856, -0.05147203803062439, 0.019243421033024788, -0.1755700707435608, -0.021410830318927765, 0.09424154460430145, 0.12876708805561066, -0.1486445665359497, -0.018640631809830666, -0.048725154250860214, -0.06339836865663528, -0.0715010017156601, -0.07038594037294388, 0.10712739825248718, 0.0513901449739933, 0.04796046018600464, -0.07435787469148636, -0.07092321664094925, 0.02726263552904129, 0.006906150374561548, -0.03382374346256256, 0.08727246522903442, 0.05199531093239784, -0.09209315478801727, 0.0756213590502739, 0.1092359870672226, 0.07177663594484329, 0.09363535046577454, 0.01574566215276718, -0.11756632477045059, -0.028492970392107964, 0.036266472190618515, 0.02740776725113392, 0.1465986967086792, -0.05952361226081848, 0.04016614332795143, 0.04494241625070572, -0.04170418903231621, 0.022319864481687546, -0.08787637203931808, 0.024075502529740334, 0.025203049182891846, -0.0034381982404738665, 0.06284574419260025, -0.02525499276816845, -0.0050758360885083675, 0.07016654312610626, 0.047779910266399384, 0.04621000960469246, 0.009655474685132504, -0.01720241829752922, -0.1047825813293457, 0.16950392723083496, -0.0951867327094078, -0.269941508769989, -0.17632324993610382, 0.026197833940386772, 0.04035249724984169, -0.022378476336598396, 0.031619444489479065, -0.07056326419115067, -0.10630585998296738, -0.1060405746102333, -0.002429972169920802, 0.01714223250746727, -0.06364088505506516, -0.0741225928068161, 0.07348573952913284, 0.04382912442088127, -0.14902326464653015, 0.038552410900592804, 0.055694397538900375, -0.057955220341682434, -0.0233661737293005, 0.09118817001581192, 0.12397737801074982, 0.14583967626094818, -0.021366750821471214, -0.028626007959246635, 0.029004426673054695, 0.19620531797409058, -0.13469526171684265, 0.10371150821447372, 0.13814030587673187, -0.04545360431075096, 0.08360563963651657, 0.1560150384902954, 0.029186224564909935, -0.08317049592733383, 0.05044832453131676, 0.04082648828625679, -0.043159641325473785, -0.2666129767894745, -0.0534592866897583, 0.012832709588110447, -0.06255637854337692, 0.09786593168973923, 0.10183793306350708, 0.11542957276105881, 0.034910861402750015, -0.07166364789009094, -0.043925940990448, -0.0058974819257855415, 0.11737963557243347, -0.05490213260054588, -0.012639665976166725, 0.07686592638492584, -0.05086168646812439, 0.005355054512619972, 0.10266812145709991, 0.02973790094256401, 0.17442677915096283, 0.020399179309606552, 0.11231429129838943, 0.06195578724145889, 0.08633565157651901, 0.0007386076031252742, 0.02951662428677082, 0.05147615820169449, 0.017203815281391144, -0.002300140680745244, -0.10421168059110641, -0.006156572140753269, 0.1449710875749588, 0.028103826567530632, 0.029669636860489845, -0.0018948549404740334, -0.005003341939300299, 0.05121048167347908, 0.1746254414319992, -0.011592294089496136, -0.22072425484657288, -0.0845772922039032, 0.06936841458082199, -0.06218599155545235, -0.12968985736370087, -0.026130788028240204, 0.045467354357242584, -0.17519839107990265, 0.026703642681241035, -0.027433741837739944, 0.0919293761253357, -0.09345759451389313, -0.02221956104040146, 0.03687324374914169, 0.084866963326931, -0.014529162086546421, 0.08703910559415817, -0.14498743414878845, 0.11886418610811234, 0.02978132851421833, 0.09024628251791, -0.11081171780824661, 0.07909037172794342, -0.007550720125436783, 0.009180475026369095, 0.19379350543022156, -0.011335089802742004, -0.03514958545565605, -0.08774717897176743, -0.11210042238235474, -0.013537433929741383, 0.12687496840953827, -0.1243172138929367, 0.08773399889469147, -0.015198243781924248, -0.044079482555389404, 0.00937260314822197, -0.12100647389888763, -0.17273177206516266, -0.19628387689590454, 0.05585884302854538, -0.09575839340686798, 0.025643249973654747, -0.11914430558681488, -0.07089093327522278, -0.02952558360993862, 0.241120383143425, -0.1745356321334839, -0.06510113179683685, -0.1468164622783661, -0.046294767409563065, 0.1662203073501587, -0.04437198117375374, 0.0718095526099205, -0.0208172257989645, 0.20345525443553925, 0.005988610442727804, -0.004939318168908358, 0.06724198162555695, -0.08892562240362167, -0.16873881220817566, -0.06771010160446167, 0.1510489284992218, 0.11680185794830322, 0.04907919466495514, -0.002248800592496991, 0.0011772146681323647, -0.016943959519267082, -0.1137804463505745, -0.0033210667315870523, 0.16037839651107788, 0.03878779336810112, 0.025986969470977783, -0.05243593826889992, -0.08797456324100494, -0.06899320334196091, -0.06853509694337845, 0.06221301481127739, 0.19590823352336884, -0.10376439243555069, 0.1700313836336136, 0.147536963224411, -0.07305635511875153, -0.23175598680973053, 0.035342130810022354, 0.04983805492520332, 0.0014306638622656465, 0.04886869341135025, -0.18252557516098022, 0.10521943867206573, 0.019543392583727837, -0.05505957826972008, 0.13485197722911835, -0.1557481735944748, -0.1552847921848297, 0.0722852572798729, 0.03904085233807564, -0.22423844039440155, -0.1354004591703415, -0.09622503817081451, -0.05825018882751465, -0.14065024256706238, 0.06054598465561867, -0.002136280992999673, 0.015948504209518433, 0.03500790148973465, -0.0015643214574083686, 0.027123261243104935, -0.058935679495334625, 0.18609118461608887, -0.004065449349582195, 0.020676052197813988, -0.060264769941568375, -0.0478842556476593, 0.09839435666799545, -0.06130504235625267, 0.12208222597837448, 0.004057085141539574, 0.01594383642077446, -0.10362856835126877, -0.048314861953258514, -0.04328322783112526, 0.05154227837920189, -0.07548051327466965, -0.10070807486772537, -0.043625857681035995, 0.08841723203659058, 0.07005169242620468, -0.03383097052574158, 0.00549331633374095, -0.07189501076936722, 0.10019614547491074, 0.17795267701148987, 0.17573626339435577, 0.009926567785441875, -0.07241068035364151, 0.01677953451871872, -0.04142116755247116, 0.044231921434402466, -0.2513144314289093, 0.03756171092391014, 0.06098250672221184, 0.029438555240631104, 0.09217222779989243, -0.020435843616724014, -0.1820858269929886, -0.04050002992153168, 0.08094815909862518, -0.05452597141265869, -0.22617179155349731, -0.019085140898823738, 0.0954197570681572, -0.2020406424999237, -0.007372708059847355, 0.03995226323604584, -0.048725228756666183, -0.023169852793216705, 0.00010950004070764408, 0.06317184865474701, 0.002471912419423461, 0.09773622453212738, 0.0735151618719101, 0.09715340286493301, -0.08337292820215225, 0.10562895983457565, 0.10150538384914398, -0.09572599828243256, 0.03605884686112404, 0.06754924356937408, -0.05300498008728027, -0.043293699622154236, 0.03665391728281975, 0.033023297786712646, 0.005234600510448217, -0.060321882367134094, 0.013913018628954887, -0.036497246474027634, 0.044923391193151474, 0.08326134830713272, 0.03754979372024536, -0.013354414142668247, 0.06462216377258301, 0.03401726484298706, -0.10898099094629288, 0.10366570204496384, 0.01731540448963642, 0.04105307161808014, -0.08384523540735245, -0.019968897104263306, 0.035425446927547455, 0.030576206743717194, -0.01765924133360386, -0.02306121215224266, -0.02860277332365513, -0.01614218018949032, -0.14299540221691132, -0.023106401786208153, -0.07243485748767853, 0.006181265693157911, 0.014656842686235905, -0.031884219497442245, -0.011233693920075893, 0.02475680410861969, -0.06979699432849884, -0.07426341623067856, -0.006949664559215307, 0.09833318740129471, -0.15115703642368317, 0.008848577737808228, 0.06907843053340912, -0.11088496446609497, 0.08190931379795074, -0.008411259390413761, 0.016245156526565552, 0.022527478635311127, -0.15448406338691711, 0.05601610988378525, 0.0008648968650959432, 0.01916889287531376, 0.025886621326208115, -0.16471809148788452, 0.004104440100491047, -0.04661374166607857, -0.02149827405810356, -0.00004464812809601426, -0.02647159807384014, -0.12325995415449142, 0.06858719140291214, -0.015622655861079693, -0.035931166261434555, -0.02701525390148163, 0.0539589487016201, 0.07888586074113846, -0.027474910020828247, 0.10445091128349304, -0.008690856397151947, 0.04941811040043831, -0.16801609098911285, -0.02470702864229679, -0.04982255399227142, 0.019377702847123146, 0.009884213097393513, -0.007693959400057793, 0.04183054715394974, -0.00976533442735672, 0.21883612871170044, -0.05075952783226967, 0.1607085019350052, 0.05847611650824547, -0.017352959141135216, -0.0007513365126214921, 0.06180921941995621, 0.05997028574347496, 0.04658793285489082, 0.009480604901909828, 0.023740366101264954, -0.022450892254710197, -0.006695089396089315, -0.15932634472846985, 0.01890849508345127, 0.14999441802501678, 0.06301083415746689, 0.024745315313339233, 0.05866100639104843, -0.12775006890296936, -0.12135478109121323, 0.09311001747846603, -0.026755332946777344, 0.00928465835750103, -0.08245618641376495, 0.1358020007610321, 0.14980104565620422, -0.14000412821769714, 0.05256148427724838, -0.06134212389588356, -0.05217423290014267, -0.10388828068971634, -0.12032219022512436, -0.05887215584516525, -0.053666237741708755, 0.002330566756427288, -0.03760887682437897, 0.054546963423490524, 0.03344334661960602, -0.009351172484457493, -0.00022941511997487396, 0.13597318530082703, -0.019751882180571556, -0.0028988157864660025, 0.048313532024621964, 0.03693558648228645, 0.02373051457107067, -0.05275435373187065, 0.02940409444272518, 0.02539868652820587, 0.032232340425252914, 0.06546790152788162, 0.033412106335163116, -0.047448933124542236, 0.03804153576493263, -0.0025254099164158106, -0.11207924783229828, 0.019641218706965446, -0.00460948096588254, -0.0742158442735672, 0.1268945336341858, 0.0407399944961071, 0.010224059224128723, -0.03741471841931343, 0.24361543357372284, -0.06653323769569397, -0.06378097087144852, -0.13251738250255585, 0.10491154342889786, -0.0027236645109951496, 0.06476365029811859, 0.023412218317389488, -0.1284150779247284, 0.005243356805294752, 0.13858191668987274, 0.12181595712900162, 0.0045748427510261536, 0.009228081442415714, 0.0518609918653965, 0.0025186820421367884, -0.06998204439878464, 0.054019294679164886, 0.06992026418447495, 0.12919506430625916, -0.07847554981708527, 0.07680778950452805, 0.0006860480643808842, -0.08370215445756912, -0.02947772853076458, 0.11312682181596756, -0.0409729965031147, 0.03491825982928276, -0.047444481402635574, 0.10916327685117722, -0.05787910893559456, -0.29412412643432617, 0.02350960113108158, -0.09588567912578583, -0.15202060341835022, -0.018367812037467957, 0.05944539234042168, -0.02624768204987049, 0.018029648810625076, 0.06971040368080139, -0.06011629104614258, 0.20098382234573364, 0.0335683599114418, -0.07864278554916382, -0.0664360448718071, 0.04837050288915634, -0.06564252078533173, 0.2949807047843933, 0.008418165147304535, 0.02863333560526371, 0.10770907253026962, -0.03253700211644173, -0.18271861970424652, 0.010723991319537163, 0.1133992001414299, -0.08056149631738663, 0.08200647681951523, 0.19000613689422607, -0.012578671798110008, 0.1209007054567337, 0.05294662341475487, -0.047376248985528946, 0.04217283055186272, -0.03389401361346245, -0.051268599927425385, -0.10752558708190918, 0.058453381061553955, -0.05909625440835953, 0.15447644889354706, 0.10152646154165268, -0.05671518296003342, -0.004550917539745569, -0.05555408447980881, 0.04875178262591362, 0.01804669201374054, 0.12263146042823792, 0.02951994352042675, -0.1865430772304535, 0.032826557755470276, -0.01144319772720337, 0.10186848044395447, -0.25588861107826233, -0.08421015739440918, 0.08833149075508118, -0.011924264021217823, -0.05105875805020332, 0.10560628771781921, 0.057650718837976456, 0.04243382066488266, -0.043439045548439026, -0.10480839014053345, -0.02186836116015911, 0.14663739502429962, -0.1469624787569046, -0.025013303384184837 ]
null
null
transformers
# Model Card for Zenith-7B-dpo-v2 Mistral-7B-v0.1 model fine-tuned on the Ultrafeedback dataset using techinques shown in the paper [Self-Rewarding Language Models](https://arxiv.org/abs/2401.10020). ## Instruction format In order to leverage instruction fine-tuning, your prompt should be surrounded by `[INST]` and `[/INST]` tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id. E.g. ``` text = "<s>[INST] What is your favourite condiment? [/INST]" "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!</s> " "[INST] Do you have mayonnaise recipes? [/INST]" ``` This format is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating) via the `apply_chat_template()` method: ```python from transformers import AutoModelForCausalLM, AutoTokenizer device = "cuda" # the device to load the model onto model = AutoModelForCausalLM.from_pretrained("Xenon1/Zenith-7B-dpo-v2") tokenizer = AutoTokenizer.from_pretrained("Xenon1/Zenith-7B-dpo-v2") messages = [ {"role": "user", "content": "What is your favourite condiment?"}, {"role": "assistant", "content": "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!"}, {"role": "user", "content": "Do you have mayonnaise recipes?"} ] encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt") model_inputs = encodeds.to(device) model.to(device) generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True) decoded = tokenizer.batch_decode(generated_ids) print(decoded[0]) ``` ## Model Architecture This instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices: - Grouped-Query Attention - Sliding-Window Attention - Byte-fallback BPE tokenizer
{"language": ["en"], "license": "apache-2.0", "tags": ["mistral", "Zenith-7B-dpo-v2"], "pipeline_tag": "text-generation"}
text-generation
Xenon1/Zenith-7B-dpo-v2
[ "transformers", "safetensors", "mistral", "text-generation", "Zenith-7B-dpo-v2", "conversational", "en", "arxiv:2401.10020", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T02:26:35+00:00
[ "2401.10020" ]
[ "en" ]
TAGS #transformers #safetensors #mistral #text-generation #Zenith-7B-dpo-v2 #conversational #en #arxiv-2401.10020 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Zenith-7B-dpo-v2 Mistral-7B-v0.1 model fine-tuned on the Ultrafeedback dataset using techinques shown in the paper Self-Rewarding Language Models. ## Instruction format In order to leverage instruction fine-tuning, your prompt should be surrounded by '[INST]' and '[/INST]' tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id. E.g. This format is available as a chat template via the 'apply_chat_template()' method: ## Model Architecture This instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices: - Grouped-Query Attention - Sliding-Window Attention - Byte-fallback BPE tokenizer
[ "# Model Card for Zenith-7B-dpo-v2\n\nMistral-7B-v0.1 model fine-tuned on the Ultrafeedback dataset using techinques shown in the paper Self-Rewarding Language Models.", "## Instruction format\n\nIn order to leverage instruction fine-tuning, your prompt should be surrounded by '[INST]' and '[/INST]' tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.\n\nE.g.\n\n\nThis format is available as a chat template via the 'apply_chat_template()' method:", "## Model Architecture\nThis instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices:\n- Grouped-Query Attention\n- Sliding-Window Attention\n- Byte-fallback BPE tokenizer" ]
[ "TAGS\n#transformers #safetensors #mistral #text-generation #Zenith-7B-dpo-v2 #conversational #en #arxiv-2401.10020 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Zenith-7B-dpo-v2\n\nMistral-7B-v0.1 model fine-tuned on the Ultrafeedback dataset using techinques shown in the paper Self-Rewarding Language Models.", "## Instruction format\n\nIn order to leverage instruction fine-tuning, your prompt should be surrounded by '[INST]' and '[/INST]' tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.\n\nE.g.\n\n\nThis format is available as a chat template via the 'apply_chat_template()' method:", "## Model Architecture\nThis instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices:\n- Grouped-Query Attention\n- Sliding-Window Attention\n- Byte-fallback BPE tokenizer" ]
[ 82, 50, 105, 56 ]
[ "passage: TAGS\n#transformers #safetensors #mistral #text-generation #Zenith-7B-dpo-v2 #conversational #en #arxiv-2401.10020 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Zenith-7B-dpo-v2\n\nMistral-7B-v0.1 model fine-tuned on the Ultrafeedback dataset using techinques shown in the paper Self-Rewarding Language Models.## Instruction format\n\nIn order to leverage instruction fine-tuning, your prompt should be surrounded by '[INST]' and '[/INST]' tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.\n\nE.g.\n\n\nThis format is available as a chat template via the 'apply_chat_template()' method:## Model Architecture\nThis instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices:\n- Grouped-Query Attention\n- Sliding-Window Attention\n- Byte-fallback BPE tokenizer" ]
[ -0.07516591250896454, 0.016913706436753273, -0.0037457954604178667, 0.009644999168813229, 0.08768381178379059, -0.003637594636529684, 0.14061953127384186, 0.09384749829769135, 0.03414345905184746, 0.01422712579369545, -0.014215992763638496, 0.07917260378599167, 0.1098560094833374, 0.17455238103866577, 0.03913615271449089, -0.2534770369529724, 0.064045749604702, -0.00441601499915123, 0.1114666536450386, 0.05709623917937279, 0.083580382168293, -0.037118326872587204, 0.04219880327582359, 0.06269991397857666, -0.09868881106376648, 0.004507307428866625, -0.01230582781136036, -0.010125748813152313, 0.08863198757171631, 0.06877651065587997, 0.042111556977033615, 0.011197609826922417, 0.00012013426021439955, -0.151293084025383, 0.024129197001457214, 0.10245496779680252, -0.025462351739406586, 0.049719247967004776, 0.05051256716251373, 0.03601228445768356, 0.02147665247321129, 0.007964755408465862, -0.02193697914481163, 0.06586412340402603, -0.06541753560304642, -0.15250037610530853, -0.06670684367418289, 0.030979325994849205, 0.16674424707889557, 0.051253002136945724, 0.0005287037929520011, 0.08484693616628647, 0.03170910105109215, 0.08180246502161026, 0.10268903523683548, -0.24128124117851257, -0.028443435207009315, 0.03203818202018738, 0.03981834277510643, 0.08330168575048447, -0.0457674041390419, -0.02339632622897625, 0.003358193440362811, 0.01534283347427845, -0.027912946417927742, -0.029429825022816658, -0.024907754734158516, -0.06837113201618195, -0.1142987310886383, -0.02803436666727066, 0.20358049869537354, 0.02031412161886692, -0.06884340196847916, -0.12819255888462067, -0.09607503563165665, 0.08328015357255936, -0.004321306478232145, -0.054044775664806366, 0.05544796958565712, 0.06329295039176941, 0.1267983317375183, -0.08483250439167023, -0.09240680187940598, -0.04212970659136772, -0.035437509417533875, 0.013804955407977104, -0.02953970432281494, 0.057372868061065674, -0.09261259436607361, 0.12336845695972443, -0.11430999636650085, -0.09109599888324738, -0.10377248376607895, -0.03862085938453674, -0.031547512859106064, -0.036898959428071976, 0.00824833381921053, -0.04665077105164528, 0.05482333153486252, 0.20189519226551056, 0.023376096040010452, 0.07131126523017883, -0.018067993223667145, 0.030356062576174736, 0.01287105679512024, 0.11543326079845428, 0.0356074757874012, -0.07946532219648361, 0.07261485606431961, 0.06482410430908203, 0.11755956709384918, -0.007453096564859152, -0.11584725230932236, -0.07638327032327652, -0.03258484974503517, 0.010112869553267956, 0.0672866702079773, 0.03546492010354996, -0.04157007485628128, -0.07129883766174316, 0.2808903753757477, -0.10169508308172226, 0.016428614035248756, -0.0017171829240396619, -0.01389472559094429, 0.047569628804922104, 0.08569217473268509, -0.04148178547620773, -0.06641221791505814, -0.09663607180118561, -0.06656099110841751, -0.035078082233667374, -0.1284170001745224, -0.06785432249307632, 0.007577318698167801, -0.04636090248823166, -0.03059777058660984, -0.10889700055122375, -0.2665465772151947, 0.0062364377081394196, 0.043403103947639465, -0.028389979153871536, -0.01190169621258974, -0.07805276662111282, -0.01715863309800625, -0.009054254740476608, -0.058438438922166824, 0.039720676839351654, -0.03846557438373566, 0.013420828618109226, 0.004465792793780565, 0.07228045165538788, -0.13054785132408142, 0.03449247404932976, -0.05528469756245613, 0.027746688574552536, -0.27738842368125916, 0.10999319702386856, -0.009523081593215466, 0.03941257670521736, -0.05772886052727699, 0.0362536646425724, 0.05182664841413498, 0.016468102112412453, 0.057499926537275314, 0.10049901902675629, -0.16935525834560394, -0.038077425211668015, 0.1846846342086792, -0.1804669201374054, -0.05028801038861275, 0.09081327170133591, 0.009782781824469566, 0.049903787672519684, 0.0676950067281723, 0.15557736158370972, 0.15893171727657318, -0.01505131833255291, -0.05209164693951607, 0.08517351746559143, -0.07969032973051071, -0.05655756965279579, -0.027754411101341248, -0.015248211100697517, 0.04154740273952484, 0.06490074843168259, 0.008892887271940708, 0.04911426082253456, 0.03530660271644592, 0.007117789704352617, -0.01708059199154377, 0.0027579227462410927, -0.02131398394703865, -0.046471886336803436, -0.04600119963288307, -0.02475578710436821, -0.08163072913885117, 0.09961391240358353, 0.12624837458133698, -0.04040573164820671, -0.012929482385516167, -0.07904671877622604, 0.09954767674207687, -0.09259691089391708, 0.028741709887981415, -0.11515061557292938, -0.0455746129155159, 0.041734565049409866, -0.019055480137467384, -0.03742944449186325, 0.0655912458896637, 0.04969863221049309, 0.030387045815587044, 0.04222235828638077, -0.06075473874807358, 0.07806409895420074, -0.04027962312102318, -0.05654500797390938, -0.056612204760313034, -0.04843701049685478, -0.06105648726224899, 0.10093608498573303, -0.11829298734664917, 0.06699760258197784, 0.06856314837932587, 0.08911643177270889, 0.03203118219971657, -0.038746997714042664, -0.007622090633958578, -0.03979681804776192, -0.036151207983493805, -0.04757177084684372, 0.04744114354252815, 0.10283991694450378, 0.01818479597568512, 0.03534187376499176, -0.20650814473628998, -0.12261789292097092, 0.0814552828669548, 0.02221541292965412, 0.02178635634481907, -0.12401134520769119, -0.029866080731153488, -0.01851510815322399, -0.016842298209667206, -0.1022600308060646, 0.19610200822353363, 0.00035835150629282, 0.1431385576725006, -0.06675617396831512, -0.05887933075428009, -0.012534167617559433, -0.03421206772327423, -0.01691533625125885, 0.022587604820728302, -0.10131426155567169, -0.04182548448443413, 0.05543475225567818, -0.0067140632309019566, 0.004719397518783808, 0.18225635588169098, -0.0024754367768764496, -0.029437262564897537, -0.005821520462632179, 0.05249519273638725, -0.008850744925439358, 0.09380020201206207, -0.1487659066915512, 0.002001249697059393, 0.0516502819955349, 0.03534230589866638, 0.03209933266043663, -0.12067576497793198, 0.07057860493659973, -0.00815269909799099, -0.057127974927425385, 0.0023038010112941265, 0.011525371111929417, -0.020717378705739975, 0.04417287930846214, -0.005546530708670616, 0.023367879912257195, 0.022714275866746902, -0.058350443840026855, -0.12020085752010345, 0.12257564067840576, -0.1186349019408226, -0.21935473382472992, -0.17722094058990479, -0.09164729714393616, -0.06714076548814774, 0.019167043268680573, 0.07608825713396072, -0.025685593485832214, -0.05235522612929344, -0.07484998553991318, -0.03277848660945892, 0.04907581955194473, -0.05559762939810753, -0.01708095520734787, -0.03978743776679039, 0.06878863275051117, -0.15991389751434326, -0.016641344875097275, -0.007604596670717001, -0.07492177933454514, 0.06575509905815125, -0.01126454584300518, 0.014502327889204025, 0.08650664240121841, -0.0006144574144855142, 0.01779565028846264, -0.001777391997165978, 0.157160684466362, -0.04402622580528259, 0.11777649074792862, 0.24947196245193481, -0.008067850954830647, 0.08505231142044067, 0.1240646094083786, -0.04494403675198555, -0.06327053159475327, 0.0031487587839365005, -0.040110230445861816, -0.02403016947209835, -0.14940319955348969, -0.04493377357721329, -0.08359295129776001, -0.040172137320041656, 0.019738800823688507, 0.038297705352306366, 0.08418203145265579, 0.05946530029177666, -0.06915285438299179, 0.08152730762958527, 0.05355377122759819, 0.10471747815608978, 0.11128303408622742, -0.020148755982518196, 0.06926585733890533, -0.0680672898888588, 0.022999674081802368, 0.08207811415195465, 0.10908426344394684, 0.1951497495174408, -0.03687316179275513, 0.14414606988430023, 0.04631027206778526, -0.021956460550427437, 0.04476754739880562, 0.069710373878479, -0.07160645723342896, -0.008995844051241875, -0.015067635104060173, -0.06280951201915741, -0.029068419709801674, 0.07158423215150833, -0.11863792687654495, 0.08088135719299316, -0.0439625009894371, 0.011472881771624088, 0.07020335644483566, 0.23344160616397858, 0.060949381440877914, -0.2242097407579422, -0.07868529111146927, 0.08022288978099823, -0.02243688702583313, -0.14407086372375488, -0.0041859946213662624, 0.12722767889499664, -0.04269237443804741, 0.061128776520490646, -0.0024264329113066196, 0.08548113703727722, -0.12572765350341797, 0.023218592628836632, -0.024331098422408104, 0.19051533937454224, -0.007521792780607939, 0.07499425858259201, -0.1352095603942871, 0.04694991558790207, 0.02875104732811451, 0.09281551837921143, -0.07799137383699417, 0.10295083373785019, 0.04511597007513046, 0.057576730847358704, 0.08370350301265717, 0.004724014084786177, 0.038413651287555695, 0.0054326304234564304, -0.07116777449846268, 0.023638702929019928, -0.0033641948830336332, -0.011097291484475136, 0.020569749176502228, -0.0476909764111042, -0.02059713378548622, 0.0003915343258995563, -0.06472421437501907, -0.05574575439095497, -0.1443479210138321, 0.01489955186843872, 0.07257336378097534, 0.04762745648622513, -0.04628433287143707, 0.027802269905805588, 0.0015153007116168737, 0.144664004445076, -0.07136531919240952, -0.10701444000005722, -0.10785078257322311, -0.02179878205060959, -0.033044371753931046, -0.018513508141040802, 0.0334479995071888, 0.0007992129540070891, 0.16314783692359924, -0.010336995124816895, -0.10356549173593521, 0.06502126902341843, -0.11453812569379807, -0.0008931901538744569, -0.03956574946641922, 0.040903639048337936, 0.057673379778862, -0.03502753749489784, 0.0274010319262743, -0.03502945974469185, -0.04489607363939285, -0.08168578147888184, 0.01173531636595726, 0.22699324786663055, 0.0253048837184906, 0.005908555816859007, -0.04629708454012871, -0.17884846031665802, -0.0006493977853097022, 0.02809734456241131, 0.03765618056058884, 0.24932986497879028, -0.046368926763534546, 0.08889564871788025, 0.13493432104587555, -0.06922446936368942, -0.15406152606010437, 0.011466539464890957, 0.035748183727264404, 0.03609289228916168, -0.09073103964328766, -0.13192978501319885, 0.1141807809472084, 0.04597087949514389, -0.02857748232781887, 0.19028814136981964, -0.14871227741241455, -0.07856137305498123, 0.09781353920698166, 0.0556243471801281, 0.13192899525165558, -0.041898515075445175, -0.04437634348869324, -0.0350320041179657, -0.03098279982805252, 0.02606935054063797, -0.07676078379154205, 0.04601307585835457, -0.02956051379442215, 0.06382029503583908, 0.017844023182988167, -0.03137412294745445, 0.08695822954177856, -0.022781342267990112, 0.05991048365831375, -0.062023647129535675, 0.06276246160268784, 0.10094444453716278, -0.08812379837036133, 0.09575625509023666, -0.04636688530445099, 0.021605903282761574, -0.11003921926021576, -0.016156012192368507, -0.06461475044488907, 0.06344761699438095, -0.0033562597818672657, -0.06918750703334808, 0.02712215855717659, -0.012906720861792564, 0.015493056736886501, 0.03732850402593613, -0.0677153542637825, -0.08488978445529938, 0.01434505544602871, 0.13989607989788055, 0.1302502453327179, -0.14755164086818695, -0.010927201248705387, -0.022083552554249763, -0.011748801916837692, 0.08169202506542206, -0.07978962361812592, 0.015501241199672222, 0.03240088000893593, -0.006156590301543474, 0.1262671798467636, 0.046047184616327286, -0.09907335788011551, -0.00041971338214352727, 0.04941263049840927, -0.12612056732177734, -0.0898091271519661, -0.05171411484479904, 0.14270660281181335, -0.10058670490980148, 0.06840579211711884, 0.1396482139825821, 0.021769240498542786, -0.02864862233400345, 0.035534847527742386, 0.02777525782585144, -0.03947170078754425, 0.0025923368521034718, -0.032598674297332764, 0.05045909062027931, -0.06633960455656052, 0.05964977666735649, 0.009474319405853748, -0.061879269778728485, 0.025182124227285385, 0.09452074021100998, -0.12671810388565063, -0.07517444342374802, -0.09268632531166077, 0.12436525523662567, 0.016836171969771385, -0.03721402585506439, -0.04629574716091156, -0.08493387699127197, 0.028082964941859245, 0.08174124360084534, 0.017330322414636612, -0.04127407446503639, -0.023582754656672478, 0.054745152592659, -0.0723615437746048, 0.09298796206712723, -0.05236077681183815, 0.06242615729570389, -0.07844585180282593, 0.0034654783084988594, 0.011043422855436802, -0.007237889803946018, -0.020758764818310738, -0.07462164014577866, -0.08989627659320831, -0.019699666649103165, -0.11747583746910095, 0.035748764872550964, -0.060238853096961975, 0.012561485171318054, 0.01332228071987629, 0.02264522947371006, -0.011349588632583618, 0.03337361291050911, -0.03124130144715309, -0.012844832614064217, -0.007169273681938648, 0.05286317691206932, -0.07421041280031204, -0.006615177728235722, -0.040512826293706894, -0.10242842137813568, 0.1213531643152237, 0.032357893884181976, -0.02008983865380287, -0.027615027502179146, -0.12514324486255646, -0.021311966702342033, 0.02798965573310852, 0.0711657777428627, 0.046538472175598145, -0.013238778337836266, 0.009727824479341507, 0.04139016568660736, -0.05634171515703201, -0.06254123151302338, 0.030969159677624702, -0.04967284947633743, 0.06415307521820068, -0.013453579507768154, -0.01907695084810257, -0.1124420017004013, 0.02580544538795948, 0.08670946955680847, 0.047601886093616486, 0.08082939684391022, -0.06515902280807495, -0.007030735723674297, -0.1698450893163681, -0.029345311224460602, 0.05379503220319748, -0.045618556439876556, 0.05475211516022682, -0.06669323146343231, 0.026536226272583008, -0.02385156787931919, 0.05793623626232147, 0.004602760076522827, 0.05122784525156021, 0.0032290315721184015, -0.052438825368881226, -0.0550876222550869, 0.00017402622324880213, 0.12594711780548096, 0.00624481588602066, 0.039938315749168396, 0.014444890432059765, 0.04535035043954849, 0.06136448308825493, 0.03624993562698364, 0.06586840003728867, 0.02178911305963993, -0.04381517693400383, 0.1046515479683876, 0.029106615111231804, -0.06070399284362793, -0.17718695104122162, 0.009557956829667091, -0.05421595647931099, 0.09209323674440384, -0.04516010731458664, 0.1205836683511734, 0.16859187185764313, -0.08166439831256866, 0.020499354228377342, 0.01529498677700758, -0.03055788390338421, -0.07509434223175049, -0.2715233862400055, 0.00014641373127233237, -0.14584894478321075, -0.05516904965043068, -0.10441185534000397, 0.028036637231707573, 0.04884514585137367, -0.023609008640050888, 0.030624013394117355, 0.08032822608947754, -0.013588267378509045, -0.022796493023633957, 0.029779165983200073, -0.045942287892103195, 0.041021931916475296, -0.047711435705423355, -0.03685500845313072, 0.09333908557891846, 0.015896791592240334, 0.043375443667173386, 0.0344640351831913, 0.10987766087055206, 0.04323162883520126, -0.0034754713997244835, -0.07134696841239929, -0.0002696456213016063, 0.040562696754932404, -0.058960579335689545, 0.07305306941270828, 0.08091661334037781, -0.03437279537320137, 0.006857456639409065, 0.11764544248580933, -0.050744254142045975, -0.10351128876209259, -0.08188661187887192, 0.13529404997825623, -0.04125659540295601, 0.014543388038873672, -0.016585851088166237, -0.1193196102976799, -0.005482813343405724, 0.16450485587120056, 0.08388902992010117, -0.004733697045594454, -0.016841305419802666, -0.013067435473203659, -0.011060996912419796, -0.033228274434804916, 0.10744369029998779, 0.08680076152086258, 0.25345754623413086, 0.006791895255446434, 0.054446108639240265, 0.043198663741350174, -0.025056712329387665, -0.03977898135781288, 0.027387313544750214, -0.047049615532159805, -0.004591326229274273, 0.01841045916080475, 0.011485107243061066, 0.005654572509229183, 0.007035036105662584, -0.0209969449788332, -0.0003619542985688895, -0.054568100720644, -0.02115379460155964, 0.07078571617603302, -0.032964255660772324, 0.05030137300491333, -0.048735007643699646, -0.008310760371387005, 0.29426273703575134, -0.0735313892364502, -0.10506799072027206, -0.08064522594213486, 0.011460809968411922, -0.17325356602668762, 0.13877710700035095, -0.023138239979743958, 0.029458604753017426, 0.0830611139535904, 0.06546461582183838, -0.12128622084856033, 0.032389286905527115, -0.05939764156937599, -0.12447760254144669, 0.03278689086437225, 0.0819881409406662, -0.030749138444662094, 0.030920924618840218, 0.02313075214624405, -0.1268329918384552, 0.014613335020840168, 0.04332779347896576, -0.04310420900583267, -0.045858342200517654, 0.01861540973186493, -0.07076383382081985, 0.1256401240825653, 0.05833729729056358, -0.019619200378656387, -0.025640130043029785, -0.03334280475974083, 0.03805120661854744, 0.04491369053721428, 0.05034281313419342, -0.014395655132830143, -0.09929823130369186, -0.011611987836658955, 0.058762650936841965, 0.012394269928336143, -0.18558675050735474, -0.061920881271362305, -0.08984252065420151, -0.04910729452967644, -0.11082524806261063, 0.06322526186704636, 0.10993436723947525, 0.08852499723434448, -0.04746459797024727, 0.02935878373682499, -0.010087703354656696, 0.11369947344064713, -0.07652126252651215, -0.11518123745918274 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
devashat/244-py-script
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T02:27:02+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 57, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.05622259899973869, 0.16002345085144043, -0.004987028427422047, 0.023115945979952812, 0.0962471067905426, 0.011845538392663002, 0.06785304099321365, 0.11496778577566147, -0.020396295934915543, 0.11142492294311523, 0.03292480856180191, 0.0972127765417099, 0.11474913358688354, 0.16215258836746216, 0.004439093638211489, -0.23455148935317993, 0.04782992601394653, -0.12695099413394928, -0.033447545021772385, 0.11785799264907837, 0.14491069316864014, -0.10402194410562515, 0.07766910642385483, -0.030544815585017204, -0.009361269883811474, -0.03290390968322754, -0.06365230679512024, -0.05152205005288124, 0.05037128925323486, 0.06932847946882248, 0.06591591984033585, 0.007509593386203051, 0.09122733771800995, -0.2655104100704193, 0.02280162274837494, 0.07630051672458649, -0.0015554219717159867, 0.07497020810842514, 0.048351652920246124, -0.08209776133298874, 0.0788840726017952, -0.05696587264537811, 0.14718368649482727, 0.08216129243373871, -0.08924587815999985, -0.1965435892343521, -0.08464295417070389, 0.10284840315580368, 0.18357418477535248, 0.05158785358071327, -0.024141347035765648, 0.10476154088973999, -0.08419200032949448, 0.008797040209174156, 0.06024181470274925, -0.06443428993225098, -0.05412506312131882, 0.06934051215648651, 0.07975570857524872, 0.07967228442430496, -0.13025140762329102, -0.014651902951300144, 0.011243549175560474, 0.007594773545861244, 0.08504551649093628, 0.022028017789125443, 0.14595499634742737, 0.04393624886870384, -0.13030564785003662, -0.044304780662059784, 0.09771761298179626, 0.04345165938138962, -0.053857799619436264, -0.2537047266960144, -0.024983759969472885, -0.03927002474665642, -0.03094942681491375, -0.038562554866075516, 0.04431856796145439, -0.011080716736614704, 0.08032315224409103, -0.01118796318769455, -0.08149448037147522, -0.041395120322704315, 0.06544242054224014, 0.062143467366695404, 0.026896316558122635, -0.01158317644149065, 0.00973866879940033, 0.1224486380815506, 0.10907839238643646, -0.12763150036334991, -0.05768941715359688, -0.06755511462688446, -0.08307720720767975, -0.04300352931022644, 0.03337155282497406, 0.044020529836416245, 0.04436098039150238, 0.2466370165348053, 0.01108562108129263, 0.05453123152256012, 0.045806169509887695, 0.010608446784317493, 0.06787561625242233, 0.11606968939304352, -0.062306761741638184, -0.09178462624549866, -0.029058339074254036, 0.09215214103460312, 0.006741520017385483, -0.042814407497644424, -0.060904473066329956, 0.06479041278362274, 0.012608112767338753, 0.12110785394906998, 0.08444269746541977, 0.0026690615341067314, -0.07305197417736053, -0.06963318586349487, 0.18848419189453125, -0.1598394364118576, 0.047875016927719116, 0.031182926148176193, -0.038971830159425735, -0.0014042917173355818, 0.008752269670367241, 0.02394084818661213, -0.020246321335434914, 0.08923295140266418, -0.05574449151754379, -0.03784004598855972, -0.11079790443181992, -0.03252100944519043, 0.030985163524746895, 0.0051483530551195145, -0.027043871581554413, -0.033837489783763885, -0.09040277451276779, -0.059588029980659485, 0.0922931432723999, -0.07471107691526413, -0.04984431713819504, -0.013726521283388138, -0.07691634446382523, 0.023329194635152817, 0.016799474135041237, 0.08357251435518265, -0.02157396264374256, 0.0384126678109169, -0.0560205839574337, 0.0631464347243309, 0.11269522458314896, 0.029363946989178658, -0.053069718182086945, 0.05750001594424248, -0.24315528571605682, 0.10326608270406723, -0.07320205867290497, 0.050549428910017014, -0.15059062838554382, -0.026000602170825005, 0.044471126049757004, 0.00805877335369587, -0.013138634152710438, 0.14088952541351318, -0.21621745824813843, -0.0323486253619194, 0.16741067171096802, -0.0939871072769165, -0.07602590322494507, 0.059108685702085495, -0.05233629792928696, 0.10869261622428894, 0.04351044446229935, -0.02232111617922783, 0.060673557221889496, -0.14475463330745697, -0.01067100279033184, -0.04139741137623787, -0.02402937039732933, 0.16397778689861298, 0.07567544281482697, -0.06286642700433731, 0.08052356541156769, 0.024165838956832886, -0.017831770703196526, -0.04484899342060089, -0.023361295461654663, -0.10819391161203384, 0.009856974706053734, -0.06032416597008705, 0.02424289658665657, -0.025761527940630913, -0.09367526322603226, -0.02868773601949215, -0.1802000105381012, -0.009223134256899357, 0.0881323292851448, -0.011722641065716743, -0.021903391927480698, -0.12039245665073395, 0.011948852799832821, 0.031212422996759415, 0.002984174294397235, -0.13029038906097412, -0.05838731303811073, 0.027675874531269073, -0.16422230005264282, 0.03272955119609833, -0.05597274377942085, 0.05056252330541611, 0.03445037454366684, -0.03187771514058113, -0.033117350190877914, 0.009550533257424831, 0.006354342680424452, -0.010578392073512077, -0.2502359449863434, -0.02440580166876316, -0.0219739843159914, 0.17386503517627716, -0.21793730556964874, 0.04213962331414223, 0.07686693966388702, 0.14929872751235962, 0.006240781396627426, -0.038500864058732986, 0.010139784775674343, -0.08222103863954544, -0.030560437589883804, -0.0643099993467331, -0.012082485482096672, -0.03717579320073128, -0.05608142167329788, 0.05165567249059677, -0.16133594512939453, -0.028727244585752487, 0.1057019829750061, 0.06860516220331192, -0.14001330733299255, -0.019125886261463165, -0.04171464592218399, -0.043496038764715195, -0.05877087265253067, -0.0552728995680809, 0.1185101792216301, 0.05596614256501198, 0.04696191847324371, -0.06956122815608978, -0.07775315642356873, 0.007865429855883121, -0.017090093344449997, -0.017978519201278687, 0.08920905739068985, 0.07311701774597168, -0.12023317068815231, 0.09247473627328873, 0.10194233059883118, 0.09365488588809967, 0.108615942299366, -0.017981963232159615, -0.08929306268692017, -0.04584396257996559, 0.02045595459640026, 0.013332244008779526, 0.14797501266002655, -0.01403066236525774, 0.056954506784677505, 0.03922648727893829, -0.01123172789812088, 0.012020308524370193, -0.09384570270776749, 0.027314940467476845, 0.034342724829912186, -0.020308034494519234, 0.03796098753809929, -0.04001156985759735, 0.019826533272862434, 0.08712323755025864, 0.04676510766148567, 0.04415108636021614, 0.011758276261389256, -0.04233846068382263, -0.10904491692781448, 0.173858180642128, -0.12615609169006348, -0.24583272635936737, -0.14115718007087708, 0.0015609683468937874, 0.04152948409318924, -0.009671499952673912, 0.003867273684591055, -0.07054664939641953, -0.11710625886917114, -0.0934595838189125, 0.018713686615228653, 0.04491026699542999, -0.07426843047142029, -0.0596279613673687, 0.059872306883335114, 0.03894329443573952, -0.14430272579193115, 0.022237464785575867, 0.047419775277376175, -0.09032250195741653, -0.006925572175532579, 0.08398029953241348, 0.06729988008737564, 0.17764869332313538, 0.009659109637141228, -0.021044570952653885, 0.03080335259437561, 0.21258224546909332, -0.14283664524555206, 0.11252175271511078, 0.14021345973014832, -0.09024007618427277, 0.08099348843097687, 0.1948828399181366, 0.039186809211969376, -0.10478170961141586, 0.03259138762950897, 0.02489176020026207, -0.028939135372638702, -0.25018003582954407, -0.0680207833647728, 0.002590036718174815, -0.04892077296972275, 0.07092583924531937, 0.0918794497847557, 0.09946957975625992, 0.015428726561367512, -0.09732488542795181, -0.08017807453870773, 0.0468163788318634, 0.10640767961740494, 0.0070237633772194386, -0.01532268337905407, 0.08905128389596939, -0.03260866180062294, 0.018378758803009987, 0.0954233929514885, 0.00412675691768527, 0.17459604144096375, 0.05586163327097893, 0.17767499387264252, 0.07751350849866867, 0.06634163856506348, 0.019167855381965637, 0.0069374511949718, 0.02067388966679573, 0.017508454620838165, -0.004214957356452942, -0.08522020280361176, -0.00457410141825676, 0.12029227614402771, 0.06321834027767181, 0.024303704500198364, 0.0137604009360075, -0.03941800817847252, 0.08438141644001007, 0.17332784831523895, 0.0020201504230499268, -0.18486954271793365, -0.07240456342697144, 0.07921045273542404, -0.0910051167011261, -0.10552998632192612, -0.03353073075413704, 0.03346012532711029, -0.1747758537530899, 0.02097497321665287, -0.017018353566527367, 0.10809773951768875, -0.13855572044849396, -0.018670624122023582, 0.06328251957893372, 0.07232730835676193, -0.0028869258239865303, 0.06308864802122116, -0.153975248336792, 0.1050168052315712, 0.016289174556732178, 0.06754438579082489, -0.09747608006000519, 0.10138221830129623, -0.006303760688751936, -0.007241528946906328, 0.13875643908977509, 0.010596190579235554, -0.05694379657506943, -0.08987913280725479, -0.10555228590965271, -0.008462639525532722, 0.12933635711669922, -0.15157614648342133, 0.0847775787115097, -0.028662750497460365, -0.043171048164367676, 0.0024383023846894503, -0.1199452206492424, -0.1302652359008789, -0.1875755488872528, 0.058235347270965576, -0.1366453617811203, 0.039557021111249924, -0.10582595318555832, -0.04340389743447304, -0.028466427698731422, 0.2041483372449875, -0.2317875325679779, -0.0682469978928566, -0.1541893482208252, -0.08429346233606339, 0.14446710050106049, -0.04730919376015663, 0.08914490789175034, -0.0013825427740812302, 0.19013537466526031, 0.024473950266838074, -0.02387205697596073, 0.10308998823165894, -0.09543927758932114, -0.19450686872005463, -0.08603953570127487, 0.15582145750522614, 0.13931062817573547, 0.03702725097537041, -0.004593946039676666, 0.029260434210300446, -0.020000332966446877, -0.12535293400287628, 0.025526588782668114, 0.1793687790632248, 0.07859015464782715, 0.023437971249222755, -0.025896867737174034, -0.10993997752666473, -0.06524094194173813, -0.0335373692214489, 0.02718053013086319, 0.18264614045619965, -0.07421271502971649, 0.1900695115327835, 0.13626199960708618, -0.05445687845349312, -0.1955246478319168, 0.018216576427221298, 0.040417760610580444, 0.010847307741641998, 0.03138056397438049, -0.2078717201948166, 0.09027513861656189, 0.0014845491386950016, -0.05172133818268776, 0.141556978225708, -0.174949511885643, -0.1512570083141327, 0.06491631269454956, 0.0364508256316185, -0.19348180294036865, -0.117862768471241, -0.08817066252231598, -0.046907443553209305, -0.17498233914375305, 0.10519181191921234, 0.016932250931859016, 0.009516867808997631, 0.03492651879787445, 0.02640140987932682, 0.011080757714807987, -0.03873949125409126, 0.19461296498775482, -0.02505207620561123, 0.029532426968216896, -0.08079101145267487, -0.06136554479598999, 0.0607450045645237, -0.05577658861875534, 0.07896649837493896, -0.020188091322779655, 0.012835816480219364, -0.1100873053073883, -0.0468425452709198, -0.027396185323596, 0.017321845516562462, -0.09195652604103088, -0.09473495930433273, -0.05146971344947815, 0.09373841434717178, 0.08845265954732895, -0.036603908985853195, -0.04043547809123993, -0.07348548620939255, 0.0325477197766304, 0.17183002829551697, 0.17659065127372742, 0.038550034165382385, -0.08084331452846527, -0.005880105309188366, -0.01188716571778059, 0.04436201974749565, -0.22519725561141968, 0.06208868324756622, 0.04557957127690315, 0.015879612416028976, 0.11362850666046143, -0.018783990293741226, -0.16298477351665497, -0.06594224274158478, 0.06143777072429657, -0.06664001196622849, -0.18599680066108704, 0.0032026967965066433, 0.058006007224321365, -0.1646854728460312, -0.037671029567718506, 0.042260222136974335, -0.0045668939128518105, -0.04300284758210182, 0.01627597212791443, 0.08071378618478775, 0.005054219625890255, 0.07112491130828857, 0.05733523517847061, 0.0842885971069336, -0.10417009145021439, 0.07519911974668503, 0.08007751405239105, -0.08229218423366547, 0.031453702598810196, 0.08910130709409714, -0.061817802488803864, -0.03069761022925377, 0.032593827694654465, 0.07753410935401917, 0.019773589447140694, -0.041717879474163055, 0.008655321784317493, -0.09745000302791595, 0.06339588761329651, 0.09504765272140503, 0.03549657016992569, 0.014742289669811726, 0.034356739372015, 0.04988397657871246, -0.07460241764783859, 0.11766603589057922, 0.022336218506097794, 0.01780087500810623, -0.044981084764003754, -0.05459042266011238, 0.032110098749399185, -0.022974027320742607, -0.010163158178329468, -0.03885438293218613, -0.07015778869390488, -0.018130742013454437, -0.15929651260375977, -0.014899281784892082, -0.04085385054349899, 0.007158880587667227, 0.02551902085542679, -0.03834335505962372, 0.007963370531797409, 0.012195355258882046, -0.07085035741329193, -0.061454467475414276, -0.022903166711330414, 0.09224231541156769, -0.16436699032783508, 0.025155464187264442, 0.08285263180732727, -0.12099926173686981, 0.09775067120790482, 0.021939631551504135, 0.0031351554207503796, 0.028338242322206497, -0.1542527824640274, 0.04096807911992073, -0.024365095421671867, 0.01272035762667656, 0.04409142583608627, -0.22033950686454773, 0.001463581225834787, -0.03818526118993759, -0.05954346805810928, -0.010227864608168602, -0.033079732209444046, -0.11291328817605972, 0.09883669763803482, 0.008058897219598293, -0.08219768106937408, -0.030809206888079643, 0.03451729565858841, 0.08243680745363235, -0.02608415111899376, 0.15152283012866974, 0.0016822130419313908, 0.07172226905822754, -0.17519205808639526, -0.021702464669942856, -0.011611736379563808, 0.02207101881504059, -0.014536668546497822, -0.015496513806283474, 0.042471300810575485, -0.02421419881284237, 0.19108575582504272, -0.026401294395327568, 0.038726791739463806, 0.06405707448720932, 0.01593620702624321, -0.014801506884396076, 0.10957890748977661, 0.05975057929754257, 0.02399693801999092, 0.022115202620625496, 0.007329683285206556, -0.039842452853918076, -0.014149460941553116, -0.19538825750350952, 0.06474217027425766, 0.1377464383840561, 0.08781574666500092, -0.01322576031088829, 0.07683692127466202, -0.10024392604827881, -0.12397097796201706, 0.11215250939130783, -0.06283260136842728, -0.007701667957007885, -0.06531554460525513, 0.13346771895885468, 0.14944057166576385, -0.18992236256599426, 0.06835456937551498, -0.06228158622980118, -0.05332518368959427, -0.11744599789381027, -0.1957325041294098, -0.055616896599531174, -0.056456826627254486, -0.014700124971568584, -0.048795297741889954, 0.07307228446006775, 0.05693497136235237, 0.012962869368493557, 0.003600025549530983, 0.0766802653670311, -0.015357231721282005, 0.0008028073934838176, 0.03077360987663269, 0.06600049883127213, 0.013312965631484985, -0.02929985709488392, 0.020537450909614563, -0.007275243755429983, 0.04005419462919235, 0.06378308683633804, 0.038119763135910034, -0.02801438421010971, 0.01591232419013977, -0.03770609200000763, -0.10940317064523697, 0.0409080907702446, -0.028551526367664337, -0.08112191408872604, 0.13721226155757904, 0.02428387477993965, 0.005870606284588575, -0.02180131897330284, 0.24582624435424805, -0.07231455296278, -0.09001907706260681, -0.1473579704761505, 0.10211005061864853, -0.04095151647925377, 0.06560079753398895, 0.04110138490796089, -0.10732010751962662, 0.013498948886990547, 0.12688814103603363, 0.15896959602832794, -0.044884394854307175, 0.020156091079115868, 0.03252736106514931, 0.003683826420456171, -0.04006262496113777, 0.05253688618540764, 0.0694650411605835, 0.14883354306221008, -0.04907030612230301, 0.08928520232439041, 0.005485867150127888, -0.10256236046552658, -0.03822692111134529, 0.11808354407548904, -0.017866896465420723, 0.018703164532780647, -0.057248231023550034, 0.11889533698558807, -0.059861693531274796, -0.23005777597427368, 0.06317704170942307, -0.0720362737774849, -0.14286935329437256, -0.021647587418556213, 0.07456772774457932, -0.017636949196457863, 0.02658887766301632, 0.07326807081699371, -0.07681973278522491, 0.19899281859397888, 0.038975972682237625, -0.05729197710752487, -0.05658522993326187, 0.0789351835846901, -0.114089734852314, 0.2792985737323761, 0.01164181251078844, 0.04984506592154503, 0.10365619510412216, -0.016686614602804184, -0.13768579065799713, 0.015234606340527534, 0.09244892746210098, -0.09004336595535278, 0.03869183734059334, 0.2132277488708496, -0.002569539239630103, 0.1152428612112999, 0.07714667171239853, -0.07265080511569977, 0.04592108353972435, -0.1130065843462944, -0.0718315914273262, -0.086885966360569, 0.09441597014665604, -0.07240451127290726, 0.14123490452766418, 0.12318195402622223, -0.053516924381256104, 0.010368985123932362, -0.031209774315357208, 0.04651070013642311, 0.007842876948416233, 0.10365527868270874, 0.010769560933113098, -0.18099099397659302, 0.022656621411442757, 0.018202748149633408, 0.10856854915618896, -0.17241089046001434, -0.09672945737838745, 0.04725200682878494, 0.001958663808181882, -0.059874359518289566, 0.1282012164592743, 0.057909298688173294, 0.04923510178923607, -0.043742597103118896, -0.017267800867557526, -0.009560109116137028, 0.13584671914577484, -0.10737434774637222, -0.0021453071385622025 ]
null
null
diffusers
# Textual inversion text2image fine-tuning - janetsw/textual_inversion_cat These are textual inversion adaption weights for runwayml/stable-diffusion-v1-5. You can find some example images in the following.
{"license": "creativeml-openrail-m", "tags": ["stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "diffusers", "textual_inversion"], "base_model": "runwayml/stable-diffusion-v1-5", "inference": true}
text-to-image
janetsw/textual_inversion_cat
[ "diffusers", "tensorboard", "safetensors", "stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "textual_inversion", "base_model:runwayml/stable-diffusion-v1-5", "license:creativeml-openrail-m", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
2024-02-15T02:28:53+00:00
[]
[]
TAGS #diffusers #tensorboard #safetensors #stable-diffusion #stable-diffusion-diffusers #text-to-image #textual_inversion #base_model-runwayml/stable-diffusion-v1-5 #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us
# Textual inversion text2image fine-tuning - janetsw/textual_inversion_cat These are textual inversion adaption weights for runwayml/stable-diffusion-v1-5. You can find some example images in the following.
[ "# Textual inversion text2image fine-tuning - janetsw/textual_inversion_cat\nThese are textual inversion adaption weights for runwayml/stable-diffusion-v1-5. You can find some example images in the following." ]
[ "TAGS\n#diffusers #tensorboard #safetensors #stable-diffusion #stable-diffusion-diffusers #text-to-image #textual_inversion #base_model-runwayml/stable-diffusion-v1-5 #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n", "# Textual inversion text2image fine-tuning - janetsw/textual_inversion_cat\nThese are textual inversion adaption weights for runwayml/stable-diffusion-v1-5. You can find some example images in the following." ]
[ 101, 61 ]
[ "passage: TAGS\n#diffusers #tensorboard #safetensors #stable-diffusion #stable-diffusion-diffusers #text-to-image #textual_inversion #base_model-runwayml/stable-diffusion-v1-5 #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n# Textual inversion text2image fine-tuning - janetsw/textual_inversion_cat\nThese are textual inversion adaption weights for runwayml/stable-diffusion-v1-5. You can find some example images in the following." ]
[ -0.11650994420051575, -0.04112287983298302, -0.0033067092299461365, 0.013822279870510101, 0.07903893291950226, -0.005528863053768873, 0.14665751159191132, 0.07733993977308273, -0.048028357326984406, 0.05199899524450302, 0.07501217722892761, 0.045700665563344955, -0.000049643309466773644, 0.1181027889251709, -0.028794966638088226, -0.18477720022201538, -0.002490448299795389, 0.004953967407345772, -0.1330234259366989, 0.06947023421525955, 0.08080638945102692, -0.03582971170544624, 0.0725679025053978, 0.030992312356829643, -0.09251996874809265, 0.057171374559402466, 0.038175854831933975, -0.05178484693169594, 0.085059754550457, 0.04214393347501755, 0.06685858964920044, 0.12833310663700104, 0.05982669070363045, -0.10927867889404297, 0.03489289805293083, 0.06722613424062729, -0.032548774033784866, 0.06718602031469345, 0.016784027218818665, -0.06061243638396263, 0.12813717126846313, -0.06569229066371918, 0.02253301814198494, 0.052835848182439804, -0.0217064768075943, -0.11868646740913391, -0.017046138644218445, 0.016188133507966995, 0.08209756761789322, 0.011071880348026752, 0.02750258333981037, 0.04086821898818016, -0.040577348321676254, 0.07282496243715286, 0.23329298198223114, -0.2795533835887909, -0.00910171214491129, 0.11434067040681839, 0.06226359307765961, 0.06872617453336716, -0.05916328728199005, 0.09137438237667084, 0.035473987460136414, -0.04037008062005043, 0.11213849484920502, -0.040914732962846756, 0.11390581727027893, -0.028791703283786774, -0.11377675831317902, 0.057772159576416016, 0.13222669064998627, -0.011191554367542267, -0.031709399074316025, -0.2042875587940216, -0.09275563061237335, 0.09347435086965561, -0.037909794598817825, -0.020489074289798737, -0.02526501566171646, 0.0235764067620039, 0.0019567508716136217, -0.054078251123428345, -0.1125982329249382, -0.05852147191762924, -0.042811620980501175, 0.08788785338401794, -0.02452508546411991, 0.005302960053086281, -0.022056221961975098, 0.12707608938217163, -0.09304628521203995, -0.14235612750053406, 0.06072784215211868, -0.06770355999469757, 0.015628980472683907, 0.05399464815855026, -0.028144633397459984, -0.2461709827184677, 0.03958786651492119, 0.05448324233293533, 0.13421083986759186, -0.011065584607422352, -0.05231243744492531, 0.09447977691888809, -0.06113320216536522, -0.043167006224393845, -0.07522263377904892, -0.06263391673564911, 0.029045064002275467, 0.05858626589179039, 0.08025778084993362, -0.027287134900689125, -0.07633309066295624, -0.05037056654691696, -0.035711873322725296, 0.049097590148448944, -0.025226593017578125, 0.06820416450500488, -0.06222890317440033, -0.014172286726534367, 0.18871928751468658, -0.07213065773248672, 0.02512243203818798, 0.0114221116527915, 0.022276559844613075, 0.09304140508174896, 0.16740921139717102, 0.010627489537000656, 0.024950142949819565, 0.08795250207185745, -0.07011586427688599, 0.01044582948088646, -0.005244072526693344, -0.10614679008722305, -0.02511717565357685, -0.1497243195772171, -0.03895779326558113, -0.12317254394292831, -0.12674067914485931, 0.05665726214647293, 0.036839086562395096, -0.0034245597198605537, 0.08326920121908188, 0.014059046283364296, -0.042595770210027695, 0.04796841740608215, 0.015556358732283115, -0.015271017327904701, -0.02758382260799408, 0.039839211851358414, -0.005909852217882872, 0.1292053908109665, -0.0003735220234375447, -0.009284268133342266, -0.07062531262636185, 0.0029593452345579863, -0.23998212814331055, 0.025664445012807846, -0.11348722130060196, 0.03478240221738815, -0.07515586167573929, -0.0303729847073555, -0.07341426610946655, 0.03462846204638481, 0.02978881634771824, 0.16837915778160095, -0.2377365529537201, -0.05098239332437515, 0.13435637950897217, -0.17191357910633087, -0.020910847932100296, 0.06112586334347725, 0.01658448949456215, 0.06689973175525665, 0.027340110391378403, 0.10327816009521484, 0.1120423674583435, -0.2299707680940628, 0.11261577159166336, 0.016007065773010254, -0.08170478045940399, -0.012941363267600536, 0.012543206103146076, 0.0017875948688015342, 0.029275527223944664, 0.01377690676599741, -0.0708480030298233, 0.09513426572084427, -0.0026395104359835386, -0.0017375651514157653, -0.06338401883840561, -0.03875948488712311, 0.12976816296577454, 0.04361991584300995, 0.02991064451634884, -0.041584618389606476, -0.07731273770332336, -0.0028708658646792173, 0.025737348943948746, -0.08344880491495132, 0.043814923614263535, -0.02229606732726097, 0.08062030375003815, -0.09897416830062866, -0.02258346788585186, -0.13241459429264069, -0.006082573439925909, -0.022777419537305832, 0.15909762680530548, -0.033808909356594086, 0.11872528493404388, 0.11277754604816437, 0.040053706616163254, -0.012617395259439945, -0.01290454063564539, 0.11446427553892136, 0.03999868780374527, -0.048830434679985046, -0.18320447206497192, 0.0444599874317646, -0.10076495260000229, -0.003290922148153186, -0.22353248298168182, 0.08737310767173767, 0.07420596480369568, 0.20390191674232483, 0.13024963438510895, -0.026763448491692543, 0.05770396068692207, -0.04957878589630127, -0.04988609626889229, -0.05274669826030731, 0.03278357908129692, -0.003802645020186901, -0.02827617898583412, 0.18978723883628845, -0.13377170264720917, 0.2595618665218353, 0.1135682612657547, 0.018330873921513557, -0.06428546458482742, -0.0648634061217308, -0.005935565568506718, 0.0036638902965933084, -0.036840472370386124, -0.03677805885672569, -0.01471850834786892, 0.008233639411628246, 0.14994344115257263, -0.047132622450590134, 0.0030132546089589596, 0.05453195422887802, -0.03150065988302231, -0.019356323406100273, 0.06134969741106033, 0.04577385634183884, -0.08948983997106552, 0.03289755433797836, 0.16314749419689178, -0.01704256422817707, 0.1123654842376709, -0.014983237721025944, -0.08704149723052979, 0.00548508670181036, 0.044288139790296555, 0.049383047968149185, 0.14095038175582886, -0.009667760692536831, -0.03439934924244881, 0.03552773967385292, -0.03467896580696106, 0.030965790152549744, -0.12559007108211517, -0.008027050644159317, 0.08087341487407684, -0.019780287519097328, 0.09884323924779892, 0.06849609315395355, -0.05492902547121048, 0.09425301104784012, -0.13811860978603363, -0.0572899729013443, -0.023008404299616814, -0.01663886569440365, -0.11881688237190247, 0.11931164562702179, -0.09265869855880737, -0.14966616034507751, -0.14277562499046326, 0.004507826175540686, -0.0051485151052474976, 0.005820761900395155, 0.04611106961965561, -0.015065588057041168, -0.09965439140796661, -0.12476794421672821, 0.04748782888054848, 0.07508385181427002, 0.03787345066666603, -0.007197114173322916, -0.027592185884714127, 0.034559279680252075, -0.13659264147281647, 0.025014648213982582, -0.021365420892834663, 0.047558754682540894, 0.04334999620914459, 0.017023220658302307, 0.1076916754245758, 0.08432526886463165, -0.01856030523777008, -0.01571047119796276, 0.004257636610418558, 0.08837387710809708, -0.0132986418902874, 0.12315529584884644, 0.14239278435707092, -0.022002769634127617, 0.061248525977134705, 0.09451889991760254, 0.06464625895023346, -0.014526695013046265, 0.018144449219107628, -0.02981032244861126, -0.08036835491657257, -0.06522558629512787, -0.08803965896368027, -0.0692964717745781, -0.007239638827741146, 0.0555756650865078, 0.024521896615624428, 0.08230825513601303, 0.08617793023586273, 0.014412470161914825, -0.013377261348068714, 0.0840502381324768, 0.07461501657962799, 0.10501651465892792, -0.014287476427853107, 0.09343680739402771, -0.0837395191192627, -0.04549872502684593, 0.07937750965356827, -0.042871516197919846, 0.17823487520217896, -0.0704914852976799, 0.045163486152887344, 0.05288147181272507, 0.052948959171772, 0.11562491953372955, 0.16990940272808075, -0.07857558876276016, -0.05645450949668884, -0.010900555178523064, -0.09904444217681885, 0.05889349803328514, 0.05437154695391655, -0.04663471877574921, -0.016376404091715813, -0.02718927524983883, 0.028947290033102036, 0.004456130787730217, -0.03411807864904404, 0.07394935190677643, -0.2202797383069992, 0.017485646530985832, 0.0030115318950265646, 0.011641990393400192, -0.014936295337975025, 0.029558241367340088, 0.20414584875106812, -0.015698950737714767, 0.06706663966178894, -0.08984372019767761, 0.050173476338386536, 0.034522656351327896, -0.0037884367629885674, -0.060090139508247375, 0.06858116388320923, -0.042319152504205704, 0.012402627617120743, -0.16335739195346832, 0.09351766854524612, 0.005801527760922909, 0.005845126695930958, -0.038862258195877075, 0.031598322093486786, 0.06320691108703613, 0.17431707680225372, 0.12582379579544067, -0.03562195226550102, 0.012674344703555107, 0.014409007504582405, -0.09894002228975296, -0.025221962481737137, 0.05958952382206917, -0.025541141629219055, 0.02645142748951912, 0.026293521746993065, -0.06088685244321823, 0.05805198848247528, 0.03228546306490898, -0.25449565052986145, -0.18885891139507294, 0.02472318708896637, 0.027943409979343414, -0.12783947587013245, -0.09817744791507721, -0.08867757022380829, -0.06347435712814331, 0.21469493210315704, -0.1271691471338272, -0.05541045963764191, -0.12852279841899872, -0.05752392113208771, 0.05652262642979622, -0.023483334109187126, 0.06897620856761932, 0.004055310040712357, 0.08278519660234451, -0.09783189743757248, -0.16375510394573212, 0.13389235734939575, -0.09492460638284683, -0.10271235555410385, -0.12469707429409027, 0.1058105081319809, -0.027451107278466225, -0.02473551034927368, 0.008155515417456627, 0.008189377374947071, 0.03684856370091438, -0.06392060220241547, 0.04399209842085838, 0.10471909493207932, -0.017625438049435616, -0.032810308039188385, -0.06455520540475845, -0.1999150961637497, 0.0031639316584914923, 0.03985040262341499, 0.1536518782377243, 0.09399090707302094, -0.11155295372009277, 0.14186590909957886, 0.09148122370243073, -0.02878520078957081, -0.19925637543201447, -0.0006914203404448926, -0.051272202283144, 0.021617773920297623, 0.060594189912080765, -0.08490566164255142, 0.15968750417232513, 0.028355486690998077, -0.020960204303264618, 0.20288395881652832, -0.3221111595630646, -0.13238824903964996, 0.07833544164896011, 0.1324746310710907, 0.1329994797706604, -0.12401135265827179, -0.07355473935604095, 0.016580522060394287, -0.2010190337896347, 0.05365941673517227, -0.07427068799734116, 0.015779312700033188, -0.03450101986527443, -0.06276556104421616, -0.00503910006955266, -0.058988306671381, 0.07739447802305222, -0.04554693400859833, 0.0543501116335392, -0.09985395520925522, 0.03987932577729225, 0.1119893416762352, -0.037892796099185944, 0.029811955988407135, -0.21851789951324463, 0.047404736280441284, -0.18896277248859406, -0.015388279221951962, 0.024249373003840446, 0.05822576954960823, -0.023170875385403633, -0.05082316696643829, -0.0335988812148571, -0.024876603856682777, 0.021662697196006775, -0.03095431812107563, 0.03665941208600998, -0.024411244317889214, 0.07574854791164398, 0.18153071403503418, 0.07684243470430374, -0.049477409571409225, -0.10092779994010925, -0.05037236958742142, -0.015827834606170654, 0.10516433417797089, -0.18959090113639832, 0.014033001847565174, 0.0752372145652771, 0.07499566674232483, 0.04183102026581764, 0.05271885171532631, -0.030421145260334015, 0.039291538298130035, 0.14506207406520844, -0.11244750767946243, 0.039886705577373505, -0.028739148750901222, 0.03751906007528305, 0.07897535711526871, 0.12020109593868256, 0.13449698686599731, -0.048256341367959976, 0.0313747376203537, 0.012895859777927399, 0.03662005066871643, -0.02621198631823063, 0.11822734028100967, 0.06452009826898575, 0.048952434211969376, -0.052710872143507004, 0.05289968103170395, -0.03676078841090202, -0.12605303525924683, -0.041818197816610336, 0.08567577600479126, -0.12387621402740479, -0.05085878446698189, 0.04895499348640442, 0.19408094882965088, -0.030880413949489594, 0.010080207139253616, -0.05209672451019287, -0.10565213114023209, -0.011418318375945091, 0.18436871469020844, 0.04323146492242813, -0.04163186997175217, -0.07766971737146378, -0.04582047462463379, -0.03575759753584862, 0.09219494462013245, 0.10882568359375, 0.09608080983161926, -0.15445007383823395, -0.04637615382671356, -0.029434934258461, -0.07028622180223465, -0.08172832429409027, 0.02262749709188938, -0.026325251907110214, -0.06529492139816284, -0.050945643335580826, 0.05344107747077942, -0.1164759173989296, -0.03757508471608162, -0.025178242474794388, -0.04684631898999214, 0.000859356252476573, 0.02885124273598194, -0.03805966675281525, 0.004584645852446556, -0.022849729284644127, 0.007355490233749151, -0.09557431936264038, -0.021304430440068245, -0.028505316004157066, -0.11525293439626694, 0.06452269107103348, -0.023261679336428642, -0.06893844157457352, -0.014037453569471836, -0.1966734081506729, -0.05580149590969086, 0.09143876284360886, -0.02232276275753975, 0.00003967407246818766, 0.03475480526685715, 0.03577182814478874, 0.004371165297925472, 0.0164451003074646, -0.04352550208568573, 0.05523144453763962, -0.07410644739866257, 0.0737767368555069, -0.07771214842796326, 0.019502051174640656, -0.06647132337093353, 0.01803230307996273, 0.15229272842407227, 0.03548242524266243, 0.16580082476139069, -0.111306332051754, 0.07165107131004333, -0.12265137583017349, 0.005336964502930641, 0.04392823949456215, -0.07917636632919312, 0.08698076009750366, 0.019758863374590874, -0.016925927251577377, -0.06332498788833618, 0.1670248955488205, -0.03042704239487648, -0.15107418596744537, 0.020175354555249214, -0.00979746226221323, 0.06233194097876549, 0.05459929630160332, 0.2309960275888443, 0.028999846428632736, 0.03355249762535095, -0.12950552999973297, 0.08102209866046906, 0.09334484487771988, -0.031209342181682587, 0.10427548736333847, 0.1176709532737732, -0.10023511201143265, 0.12660899758338928, 0.011174464598298073, -0.03637632355093956, -0.06625664234161377, 0.09883630275726318, -0.06531543284654617, 0.1360759437084198, -0.01797243021428585, -0.0008625364862382412, 0.2307599037885666, -0.03292340785264969, -0.034581929445266724, 0.08261820673942566, -0.022399958223104477, -0.05466804280877113, -0.20360608398914337, -0.03682549297809601, -0.18202021718025208, -0.003943464253097773, -0.038872379809617996, -0.0009651324362494051, 0.016673726961016655, 0.05684688314795494, 0.07882462441921234, 0.1220003068447113, 0.01422656886279583, -0.05830354988574982, 0.0958084911108017, 0.00405182084068656, -0.06480099260807037, 0.04922693967819214, 0.04800250381231308, 0.027492042630910873, 0.0023536663502454758, -0.025665801018476486, 0.0420183464884758, 0.048828091472387314, 0.028237026184797287, -0.023320559412240982, -0.043999165296554565, -0.018019387498497963, 0.003793340176343918, -0.04185221716761589, 0.17466332018375397, 0.06297262012958527, -0.014828470535576344, -0.028722429648041725, 0.12422274053096771, -0.05153033137321472, -0.12166552990674973, -0.09704237431287766, 0.03417186439037323, -0.04260888695716858, 0.0684586688876152, -0.06761986017227173, -0.12166725099086761, 0.03940432518720627, 0.21198725700378418, 0.2370786964893341, -0.017571967095136642, 0.05625651404261589, -0.018310414627194405, -0.014552551321685314, 0.015898989513516426, 0.03187105059623718, 0.030747121199965477, 0.27122750878334045, -0.02249602973461151, -0.09191064536571503, -0.06968745589256287, -0.07595665752887726, -0.11001533269882202, 0.000729462830349803, 0.021241143345832825, -0.05762775242328644, -0.04874037951231003, 0.08575087040662766, 0.0362432524561882, -0.09259587526321411, 0.028775490820407867, -0.09358510375022888, -0.0070711602456867695, -0.05602116137742996, 0.15151894092559814, -0.03081582859158516, -0.020169392228126526, -0.030318202450871468, -0.02167653664946556, 0.041716303676366806, 0.009700906462967396, -0.1165509968996048, 0.07204116135835648, -0.07954688370227814, -0.2584337890148163, 0.07899300754070282, -0.023210059851408005, 0.017059991136193275, 0.038234349340200424, 0.024619918316602707, -0.08066151291131973, 0.030026881024241447, -0.0013645068975165486, -0.061009567230939865, -0.04193279147148132, 0.1194302886724472, 0.016692321747541428, 0.0432901494204998, 0.026495736092329025, -0.16599583625793457, 0.06538774073123932, 0.019888339564204216, -0.09937460720539093, -0.07490114122629166, -0.0035816561430692673, -0.04246171563863754, 0.09515956789255142, 0.010530713945627213, 0.0006043332978151739, 0.019486330449581146, -0.01046405453234911, 0.013644523918628693, 0.023933934047818184, -0.009181167930364609, 0.037029288709163666, -0.11198889464139938, 0.005611301399767399, 0.006759509909898043, 0.051891569048166275, -0.20346137881278992, -0.058046210557222366, -0.11491620540618896, -0.022814318537712097, -0.0442802794277668, 0.0636604055762291, 0.2524619400501251, 0.044366467744112015, -0.005976727232336998, -0.2593748867511749, 0.06653732061386108, 0.08156564086675644, -0.09216499328613281, -0.03473460674285889 ]
null
null
null
# Lora of oklahoma/オクラホマ/俄克拉荷马 (Azur Lane) ## What Is This? This is the LoRA model of waifu oklahoma/オクラホマ/俄克拉荷马 (Azur Lane). ## How Is It Trained? * This model is trained with [HCP-Diffusion](https://github.com/7eu7d7/HCP-Diffusion). * The [auto-training framework](https://github.com/deepghs/cyberharem) is maintained by [DeepGHS Team](https://huggingface.co/deepghs). * The base model used for training is [deepghs/animefull-latest](https://huggingface.co/deepghs/animefull-latest). * Dataset used for training is the `stage3-p480-800` in [CyberHarem/oklahoma_azurlane](https://huggingface.co/datasets/CyberHarem/oklahoma_azurlane), which contains 65 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 800 steps, 40 checkpoints were saved and evaluated. * **Trigger word is `oklahoma_azurlane`.** * Pruned core tags for this waifu are `ahoge, blue_eyes, breasts, hair_between_eyes, blonde_hair, short_hair, bangs, large_breasts`. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. ## How to Use It? ### If You Are Using A1111 WebUI v1.7+ **Just use it like the classic LoRA**. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 500, you need to download [`500/oklahoma_azurlane.pt`](https://huggingface.co/CyberHarem/oklahoma_azurlane/resolve/main/500/oklahoma_azurlane.pt) as the embedding and [`500/oklahoma_azurlane.safetensors`](https://huggingface.co/CyberHarem/oklahoma_azurlane/resolve/main/500/oklahoma_azurlane.safetensors) for loading Lora. By using both files together, you can generate images for the desired characters. ## Which Step Should I Use? We selected 5 good steps for you to choose. The best one is step 500. 1440 images (1.43 GiB) were generated for auto-testing. ![Metrics Plot](metrics_plot.png) The base model used for generating preview images is [Meina/MeinaMix_V11](https://huggingface.co/Meina/MeinaMix_V11). Here are the preview of the recommended steps: | Step | Epoch | CCIP | AI Corrupt | Bikini Plus | Score | Download | pattern_0 | portrait_0 | portrait_1 | portrait_2 | full_body_0 | full_body_1 | profile_0 | profile_1 | free_0 | free_1 | shorts | maid_0 | maid_1 | miko | yukata | suit | china | bikini_0 | bikini_1 | bikini_2 | sit | squat | kneel | jump | crossed_arms | angry | smile | cry | grin | n_lie_0 | n_lie_1 | n_stand_0 | n_stand_1 | n_stand_2 | n_sex_0 | n_sex_1 | |-------:|--------:|:----------|:-------------|:--------------|:----------|:-------------------------------------------------------------------------------------------------------|:-----------------------------------------|:-------------------------------------------|:-------------------------------------------|:-------------------------------------------|:---------------------------------------------|:---------------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-------------------------------|:-----------------------------------|:-------------------------------|:---------------------------------|:---------------------------------------|:---------------------------------------|:---------------------------------------|:-----------------------------|:---------------------------------|:---------------------------------|:-------------------------------|:-----------------------------------------------|:---------------------------------|:---------------------------------|:-----------------------------|:-------------------------------|:-------------------------------------|:-------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------------|:-------------------------------------|:-------------------------------------| | 500 | 31 | **0.774** | 0.954 | 0.849 | **0.782** | [Download](https://huggingface.co/CyberHarem/oklahoma_azurlane/resolve/main/500/oklahoma_azurlane.zip) | ![pattern_0](500/previews/pattern_0.png) | ![portrait_0](500/previews/portrait_0.png) | ![portrait_1](500/previews/portrait_1.png) | ![portrait_2](500/previews/portrait_2.png) | ![full_body_0](500/previews/full_body_0.png) | ![full_body_1](500/previews/full_body_1.png) | ![profile_0](500/previews/profile_0.png) | ![profile_1](500/previews/profile_1.png) | ![free_0](500/previews/free_0.png) | ![free_1](500/previews/free_1.png) | ![shorts](500/previews/shorts.png) | ![maid_0](500/previews/maid_0.png) | ![maid_1](500/previews/maid_1.png) | ![miko](500/previews/miko.png) | ![yukata](500/previews/yukata.png) | ![suit](500/previews/suit.png) | ![china](500/previews/china.png) | ![bikini_0](500/previews/bikini_0.png) | ![bikini_1](500/previews/bikini_1.png) | ![bikini_2](500/previews/bikini_2.png) | ![sit](500/previews/sit.png) | ![squat](500/previews/squat.png) | ![kneel](500/previews/kneel.png) | ![jump](500/previews/jump.png) | ![crossed_arms](500/previews/crossed_arms.png) | ![angry](500/previews/angry.png) | ![smile](500/previews/smile.png) | ![cry](500/previews/cry.png) | ![grin](500/previews/grin.png) | ![n_lie_0](500/previews/n_lie_0.png) | ![n_lie_1](500/previews/n_lie_1.png) | ![n_stand_0](500/previews/n_stand_0.png) | ![n_stand_1](500/previews/n_stand_1.png) | ![n_stand_2](500/previews/n_stand_2.png) | ![n_sex_0](500/previews/n_sex_0.png) | ![n_sex_1](500/previews/n_sex_1.png) | | 680 | 42 | 0.690 | 0.970 | **0.858** | 0.716 | [Download](https://huggingface.co/CyberHarem/oklahoma_azurlane/resolve/main/680/oklahoma_azurlane.zip) | ![pattern_0](680/previews/pattern_0.png) | ![portrait_0](680/previews/portrait_0.png) | ![portrait_1](680/previews/portrait_1.png) | ![portrait_2](680/previews/portrait_2.png) | ![full_body_0](680/previews/full_body_0.png) | ![full_body_1](680/previews/full_body_1.png) | ![profile_0](680/previews/profile_0.png) | ![profile_1](680/previews/profile_1.png) | ![free_0](680/previews/free_0.png) | ![free_1](680/previews/free_1.png) | ![shorts](680/previews/shorts.png) | ![maid_0](680/previews/maid_0.png) | ![maid_1](680/previews/maid_1.png) | ![miko](680/previews/miko.png) | ![yukata](680/previews/yukata.png) | ![suit](680/previews/suit.png) | ![china](680/previews/china.png) | ![bikini_0](680/previews/bikini_0.png) | ![bikini_1](680/previews/bikini_1.png) | ![bikini_2](680/previews/bikini_2.png) | ![sit](680/previews/sit.png) | ![squat](680/previews/squat.png) | ![kneel](680/previews/kneel.png) | ![jump](680/previews/jump.png) | ![crossed_arms](680/previews/crossed_arms.png) | ![angry](680/previews/angry.png) | ![smile](680/previews/smile.png) | ![cry](680/previews/cry.png) | ![grin](680/previews/grin.png) | ![n_lie_0](680/previews/n_lie_0.png) | ![n_lie_1](680/previews/n_lie_1.png) | ![n_stand_0](680/previews/n_stand_0.png) | ![n_stand_1](680/previews/n_stand_1.png) | ![n_stand_2](680/previews/n_stand_2.png) | ![n_sex_0](680/previews/n_sex_0.png) | ![n_sex_1](680/previews/n_sex_1.png) | | 300 | 19 | 0.688 | **0.977** | 0.858 | 0.713 | [Download](https://huggingface.co/CyberHarem/oklahoma_azurlane/resolve/main/300/oklahoma_azurlane.zip) | ![pattern_0](300/previews/pattern_0.png) | ![portrait_0](300/previews/portrait_0.png) | ![portrait_1](300/previews/portrait_1.png) | ![portrait_2](300/previews/portrait_2.png) | ![full_body_0](300/previews/full_body_0.png) | ![full_body_1](300/previews/full_body_1.png) | ![profile_0](300/previews/profile_0.png) | ![profile_1](300/previews/profile_1.png) | ![free_0](300/previews/free_0.png) | ![free_1](300/previews/free_1.png) | ![shorts](300/previews/shorts.png) | ![maid_0](300/previews/maid_0.png) | ![maid_1](300/previews/maid_1.png) | ![miko](300/previews/miko.png) | ![yukata](300/previews/yukata.png) | ![suit](300/previews/suit.png) | ![china](300/previews/china.png) | ![bikini_0](300/previews/bikini_0.png) | ![bikini_1](300/previews/bikini_1.png) | ![bikini_2](300/previews/bikini_2.png) | ![sit](300/previews/sit.png) | ![squat](300/previews/squat.png) | ![kneel](300/previews/kneel.png) | ![jump](300/previews/jump.png) | ![crossed_arms](300/previews/crossed_arms.png) | ![angry](300/previews/angry.png) | ![smile](300/previews/smile.png) | ![cry](300/previews/cry.png) | ![grin](300/previews/grin.png) | ![n_lie_0](300/previews/n_lie_0.png) | ![n_lie_1](300/previews/n_lie_1.png) | ![n_stand_0](300/previews/n_stand_0.png) | ![n_stand_1](300/previews/n_stand_1.png) | ![n_stand_2](300/previews/n_stand_2.png) | ![n_sex_0](300/previews/n_sex_0.png) | ![n_sex_1](300/previews/n_sex_1.png) | | 440 | 28 | 0.683 | 0.960 | 0.852 | 0.700 | [Download](https://huggingface.co/CyberHarem/oklahoma_azurlane/resolve/main/440/oklahoma_azurlane.zip) | ![pattern_0](440/previews/pattern_0.png) | ![portrait_0](440/previews/portrait_0.png) | ![portrait_1](440/previews/portrait_1.png) | ![portrait_2](440/previews/portrait_2.png) | ![full_body_0](440/previews/full_body_0.png) | ![full_body_1](440/previews/full_body_1.png) | ![profile_0](440/previews/profile_0.png) | ![profile_1](440/previews/profile_1.png) | ![free_0](440/previews/free_0.png) | ![free_1](440/previews/free_1.png) | ![shorts](440/previews/shorts.png) | ![maid_0](440/previews/maid_0.png) | ![maid_1](440/previews/maid_1.png) | ![miko](440/previews/miko.png) | ![yukata](440/previews/yukata.png) | ![suit](440/previews/suit.png) | ![china](440/previews/china.png) | ![bikini_0](440/previews/bikini_0.png) | ![bikini_1](440/previews/bikini_1.png) | ![bikini_2](440/previews/bikini_2.png) | ![sit](440/previews/sit.png) | ![squat](440/previews/squat.png) | ![kneel](440/previews/kneel.png) | ![jump](440/previews/jump.png) | ![crossed_arms](440/previews/crossed_arms.png) | ![angry](440/previews/angry.png) | ![smile](440/previews/smile.png) | ![cry](440/previews/cry.png) | ![grin](440/previews/grin.png) | ![n_lie_0](440/previews/n_lie_0.png) | ![n_lie_1](440/previews/n_lie_1.png) | ![n_stand_0](440/previews/n_stand_0.png) | ![n_stand_1](440/previews/n_stand_1.png) | ![n_stand_2](440/previews/n_stand_2.png) | ![n_sex_0](440/previews/n_sex_0.png) | ![n_sex_1](440/previews/n_sex_1.png) | | 700 | 44 | 0.676 | 0.953 | 0.856 | 0.699 | [Download](https://huggingface.co/CyberHarem/oklahoma_azurlane/resolve/main/700/oklahoma_azurlane.zip) | ![pattern_0](700/previews/pattern_0.png) | ![portrait_0](700/previews/portrait_0.png) | ![portrait_1](700/previews/portrait_1.png) | ![portrait_2](700/previews/portrait_2.png) | ![full_body_0](700/previews/full_body_0.png) | ![full_body_1](700/previews/full_body_1.png) | ![profile_0](700/previews/profile_0.png) | ![profile_1](700/previews/profile_1.png) | ![free_0](700/previews/free_0.png) | ![free_1](700/previews/free_1.png) | ![shorts](700/previews/shorts.png) | ![maid_0](700/previews/maid_0.png) | ![maid_1](700/previews/maid_1.png) | ![miko](700/previews/miko.png) | ![yukata](700/previews/yukata.png) | ![suit](700/previews/suit.png) | ![china](700/previews/china.png) | ![bikini_0](700/previews/bikini_0.png) | ![bikini_1](700/previews/bikini_1.png) | ![bikini_2](700/previews/bikini_2.png) | ![sit](700/previews/sit.png) | ![squat](700/previews/squat.png) | ![kneel](700/previews/kneel.png) | ![jump](700/previews/jump.png) | ![crossed_arms](700/previews/crossed_arms.png) | ![angry](700/previews/angry.png) | ![smile](700/previews/smile.png) | ![cry](700/previews/cry.png) | ![grin](700/previews/grin.png) | ![n_lie_0](700/previews/n_lie_0.png) | ![n_lie_1](700/previews/n_lie_1.png) | ![n_stand_0](700/previews/n_stand_0.png) | ![n_stand_1](700/previews/n_stand_1.png) | ![n_stand_2](700/previews/n_stand_2.png) | ![n_sex_0](700/previews/n_sex_0.png) | ![n_sex_1](700/previews/n_sex_1.png) | ## Anything Else? Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. ## All Steps We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * [Steps From 620 to 800](all/0.md) * [Steps From 420 to 600](all/1.md) * [Steps From 220 to 400](all/2.md) * [Steps From 20 to 200](all/3.md)
{"license": "mit", "tags": ["art", "not-for-all-audiences"], "datasets": ["CyberHarem/oklahoma_azurlane"], "pipeline_tag": "text-to-image"}
text-to-image
CyberHarem/oklahoma_azurlane
[ "art", "not-for-all-audiences", "text-to-image", "dataset:CyberHarem/oklahoma_azurlane", "license:mit", "region:us" ]
2024-02-15T02:29:28+00:00
[]
[]
TAGS #art #not-for-all-audiences #text-to-image #dataset-CyberHarem/oklahoma_azurlane #license-mit #region-us
Lora of oklahoma/オクラホマ/俄克拉荷马 (Azur Lane) ======================================== What Is This? ------------- This is the LoRA model of waifu oklahoma/オクラホマ/俄克拉荷马 (Azur Lane). How Is It Trained? ------------------ * This model is trained with HCP-Diffusion. * The auto-training framework is maintained by DeepGHS Team. * The base model used for training is deepghs/animefull-latest. * Dataset used for training is the 'stage3-p480-800' in CyberHarem/oklahoma\_azurlane, which contains 65 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 800 steps, 40 checkpoints were saved and evaluated. * Trigger word is 'oklahoma\_azurlane'. * Pruned core tags for this waifu are 'ahoge, blue\_eyes, breasts, hair\_between\_eyes, blonde\_hair, short\_hair, bangs, large\_breasts'. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. How to Use It? -------------- ### If You Are Using A1111 WebUI v1.7+ Just use it like the classic LoRA. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 500, you need to download '500/oklahoma\_azurlane.pt' as the embedding and '500/oklahoma\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters. Which Step Should I Use? ------------------------ We selected 5 good steps for you to choose. The best one is step 500. 1440 images (1.43 GiB) were generated for auto-testing. !Metrics Plot The base model used for generating preview images is Meina/MeinaMix\_V11. Here are the preview of the recommended steps: Anything Else? -------------- Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. All Steps --------- We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * Steps From 620 to 800 * Steps From 420 to 600 * Steps From 220 to 400 * Steps From 20 to 200
[ "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 500, you need to download '500/oklahoma\\_azurlane.pt' as the embedding and '500/oklahoma\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 500.\n\n\n1440 images (1.43 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 620 to 800\n* Steps From 420 to 600\n* Steps From 220 to 400\n* Steps From 20 to 200" ]
[ "TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/oklahoma_azurlane #license-mit #region-us \n", "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 500, you need to download '500/oklahoma\\_azurlane.pt' as the embedding and '500/oklahoma\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 500.\n\n\n1440 images (1.43 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 620 to 800\n* Steps From 420 to 600\n* Steps From 220 to 400\n* Steps From 20 to 200" ]
[ 45, 38, 467 ]
[ "passage: TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/oklahoma_azurlane #license-mit #region-us \n### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file." ]
[ 0.006692702881991863, 0.004087130073457956, -0.0036481625866144896, 0.07213525474071503, 0.06568913161754608, 0.08926716446876526, 0.2237376719713211, 0.07803715020418167, 0.12912128865718842, -0.06963061541318893, 0.09379828721284866, 0.057027604430913925, 0.013374964706599712, 0.043467357754707336, -0.0417439267039299, -0.1329069435596466, -0.08015776425600052, -0.030533280223608017, 0.02950366400182247, 0.0229696836322546, 0.06529232859611511, 0.0032425543759018183, 0.0866897851228714, -0.04903370887041092, -0.052354808896780014, 0.04691919684410095, -0.04150934889912605, -0.044186949729919434, 0.026289528235793114, 0.06341180205345154, 0.11818256229162216, 0.008086230605840683, 0.0639183297753334, -0.16467612981796265, 0.06935268640518188, -0.022651951760053635, -0.11026131361722946, -0.009574015624821186, 0.017691195011138916, -0.029136136174201965, 0.11737314611673355, 0.026803860440850258, -0.11642935127019882, 0.0379645936191082, -0.13124153017997742, -0.011745767667889595, -0.047863613814115524, 0.06570599228143692, 0.1630731076002121, 0.05975916609168053, 0.028332436457276344, 0.04080989211797714, -0.038823921233415604, 0.08234670013189316, 0.11939732730388641, -0.09527364373207092, -0.06816790252923965, 0.08772905170917511, 0.019570326432585716, 0.12218863517045975, -0.10180636495351791, 0.08764804154634476, 0.06571463495492935, -0.06260471045970917, -0.16164961457252502, -0.1053818091750145, -0.2233305275440216, -0.01120854914188385, 0.02163121849298477, 0.025122614577412605, 0.4065696895122528, 0.06519164890050888, 0.037418432533741, 0.06195403262972832, -0.06811556965112686, 0.03883838653564453, -0.10197582095861435, 0.13973523676395416, 0.04649505391716957, 0.09099551290273666, -0.046245504170656204, -0.1039925068616867, -0.1223868802189827, -0.06749463081359863, -0.08550986647605896, -0.01556003000587225, 0.013162549585103989, 0.10703156888484955, -0.21297664940357208, 0.014078164473176003, -0.05813255533576012, -0.12478671967983246, 0.0223444364964962, -0.08651591837406158, 0.1690269261598587, 0.06188342347741127, -0.016314387321472168, 0.028046952560544014, 0.23954734206199646, 0.1203414797782898, 0.20228368043899536, 0.05973133444786072, -0.09755881130695343, 0.12419334799051285, 0.03414778411388397, -0.09668869525194168, 0.02889551781117916, -0.10158856958150864, 0.13944603502750397, -0.033014971762895584, 0.10520657896995544, -0.05347812548279762, -0.09841611981391907, 0.02278078906238079, -0.10210432857275009, 0.0682387426495552, 0.041339848190546036, -0.0020156719256192446, -0.06378839910030365, 0.05819722265005112, 0.022045329213142395, -0.040614817291498184, -0.004469018429517746, -0.010985679924488068, -0.052588243037462234, 0.01086544431746006, 0.11329983174800873, 0.04188496619462967, 0.058823972940444946, -0.01994420774281025, -0.02247631549835205, 0.009820161387324333, -0.035338085144758224, 0.03233189880847931, 0.03645796328783035, 0.024148762226104736, 0.08904790133237839, -0.1618742197751999, -0.09800654649734497, -0.011151956394314766, 0.049522168934345245, 0.015338589437305927, 0.07445311546325684, -0.0016489652916789055, 0.06539347767829895, 0.013673079200088978, -0.03388666734099388, 0.040764667093753815, -0.10774989426136017, 0.08380033075809479, -0.008138186298310757, 0.09409408271312714, -0.17166312038898468, -0.0005179242580197752, -0.03923787549138069, 0.005472427234053612, 0.06499845534563065, -0.007957624271512032, -0.10023205727338791, 0.14295583963394165, -0.00836175587028265, 0.08771495521068573, -0.11158847063779831, 0.03992585465312004, 0.01761113479733467, 0.09389220923185349, -0.08786003291606903, -0.006120047532021999, 0.10746151208877563, -0.12846168875694275, -0.18071633577346802, 0.07065338641405106, -0.018436919897794724, 0.04360704496502876, 0.041967280209064484, 0.17486980557441711, 0.18967507779598236, -0.204450324177742, -0.02259448729455471, 0.07833144813776016, -0.017611777409911156, -0.09435397386550903, -0.01937372051179409, 0.09039682894945145, 0.015480943024158478, 0.025782138109207153, -0.007931165397167206, 0.12734481692314148, -0.03534221276640892, -0.08459372073411942, -0.02860136143863201, -0.08226529508829117, -0.0904066413640976, 0.0638069435954094, -0.009817449375987053, -0.0568326935172081, 0.015857651829719543, -0.17043866217136383, 0.1568862944841385, 0.03226489946246147, 0.024067005142569542, -0.07472655922174454, 0.0953618660569191, 0.01132167037576437, -0.0012229025596752763, 0.008557346649467945, -0.06330590695142746, -0.11387469619512558, 0.2224525809288025, 0.08274073153734207, 0.09421418607234955, 0.0509939081966877, -0.0541483610868454, -0.06707530468702316, -0.0018780806567519903, 0.023485036566853523, -0.030995763838291168, -0.002853930229321122, -0.08588439226150513, 0.05470447614789009, -0.011897599324584007, 0.04545694217085838, 0.001374058541841805, -0.016525276005268097, 0.0659300908446312, 0.017553605139255524, -0.018671421334147453, 0.08599016815423965, 0.05060445889830589, -0.024558020755648613, -0.06065332517027855, 0.0002433671324979514, 0.07717496901750565, -0.00255896826274693, -0.08876790851354599, 0.03137455880641937, -0.011665122583508492, 0.037213876843452454, 0.19045844674110413, -0.1949349343776703, 0.04722706601023674, 0.005911662708967924, 0.04209480062127113, 0.03357557952404022, 0.030013618990778923, -0.025878239423036575, 0.0442940928041935, -0.0280221626162529, 0.07390522956848145, -0.003873666049912572, 0.05020618811249733, -0.03762796148657799, -0.1288936585187912, -0.017915889620780945, -0.032027266919612885, 0.16619646549224854, -0.1679241955280304, 0.06382684409618378, 0.18888260424137115, -0.09333737939596176, 0.12294434756040573, -0.0046601914800703526, -0.01043892465531826, 0.0051630106754601, 0.011493302881717682, -0.002582333981990814, 0.09625157713890076, -0.07782196998596191, -0.037971995770931244, 0.01781267672777176, -0.09407740831375122, 0.0335608571767807, -0.12367507070302963, -0.1152372807264328, -0.0684879943728447, -0.019154999405145645, -0.036038029938936234, 0.028149623423814774, -0.042793888598680496, 0.07656269520521164, -0.09965431690216064, -0.08037824183702469, -0.019068630412220955, -0.08490816503763199, 0.014993315562605858, 0.012321839109063148, -0.054439298808574677, -0.15605050325393677, -0.11636374145746231, -0.10487156361341476, -0.12497390806674957, -0.008377443999052048, 0.07758811116218567, -0.10892345011234283, -0.03258444368839264, -0.0004956235643476248, -0.06579776853322983, 0.09594159573316574, -0.06712790578603745, 0.02714589238166809, 0.03786946460604668, -0.019153665751218796, -0.17760099470615387, -0.0040347944013774395, -0.062257926911115646, -0.06324195861816406, 0.16994625329971313, -0.14045046269893646, 0.17464156448841095, -0.03964632749557495, 0.05460752174258232, 0.057171378284692764, 0.02992633543908596, 0.09759508073329926, -0.11264053732156754, 0.08122523874044418, 0.17554418742656708, 0.0421445406973362, 0.08706808090209961, 0.1238112524151802, 0.07760842144489288, -0.11415113508701324, 0.04279014840722084, 0.09036856889724731, -0.08545123785734177, -0.0735291913151741, -0.04247434064745903, -0.10856573283672333, -0.06945805996656418, 0.05748920515179634, 0.06787199527025223, 0.05519409850239754, 0.1138487309217453, -0.056338705122470856, -0.004647166468203068, 0.09612427651882172, 0.046258483082056046, 0.06574392318725586, 0.015642723068594933, 0.050051070749759674, -0.1478603333234787, -0.03461775556206703, 0.15172933042049408, 0.20571035146713257, 0.23090673983097076, 0.024227473884820938, 0.061433397233486176, 0.12169548869132996, 0.0887833833694458, 0.10017585009336472, 0.050808072090148926, 0.013202184811234474, 0.010843724943697453, -0.06644432246685028, -0.031290795654058456, 0.018592260777950287, 0.00024294716422446072, -0.0266238022595644, -0.13655993342399597, 0.0959259420633316, -0.0010341478046029806, 0.07270798087120056, 0.1446547955274582, 0.044970620423555374, -0.10895024985074997, 0.16806264221668243, 0.10009481757879257, 0.05498909205198288, -0.06490935385227203, 0.12731024622917175, 0.04464738070964813, -0.0002579526335466653, 0.17331528663635254, 0.030245117843151093, 0.14792928099632263, -0.008638481609523296, -0.08137445896863937, -0.08054346591234207, -0.07350132614374161, -0.005881382618099451, 0.03411662578582764, -0.2191326916217804, 0.0896487906575203, 0.0604303739964962, 0.021804844960570335, -0.00077293865615502, -0.055055294185876846, 0.1952269971370697, 0.14135028421878815, 0.06924816220998764, 0.021876519545912743, -0.012503714300692081, -0.022391876205801964, -0.10482471436262131, 0.061074476689100266, 0.014595448039472103, 0.04289760813117027, -0.038418423384428024, -0.09855707734823227, -0.012196806259453297, -0.00510416692122817, 0.02611609920859337, -0.09166918694972992, -0.10333003103733063, -0.04090040177106857, 0.23472391068935394, -0.045337822288274765, 0.03422531113028526, 0.055548425763845444, 0.015017928555607796, -0.050148170441389084, 0.03150701895356178, -0.022050924599170685, -0.012971458956599236, -0.0481657013297081, -0.0028889651875942945, 0.002897779457271099, -0.04912123084068298, -0.04850861057639122, -0.03667053207755089, -0.09842579811811447, -0.10553104430437088, -0.004902814514935017, -0.04817444831132889, 0.011407994665205479, -0.03000563383102417, 0.026824789121747017, -0.09161262214183807, -0.02540098875761032, 0.01699763536453247, 0.0335480161011219, -0.0845850333571434, -0.1256553679704666, 0.0002555333194322884, -0.024548379704356194, -0.03556660935282707, 0.03941148892045021, -0.07812945544719696, -0.10369167476892471, -0.05260984227061272, -0.034237995743751526, 0.11454319208860397, 0.23266486823558807, -0.0053529744036495686, 0.008094602264463902, 0.15024800598621368, -0.10530199855566025, -0.3159102201461792, -0.16872693598270416, -0.1608974039554596, -0.10550367087125778, 0.04317975044250488, -0.06183928623795509, 0.041150759905576706, 0.08336684107780457, -0.034626808017492294, 0.22448177635669708, -0.1913328319787979, -0.10253766924142838, 0.10064864158630371, 0.08311009407043457, 0.3112192153930664, -0.27136558294296265, 0.016435950994491577, -0.10315602272748947, -0.024800138548016548, 0.006759640295058489, -0.09363816678524017, 0.1163598895072937, 0.024328280240297318, 0.07939831167459488, -0.007010666187852621, -0.00017476548964623362, 0.15731678903102875, -0.07189951092004776, 0.13785912096500397, -0.13653257489204407, -0.09882776439189911, 0.20308862626552582, -0.046454254537820816, -0.0015007267938926816, -0.22551749646663666, -0.03321258723735809, -0.05982472002506256, 0.04024340212345123, 0.0027878449764102697, 0.06719247251749039, -0.008226688019931316, 0.002822352573275566, -0.13225102424621582, -0.029319413006305695, -0.03099038451910019, 0.06107557192444801, 0.22895507514476776, -0.054443832486867905, -0.040589142590761185, 0.018692994490265846, 0.010803352110087872, 0.09402850270271301, 0.0035449988208711147, -0.0720023363828659, -0.03884956240653992, 0.08365615457296371, -0.19413132965564728, 0.04672040045261383, 0.012412887997925282, 0.0031414981931447983, 0.02912435308098793, 0.002764602191746235, 0.03678131103515625, 0.13566230237483978, 0.19145551323890686, 0.002808597171679139, -0.04508770629763603, -0.01780153624713421, 0.0318167470395565, 0.09854205697774887, -0.022922508418560028, 0.09548655152320862, 0.039966877549886703, 0.038836490362882614, 0.005797508638352156, 0.06976225972175598, -0.07408931851387024, -0.09579062461853027, 0.09860596805810928, -0.03279602527618408, -0.07736121863126755, 0.09340912103652954, 0.04401114210486412, 0.07063127309083939, -0.00007655006629647687, 0.04948199912905693, 0.03029063157737255, -0.12540782988071442, 0.026346178725361824, 0.21408438682556152, -0.032108280807733536, -0.06044458970427513, -0.05992816388607025, 0.012083792127668858, -0.11639054864645004, 0.08108318597078323, 0.006943587679415941, -0.029466237872838974, 0.12116339057683945, -0.03543855994939804, -0.021541370078921318, 0.00877327285706997, -0.06078338623046875, 0.02918032743036747, -0.152312234044075, -0.1903902292251587, 0.048028796911239624, 0.0055532376281917095, -0.05834135040640831, -0.09642697125673294, -0.07434288412332535, 0.07120288908481598, -0.1630459427833557, 0.14494770765304565, -0.08815119415521622, 0.07037974148988724, -0.03049355000257492, -0.052541445940732956, -0.10981608927249908, -0.024001585319638252, -0.04858144745230675, -0.012408820912241936, 0.07151483744382858, 0.020272381603717804, -0.11631952226161957, -0.11355572938919067, 0.05891483277082443, -0.0014254992129281163, 0.001000546501018107, 0.022249549627304077, -0.07224767655134201, 0.024434693157672882, -0.2292425036430359, -0.06916221976280212, 0.08997131884098053, 0.033414456993341446, -0.10436353832483292, 0.13363118469715118, 0.04039379954338074, -0.022830909118056297, 0.034874122589826584, 0.003378600813448429, 0.1454102247953415, -0.07227960228919983, 0.02375819720327854, -0.12887001037597656, -0.16470883786678314, -0.04007686302065849, 0.024233026430010796, 0.23389461636543274, 0.08217122405767441, 0.13164857029914856, -0.05038519948720932, 0.016331469640135765, -0.02746153436601162, 0.0698884129524231, 0.018012747168540955, -0.09616870433092117, -0.028434492647647858, -0.17970944941043854, -0.0633813738822937, -0.0586116723716259, 0.15162789821624756, 0.02958889864385128, -0.15834975242614746, -0.007827523164451122, 0.11210906505584717, -0.1620689332485199, -0.015958845615386963, 0.17837472259998322, -0.045588813722133636, 0.023913701996207237, -0.14687511324882507, 0.02512546256184578, 0.0818490982055664, -0.06200915202498436, 0.005282425321638584, 0.11211881786584854, -0.002698090858757496, 0.008162944577634335, 0.022293642163276672, -0.05329760164022446, 0.09175368398427963, -0.07208758592605591, 0.03837888687849045, -0.0038529010489583015, -0.032284580171108246, -0.12252283841371536, 0.19642825424671173, -0.008463453501462936, -0.001098055625334382, -0.0624738447368145, -0.0027094418182969093, -0.08954428881406784, -0.08255801349878311, -0.06437448412179947, -0.11469209939241409, 0.07895464450120926, -0.05737374722957611, 0.007373354863375425, 0.02171393856406212, 0.00459452485665679, -0.0732223317027092, 0.014906231313943863, -0.15413743257522583, -0.055748242884874344, 0.019973736256361008, -0.015595712698996067, -0.03196880593895912, -0.06834243983030319, -0.037533558905124664, 0.02502247877418995, -0.06341508775949478, -0.06401525437831879, 0.06088169664144516, 0.07342492043972015, 0.06911821663379669, -0.1766795814037323, -0.11132880300283432, -0.06983437389135361, 0.03562913089990616, 0.0729433000087738, 0.17083892226219177, 0.0440654456615448, -0.0005674146232195199, 0.04158288240432739, 0.11392254382371902, 0.02314162440598011, -0.07906758785247803, -0.03959902748465538, -0.1486448049545288, -0.14339366555213928, -0.03422936797142029, -0.0539914071559906, -0.029513830319046974, 0.03177621588110924, 0.1982426941394806, 0.16458292305469513, -0.14140747487545013, 0.05165554955601692, -0.08477790653705597, 0.035340193659067154, -0.03075149655342102, 0.15260009467601776, 0.0541004054248333, 0.16007106006145477, -0.03853457048535347, -0.02835242822766304, -0.045488353818655014, 0.02142116241157055, -0.09244811534881592, 0.027460284531116486, -0.026593759655952454, -0.057313431054353714, -0.057478953152894974, 0.08998880535364151, -0.1097152829170227, 0.06131353974342346, 0.19896642863750458, -0.13617941737174988, -0.007373847533017397, -0.038018327206373215, 0.06058940291404724, 0.10877835005521774, 0.011300358921289444, -0.07900355756282806, -0.01949603296816349, -0.007446878124028444, 0.019965477287769318, -0.17817233502864838, -0.11514389514923096, 0.0003785028529819101, -0.14833985269069672, 0.1399853378534317, -0.017173631116747856, 0.012703950516879559, 0.02675873599946499, -0.06564029306173325, -0.005413688253611326, 0.18332046270370483, 0.02325146272778511, -0.03387359529733658, -0.017498698085546494, -0.06884218007326126, -0.092921681702137, 0.06581158936023712, 0.09896359592676163, 0.07521508634090424, -0.00688514718785882, 0.15841321647167206, -0.018110686913132668, -0.037726834416389465, 0.1285899430513382, -0.16903764009475708, 0.09863590449094772, -0.017370857298374176, -0.025000659748911858, -0.06439342349767685, -0.0360243134200573, 0.02847721427679062, 0.0770995169878006, -0.17078514397144318, -0.039114803075790405, 0.059607066214084625, -0.08592666685581207, 0.051528893411159515, 0.041348088532686234, -0.09049802273511887, 0.00853903591632843, -0.11976923048496246, -0.010206938721239567, -0.09900081902742386, 0.04820645973086357, 0.19825592637062073, -0.0328022837638855, 0.013384941965341568, -0.14509250223636627, 0.04888484254479408, -0.023775439709424973, -0.04975837841629982, -0.06960371881723404 ]
null
null
transformers
# ultra0-reshaped1 ultra0-reshaped1 is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [liminerity/ultra0-reshaped](https://huggingface.co/liminerity/ultra0-reshaped) * [eren23/dpo-binarized-NeutrixOmnibe-7b](https://huggingface.co/eren23/dpo-binarized-NeutrixOmnibe-7b) ## 🧩 Configuration ```yaml slices: - sources: - model: liminerity/ultra0-reshaped layer_range: [0, 24] - model: eren23/dpo-binarized-NeutrixOmnibe-7b layer_range: [0, 24] merge_method: slerp base_model: liminerity/ultra0-reshaped1 parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ``` ## 💻 Usage ```python !pip install -qU transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "liminerity/ultra0-reshaped1" messages = [{"role": "user", "content": "What is a large language model?"}] tokenizer = AutoTokenizer.from_pretrained(model) prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ```
{"tags": ["merge", "mergekit", "lazymergekit", "liminerity/ultra0-reshaped", "eren23/dpo-binarized-NeutrixOmnibe-7b"], "base_model": ["liminerity/ultra0-reshaped", "eren23/dpo-binarized-NeutrixOmnibe-7b"]}
text-generation
liminerity/ultra0-reshaped1
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "liminerity/ultra0-reshaped", "eren23/dpo-binarized-NeutrixOmnibe-7b", "conversational", "base_model:liminerity/ultra0-reshaped", "base_model:eren23/dpo-binarized-NeutrixOmnibe-7b", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T02:30:33+00:00
[]
[]
TAGS #transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #liminerity/ultra0-reshaped #eren23/dpo-binarized-NeutrixOmnibe-7b #conversational #base_model-liminerity/ultra0-reshaped #base_model-eren23/dpo-binarized-NeutrixOmnibe-7b #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# ultra0-reshaped1 ultra0-reshaped1 is a merge of the following models using LazyMergekit: * liminerity/ultra0-reshaped * eren23/dpo-binarized-NeutrixOmnibe-7b ## Configuration ## Usage
[ "# ultra0-reshaped1\n\nultra0-reshaped1 is a merge of the following models using LazyMergekit:\n* liminerity/ultra0-reshaped\n* eren23/dpo-binarized-NeutrixOmnibe-7b", "## Configuration", "## Usage" ]
[ "TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #liminerity/ultra0-reshaped #eren23/dpo-binarized-NeutrixOmnibe-7b #conversational #base_model-liminerity/ultra0-reshaped #base_model-eren23/dpo-binarized-NeutrixOmnibe-7b #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# ultra0-reshaped1\n\nultra0-reshaped1 is a merge of the following models using LazyMergekit:\n* liminerity/ultra0-reshaped\n* eren23/dpo-binarized-NeutrixOmnibe-7b", "## Configuration", "## Usage" ]
[ 134, 61, 4, 3 ]
[ "passage: TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #liminerity/ultra0-reshaped #eren23/dpo-binarized-NeutrixOmnibe-7b #conversational #base_model-liminerity/ultra0-reshaped #base_model-eren23/dpo-binarized-NeutrixOmnibe-7b #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# ultra0-reshaped1\n\nultra0-reshaped1 is a merge of the following models using LazyMergekit:\n* liminerity/ultra0-reshaped\n* eren23/dpo-binarized-NeutrixOmnibe-7b## Configuration## Usage" ]
[ -0.01586969755589962, -0.05186586081981659, -0.005776014644652605, 0.007759540807455778, 0.0815962478518486, 0.03855487331748009, 0.14573919773101807, 0.06166788563132286, -0.1195208951830864, 0.030126839876174927, 0.05515863001346588, 0.08656828850507736, -0.014299667440354824, 0.09098206460475922, -0.0736972838640213, -0.24602286517620087, 0.0998581126332283, 0.03417263180017471, -0.08159375190734863, 0.062240395694971085, 0.09201110154390335, -0.05604761466383934, 0.08934951573610306, -0.0023190942592918873, -0.08855544030666351, 0.0275945533066988, -0.002019519917666912, -0.015233374200761318, 0.07709240168333054, 0.11029832810163498, 0.16155067086219788, 0.08919040858745575, -0.010121659375727177, -0.14193902909755707, 0.037659164518117905, 0.04021185263991356, -0.04753587022423744, 0.05840457230806351, 0.021760491654276848, -0.03448710963129997, 0.063435859978199, -0.10052838921546936, 0.019515812397003174, 0.04808272421360016, -0.11659229546785355, -0.11131695657968521, -0.06223335489630699, 0.07980075478553772, 0.06278087198734283, 0.0522947758436203, -0.03335968032479286, 0.07928473502397537, -0.019167887046933174, 0.07166014611721039, 0.19798463582992554, -0.23495882749557495, -0.03697113320231438, 0.07747912406921387, 0.03295489028096199, 0.02115294709801674, -0.0033879168331623077, 0.05953280255198479, 0.006133097689598799, 0.024703560397028923, 0.055839404463768005, -0.10173123329877853, 0.07084985822439194, -0.06008822098374367, -0.13276608288288116, 0.07601939886808395, 0.054931242018938065, 0.01752205565571785, -0.039094265550374985, -0.091131791472435, -0.14015275239944458, 0.035024769604206085, -0.07977230101823807, -0.08761553466320038, 0.050326403230428696, -0.0038294270634651184, -0.03563558682799339, -0.07466775923967361, -0.03101384826004505, -0.03441278636455536, -0.1445896327495575, 0.25723740458488464, 0.0019333282252773643, -0.02128048799932003, -0.056162212044000626, 0.0353257916867733, -0.09949210286140442, -0.09998304396867752, -0.041875097900629044, -0.05561020225286484, 0.040463726967573166, -0.01451929286122322, -0.1161242425441742, -0.1407596617937088, 0.0890916958451271, 0.2100081741809845, -0.09331554919481277, 0.0381765142083168, -0.003938902635127306, 0.04191878065466881, -0.02959982492029667, 0.059261322021484375, -0.06905848532915115, -0.08959158509969711, 0.021310899406671524, 0.06862224638462067, 0.10603078454732895, -0.03333557769656181, -0.1011287271976471, -0.027693569660186768, 0.03392142057418823, -0.0037765561137348413, 0.018218092620372772, 0.1529400795698166, -0.0420568510890007, -0.06614716351032257, 0.15116603672504425, -0.07159631699323654, -0.0042183781042695045, -0.018541375175118446, -0.0030323374085128307, 0.08412456512451172, 0.08837315440177917, 0.03946914151310921, -0.018453005701303482, 0.08528453856706619, -0.06613414734601974, -0.032554224133491516, -0.05329132080078125, -0.13773347437381744, -0.01694241166114807, -0.04032406955957413, -0.01493010576814413, -0.143591970205307, -0.15922600030899048, -0.001207371591590345, 0.02735789306461811, 0.004596560727804899, 0.019288625568151474, -0.07138434797525406, -0.01781950332224369, -0.010167145170271397, 0.0008312324644066393, -0.0549992099404335, 0.0250247735530138, -0.017792057245969772, 0.03443080559372902, 0.05582863837480545, -0.2048300951719284, 0.04762428253889084, -0.11525134742259979, 0.07920736819505692, -0.20633742213249207, 0.06693630665540695, -0.06200922280550003, 0.01950105093419552, -0.05273424834012985, -0.0296317245811224, -0.08492700755596161, 0.07007920742034912, 0.041285254061222076, 0.10710583627223969, -0.06220703572034836, -0.12054509669542313, 0.11373046785593033, -0.16949430108070374, -0.10597455501556396, 0.07037577033042908, 0.0017031242605298758, 0.047263577580451965, 0.07015656679868698, 0.14452669024467468, 0.07996734231710434, 0.036813199520111084, -0.06628973037004471, -0.025905363261699677, -0.027855753898620605, 0.0981748104095459, 0.042568013072013855, -0.022346116602420807, -0.030269844457507133, 0.029466601088643074, -0.06043853983283043, 0.010861096903681755, -0.02487821690738201, -0.04585431143641472, -0.020314019173383713, -0.028719063848257065, 0.16650667786598206, -0.03145139291882515, 0.04186243563890457, -0.05062977969646454, -0.07252804934978485, 0.1322222352027893, 0.08416378498077393, -0.06672126799821854, 0.05862083286046982, -0.09738250821828842, 0.028565850108861923, -0.018312089145183563, 0.08122044056653976, -0.14449121057987213, -0.18799346685409546, -0.03814810886979103, -0.03400163725018501, 0.014959861524403095, -0.002320549450814724, 0.0720883458852768, 0.02537376992404461, -0.09863577783107758, -0.038876209408044815, 0.027075868099927902, 0.04009397327899933, -0.025268785655498505, -0.15080326795578003, -0.030549107119441032, -0.07169176638126373, 0.11855089664459229, -0.09791257977485657, 0.08809670060873032, -0.01509333960711956, 0.1254095435142517, 0.060901861637830734, 0.0375593863427639, -0.03949733451008797, 0.06615813076496124, -0.043098773807287216, -0.003553576534613967, 0.09841315448284149, -0.023342035710811615, -0.15269631147384644, 0.06795904040336609, -0.17143449187278748, 0.18513360619544983, 0.11738470196723938, -0.07518621534109116, -0.05045595020055771, -0.030919820070266724, 0.01383175514638424, -0.04202066361904144, 0.04137421026825905, -0.058212243020534515, 0.04740586504340172, 0.005628703162074089, 0.09944678843021393, -0.050394780933856964, 0.0252765491604805, -0.013786531053483486, -0.06532567739486694, -0.02754618041217327, 0.05066772550344467, -0.03222827613353729, -0.23016779124736786, 0.10386840254068375, 0.20447519421577454, -0.05013522878289223, 0.10189957171678543, 0.013823877088725567, -0.0011009488953277469, -0.01820993982255459, -0.006353976204991341, 0.022817164659500122, -0.0401921272277832, -0.053575560450553894, 0.004726144019514322, 0.03179343417286873, 0.009930361062288284, 0.0484456829726696, -0.057742927223443985, 0.024099446833133698, 0.03510858863592148, -0.023891843855381012, 0.005699628032743931, 0.07276283949613571, -0.010611982084810734, 0.05013338848948479, 0.026547342538833618, -0.03774789348244667, 0.0641811192035675, -0.002776667708531022, -0.05289340391755104, 0.15986356139183044, -0.08381855487823486, -0.2142634242773056, -0.13957564532756805, -0.1253364235162735, -0.12063620984554291, 0.01869715377688408, 0.06197747960686684, -0.046965207904577255, -0.04488011449575424, -0.08929944038391113, 0.1926974207162857, 0.055733341723680496, 0.011500324122607708, -0.005007416475564241, -0.00895050261169672, 0.026840319857001305, -0.07823571562767029, -0.03134649246931076, -0.03549060598015785, -0.0708324983716011, 0.0698380395770073, -0.06599574536085129, 0.12077336758375168, 0.1024666503071785, 0.006876097992062569, -0.04206068813800812, -0.021979179233312607, 0.12483803927898407, -0.027299076318740845, 0.0404549241065979, 0.15998288989067078, -0.061595503240823746, 0.08463501185178757, 0.1896473616361618, 0.03969993814826012, -0.028176458552479744, 0.03989836201071739, 0.04926924407482147, -0.026272455230355263, -0.11954598873853683, -0.09189840406179428, -0.026738759130239487, 0.061012785881757736, 0.017714161425828934, 0.025740478187799454, 0.08645467460155487, 0.08789493143558502, -0.05195540562272072, 0.012323773466050625, 0.04941721633076668, 0.06651422381401062, 0.24660757184028625, 0.009993941523134708, 0.1707870364189148, -0.01983359083533287, -0.11102326959371567, 0.03387909755110741, -0.02021251805126667, 0.1433085799217224, 0.060694172978401184, 0.17864862084388733, 0.0591757670044899, -0.002013006480410695, 0.08589589595794678, 0.0736641064286232, 0.011516277678310871, -0.042357150465250015, -0.0238272026181221, -0.07545202225446701, 0.018254294991493225, 0.032220806926488876, -0.07534977793693542, 0.026647936552762985, -0.07431767880916595, 0.03805151209235191, 0.0800967738032341, 0.11662193387746811, 0.02992197684943676, -0.25626981258392334, -0.07476352900266647, 0.044683512300252914, 0.03185071051120758, -0.007162519730627537, -0.0349072590470314, 0.03329611197113991, -0.03331036865711212, 0.1175069734454155, -0.06170554459095001, 0.08770762383937836, 0.013310239650309086, 0.09000679105520248, -0.017354879528284073, 0.12219585478305817, -0.006028291769325733, 0.02013215608894825, -0.1934293955564499, 0.11120392382144928, 0.03348778933286667, 0.013646409846842289, 0.04817119985818863, -0.0009074422996491194, 0.06346996128559113, 0.2544404864311218, 0.07723730057477951, -0.017702240496873856, 0.008636566810309887, -0.0518314391374588, -0.007027511950582266, 0.027928326278924942, 0.11720918864011765, -0.04936845973134041, 0.12055762112140656, -0.05880317464470863, -0.02455654926598072, 0.0746234804391861, 0.10387139022350311, -0.13244429230690002, -0.10019365698099136, 0.07843872159719467, 0.024311505258083344, 0.06430888921022415, -0.09339818358421326, -0.049870602786540985, -0.08043170720338821, 0.26675888895988464, 0.05255939066410065, -0.04592384770512581, -0.10832583159208298, -0.03451479971408844, 0.0566709041595459, -0.11265357583761215, 0.05602111294865608, -0.02177894301712513, 0.06184786185622215, -0.025436611846089363, -0.17964275181293488, 0.11854849755764008, -0.06469815224409103, -0.09741120785474777, -0.015249183401465416, 0.014072908088564873, -0.0447695329785347, -0.010733252391219139, -0.01786446012556553, 0.028588714078068733, -0.015867061913013458, -0.05896490067243576, 0.0016158848302438855, 0.08217103034257889, -0.05239344388246536, 0.08796605467796326, -0.07750585675239563, -0.08906012028455734, -0.028989188373088837, 0.0494755320250988, 0.137898787856102, 0.20525357127189636, -0.017566148191690445, 0.00342228589579463, 0.17293573915958405, -0.03688860684633255, -0.303289532661438, -0.07286114245653152, -0.033055100589990616, 0.003547599771991372, 0.034573182463645935, -0.16311407089233398, 0.09158758819103241, 0.11120268702507019, -0.002521313726902008, 0.06420797854661942, -0.23746727406978607, -0.12251164019107819, 0.154973104596138, 0.06998569518327713, 0.22001934051513672, -0.13073961436748505, -0.07340643554925919, -0.08712886273860931, -0.10771533101797104, 0.10815794765949249, -0.20686370134353638, 0.11093483865261078, -0.01592717319726944, 0.00864377897232771, 0.009013528935611248, 0.009517272934317589, 0.08835528790950775, -0.049933526664972305, 0.053097035735845566, -0.07008698582649231, -0.08308791369199753, 0.08456500619649887, -0.04197371006011963, 0.03862392157316208, -0.09628275781869888, 0.04258808121085167, 0.09620801359415054, -0.05588532239198685, -0.04152718558907509, 0.0843762531876564, -0.02618139237165451, -0.0819929763674736, -0.02913774736225605, 0.03402691334486008, -0.0003283371916040778, 0.019507355988025665, 0.24344755709171295, -0.065004363656044, 0.08472760766744614, 0.15656068921089172, 0.09750836342573166, -0.09563291072845459, 0.03128023445606232, 0.008586039766669273, -0.07536383718252182, 0.09401071816682816, -0.018210519105196, 0.03747229650616646, 0.07083746045827866, -0.009753123857080936, 0.09752468764781952, 0.08733668923377991, 0.010847842320799828, -0.01917148567736149, 0.08525509387254715, -0.12260092049837112, -0.21175406873226166, -0.03033246472477913, -0.07577553391456604, -0.09709449857473373, 0.16706910729408264, 0.21857067942619324, -0.02954518422484398, 0.015161383897066116, 0.014905616641044617, 0.009326093830168247, -0.0799049437046051, 0.07638230919837952, 0.033256590366363525, 0.016670258715748787, -0.07654915750026703, 0.07456778734922409, 0.036228932440280914, -0.09351374208927155, -0.036630526185035706, 0.07299137860536575, -0.11153975874185562, -0.10971853137016296, -0.10157366842031479, 0.2074824869632721, -0.09713190048933029, -0.043281491845846176, -0.12764185667037964, -0.15853993594646454, 0.01950770802795887, 0.14220933616161346, 0.09021387994289398, 0.03890392929315567, -0.01806059665977955, 0.007109934464097023, -0.01810806430876255, 0.043450500816106796, 0.00291992980055511, 0.11909803003072739, -0.10432957857847214, 0.0317959189414978, -0.020537249743938446, 0.023295067250728607, -0.0789911225438118, 0.019507387652993202, -0.1252266764640808, -0.06418950110673904, -0.20351675152778625, -0.06082381680607796, -0.11067598313093185, -0.02672598883509636, -0.0018077777931466699, 0.021902790293097496, -0.03847673907876015, -0.014936178922653198, -0.02354625053703785, -0.06847383081912994, -0.006032347213476896, 0.013503530994057655, -0.08710472285747528, 0.02127862349152565, 0.0011766048846766353, -0.05328000336885452, 0.06698388606309891, 0.010743562132120132, -0.0030987276695668697, 0.01719779334962368, -0.12719348073005676, -0.044285163283348083, 0.06162039563059807, -0.005247860681265593, 0.02669687382876873, -0.041128892451524734, -0.058984462171792984, -0.010114407166838646, -0.004737033974379301, 0.0023140637204051018, 0.11239108443260193, -0.08487454056739807, 0.0775512158870697, -0.07954971492290497, -0.029075613245368004, -0.056307103484869, -0.0027445820160210133, 0.01411785650998354, 0.06029975041747093, 0.19184724986553192, -0.06313388794660568, 0.02358737587928772, -0.14754043519496918, 0.02123529650270939, 0.026129581034183502, -0.1302516758441925, 0.023154407739639282, -0.057201825082302094, -0.005375499837100506, -0.052859727293252945, 0.11107286810874939, -0.09242776781320572, -0.11373043060302734, 0.03418418765068054, -0.023355236276984215, -0.03193230554461479, 0.04822966083884239, 0.21364115178585052, 0.09813238680362701, -0.06331422179937363, -0.08741194754838943, 0.03972293436527252, 0.05364425480365753, 0.01496117189526558, 0.132957324385643, 0.12416546791791916, -0.004682296421378851, 0.12723036110401154, 0.05006859451532364, 0.02256833203136921, -0.08741389214992523, 0.046673599630594254, -0.0004526970151346177, 0.02548336610198021, -0.0037929078098386526, 0.12356607615947723, 0.15354153513908386, -0.03969205915927887, 0.05143674463033676, 0.022962598130106926, -0.028849540278315544, -0.10603421181440353, -0.09906162321567535, -0.13605141639709473, -0.16233490407466888, -0.06303567439317703, -0.10328146815299988, -0.060648106038570404, 0.05524726212024689, -0.013643583282828331, -0.002789693418890238, 0.2208607792854309, -0.07582655549049377, -0.053401198238134384, 0.04494215548038483, -0.05252659693360329, -0.02260461635887623, 0.06356897950172424, -0.0640791729092598, 0.041595105081796646, 0.08882495015859604, -0.002456996589899063, 0.0039782351814210415, -0.015445076860487461, 0.052666034549474716, -0.0723930075764656, -0.10669439285993576, -0.006983543746173382, 0.0810527354478836, -0.005755179561674595, 0.04296044632792473, 0.008383473381400108, -0.051514122635126114, 0.006593847647309303, 0.13753703236579895, -0.04774530604481697, -0.13137665390968323, -0.06624112278223038, 0.1929381638765335, 0.031084047630429268, 0.11577026546001434, -0.02335824817419052, -0.09697805345058441, -0.014657283201813698, 0.13406582176685333, 0.2834095358848572, -0.08738720417022705, 0.02851327508687973, 0.046646203845739365, 0.012079187668859959, 0.019804242998361588, 0.05254168063402176, 0.020216871052980423, 0.1045483648777008, -0.02057715319097042, -0.0045287045650184155, -0.012514988891780376, -0.03903214633464813, -0.06713659316301346, 0.015335251577198505, 0.07360298931598663, 0.021005723625421524, -0.046530842781066895, 0.030554192140698433, -0.08798301219940186, 0.044085536152124405, -0.006161903031170368, -0.11191805452108383, -0.07274763286113739, -0.07443632185459137, 0.0695113092660904, -0.00018088609795086086, 0.0943358838558197, -0.018150009214878082, -0.047444719821214676, 0.10406274348497391, -0.014647225849330425, -0.025549283251166344, -0.017366399988532066, 0.040565166622400284, -0.07309076189994812, 0.056645479053258896, -0.018585754558444023, -0.033929493278265, 0.11531127244234085, 0.01505032368004322, -0.04383017495274544, 0.11517328768968582, -0.026104895398020744, -0.04931185021996498, 0.0481019951403141, 0.06096445024013519, -0.03776141628623009, 0.08864513784646988, 0.061197444796562195, -0.22051264345645905, 0.0466594472527504, 0.1257871836423874, -0.11911111325025558, -0.017527135089039803, 0.022034727036952972, -0.07361915707588196, 0.0679539367556572, 0.2043428272008896, 0.004955303855240345, 0.021614613011479378, 0.0055939084850251675, 0.02772257663309574, 0.09989277273416519, 0.08357629179954529, -0.05036348104476929, -0.13547268509864807, -0.005994448903948069, 0.12115813046693802, 0.0002843922411557287, -0.30273938179016113, -0.0665581226348877, -0.08061199635267258, 0.031191904097795486, -0.06350843608379364, 0.03370126336812973, 0.16614648699760437, 0.014505384489893913, -0.02111678570508957, -0.12840935587882996, -0.008825761266052723, 0.0990893691778183, -0.09436238557100296, -0.11175265908241272 ]
null
null
transformers
# Model Card for Zenith-7B-dpo-v3 Mistral-7B-v0.1 model fine-tuned on the Ultrafeedback dataset using techinques shown in the paper [Self-Rewarding Language Models](https://arxiv.org/abs/2401.10020). ## Results | model_name | Average | arc_challenge | hellaswag | truthfulqa_mc2 | winogrande | |:-----------------|----------:|----------------:|------------:|-----------------:|-------------:| | Zenith-7B-dpo-v3 | 0.707576 | 0.613481 | 0.848337 | 0.602897 | 0.765588 | ## Instruction format In order to leverage instruction fine-tuning, your prompt should be surrounded by `[INST]` and `[/INST]` tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id. E.g. ``` text = "<s>[INST] What is your favourite condiment? [/INST]" "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!</s> " "[INST] Do you have mayonnaise recipes? [/INST]" ``` This format is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating) via the `apply_chat_template()` method: ```python from transformers import AutoModelForCausalLM, AutoTokenizer device = "cuda" # the device to load the model onto model = AutoModelForCausalLM.from_pretrained("Xenon1/Zenith-7B-dpo-v3") tokenizer = AutoTokenizer.from_pretrained("Xenon1/Zenith-7B-dpo-v3") messages = [ {"role": "user", "content": "What is your favourite condiment?"}, {"role": "assistant", "content": "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!"}, {"role": "user", "content": "Do you have mayonnaise recipes?"} ] encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt") model_inputs = encodeds.to(device) model.to(device) generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True) decoded = tokenizer.batch_decode(generated_ids) print(decoded[0]) ``` ## Model Architecture This instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices: - Grouped-Query Attention - Sliding-Window Attention - Byte-fallback BPE tokenizer
{"language": ["en"], "license": "apache-2.0", "tags": ["mistral", "Zenith-7B-dpo-v3"], "pipeline_tag": "text-generation"}
text-generation
Xenon1/Zenith-7B-dpo-v3
[ "transformers", "safetensors", "mistral", "text-generation", "Zenith-7B-dpo-v3", "conversational", "en", "arxiv:2401.10020", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T02:33:05+00:00
[ "2401.10020" ]
[ "en" ]
TAGS #transformers #safetensors #mistral #text-generation #Zenith-7B-dpo-v3 #conversational #en #arxiv-2401.10020 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Model Card for Zenith-7B-dpo-v3 =============================== Mistral-7B-v0.1 model fine-tuned on the Ultrafeedback dataset using techinques shown in the paper Self-Rewarding Language Models. Results ------- Instruction format ------------------ In order to leverage instruction fine-tuning, your prompt should be surrounded by '[INST]' and '[/INST]' tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id. E.g. This format is available as a chat template via the 'apply\_chat\_template()' method: Model Architecture ------------------ This instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices: * Grouped-Query Attention * Sliding-Window Attention * Byte-fallback BPE tokenizer
[]
[ "TAGS\n#transformers #safetensors #mistral #text-generation #Zenith-7B-dpo-v3 #conversational #en #arxiv-2401.10020 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 82 ]
[ "passage: TAGS\n#transformers #safetensors #mistral #text-generation #Zenith-7B-dpo-v3 #conversational #en #arxiv-2401.10020 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.05069350823760033, 0.05730914697051048, -0.006329437252134085, 0.01361668948084116, 0.05781414732336998, -0.03514399006962776, 0.17721234261989594, 0.10540158301591873, -0.05954192578792572, -0.037309758365154266, 0.18325546383857727, 0.1756470650434494, 0.012341349385678768, 0.08124203234910965, -0.09476080536842346, -0.14070282876491547, 0.10085847973823547, -0.0010928385891020298, -0.026909630745649338, 0.07588633894920349, 0.12740303575992584, -0.03256884217262268, 0.05856954678893089, -0.05372265353798866, -0.057106684893369675, 0.01255191769450903, 0.03103136457502842, -0.10696669667959213, 0.10545014590024948, 0.04547407478094101, 0.060108937323093414, 0.09031513333320618, 0.002776659093797207, -0.18262241780757904, 0.016830408945679665, 0.0283366609364748, -0.04175770282745361, 0.03863988071680069, 0.019096476957201958, 0.016072120517492294, 0.024505650624632835, 0.008319866843521595, -0.03649988770484924, 0.054472096264362335, -0.06692073494195938, -0.11775067448616028, -0.08723296225070953, 0.037994708865880966, 0.0784880593419075, 0.10795504599809647, 0.028171129524707794, 0.1433272510766983, 0.004105743486434221, 0.08026522397994995, 0.14674648642539978, -0.3378308415412903, 0.012266930192708969, 0.07846001535654068, 0.045230913907289505, 0.09417890757322311, -0.011922995559871197, 0.042225148528814316, 0.07475605607032776, -0.015149076469242573, 0.025768756866455078, -0.053730227053165436, -0.05444108694791794, 0.03024599514901638, -0.08542311191558838, -0.032363854348659515, 0.3219785690307617, 0.011704698204994202, 0.007413832936435938, -0.015013470314443111, -0.09549953788518906, -0.016821250319480896, -0.0013863638741895556, -0.04619714990258217, 0.030149852856993675, 0.09391744434833527, 0.07546070963144302, -0.013610629364848137, -0.13740602135658264, -0.028041180223226547, -0.1601375788450241, 0.05664752423763275, 0.006282232701778412, 0.05061793327331543, -0.1558530330657959, 0.048020895570516586, 0.026876013725996017, -0.13520564138889313, -0.010903540998697281, -0.0805981233716011, 0.12748585641384125, 0.016198432072997093, -0.0448722168803215, -0.04875548928976059, 0.16013991832733154, 0.17468997836112976, 0.027421867474913597, 0.013252267614006996, -0.09646514058113098, 0.05741600692272186, -0.056795135140419006, 0.02409469522535801, 0.0023332529235631227, -0.048848867416381836, 0.1172669529914856, -0.019196292385458946, 0.10977132618427277, -0.0188697911798954, -0.141632542014122, -0.008657548576593399, 0.04927680268883705, 0.12849588692188263, 0.031799107789993286, 0.07442596554756165, -0.0023105337750166655, 0.04219329357147217, 0.10447895526885986, -0.07010187208652496, -0.052331969141960144, -0.0014579140115529299, 0.02315087616443634, 0.06239396333694458, 0.028562624007463455, 0.05022663250565529, -0.05772761255502701, 0.00824288371950388, -0.03195352107286453, -0.05895742028951645, -0.031898509711027145, -0.03720702975988388, 0.07118783891201019, -0.00902191735804081, 0.03155291825532913, -0.18961192667484283, -0.18109677731990814, 0.030650349333882332, 0.021703435108065605, 0.0003175312594976276, -0.05885763093829155, -0.020847661420702934, -0.03843904659152031, 0.03346531465649605, -0.06595263630151749, -0.015400818549096584, -0.07260128855705261, 0.0697832778096199, -0.02142845094203949, 0.043688129633665085, -0.1701056957244873, 0.041492193937301636, -0.11019691079854965, 0.011891921050846577, -0.00037961575435474515, -0.032097529619932175, -0.06102202832698822, 0.16949902474880219, -0.056748051196336746, 0.031059620901942253, -0.02853267453610897, -0.0005684952484443784, 0.008473947644233704, 0.1418720781803131, -0.15191124379634857, -0.007818995043635368, 0.18377816677093506, -0.1198461577296257, -0.22531360387802124, 0.09354237467050552, 0.036388348788022995, 0.01868331804871559, 0.09100466221570969, 0.14851096272468567, 0.004050747957080603, -0.08490819483995438, 0.058142147958278656, 0.11232134699821472, -0.07061358541250229, -0.1035182848572731, 0.042387012392282486, -0.03095117025077343, -0.09545174241065979, 0.049907878041267395, 0.02683389186859131, 0.060486942529678345, 0.014867328107357025, -0.07233118265867233, -0.07308800518512726, -0.06414370238780975, -0.02387523464858532, -0.04328033700585365, 0.02627069503068924, -0.0999690443277359, -0.005844921339303255, -0.06684091687202454, 0.04985538870096207, 0.011152741499245167, 0.02272682636976242, -0.04830138012766838, 0.04775889590382576, -0.0015222820220515132, 0.07773352414369583, -0.0709921270608902, -0.010181417688727379, 0.004152078181505203, 0.048269256949424744, -0.022312935441732407, 0.04550018161535263, 0.05194925144314766, -0.026244645938277245, -0.013874990865588188, -0.007363170385360718, 0.12988528609275818, 0.05543753132224083, -0.03973725810647011, -0.11900104582309723, 0.11086951941251755, -0.04713945463299751, 0.05893460661172867, -0.0770006775856018, 0.03357478976249695, 0.031003674492239952, 0.09109432250261307, -0.01433494221419096, 0.10374926775693893, -0.029801415279507637, -0.02252180315554142, -0.054864224046468735, -0.007791709154844284, 0.09315180033445358, 0.04501352086663246, -0.08047056943178177, 0.17499330639839172, -0.1575157195329666, 0.24490247666835785, 0.18744467198848724, -0.13826346397399902, 0.04710189625620842, -0.05109657347202301, 0.00007976154302014038, -0.027020396664738655, 0.04984975978732109, -0.04781002178788185, -0.05693434178829193, -0.015573081560432911, 0.13780418038368225, -0.07429826259613037, 0.0005697543965652585, -0.003273837501183152, -0.07146159559488297, -0.038010988384485245, 0.04023301228880882, 0.04417131841182709, -0.17964889109134674, 0.17421108484268188, 0.31176069378852844, -0.025040501728653908, 0.10265383869409561, -0.08105754107236862, -0.024342168122529984, 0.022958852350711823, 0.03677341714501381, -0.007488820236176252, 0.02095276489853859, -0.11302658915519714, 0.0189374890178442, 0.0745977908372879, 0.04340383782982826, 0.044719398021698, -0.11969180405139923, -0.04163104295730591, 0.005199607461690903, -0.04417697340250015, 0.011326327919960022, 0.05925979092717171, -0.03668249770998955, 0.13190722465515137, -0.029933683574199677, -0.11198922246694565, 0.11783425509929657, -0.004186525009572506, -0.09462865442037582, 0.15147589147090912, -0.17643846571445465, -0.23174557089805603, -0.11690717190504074, -0.09610963612794876, -0.08928973227739334, 0.043148331344127655, 0.13052725791931152, -0.059757448732852936, -0.05318591371178627, -0.05996321141719818, -0.043633315712213516, -0.0217841025441885, 0.014684082008898258, 0.032871063798666, 0.059226419776678085, -0.0015346705913543701, -0.14053112268447876, -0.034378089010715485, 0.024561939761042595, -0.012701192870736122, 0.07860109955072403, -0.06612012535333633, 0.09152550995349884, 0.1153392344713211, 0.03907983377575874, -0.00440975883975625, -0.02713482826948166, 0.08129694312810898, -0.020451096817851067, 0.0026826730463653803, 0.20226874947547913, -0.013219893909990788, 0.062299929559230804, 0.18268756568431854, -0.0024892084766179323, -0.08234460651874542, 0.06427785754203796, -0.052763327956199646, -0.04850076511502266, -0.23721659183502197, -0.077837273478508, -0.07374335080385208, 0.11760637164115906, -0.010790437459945679, 0.06689457595348358, 0.16062088310718536, 0.06306283175945282, -0.05246427282691002, 0.025443388149142265, 0.05958634242415428, 0.07847008854150772, 0.23903687298297882, -0.03258626535534859, 0.11496798694133759, -0.10693765431642532, -0.06861179322004318, 0.08569016307592392, 0.0898762047290802, 0.09484460949897766, 0.11178495734930038, 0.10633721202611923, 0.039168424904346466, 0.1139603778719902, 0.07244516909122467, 0.10744843631982803, 0.05993754416704178, -0.034892305731773376, -0.04428298398852348, -0.061034467071294785, -0.013303271494805813, 0.05755877494812012, -0.12329978495836258, -0.06269050389528275, -0.002563206013292074, -0.010444381274282932, 0.0739123523235321, 0.1778343766927719, 0.03182847425341606, -0.1679360419511795, 0.01375346165150404, 0.09142254292964935, 0.022156784310936928, -0.0548231340944767, 0.11468812823295593, -0.03816353157162666, -0.022335773333907127, 0.08892791718244553, -0.018147563561797142, 0.08732274174690247, 0.014474540017545223, 0.07259923964738846, -0.07885712385177612, -0.03456816077232361, 0.023984845727682114, 0.11685827374458313, -0.38980600237846375, 0.1692228615283966, 0.0007400436443276703, 0.0018197663594037294, -0.07633964717388153, 0.02689305506646633, 0.0331539660692215, 0.16609765589237213, 0.10600528866052628, -0.019213842228055, -0.08751624077558517, 0.040801454335451126, -0.0599256232380867, 0.04810071364045143, 0.05353282392024994, 0.04059114679694176, -0.030511099845170975, -0.05553261563181877, -0.007738209795206785, 0.03060794062912464, 0.04947047308087349, -0.11782348901033401, -0.11991837620735168, 0.07764347642660141, 0.12629839777946472, 0.08036118000745773, -0.06486830860376358, 0.010939358733594418, -0.13036571443080902, 0.12348455935716629, -0.07134396582841873, -0.05216897279024124, -0.06125950813293457, -0.12479627877473831, -0.007816312834620476, -0.02169625461101532, 0.02376932092010975, -0.05766981095075607, 0.026889322325587273, -0.06226729229092598, -0.1682848036289215, 0.1141713410615921, -0.12000355869531631, -0.07092498987913132, -0.0477592796087265, 0.12539558112621307, -0.08711464703083038, -0.02855018712580204, 0.018547723069787025, -0.00808898638933897, -0.07851013541221619, -0.08730418235063553, 0.012861918658018112, 0.04509595409035683, 0.03699781000614166, -0.04770739749073982, -0.08160242438316345, -0.11760473996400833, -0.013880660757422447, -0.10008520632982254, 0.19554287195205688, 0.23589901626110077, -0.04474058747291565, 0.1184108629822731, 0.22802142798900604, -0.0675269141793251, -0.3128252923488617, -0.14885739982128143, -0.15664127469062805, -0.07541468739509583, -0.04079491272568703, -0.0810176357626915, 0.07925936579704285, 0.0742160752415657, -0.07916652411222458, 0.10391534864902496, -0.21005195379257202, -0.09808897227048874, 0.21884490549564362, 0.015574327670037746, 0.3088034987449646, -0.17669247090816498, -0.07241649180650711, -0.13969331979751587, -0.16242219507694244, 0.13386110961437225, -0.18780694901943207, 0.03835369646549225, -0.008939377032220364, -0.0028377340640872717, -0.001488825655542314, -0.03658326715230942, 0.11831759661436081, -0.048667579889297485, 0.053147248923778534, -0.1451740264892578, 0.046406470239162445, 0.0843018889427185, -0.02737189643085003, 0.07300671190023422, -0.1696978658437729, 0.04803057014942169, 0.01610388606786728, -0.0127260722219944, -0.016084162518382072, 0.05759143456816673, -0.007825535722076893, -0.08568592369556427, -0.036505915224552155, -0.06701730936765671, 0.02705017663538456, -0.0327821709215641, 0.23503413796424866, 0.0005499277613125741, 0.07718956470489502, 0.19350270926952362, 0.10495518893003464, -0.17565500736236572, 0.09928767383098602, -0.04233095422387123, -0.08186893165111542, 0.08885933458805084, -0.12152665108442307, 0.05857124924659729, 0.07526960223913193, -0.05745934322476387, 0.06566252559423447, 0.058421917259693146, -0.010105354711413383, -0.049302779138088226, 0.09954743087291718, -0.18556897342205048, -0.0612308643758297, -0.03170410171151161, 0.1554751694202423, 0.04669296368956566, 0.13753782212734222, 0.16112031042575836, -0.025629105046391487, 0.006517055444419384, 0.01619529165327549, 0.04352192208170891, -0.05904380604624748, 0.062009140849113464, 0.03507893159985542, 0.011441156268119812, -0.0884118527173996, 0.15724694728851318, -0.009447216056287289, -0.13571001589298248, 0.006020327564328909, 0.07579226046800613, -0.151741161942482, -0.12472479790449142, -0.0021852198988199234, 0.09181304275989532, -0.08065132051706314, -0.12035857886075974, -0.053763676434755325, -0.15814518928527832, 0.04054553434252739, 0.12606582045555115, 0.055832646787166595, 0.0596415176987648, 0.016202835366129875, -0.04622466117143631, -0.0038273490499705076, 0.05233331769704819, -0.07425349205732346, 0.04167962446808815, -0.1192220076918602, -0.07945458590984344, -0.048192545771598816, 0.02881321869790554, -0.050657544285058975, 0.008334971964359283, -0.10099589824676514, 0.0033512094523757696, -0.19277571141719818, -0.021869216114282608, -0.0767909437417984, -0.01717066951096058, 0.00639471085742116, -0.02692372538149357, -0.010750701650977135, 0.008249773643910885, -0.06518521159887314, -0.012825271114706993, -0.013708007521927357, 0.05675807595252991, -0.09223628044128418, -0.03978295251727104, 0.024577923119068146, -0.026716362684965134, 0.10846905410289764, 0.09366809576749802, -0.06526488065719604, 0.06085818260908127, -0.1802334040403366, -0.03181949257850647, 0.09055044502019882, 0.03935092315077782, -0.0013749967329204082, -0.0018992749974131584, -0.04911499843001366, 0.1324012279510498, -0.027446402236819267, 0.026817219331860542, 0.04014918580651283, -0.09720184653997421, -0.03103453665971756, -0.028606152161955833, -0.06926868110895157, -0.033126700669527054, -0.10507731139659882, 0.1114732176065445, 0.016060061752796173, 0.15612760186195374, -0.059424448758363724, -0.0007155569037422538, -0.047714993357658386, 0.02810710296034813, 0.012617584317922592, -0.14566367864608765, -0.10955490916967392, -0.02956734225153923, -0.022351782768964767, -0.03249382600188255, 0.20298504829406738, -0.06062578782439232, -0.06627342104911804, 0.07979080080986023, 0.011097819544374943, 0.03630591928958893, 0.031675610691308975, 0.26060670614242554, 0.07006891071796417, -0.01172675471752882, -0.1454879492521286, 0.007216510362923145, 0.051195111125707626, -0.018948912620544434, 0.03586040064692497, 0.10293896496295929, 0.013663933612406254, 0.1021728515625, 0.035311806946992874, -0.0255266185849905, -0.017875978723168373, -0.06461410224437714, -0.006095385178923607, 0.05850205942988396, -0.01668347604572773, 0.12649713456630707, 0.21914908289909363, 0.012097501195967197, -0.040907472372055054, -0.08794946223497391, -0.021224191412329674, -0.13796979188919067, -0.1265607625246048, -0.08371798694133759, -0.15174755454063416, -0.017766403034329414, -0.06373713910579681, 0.0364217683672905, 0.04659445956349373, 0.05110593140125275, -0.06379440426826477, 0.040924716740846634, 0.006670293398201466, -0.06270880252122879, 0.0216841958463192, -0.02233869768679142, -0.02267717942595482, -0.00897753145545721, -0.06873948127031326, -0.022843923419713974, -0.045638326555490494, -0.04831600561738014, 0.029286066070199013, 0.02902418188750744, 0.046263065189123154, -0.0907863900065422, -0.04007731005549431, -0.020603928714990616, 0.05377013236284256, -0.023632382974028587, 0.16928701102733612, 0.027566516771912575, -0.01622980646789074, 0.10755867511034012, 0.182614266872406, -0.06455401331186295, -0.17740322649478912, -0.05573352426290512, 0.09456516057252884, 0.015955664217472076, 0.11176415532827377, -0.020501762628555298, -0.0012483377940952778, -0.03121393546462059, 0.253994345664978, 0.24765320122241974, -0.029373373836278915, 0.020433716475963593, -0.06103609874844551, 0.021626204252243042, -0.00821808073669672, 0.08098629862070084, 0.160279780626297, 0.1544075608253479, -0.03591197729110718, 0.01743444800376892, -0.024682460352778435, 0.03013731725513935, -0.14719568192958832, 0.09864775836467743, -0.012147780507802963, -0.07232272624969482, 0.021060079336166382, 0.07369150221347809, -0.06260932981967926, 0.06111467257142067, -0.1205621063709259, -0.06923412531614304, -0.014734177850186825, -0.0024719969369471073, 0.16241930425167084, -0.023529868572950363, -0.0027199757751077414, -0.03404951095581055, -0.026901444420218468, 0.11341457813978195, -0.027289213612675667, -0.14804309606552124, -0.02421795204281807, 0.03159596025943756, -0.05844179913401604, 0.07668861746788025, 0.007559497840702534, 0.07413666695356369, 0.09306203573942184, 0.06645538657903671, -0.09208913147449493, 0.15963871777057648, 0.030567945912480354, -0.054885316640138626, 0.03325112909078598, -0.09109950810670853, -0.025082871317863464, 0.009246721863746643, 0.04652312770485878, -0.0815407931804657, 0.027681749314069748, 0.09304553270339966, -0.08340401947498322, -0.043890580534935, 0.06764139235019684, -0.07927206158638, 0.07165279239416122, -0.0074235559441149235, -0.04364260286092758, 0.0050326185300946236, -0.03935840725898743, -0.0009297485812567174, 0.0009655929752625525, -0.14954552054405212, -0.0037939585745334625, -0.08122622966766357, -0.0337984599173069, 0.11337511986494064, 0.034186430275440216, -0.1332302689552307, -0.03308736905455589, -0.08688810467720032, 0.014869429171085358, -0.1402810662984848, -0.00996239110827446, 0.13046503067016602, 0.0016892676940187812, -0.029498720541596413, -0.07646221667528152, 0.01362898014485836, 0.07009202241897583, -0.06371202319860458, -0.12394685298204422 ]
null
null
null
# Lora of hornet/ホーネット/大黄蜂 (Azur Lane) ## What Is This? This is the LoRA model of waifu hornet/ホーネット/大黄蜂 (Azur Lane). ## How Is It Trained? * This model is trained with [HCP-Diffusion](https://github.com/7eu7d7/HCP-Diffusion). * The [auto-training framework](https://github.com/deepghs/cyberharem) is maintained by [DeepGHS Team](https://huggingface.co/deepghs). * The base model used for training is [deepghs/animefull-latest](https://huggingface.co/deepghs/animefull-latest). * Dataset used for training is the `stage3-p480-800` in [CyberHarem/hornet_azurlane](https://huggingface.co/datasets/CyberHarem/hornet_azurlane), which contains 361 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 10, resolution is 720x720, clustering into 20 buckets. * Trained for 3640 steps, 40 checkpoints were saved and evaluated. * **Trigger word is `hornet_azurlane`.** * Pruned core tags for this waifu are `blonde_hair, long_hair, green_eyes, twintails, breasts, large_breasts, bangs, sidelocks, very_long_hair, hat, cowboy_hat, black_headwear`. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. ## How to Use It? ### If You Are Using A1111 WebUI v1.7+ **Just use it like the classic LoRA**. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 546, you need to download [`546/hornet_azurlane.pt`](https://huggingface.co/CyberHarem/hornet_azurlane/resolve/main/546/hornet_azurlane.pt) as the embedding and [`546/hornet_azurlane.safetensors`](https://huggingface.co/CyberHarem/hornet_azurlane/resolve/main/546/hornet_azurlane.safetensors) for loading Lora. By using both files together, you can generate images for the desired characters. ## Which Step Should I Use? We selected 5 good steps for you to choose. The best one is step 546. 1800 images (1.84 GiB) were generated for auto-testing. ![Metrics Plot](metrics_plot.png) The base model used for generating preview images is [Meina/MeinaMix_V11](https://huggingface.co/Meina/MeinaMix_V11). Here are the preview of the recommended steps: | Step | Epoch | CCIP | AI Corrupt | Bikini Plus | Score | Download | pattern_0_0 | pattern_0_1 | pattern_1 | pattern_2 | pattern_3_0 | pattern_3_1 | pattern_4_0 | pattern_4_1 | pattern_5 | pattern_6 | portrait_0 | portrait_1 | portrait_2 | full_body_0 | full_body_1 | profile_0 | profile_1 | free_0 | free_1 | shorts | maid_0 | maid_1 | miko | yukata | suit | china | bikini_0 | bikini_1 | bikini_2 | sit | squat | kneel | jump | crossed_arms | angry | smile | cry | grin | n_lie_0 | n_lie_1 | n_stand_0 | n_stand_1 | n_stand_2 | n_sex_0 | n_sex_1 | |-------:|--------:|:----------|:-------------|:--------------|:----------|:----------------------------------------------------------------------------------------------------|:----------------------------------------------|:----------------------------------------------|:------------------------------------------|:------------------------------------------|:----------------------------------------------|:----------------------------------------------|:----------------------------------------------|:----------------------------------------------|:------------------------------------------|:------------------------------------------|:--------------------------------------------|:--------------------------------------------|:--------------------------------------------|:----------------------------------------------|:----------------------------------------------|:------------------------------------------|:------------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:--------------------------------|:------------------------------------|:--------------------------------|:----------------------------------|:----------------------------------------|:----------------------------------------|:----------------------------------------|:------------------------------|:----------------------------------|:----------------------------------|:--------------------------------|:------------------------------------------------|:----------------------------------|:----------------------------------|:------------------------------|:--------------------------------|:--------------------------------------|:--------------------------------------|:------------------------------------------|:------------------------------------------|:------------------------------------------|:--------------------------------------|:--------------------------------------| | 546 | 7 | 0.772 | **0.965** | **0.836** | **0.773** | [Download](https://huggingface.co/CyberHarem/hornet_azurlane/resolve/main/546/hornet_azurlane.zip) | ![pattern_0_0](546/previews/pattern_0_0.png) | ![pattern_0_1](546/previews/pattern_0_1.png) | ![pattern_1](546/previews/pattern_1.png) | ![pattern_2](546/previews/pattern_2.png) | ![pattern_3_0](546/previews/pattern_3_0.png) | ![pattern_3_1](546/previews/pattern_3_1.png) | ![pattern_4_0](546/previews/pattern_4_0.png) | ![pattern_4_1](546/previews/pattern_4_1.png) | ![pattern_5](546/previews/pattern_5.png) | ![pattern_6](546/previews/pattern_6.png) | ![portrait_0](546/previews/portrait_0.png) | ![portrait_1](546/previews/portrait_1.png) | ![portrait_2](546/previews/portrait_2.png) | ![full_body_0](546/previews/full_body_0.png) | ![full_body_1](546/previews/full_body_1.png) | ![profile_0](546/previews/profile_0.png) | ![profile_1](546/previews/profile_1.png) | ![free_0](546/previews/free_0.png) | ![free_1](546/previews/free_1.png) | ![shorts](546/previews/shorts.png) | ![maid_0](546/previews/maid_0.png) | ![maid_1](546/previews/maid_1.png) | ![miko](546/previews/miko.png) | ![yukata](546/previews/yukata.png) | ![suit](546/previews/suit.png) | ![china](546/previews/china.png) | ![bikini_0](546/previews/bikini_0.png) | ![bikini_1](546/previews/bikini_1.png) | ![bikini_2](546/previews/bikini_2.png) | ![sit](546/previews/sit.png) | ![squat](546/previews/squat.png) | ![kneel](546/previews/kneel.png) | ![jump](546/previews/jump.png) | ![crossed_arms](546/previews/crossed_arms.png) | ![angry](546/previews/angry.png) | ![smile](546/previews/smile.png) | ![cry](546/previews/cry.png) | ![grin](546/previews/grin.png) | ![n_lie_0](546/previews/n_lie_0.png) | ![n_lie_1](546/previews/n_lie_1.png) | ![n_stand_0](546/previews/n_stand_0.png) | ![n_stand_1](546/previews/n_stand_1.png) | ![n_stand_2](546/previews/n_stand_2.png) | ![n_sex_0](546/previews/n_sex_0.png) | ![n_sex_1](546/previews/n_sex_1.png) | | 1365 | 16 | 0.770 | 0.964 | 0.835 | 0.771 | [Download](https://huggingface.co/CyberHarem/hornet_azurlane/resolve/main/1365/hornet_azurlane.zip) | ![pattern_0_0](1365/previews/pattern_0_0.png) | ![pattern_0_1](1365/previews/pattern_0_1.png) | ![pattern_1](1365/previews/pattern_1.png) | ![pattern_2](1365/previews/pattern_2.png) | ![pattern_3_0](1365/previews/pattern_3_0.png) | ![pattern_3_1](1365/previews/pattern_3_1.png) | ![pattern_4_0](1365/previews/pattern_4_0.png) | ![pattern_4_1](1365/previews/pattern_4_1.png) | ![pattern_5](1365/previews/pattern_5.png) | ![pattern_6](1365/previews/pattern_6.png) | ![portrait_0](1365/previews/portrait_0.png) | ![portrait_1](1365/previews/portrait_1.png) | ![portrait_2](1365/previews/portrait_2.png) | ![full_body_0](1365/previews/full_body_0.png) | ![full_body_1](1365/previews/full_body_1.png) | ![profile_0](1365/previews/profile_0.png) | ![profile_1](1365/previews/profile_1.png) | ![free_0](1365/previews/free_0.png) | ![free_1](1365/previews/free_1.png) | ![shorts](1365/previews/shorts.png) | ![maid_0](1365/previews/maid_0.png) | ![maid_1](1365/previews/maid_1.png) | ![miko](1365/previews/miko.png) | ![yukata](1365/previews/yukata.png) | ![suit](1365/previews/suit.png) | ![china](1365/previews/china.png) | ![bikini_0](1365/previews/bikini_0.png) | ![bikini_1](1365/previews/bikini_1.png) | ![bikini_2](1365/previews/bikini_2.png) | ![sit](1365/previews/sit.png) | ![squat](1365/previews/squat.png) | ![kneel](1365/previews/kneel.png) | ![jump](1365/previews/jump.png) | ![crossed_arms](1365/previews/crossed_arms.png) | ![angry](1365/previews/angry.png) | ![smile](1365/previews/smile.png) | ![cry](1365/previews/cry.png) | ![grin](1365/previews/grin.png) | ![n_lie_0](1365/previews/n_lie_0.png) | ![n_lie_1](1365/previews/n_lie_1.png) | ![n_stand_0](1365/previews/n_stand_0.png) | ![n_stand_1](1365/previews/n_stand_1.png) | ![n_stand_2](1365/previews/n_stand_2.png) | ![n_sex_0](1365/previews/n_sex_0.png) | ![n_sex_1](1365/previews/n_sex_1.png) | | 1729 | 20 | 0.763 | 0.933 | 0.831 | 0.759 | [Download](https://huggingface.co/CyberHarem/hornet_azurlane/resolve/main/1729/hornet_azurlane.zip) | ![pattern_0_0](1729/previews/pattern_0_0.png) | ![pattern_0_1](1729/previews/pattern_0_1.png) | ![pattern_1](1729/previews/pattern_1.png) | ![pattern_2](1729/previews/pattern_2.png) | ![pattern_3_0](1729/previews/pattern_3_0.png) | ![pattern_3_1](1729/previews/pattern_3_1.png) | ![pattern_4_0](1729/previews/pattern_4_0.png) | ![pattern_4_1](1729/previews/pattern_4_1.png) | ![pattern_5](1729/previews/pattern_5.png) | ![pattern_6](1729/previews/pattern_6.png) | ![portrait_0](1729/previews/portrait_0.png) | ![portrait_1](1729/previews/portrait_1.png) | ![portrait_2](1729/previews/portrait_2.png) | ![full_body_0](1729/previews/full_body_0.png) | ![full_body_1](1729/previews/full_body_1.png) | ![profile_0](1729/previews/profile_0.png) | ![profile_1](1729/previews/profile_1.png) | ![free_0](1729/previews/free_0.png) | ![free_1](1729/previews/free_1.png) | ![shorts](1729/previews/shorts.png) | ![maid_0](1729/previews/maid_0.png) | ![maid_1](1729/previews/maid_1.png) | ![miko](1729/previews/miko.png) | ![yukata](1729/previews/yukata.png) | ![suit](1729/previews/suit.png) | ![china](1729/previews/china.png) | ![bikini_0](1729/previews/bikini_0.png) | ![bikini_1](1729/previews/bikini_1.png) | ![bikini_2](1729/previews/bikini_2.png) | ![sit](1729/previews/sit.png) | ![squat](1729/previews/squat.png) | ![kneel](1729/previews/kneel.png) | ![jump](1729/previews/jump.png) | ![crossed_arms](1729/previews/crossed_arms.png) | ![angry](1729/previews/angry.png) | ![smile](1729/previews/smile.png) | ![cry](1729/previews/cry.png) | ![grin](1729/previews/grin.png) | ![n_lie_0](1729/previews/n_lie_0.png) | ![n_lie_1](1729/previews/n_lie_1.png) | ![n_stand_0](1729/previews/n_stand_0.png) | ![n_stand_1](1729/previews/n_stand_1.png) | ![n_stand_2](1729/previews/n_stand_2.png) | ![n_sex_0](1729/previews/n_sex_0.png) | ![n_sex_1](1729/previews/n_sex_1.png) | | 2821 | 32 | 0.766 | 0.904 | 0.821 | 0.748 | [Download](https://huggingface.co/CyberHarem/hornet_azurlane/resolve/main/2821/hornet_azurlane.zip) | ![pattern_0_0](2821/previews/pattern_0_0.png) | ![pattern_0_1](2821/previews/pattern_0_1.png) | ![pattern_1](2821/previews/pattern_1.png) | ![pattern_2](2821/previews/pattern_2.png) | ![pattern_3_0](2821/previews/pattern_3_0.png) | ![pattern_3_1](2821/previews/pattern_3_1.png) | ![pattern_4_0](2821/previews/pattern_4_0.png) | ![pattern_4_1](2821/previews/pattern_4_1.png) | ![pattern_5](2821/previews/pattern_5.png) | ![pattern_6](2821/previews/pattern_6.png) | ![portrait_0](2821/previews/portrait_0.png) | ![portrait_1](2821/previews/portrait_1.png) | ![portrait_2](2821/previews/portrait_2.png) | ![full_body_0](2821/previews/full_body_0.png) | ![full_body_1](2821/previews/full_body_1.png) | ![profile_0](2821/previews/profile_0.png) | ![profile_1](2821/previews/profile_1.png) | ![free_0](2821/previews/free_0.png) | ![free_1](2821/previews/free_1.png) | ![shorts](2821/previews/shorts.png) | ![maid_0](2821/previews/maid_0.png) | ![maid_1](2821/previews/maid_1.png) | ![miko](2821/previews/miko.png) | ![yukata](2821/previews/yukata.png) | ![suit](2821/previews/suit.png) | ![china](2821/previews/china.png) | ![bikini_0](2821/previews/bikini_0.png) | ![bikini_1](2821/previews/bikini_1.png) | ![bikini_2](2821/previews/bikini_2.png) | ![sit](2821/previews/sit.png) | ![squat](2821/previews/squat.png) | ![kneel](2821/previews/kneel.png) | ![jump](2821/previews/jump.png) | ![crossed_arms](2821/previews/crossed_arms.png) | ![angry](2821/previews/angry.png) | ![smile](2821/previews/smile.png) | ![cry](2821/previews/cry.png) | ![grin](2821/previews/grin.png) | ![n_lie_0](2821/previews/n_lie_0.png) | ![n_lie_1](2821/previews/n_lie_1.png) | ![n_stand_0](2821/previews/n_stand_0.png) | ![n_stand_1](2821/previews/n_stand_1.png) | ![n_stand_2](2821/previews/n_stand_2.png) | ![n_sex_0](2821/previews/n_sex_0.png) | ![n_sex_1](2821/previews/n_sex_1.png) | | 2002 | 23 | **0.777** | 0.905 | 0.809 | 0.741 | [Download](https://huggingface.co/CyberHarem/hornet_azurlane/resolve/main/2002/hornet_azurlane.zip) | ![pattern_0_0](2002/previews/pattern_0_0.png) | ![pattern_0_1](2002/previews/pattern_0_1.png) | ![pattern_1](2002/previews/pattern_1.png) | ![pattern_2](2002/previews/pattern_2.png) | ![pattern_3_0](2002/previews/pattern_3_0.png) | ![pattern_3_1](2002/previews/pattern_3_1.png) | ![pattern_4_0](2002/previews/pattern_4_0.png) | ![pattern_4_1](2002/previews/pattern_4_1.png) | ![pattern_5](2002/previews/pattern_5.png) | ![pattern_6](2002/previews/pattern_6.png) | ![portrait_0](2002/previews/portrait_0.png) | ![portrait_1](2002/previews/portrait_1.png) | ![portrait_2](2002/previews/portrait_2.png) | ![full_body_0](2002/previews/full_body_0.png) | ![full_body_1](2002/previews/full_body_1.png) | ![profile_0](2002/previews/profile_0.png) | ![profile_1](2002/previews/profile_1.png) | ![free_0](2002/previews/free_0.png) | ![free_1](2002/previews/free_1.png) | ![shorts](2002/previews/shorts.png) | ![maid_0](2002/previews/maid_0.png) | ![maid_1](2002/previews/maid_1.png) | ![miko](2002/previews/miko.png) | ![yukata](2002/previews/yukata.png) | ![suit](2002/previews/suit.png) | ![china](2002/previews/china.png) | ![bikini_0](2002/previews/bikini_0.png) | ![bikini_1](2002/previews/bikini_1.png) | ![bikini_2](2002/previews/bikini_2.png) | ![sit](2002/previews/sit.png) | ![squat](2002/previews/squat.png) | ![kneel](2002/previews/kneel.png) | ![jump](2002/previews/jump.png) | ![crossed_arms](2002/previews/crossed_arms.png) | ![angry](2002/previews/angry.png) | ![smile](2002/previews/smile.png) | ![cry](2002/previews/cry.png) | ![grin](2002/previews/grin.png) | ![n_lie_0](2002/previews/n_lie_0.png) | ![n_lie_1](2002/previews/n_lie_1.png) | ![n_stand_0](2002/previews/n_stand_0.png) | ![n_stand_1](2002/previews/n_stand_1.png) | ![n_stand_2](2002/previews/n_stand_2.png) | ![n_sex_0](2002/previews/n_sex_0.png) | ![n_sex_1](2002/previews/n_sex_1.png) | ## Anything Else? Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. ## All Steps We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * [Steps From 2821 to 3640](all/0.md) * [Steps From 1911 to 2730](all/1.md) * [Steps From 1001 to 1820](all/2.md) * [Steps From 91 to 910](all/3.md)
{"license": "mit", "tags": ["art", "not-for-all-audiences"], "datasets": ["CyberHarem/hornet_azurlane"], "pipeline_tag": "text-to-image"}
text-to-image
CyberHarem/hornet_azurlane
[ "art", "not-for-all-audiences", "text-to-image", "dataset:CyberHarem/hornet_azurlane", "license:mit", "region:us" ]
2024-02-15T02:36:48+00:00
[]
[]
TAGS #art #not-for-all-audiences #text-to-image #dataset-CyberHarem/hornet_azurlane #license-mit #region-us
Lora of hornet/ホーネット/大黄蜂 (Azur Lane) ==================================== What Is This? ------------- This is the LoRA model of waifu hornet/ホーネット/大黄蜂 (Azur Lane). How Is It Trained? ------------------ * This model is trained with HCP-Diffusion. * The auto-training framework is maintained by DeepGHS Team. * The base model used for training is deepghs/animefull-latest. * Dataset used for training is the 'stage3-p480-800' in CyberHarem/hornet\_azurlane, which contains 361 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 10, resolution is 720x720, clustering into 20 buckets. * Trained for 3640 steps, 40 checkpoints were saved and evaluated. * Trigger word is 'hornet\_azurlane'. * Pruned core tags for this waifu are 'blonde\_hair, long\_hair, green\_eyes, twintails, breasts, large\_breasts, bangs, sidelocks, very\_long\_hair, hat, cowboy\_hat, black\_headwear'. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. How to Use It? -------------- ### If You Are Using A1111 WebUI v1.7+ Just use it like the classic LoRA. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 546, you need to download '546/hornet\_azurlane.pt' as the embedding and '546/hornet\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters. Which Step Should I Use? ------------------------ We selected 5 good steps for you to choose. The best one is step 546. 1800 images (1.84 GiB) were generated for auto-testing. !Metrics Plot The base model used for generating preview images is Meina/MeinaMix\_V11. Here are the preview of the recommended steps: Anything Else? -------------- Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. All Steps --------- We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * Steps From 2821 to 3640 * Steps From 1911 to 2730 * Steps From 1001 to 1820 * Steps From 91 to 910
[ "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 546, you need to download '546/hornet\\_azurlane.pt' as the embedding and '546/hornet\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 546.\n\n\n1800 images (1.84 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 2821 to 3640\n* Steps From 1911 to 2730\n* Steps From 1001 to 1820\n* Steps From 91 to 910" ]
[ "TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/hornet_azurlane #license-mit #region-us \n", "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 546, you need to download '546/hornet\\_azurlane.pt' as the embedding and '546/hornet\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 546.\n\n\n1800 images (1.84 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 2821 to 3640\n* Steps From 1911 to 2730\n* Steps From 1001 to 1820\n* Steps From 91 to 910" ]
[ 44, 38, 471 ]
[ "passage: TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/hornet_azurlane #license-mit #region-us \n### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file." ]
[ 0.005650633480399847, -0.02317667566239834, -0.003926462959498167, 0.083487369120121, 0.06717868149280548, 0.07399158924818039, 0.23120683431625366, 0.07470877468585968, 0.1207495853304863, -0.06961455196142197, 0.08283144235610962, 0.055216483771800995, -0.006182174198329449, 0.020241769030690193, -0.026686446741223335, -0.14824791252613068, -0.06751326471567154, -0.032944124191999435, -0.006183266639709473, 0.01911224238574505, 0.07707100361585617, 0.011756956577301025, 0.09781523793935776, -0.04330062493681908, -0.0373963825404644, 0.06093563511967659, -0.026431558653712273, -0.04233365133404732, 0.04170899838209152, 0.08339741080999374, 0.12411440163850784, 0.004078404512256384, 0.057820092886686325, -0.16168519854545593, 0.06971904635429382, -0.004294062033295631, -0.11066815257072449, -0.0064407349564135075, 0.02781120501458645, -0.029432523995637894, 0.1299070417881012, 0.039202068001031876, -0.11129158735275269, 0.04298844188451767, -0.12823635339736938, -0.02616209164261818, -0.059638090431690216, 0.04747448489069939, 0.14260506629943848, 0.05633718520402908, 0.022389274090528488, 0.05759309604763985, -0.04439742490649223, 0.08657360076904297, 0.12582845985889435, -0.12673741579055786, -0.0634472593665123, 0.11118804663419724, 0.017603633925318718, 0.14033661782741547, -0.09803692251443863, 0.09280058741569519, 0.07209505885839462, -0.04370202124118805, -0.1519760936498642, -0.09652846306562424, -0.2197195291519165, -0.014667848125100136, 0.013388844206929207, 0.015197869390249252, 0.39943429827690125, 0.05112925171852112, 0.030397607013583183, 0.07006149739027023, -0.07133520394563675, 0.016476912423968315, -0.09676939994096756, 0.13027644157409668, 0.038806118071079254, 0.09627224504947662, -0.04065149277448654, -0.10604715347290039, -0.10911989212036133, -0.07452385872602463, -0.08528237789869308, -0.03035948984324932, 0.021626126021146774, 0.11121327430009842, -0.19520674645900726, 0.003135683946311474, -0.07030379772186279, -0.11648755520582199, 0.01640959456562996, -0.09081468731164932, 0.1598186045885086, 0.06394515931606293, -0.011828501708805561, -0.0017501722322776914, 0.24945835769176483, 0.11914722621440887, 0.16845017671585083, 0.05091322958469391, -0.10625970363616943, 0.1274893581867218, 0.034081943333148956, -0.10777582973241806, 0.001072225859388709, -0.12062706798315048, 0.1434556096792221, -0.058870818465948105, 0.10652414709329605, -0.06285575032234192, -0.1100640743970871, 0.01279187761247158, -0.09128142148256302, 0.0638190358877182, 0.037877585738897324, -0.006775861606001854, -0.04785702005028725, 0.04289652779698372, 0.05351148545742035, -0.036081377416849136, -0.008111461997032166, -0.01607949659228325, -0.03451118245720863, 0.0393955335021019, 0.10741056501865387, 0.041894130408763885, 0.059955403208732605, 0.006295111496001482, -0.015439752489328384, -0.00786981638520956, -0.040416743606328964, 0.0008846400305628777, 0.0409516803920269, 0.0491221621632576, 0.08882569521665573, -0.1511857956647873, -0.08011064678430557, -0.009365766309201717, 0.06331224739551544, 0.01710236631333828, 0.08482849597930908, -0.002640767954289913, 0.05423356592655182, 0.010647183284163475, -0.024095484986901283, 0.034688834100961685, -0.10287434607744217, 0.08450262993574142, -0.02652975358068943, 0.08985564857721329, -0.20711645483970642, 0.0024263765662908554, -0.044635895639657974, 0.016226639971137047, 0.050195157527923584, 0.00294245476834476, -0.10781461000442505, 0.10906048864126205, -0.013989774510264397, 0.061232637614011765, -0.10445292294025421, 0.04814012721180916, 0.02766837365925312, 0.0630776435136795, -0.10562437772750854, 0.0023706802166998386, 0.12203538417816162, -0.14172253012657166, -0.16667579114437103, 0.10008585453033447, -0.02837991528213024, 0.018582962453365326, 0.03845491632819176, 0.17437219619750977, 0.2066616714000702, -0.2087770253419876, -0.01649622991681099, 0.06093078479170799, -0.013565469533205032, -0.088918536901474, -0.00929480604827404, 0.09468480199575424, 0.012534521520137787, 0.023711631074547768, -0.02612270601093769, 0.12020915001630783, -0.023389389738440514, -0.08716810494661331, -0.032716866582632065, -0.07640048116445541, -0.0699312835931778, 0.04872513189911842, -0.012054690159857273, -0.05829243361949921, 0.024831850081682205, -0.1305171251296997, 0.1668328046798706, 0.02097274735569954, 0.023486336693167686, -0.0661667063832283, 0.11615706235170364, -0.0037158913910388947, 0.0012590609258040786, -0.00016783576575107872, -0.05494418740272522, -0.10877569019794464, 0.23273035883903503, 0.1006111353635788, 0.10307875275611877, 0.058997370302677155, -0.04003666341304779, -0.06725507229566574, 0.014755643904209137, 0.015166639350354671, -0.04201890528202057, 0.013152627274394035, -0.10921871662139893, 0.0556124709546566, -0.019864529371261597, 0.007891572080552578, 0.0031209103763103485, -0.03226698935031891, 0.07275744527578354, 0.0031719710677862167, -0.0171702578663826, 0.0826399028301239, 0.04377549886703491, -0.017217403277754784, -0.07267500460147858, 0.006973776035010815, 0.07116883993148804, -0.005475902929902077, -0.06905180960893631, 0.018907887861132622, -0.010927517898380756, 0.0568234883248806, 0.20508979260921478, -0.21387240290641785, 0.04074857756495476, 0.003779187798500061, 0.044580698013305664, 0.031230272725224495, -0.007764881942421198, -0.02635226398706436, 0.05828769877552986, -0.024920430034399033, 0.06830193102359772, -0.022331340238451958, 0.06304098665714264, -0.025908716022968292, -0.13370533287525177, -0.0032435408793389797, -0.03732586279511452, 0.1448461264371872, -0.16247545182704926, 0.06816428899765015, 0.19247418642044067, -0.12220687419176102, 0.15187454223632812, -0.0028003999032080173, -0.010144173167645931, 0.01319569069892168, 0.03327993303537369, 0.0025469420943409204, 0.10394226014614105, -0.07141224294900894, -0.0244926568120718, 0.025864994153380394, -0.08044469356536865, 0.026546048000454903, -0.12817959487438202, -0.09869913756847382, -0.06374607980251312, -0.038656316697597504, -0.03790711984038353, 0.03453190252184868, -0.050912030041217804, 0.07388190180063248, -0.09536196291446686, -0.09074699878692627, -0.026590289548039436, -0.08697901666164398, 0.016397109255194664, 0.005418791435658932, -0.06683685630559921, -0.13337333500385284, -0.1361449956893921, -0.09774305671453476, -0.1589365452528, 0.0013158466899767518, 0.07352982461452484, -0.10491493344306946, -0.03455720096826553, 0.015931176021695137, -0.04695046320557594, 0.09444509446620941, -0.06947462260723114, 0.00793424341827631, 0.047794729471206665, -0.023323100060224533, -0.17478083074092865, -0.0017463401891291142, -0.05786987766623497, -0.06174726411700249, 0.13953232765197754, -0.13916568458080292, 0.17470474541187286, -0.030191609635949135, 0.049096301198005676, 0.05995706468820572, 0.026375385001301765, 0.13415859639644623, -0.1134817972779274, 0.08146735280752182, 0.17830604314804077, 0.045593973249197006, 0.08247723430395126, 0.12871013581752777, 0.08637334406375885, -0.10093362629413605, 0.038715120404958725, 0.08204515278339386, -0.11039711534976959, -0.09295763075351715, -0.05662780627608299, -0.11683246493339539, -0.056182995438575745, 0.05629037693142891, 0.06279227137565613, 0.0833386555314064, 0.12274966388940811, -0.058614764362573624, 0.008978035300970078, 0.10101927071809769, 0.04183579981327057, 0.07535434514284134, 0.022195683792233467, 0.06115672364830971, -0.14900408685207367, -0.04233996942639351, 0.16655048727989197, 0.23158898949623108, 0.23625469207763672, 0.024057287722826004, 0.05455390363931656, 0.12556380033493042, 0.0872282013297081, 0.09885960072278976, 0.05782761797308922, -0.005338030401617289, 0.0209178663790226, -0.07221198827028275, -0.05413583293557167, 0.02200102061033249, 0.005598335526883602, -0.03447624295949936, -0.13207414746284485, 0.10005801171064377, 0.010498286224901676, 0.07313495129346848, 0.14340266585350037, 0.03663502633571625, -0.10096585750579834, 0.1480381041765213, 0.09257335215806961, 0.08149224519729614, -0.06412645429372787, 0.14425846934318542, 0.04837692528963089, -0.005623920354992151, 0.16388751566410065, 0.023004816845059395, 0.15193335711956024, -0.02413162589073181, -0.07199433445930481, -0.057932693511247635, -0.043435536324977875, 0.0017835709732025862, 0.029733577743172646, -0.21350547671318054, 0.10316440463066101, 0.05240558087825775, 0.01775478571653366, 0.0056209624744951725, -0.04942844808101654, 0.18986870348453522, 0.15097711980342865, 0.08168423920869827, 0.02934037335216999, -0.04441791772842407, 0.007814139127731323, -0.0763053447008133, 0.04457271471619606, 0.008359366096556187, 0.08157671988010406, -0.04323792830109596, -0.10729856044054031, -0.020720474421977997, -0.0021283328533172607, 0.020149661228060722, -0.07455061376094818, -0.11068344861268997, -0.05097411572933197, 0.2646317183971405, -0.060266412794589996, 0.04202900826931, 0.05426128953695297, 0.015109539963304996, -0.033730123192071915, 0.036832332611083984, -0.02913818508386612, -0.015423874370753765, -0.053873900324106216, 0.006246950011700392, 0.006289177108556032, -0.03811033442616463, -0.06118824705481529, -0.027972420677542686, -0.10639087855815887, -0.1043190211057663, 0.009558849036693573, -0.05705729126930237, 0.020947612822055817, -0.021625127643346786, 0.031469039618968964, -0.09804528206586838, -0.01819068007171154, 0.026786236092448235, 0.038371432572603226, -0.07620329409837723, -0.1342667043209076, -0.009645508602261543, -0.039274998009204865, -0.06865250319242477, 0.03198614716529846, -0.08882475644350052, -0.09757750481367111, -0.04671066999435425, -0.031424716114997864, 0.1314018815755844, 0.2184426486492157, -0.024736199527978897, 0.02970903366804123, 0.15766513347625732, -0.10413673520088196, -0.33188313245773315, -0.16377633810043335, -0.15203449130058289, -0.10411782562732697, 0.03694227337837219, -0.06898105144500732, 0.032767560333013535, 0.07209848612546921, -0.03810913860797882, 0.19123005867004395, -0.19545038044452667, -0.10659407824277878, 0.1005912572145462, 0.09182602167129517, 0.3099201023578644, -0.24158360064029694, 0.009776189923286438, -0.11614358425140381, -0.03310811519622803, 0.009659110568463802, -0.09172939509153366, 0.1213570386171341, 0.03547987714409828, 0.08359432965517044, -0.009625278413295746, -0.006687530782073736, 0.14291633665561676, -0.05908801779150963, 0.13830837607383728, -0.11497581750154495, -0.07858887314796448, 0.20277686417102814, -0.030571680516004562, -0.004787239711731672, -0.20513774454593658, -0.03070971556007862, -0.033792644739151, 0.03940649330615997, -0.0074677849188447, 0.060479115694761276, -0.005841669626533985, -0.020688964053988457, -0.13458311557769775, -0.02628062292933464, -0.02455463446676731, 0.06054951995611191, 0.22970882058143616, -0.054924167692661285, -0.05108992010354996, 0.04008102044463158, -0.001709735719487071, 0.10588634759187698, -0.021171720698475838, -0.04874242842197418, -0.04483342543244362, 0.09304806590080261, -0.19532427191734314, 0.05930668115615845, 0.010442166589200497, 0.003621891373768449, 0.009947342798113823, 0.014246624894440174, 0.02867996133863926, 0.11882852017879486, 0.1726941466331482, -0.009688196703791618, -0.04394422098994255, -0.0226314477622509, 0.03905804827809334, 0.12387416511774063, -0.018306175246834755, 0.11288506537675858, 0.01937180571258068, 0.03635832294821739, 0.0125629473477602, 0.053975578397512436, -0.09306199848651886, -0.08145494014024734, 0.09757599979639053, -0.04116271063685417, -0.08423026651144028, 0.08219940960407257, 0.04729596897959709, 0.054050568491220474, 0.005914319306612015, 0.04213394224643707, 0.02025957964360714, -0.12675704061985016, 0.027047263458371162, 0.2063998281955719, -0.08552666008472443, -0.0687454342842102, -0.05243530496954918, 0.010071132332086563, -0.11569666117429733, 0.07277298718690872, 0.03729959949851036, -0.03876679390668869, 0.10129375010728836, -0.04687846451997757, -0.036662716418504715, 0.016206463798880577, -0.060321591794490814, 0.049512967467308044, -0.15472383797168732, -0.21506857872009277, 0.042145177721977234, -0.008688242174685001, -0.06315653026103973, -0.09427366405725479, -0.0760316550731659, 0.06462062895298004, -0.1672394573688507, 0.1408080905675888, -0.06389456242322922, 0.06141505017876625, -0.03444822132587433, -0.05312303453683853, -0.10416950285434723, -0.010036314837634563, -0.050576288253068924, -0.021450987085700035, 0.06099309027194977, 0.022759612649679184, -0.12582387030124664, -0.11050257831811905, 0.06335346400737762, 0.002676207572221756, -0.010067676194012165, 0.018454182893037796, -0.06795626133680344, 0.013460655696690083, -0.22970838844776154, -0.06747334450483322, 0.08369316905736923, 0.041804708540439606, -0.08776582032442093, 0.10727212578058243, 0.046285104006528854, -0.021395079791545868, 0.037484996020793915, 0.0024103315081447363, 0.16679954528808594, -0.07273411750793457, 0.031200384721159935, -0.1321967989206314, -0.16707703471183777, -0.0341583788394928, 0.02925405092537403, 0.24156779050827026, 0.0946195051074028, 0.12574052810668945, -0.05767448619008064, 0.02264838106930256, -0.016537144780158997, 0.07074843347072601, 0.011917395517230034, -0.10235659778118134, -0.06127281114459038, -0.17760851979255676, -0.06905092298984528, -0.057009462267160416, 0.1608179658651352, 0.038572702556848526, -0.12709182500839233, -0.003780289087444544, 0.10805682092905045, -0.18091772496700287, -0.018518168479204178, 0.17871423065662384, -0.05147792398929596, 0.02028712071478367, -0.14549455046653748, 0.021745415404438972, 0.08287466317415237, -0.01667012646794319, -0.002791592851281166, 0.12268202751874924, -0.004103512968868017, -0.008283591829240322, 0.030659789219498634, -0.031707197427749634, 0.06783744692802429, -0.08166196197271347, 0.0712810680270195, -0.0012058509746566415, -0.04131850600242615, -0.10008908063173294, 0.1895044893026352, -0.031836144626140594, 0.028593778610229492, -0.059879567474126816, -0.004852105863392353, -0.1005149856209755, -0.07597308605909348, -0.0637180283665657, -0.13488996028900146, 0.06880609691143036, -0.06292682141065598, 0.02149261347949505, -0.012275110930204391, 0.02157653495669365, -0.06466672569513321, 0.03834449499845505, -0.21294713020324707, -0.041827838867902756, 0.006899907719343901, -0.018509037792682648, -0.023274056613445282, -0.04239475354552269, -0.03631700575351715, 0.027837755158543587, -0.07015104591846466, -0.06681926548480988, 0.05857250466942787, 0.0884050577878952, 0.05855568125844002, -0.16181403398513794, -0.10101468861103058, -0.0692986249923706, 0.028531357645988464, 0.07426019012928009, 0.19231273233890533, 0.04579465463757515, -0.010646497830748558, 0.04611707106232643, 0.12344201654195786, 0.00600169412791729, -0.09933613985776901, -0.06078050285577774, -0.12096293270587921, -0.13944292068481445, -0.012007853016257286, -0.06075530871748924, -0.020624857395887375, 0.022453036159276962, 0.21837753057479858, 0.20822298526763916, -0.13927145302295685, 0.04190083593130112, -0.07425857335329056, 0.03936176002025604, -0.029936768114566803, 0.1536542773246765, 0.0491301529109478, 0.1654321402311325, -0.026438580825924873, -0.03010149858891964, -0.06991671770811081, 0.016238924115896225, -0.09777514636516571, 0.04305262118577957, -0.013373976573348045, -0.06903232634067535, -0.048141419887542725, 0.10511866211891174, -0.11411741375923157, 0.03972442075610161, 0.18026982247829437, -0.14000405371189117, -0.016330266371369362, -0.039823371917009354, 0.05681310594081879, 0.11407983303070068, 0.005172284785658121, -0.07778220623731613, -0.0225257258862257, 0.01249008160084486, 0.027332330122590065, -0.1827600747346878, -0.1064288541674614, -0.003504401072859764, -0.1318865269422531, 0.12943872809410095, -0.010909314267337322, 0.007720216177403927, 0.03373350203037262, -0.06585480272769928, 0.0010327703785151243, 0.16067351400852203, 0.02549903839826584, -0.03906607627868652, -0.030222050845623016, -0.049361828714609146, -0.09871076047420502, 0.07286878675222397, 0.08199316263198853, 0.05240456387400627, -0.010310178622603416, 0.1498558670282364, -0.02311316318809986, -0.03786246106028557, 0.12948958575725555, -0.17153669893741608, 0.09390788525342941, -0.01774759404361248, -0.012433755211532116, -0.06416571140289307, -0.047873448580503464, 0.033877428621053696, 0.08830594271421432, -0.16420593857765198, -0.03903132677078247, 0.05992860719561577, -0.0912342369556427, 0.05287342891097069, 0.04186628758907318, -0.11641191691160202, 0.005319953430444002, -0.12113900482654572, 0.0003364605945535004, -0.10834424197673798, 0.05358773097395897, 0.19188743829727173, -0.034771088510751724, 0.015470720827579498, -0.11246582865715027, 0.04597301781177521, -0.032257504761219025, -0.041071128100156784, -0.07580886781215668 ]
null
null
ml-agents
# **poca** Agent playing **SoccerTwos** This is a trained model of a **poca** agent playing **SoccerTwos** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: mathreader/poca-SoccerTwos 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
{"library_name": "ml-agents", "tags": ["SoccerTwos", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-SoccerTwos"]}
reinforcement-learning
mathreader/poca-SoccerTwos
[ "ml-agents", "tensorboard", "onnx", "SoccerTwos", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-SoccerTwos", "region:us" ]
2024-02-15T02:40:25+00:00
[]
[]
TAGS #ml-agents #tensorboard #onnx #SoccerTwos #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SoccerTwos #region-us
# poca Agent playing SoccerTwos This is a trained model of a poca agent playing SoccerTwos using the Unity ML-Agents Library. ## Usage (with ML-Agents) The Documentation: URL We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your browser: URL - A *longer tutorial* to understand how works ML-Agents: URL ### Resume the training ### Watch your Agent play You can watch your agent playing directly in your browser 1. If the environment is part of ML-Agents official environments, go to URL 2. Step 1: Find your model_id: mathreader/poca-SoccerTwos 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play
[ "# poca Agent playing SoccerTwos\n This is a trained model of a poca agent playing SoccerTwos\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: mathreader/poca-SoccerTwos\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play" ]
[ "TAGS\n#ml-agents #tensorboard #onnx #SoccerTwos #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SoccerTwos #region-us \n", "# poca Agent playing SoccerTwos\n This is a trained model of a poca agent playing SoccerTwos\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: mathreader/poca-SoccerTwos\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play" ]
[ 52, 205 ]
[ "passage: TAGS\n#ml-agents #tensorboard #onnx #SoccerTwos #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SoccerTwos #region-us \n# poca Agent playing SoccerTwos\n This is a trained model of a poca agent playing SoccerTwos\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: mathreader/poca-SoccerTwos\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play" ]
[ -0.01226712018251419, -0.005746318027377129, -0.004894532263278961, 0.05903474986553192, 0.17652906477451324, -0.020982718095183372, 0.10837912559509277, 0.1027909442782402, 0.13139405846595764, 0.09756775200366974, 0.055060695856809616, 0.06170285865664482, 0.07298478484153748, 0.1303529441356659, 0.0716925710439682, -0.1488337516784668, -0.03142900764942169, -0.10802070051431656, 0.034184351563453674, 0.05899312347173691, 0.07253029942512512, -0.051882483065128326, 0.05681697279214859, 0.03725259378552437, -0.0689048245549202, 0.00684483302757144, -0.05814262852072716, -0.05138517543673515, 0.015378674492239952, 0.015215292572975159, 0.011984741315245628, -0.06337209045886993, 0.09210718423128128, -0.17212331295013428, 0.020721353590488434, 0.04002023860812187, -0.008452385663986206, -0.06229200214147568, 0.1276315450668335, 0.043351903557777405, 0.10438428074121475, -0.07030325382947922, 0.10360065847635269, 0.0424841046333313, -0.08258668333292007, 0.0824010968208313, -0.1089855208992958, -0.006745335645973682, 0.21465235948562622, 0.14936475455760956, 0.010182777419686317, 0.09396836161613464, -0.0387948714196682, 0.007498238235712051, 0.14487196505069733, -0.27726438641548157, -0.07939694821834564, 0.13027547299861908, -0.02145846001803875, 0.08833801001310349, -0.028231414034962654, 0.04201333597302437, -0.009912663139402866, 0.03272752836346626, -0.013224991038441658, 0.018338391557335854, 0.212917298078537, -0.012817560695111752, -0.04448915645480156, -0.13246968388557434, 0.0061857933178544044, 0.04584157466888428, -0.05280997231602669, -0.16853296756744385, 0.02943428047001362, 0.10165306180715561, -0.04925490543246269, 0.014410346746444702, 0.07099152356386185, 0.006620672065764666, -0.042715251445770264, -0.11542780697345734, -0.04367780685424805, -0.04683765769004822, 0.019819997251033783, 0.10760295391082764, -0.03325663506984711, -0.05027620121836662, 0.06097789481282234, 0.08764400333166122, 0.06268270313739777, -0.04329489916563034, -0.015519820153713226, 0.006696786731481552, -0.16380831599235535, -0.08631724864244461, -0.028309321030974388, -0.04418984800577164, 0.041483260691165924, 0.11256884783506393, 0.11081676185131073, 0.004338396713137627, 0.009150825440883636, 0.061725594103336334, -0.012902549467980862, 0.06980054825544357, 0.0028522571083158255, 0.012090257368981838, 0.02819339744746685, 0.03626690059900284, 0.03746730089187622, -0.09237686544656754, -0.0949978455901146, 0.09882912039756775, -0.11034448444843292, 0.11206621676683426, 0.12235268205404282, -0.0267629437148571, -0.03494509682059288, -0.05074550583958626, 0.003843884449452162, -0.14146773517131805, 0.08371923863887787, 0.03303227201104164, -0.04981617629528046, -0.10836852341890335, -0.046147286891937256, 0.03394258767366409, -0.08627910912036896, 0.01887141540646553, -0.014409005641937256, 0.06029488518834114, -0.027241650968790054, -0.029249709099531174, 0.07890737056732178, -0.1128319576382637, -0.013247347436845303, -0.1515287607908249, -0.09656493365764618, -0.08996892720460892, 0.04666753113269806, -0.09948772937059402, -0.09887849539518356, -0.08679677546024323, 0.019346369430422783, -0.08547431975603104, 0.03922509774565697, -0.057943232357501984, -0.06756136566400528, -0.008376647718250751, -0.0640132799744606, 0.08049138635396957, 0.08408407866954803, 0.04878107085824013, -0.03937412053346634, 0.029872553423047066, -0.20567715167999268, 0.16055412590503693, -0.10186789184808731, 0.13675880432128906, -0.07645774632692337, 0.0838986188173294, 0.022465040907263756, 0.026301000267267227, 0.07142274081707001, 0.12616604566574097, -0.06325618177652359, -0.10014146566390991, 0.1260327249765396, -0.05492835119366646, -0.1828690618276596, 0.0668938010931015, 0.03159705549478531, 0.07407499104738235, 0.050498560070991516, 0.2275787889957428, 0.18897472321987152, -0.2812601327896118, 0.09412664920091629, 0.002852521138265729, -0.10588090121746063, -0.013594900257885456, 0.1085233986377716, -0.10343027114868164, 0.06901288032531738, -0.04072486609220505, -0.1810113787651062, 0.14027145504951477, -0.03158870339393616, -0.06755724549293518, 0.04458579793572426, -0.07635199278593063, -0.06034095957875252, 0.010811394080519676, 0.05002839118242264, -0.04206789284944534, -0.005107611417770386, 0.011459735222160816, 0.03514920547604561, -0.02445279434323311, 0.04466353356838226, -0.07713731378316879, 0.15146160125732422, -0.030595751479268074, 0.020160408690571785, -0.15015585720539093, -0.13933275640010834, 0.004976681433618069, 0.08274252712726593, 0.10293864458799362, -0.0960969403386116, 0.04558054730296135, 0.09184687584638596, 0.03028765320777893, -0.06409579515457153, -0.10769177228212357, 0.006067073903977871, -0.06066508963704109, -0.10513833165168762, -0.036580752581357956, -0.055650852620601654, 0.09350579977035522, -0.10942882299423218, 0.05049751326441765, -0.10463151335716248, 0.09107942134141922, -0.005507107824087143, -0.055387239903211594, -0.01272506918758154, 0.036271803081035614, 0.05017339065670967, -0.07834452390670776, 0.11126831918954849, 0.04296024516224861, -0.06192056089639664, 0.0281904935836792, 0.014511464163661003, -0.0438787117600441, 0.1251068264245987, 0.033049747347831726, -0.016845108941197395, 0.01663798838853836, -0.04071010649204254, 0.002407694235444069, -0.11795562505722046, -0.034569356590509415, 0.18131645023822784, 0.09980955719947815, 0.1173878088593483, -0.08714602142572403, -0.024689627811312675, 0.018574727699160576, -0.06502856314182281, -0.04868112504482269, 0.0743192657828331, 0.054217271506786346, -0.03042861446738243, 0.04761840030550957, 0.04552142694592476, 0.117622010409832, 0.13927343487739563, -0.0031882498878985643, -0.11663798242807388, 0.030110985040664673, 0.12299399822950363, 0.03347582742571831, 0.003102536080405116, 0.020240487530827522, -0.045143693685531616, -0.01541670598089695, -0.03277651593089104, -0.030123431235551834, -0.10606423765420914, -0.0685635656118393, 0.05701896548271179, -0.015294731594622135, 0.010999321937561035, -0.043369319289922714, -0.010131738148629665, 0.07561443001031876, 0.08474478125572205, -0.00032775342697277665, 0.01211236696690321, -0.05191362276673317, -0.12862372398376465, 0.06168781965970993, -0.08136362582445145, -0.20405526459217072, -0.1237589493393898, -0.05251871794462204, -0.0705445185303688, 0.04697805643081665, 0.06600358337163925, -0.1063452661037445, 0.012575731612741947, -0.0960071012377739, -0.03958818316459656, 0.04874784126877785, -0.07307896018028259, 0.1913866549730301, 0.1013336181640625, -0.014447901397943497, -0.07427322864532471, -0.013373201712965965, 0.00936065148562193, -0.09526906162500381, -0.013190865516662598, 0.007682847790420055, 0.13453561067581177, 0.10854537785053253, 0.003683822462335229, 0.05123621225357056, -0.03293889760971069, 0.10679891705513, -0.09239130467176437, 0.008774252608418465, 0.07371202111244202, -0.0031455736607313156, 0.0883711576461792, 0.03290828689932823, 0.0322544127702713, -0.023302504792809486, 0.040202271193265915, 0.013933992944657803, -0.06744484603404999, -0.2006303369998932, -0.12152828276157379, -0.0220841895788908, 0.10253938287496567, 0.11820917576551437, 0.07340040057897568, -0.07036536186933517, 0.0014882752439007163, 0.008603732101619244, -0.07112067192792892, 0.13719682395458221, 0.11738196015357971, -0.07218950986862183, -0.010064477100968361, 0.015614494681358337, -0.0613895058631897, 0.014479543082416058, 0.0876862183213234, -0.029801225289702415, 0.09589127451181412, 0.06028527393937111, 0.02509845234453678, 0.02296304702758789, -0.09831957519054413, -0.08150812238454819, 0.11881265789270401, 0.0473497211933136, 0.0007999254739843309, -0.029042672365903854, -0.059506259858608246, -0.0708668902516365, 0.06676384806632996, 0.13542945683002472, -0.06069793179631233, -0.1418142020702362, 0.0853998139500618, 0.10542146116495132, 0.18975824117660522, -0.002521968213841319, -0.14192232489585876, -0.04454575479030609, -0.00815159734338522, -0.1020878329873085, 0.016391463577747345, -0.0014126986498013139, 0.04785284399986267, -0.16071763634681702, 0.039481766521930695, 0.06135094165802002, 0.12946443259716034, 0.029116010293364525, -0.009517540223896503, 0.03355773910880089, 0.026819372549653053, -0.0111848721280694, 0.042937297374010086, -0.13093037903308868, 0.048090629279613495, -0.009138478897511959, 0.10814440995454788, -0.05116758868098259, 0.005678110755980015, 0.057068731635808945, -0.03502454608678818, 0.15971489250659943, 0.06462935358285904, -0.02968773804605007, -0.16861683130264282, -0.11803159862756729, -0.08282005041837692, -0.013800596818327904, -0.07612379640340805, 0.08477037400007248, 0.017523502930998802, -0.015191911719739437, -0.09796226769685745, 0.05008696764707565, -0.028106577694416046, -0.0950402021408081, -0.028666887432336807, -0.07829011231660843, 0.03819037973880768, -0.04261692985892296, -0.008879177272319794, -0.0746157243847847, 0.16781775653362274, 0.1243710145354271, -0.04749297723174095, -0.08565152436494827, 0.009445455856621265, -0.0915859118103981, -0.025704532861709595, 0.04806908220052719, 0.017527444288134575, 0.09150344878435135, -0.11194351315498352, 0.0014112445060163736, 0.006861518602818251, -0.11402852088212967, -0.04089431092143059, -0.007167094852775335, 0.17889608442783356, 0.04266286641359329, 0.04268713295459747, 0.02729089744389057, 0.03658423572778702, 0.02382800541818142, -0.09164857864379883, 0.1637999713420868, 0.15893897414207458, -0.04773499444127083, 0.040300723165273666, -0.03012506477534771, 0.010322562418878078, -0.06879031658172607, -0.028509587049484253, 0.1978452205657959, 0.2608974874019623, -0.05886983126401901, 0.19731883704662323, 0.01589803211390972, -0.09286101162433624, -0.19259892404079437, -0.04529835656285286, 0.0518985353410244, -0.03733350336551666, 0.15250547230243683, -0.13953229784965515, 0.07488813251256943, 0.025387194007635117, -0.0024160563480108976, -0.005533088929951191, -0.2143835425376892, -0.08810751140117645, 0.0004242504364810884, 0.08109671622514725, -0.021137353032827377, -0.08207269012928009, -0.05655297636985779, -0.01134414505213499, -0.20327024161815643, 0.0530456118285656, -0.1519928127527237, 0.048144131898880005, 0.028468720614910126, 0.030781863257288933, 0.05726462975144386, -0.011954660527408123, 0.15007081627845764, -0.00858539529144764, -0.03966827318072319, -0.05811900272965431, 0.004099761135876179, 0.09150728583335876, -0.07372235506772995, 0.062378980219364166, 0.07153847068548203, -0.0350908525288105, -0.2126406580209732, -0.0077478354796767235, -0.005655392073094845, 0.018163882195949554, -0.034415703266859055, 0.013080869801342487, 0.0008334999438375235, 0.0710378959774971, 0.08401238173246384, 0.04772563651204109, 0.08905669301748276, -0.019092371687293053, -0.0046347458846867085, 0.07620229572057724, 0.09256819635629654, 0.04535599425435066, -0.10080385953187943, -0.051723118871450424, -0.06401248276233673, 0.010569521225988865, -0.04260167479515076, 0.011139703914523125, 0.04726839438080788, 0.032714370638132095, -0.026635434478521347, 0.04503455385565758, -0.09823428094387054, 0.027398686856031418, 0.06421341747045517, -0.020140929147601128, -0.08972740173339844, -0.0585971437394619, -0.03857392817735672, 0.018681194633245468, -0.11123058199882507, 0.0536271333694458, -0.019343065097928047, -0.016204455867409706, 0.04566416144371033, -0.007608185987919569, -0.048748936504125595, 0.028907611966133118, -0.023948896676301956, 0.026264892891049385, -0.05003083869814873, 0.16075630486011505, 0.026797134429216385, -0.08022381365299225, 0.01645229198038578, 0.15062426030635834, -0.10534802079200745, -0.0842362493276596, -0.020551161840558052, 0.08149077743291855, 0.040152836591005325, -0.02682304009795189, -0.0011450625024735928, -0.08204767107963562, 0.10275260359048843, -0.07473482191562653, -0.015928836539387703, -0.11864254623651505, 0.04400157555937767, 0.06374483555555344, -0.028268765658140182, 0.08488325029611588, 0.009243501350283623, -0.043634094297885895, -0.06676051765680313, 0.015962552279233932, 0.03320987522602081, 0.10495772212743759, -0.0077238366939127445, -0.03543270006775856, -0.15824048221111298, 0.03210821747779846, -0.03268231451511383, -0.015729503706097603, -0.17256566882133484, -0.00898510217666626, -0.02527804486453533, 0.02674926072359085, 0.03381733596324921, 0.03410511463880539, -0.057740602642297745, -0.07948723435401917, -0.04642842337489128, 0.12327951192855835, -0.07242128252983093, -0.01674543507397175, -0.02103704772889614, -0.05384449288249016, 0.056706514209508896, 0.06659272313117981, 0.00573004549369216, -0.028833450749516487, -0.10698968172073364, -0.005927316844463348, -0.03597713261842728, -0.05157899856567383, 0.07233209908008575, -0.14296716451644897, 0.04668724909424782, -0.014253860339522362, -0.09637690335512161, 0.028421776369214058, 0.11106366664171219, -0.052924443036317825, 0.09970445185899734, 0.03264417499303818, -0.11330494284629822, -0.08598890900611877, 0.031940482556819916, 0.09737661480903625, 0.04820340499281883, 0.06837639957666397, -0.10290943086147308, 0.16079244017601013, -0.1379258781671524, -0.011445960961282253, 0.0077020227909088135, 0.06346865743398666, -0.0032363831996917725, -0.14120575785636902, 0.03269597887992859, -0.011859850026667118, 0.06956673413515091, 0.09377656877040863, 0.07321697473526001, 0.027490705251693726, 0.017586763948202133, 0.12671339511871338, 0.03945005312561989, 0.059142742305994034, -0.035519376397132874, 0.022026192396879196, 0.07610165327787399, 0.0046011474914848804, 0.03455747663974762, -0.09418950974941254, 0.09783066809177399, 0.07527720928192139, 0.10980639606714249, 0.053006045520305634, 0.0626685693860054, -0.0750739797949791, -0.16305716335773468, -0.04144623875617981, 0.0692954957485199, -0.0303922351449728, -0.06056739389896393, 0.13169142603874207, 0.15780417621135712, -0.24571602046489716, 0.0489719957113266, -0.019154498353600502, 0.05173574760556221, -0.05169704556465149, -0.08190038800239563, 0.019741196185350418, -0.19527272880077362, 0.06655962765216827, -0.053181860595941544, 0.008648998104035854, -0.08883757889270782, -0.015509863384068012, 0.008882070891559124, 0.0701121836900711, -0.08245968073606491, -0.06396008282899857, 0.07723700255155563, -0.04085921123623848, 0.06111627072095871, -0.07602959871292114, -0.033276014029979706, -0.04449925571680069, -0.04226778447628021, 0.00018474254466127604, 0.08662362396717072, 0.0091784643009305, 0.052489884197711945, -0.05359077453613281, -0.07033034414052963, 0.08332683145999908, -0.017171667888760567, 0.0020489785820245743, 0.11058209836483002, 0.08183114230632782, -0.09119584411382675, -0.03227199614048004, 0.16845138370990753, -0.05287466198205948, -0.06749428063631058, -0.08087927848100662, 0.14692802727222443, -0.0047566257417202, -0.004226958844810724, -0.017901858314871788, -0.15032850205898285, -0.0460156612098217, 0.20681925117969513, 0.10558823496103287, -0.03170701488852501, 0.01563154347240925, -0.06126134842634201, 0.008936386555433273, 0.025433015078306198, 0.10141222178936005, 0.046797361224889755, 0.09990613907575607, -0.08546845614910126, 0.004189658910036087, -0.06495550274848938, -0.06411591917276382, -0.1557980179786682, 0.045626211911439896, 0.04572034254670143, -0.003899072762578726, -0.03847186267375946, 0.12998704612255096, -0.10588870197534561, -0.0761663094162941, 0.1519303172826767, -0.07538477331399918, -0.05524703115224838, -0.019603470340371132, -0.034806303679943085, 0.033754654228687286, 0.09875240921974182, 0.05805382505059242, 0.041081055998802185, 0.061336442828178406, -0.017209090292453766, -0.07237852364778519, -0.023415720090270042, 0.03202846273779869, -0.11685533076524734, 0.19504134356975555, -0.034751228988170624, 0.03300552815198898, 0.05918801948428154, 0.07132086157798767, -0.11961985379457474, 0.025354333221912384, 0.035550545901060104, -0.08317245543003082, 0.04372725263237953, 0.02118765003979206, -0.06069882959127426, 0.03389103338122368, 0.07349816709756851, -0.054767340421676636, 0.020215461030602455, 0.07484057545661926, 0.002525019459426403, -0.053082726895809174, 0.09499107301235199, -0.1360720694065094, 0.10494062304496765, 0.09439308941364288, -0.05259300395846367, 0.03268912062048912, -0.009081338532269001, 0.050656113773584366, 0.029477005824446678, 0.08890054374933243, -0.0416448675096035, -0.14511547982692719, 0.012881864793598652, 0.030177487060427666, 0.03264656290411949, -0.2436666488647461, -0.09194885194301605, -0.01678590103983879, -0.050015371292829514, -0.024131374433636665, 0.11869364976882935, 0.10189631581306458, -0.05173017084598541, -0.023391202092170715, -0.19716443121433258, 0.04058321937918663, 0.19753305613994598, -0.03515293076634407, -0.018165014684200287 ]
null
null
transformers
# Model Trained Using AutoTrain This model was trained using AutoTrain. For more information, please visit [AutoTrain](https://hf.co/docs/autotrain). # Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_path = "PATH_TO_THIS_REPO" tokenizer = AutoTokenizer.from_pretrained(model_path) model = AutoModelForCausalLM.from_pretrained( model_path, device_map="auto", torch_dtype='auto' ).eval() # Prompt content: "hi" messages = [ {"role": "user", "content": "hi"} ] input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors='pt') output_ids = model.generate(input_ids.to('cuda')) response = tokenizer.decode(output_ids[0][input_ids.shape[1]:], skip_special_tokens=True) # Model response: "Hello! How can I assist you today?" print(response) ```
{"license": "other", "tags": ["autotrain", "text-generation"], "widget": [{"text": "I love AutoTrain because "}]}
text-generation
adarshheg/llama-7b-chat-finetuned-4bit-std
[ "transformers", "safetensors", "llama", "text-generation", "autotrain", "conversational", "license:other", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T02:40:46+00:00
[]
[]
TAGS #transformers #safetensors #llama #text-generation #autotrain #conversational #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Trained Using AutoTrain This model was trained using AutoTrain. For more information, please visit AutoTrain. # Usage
[ "# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.", "# Usage" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #autotrain #conversational #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.", "# Usage" ]
[ 60, 29, 3 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #autotrain #conversational #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.# Usage" ]
[ -0.019985521212220192, 0.041816771030426025, -0.0015093530528247356, 0.04983396455645561, 0.12702052295207977, -0.03985721990466118, 0.2503778636455536, 0.05486364662647247, -0.07805269211530685, -0.09162702411413193, 0.17846965789794922, 0.1966569423675537, -0.04739704728126526, 0.1650119423866272, -0.049305956810712814, -0.24425874650478363, 0.0308239683508873, -0.027378594502806664, 0.07924526929855347, 0.1134583130478859, 0.1459313929080963, -0.069073885679245, 0.07192834466695786, 0.03350326791405678, -0.19196271896362305, 0.03607437387108803, 0.06325650960206985, -0.13595952093601227, 0.1781395524740219, 0.0730607882142067, 0.09795299917459488, 0.05176535248756409, 0.11600667238235474, -0.13303197920322418, 0.02148369513452053, 0.0012429648777469993, -0.01376256626099348, 0.07068803161382675, 0.059910088777542114, -0.05214822664856911, 0.0754503607749939, 0.16066469252109528, 0.10474828630685806, 0.04625413194298744, -0.10307525843381882, 0.03606079891324043, -0.007664162199944258, 0.01607259176671505, 0.11656992882490158, 0.11167781800031662, -0.011104828678071499, 0.1692391037940979, -0.12352225184440613, 0.0927673876285553, -0.06312526017427444, -0.27212998270988464, -0.01953640766441822, 0.1782287359237671, 0.05536120757460594, 0.003713849000632763, -0.11036768555641174, 0.08918768912553787, 0.10817500948905945, -0.013457630760967731, 0.07449526339769363, -0.02930355817079544, -0.0747155100107193, -0.010549373924732208, -0.07913186401128769, 0.021456321701407433, 0.19510410726070404, -0.07658466696739197, -0.027360891923308372, -0.12176856398582458, -0.028924891725182533, 0.019943704828619957, -0.0005201206658966839, -0.09924954175949097, -0.020720934495329857, 0.08925458788871765, -0.03742996230721474, -0.03304271399974823, -0.12941165268421173, -0.05689829960465431, -0.10432478040456772, 0.080483578145504, 0.004176578018814325, -0.010131161659955978, -0.10812576860189438, 0.10831553488969803, 0.036837704479694366, -0.10601683706045151, 0.04897525534033775, -0.08127395063638687, 0.033307481557130814, -0.08884010463953018, -0.020026516169309616, -0.12393411993980408, 0.03442299738526344, 0.1911761611700058, 0.17161144316196442, -0.007555702235549688, -0.09846625477075577, 0.037637125700712204, 0.009408433921635151, 0.11369463801383972, 0.046487849205732346, -0.045776110142469406, 0.06709876656532288, -0.04264026880264282, -0.007700521033257246, -0.0333063006401062, -0.19418306648731232, 0.0378468781709671, 0.021328622475266457, 0.07297981530427933, -0.06390932202339172, 0.10803848505020142, -0.02806558459997177, 0.036218613386154175, 0.013528588227927685, -0.050648901611566544, 0.027312688529491425, -0.06237490847706795, -0.0030069053173065186, -0.05522473528981209, 0.05534237623214722, 0.11048513650894165, 0.02016965113580227, 0.09672053903341293, -0.07597221434116364, -0.042308881878852844, -0.10714275389909744, -0.0614769272506237, 0.004838209133595228, 0.041153065860271454, 0.04866805672645569, -0.20337146520614624, -0.28446313738822937, -0.01588602364063263, 0.0521208830177784, -0.021559109911322594, -0.08046071231365204, -0.10747288912534714, 0.00881670881062746, 0.04961434006690979, -0.041250426322221756, 0.050605256110429764, -0.0236679557710886, 0.03517379239201546, -0.05794162675738335, 0.028720907866954803, -0.06626300513744354, 0.022746717557311058, -0.13074947893619537, -0.012705030851066113, -0.005872172769159079, 0.0556497685611248, -0.029873713850975037, 0.14915479719638824, -0.01976943202316761, 0.04168194904923439, -0.032731082290410995, 0.06638596206903458, 0.011397413909435272, 0.15701933205127716, -0.14462588727474213, -0.026569174602627754, 0.15484152734279633, -0.10750064253807068, -0.12428124994039536, 0.10975891351699829, -0.10399746894836426, 0.2686645984649658, 0.12837029993534088, 0.11094734072685242, 0.052566420286893845, -0.09157175570726395, 0.10807830095291138, 0.01006117183715105, -0.07926932722330093, -0.033996015787124634, -0.0030480532441288233, 0.021757712587714195, -0.2016487568616867, 0.042763303965330124, 0.1273248940706253, 0.07622935622930527, -0.042705241590738297, -0.08762288838624954, -0.013048309832811356, -0.06594497710466385, 0.05332009866833687, -0.0165407694876194, 0.12763361632823944, -0.06454508751630783, -0.03124934248626232, 0.06888130307197571, 0.05346184968948364, 0.03705662116408348, -0.04548288881778717, -0.09575856477022171, -0.036537256091833115, -0.02627917379140854, 0.02103262208402157, -0.08334317058324814, -0.058662403374910355, -0.029133448377251625, 0.1217503622174263, 0.06030846759676933, 0.08229882270097733, 0.02534819208085537, 0.042578503489494324, -0.018014518544077873, 0.01894560270011425, 0.18454627692699432, 0.032461415976285934, -0.12112367898225784, -0.11247295141220093, 0.11623511463403702, -0.07412275671958923, 0.1516924351453781, -0.23376472294330597, 0.03449954092502594, -0.10606535524129868, 0.08345958590507507, 0.0017077438533306122, 0.08513399958610535, -0.07354991883039474, 0.028825750574469566, -0.10780332237482071, 0.006710950750857592, 0.06633526086807251, 0.035992223769426346, -0.053377557545900345, 0.15349788963794708, -0.16296741366386414, 0.2517978847026825, 0.12181705981492996, -0.13286034762859344, -0.08410528302192688, -0.10299656540155411, 0.008393438532948494, -0.014958925545215607, -0.09269360452890396, -0.0077692181803286076, 0.11051994562149048, -0.0386928990483284, 0.19442617893218994, -0.022682785987854004, -0.023846805095672607, -0.018296057358384132, -0.0968841016292572, -0.0053712609224021435, 0.01993531733751297, 0.09277937561273575, -0.19608037173748016, 0.13585902750492096, 0.13785377144813538, -0.03008674643933773, 0.19921445846557617, 0.03519749641418457, 0.027803806588053703, 0.007066489662975073, -0.05108575150370598, 0.004615103360265493, -0.016708610579371452, -0.033214226365089417, -0.03719793260097504, 0.015435577370226383, -0.0017268992960453033, 0.032471369951963425, -0.1292077600955963, -0.043498411774635315, 0.020972639322280884, 0.04506610706448555, 0.04959072545170784, 0.06424178183078766, -0.08114660531282425, 0.08710927516222, -0.04135018214583397, -0.15011906623840332, 0.12193439155817032, 0.009275504387915134, -0.11306660622358322, 0.16927401721477509, -0.08475425839424133, -0.23736794292926788, -0.20089298486709595, -0.1772872656583786, -0.019119156524538994, 0.07307472079992294, 0.06685910373926163, -0.06804469972848892, -0.07033377140760422, -0.015055990777909756, -0.06428932398557663, 0.01824975572526455, -0.017032640054821968, -0.08446785062551498, 0.0451301671564579, -0.013684476725757122, -0.11405912786722183, -0.04632818326354027, 0.014144818298518658, -0.06867410987615585, 0.0687858834862709, -0.058338385075330734, 0.06045600399374962, 0.15961861610412598, -0.019494086503982544, 0.0206222515553236, -0.03382864594459534, 0.14078357815742493, -0.07376300543546677, -0.0012525641359388828, 0.11309554427862167, -0.062387753278017044, 0.028928736224770546, 0.20800267159938812, 0.022075070068240166, -0.08314149081707001, 0.08971244096755981, -0.030545884743332863, -0.06756961345672607, -0.20312197506427765, -0.11030048131942749, -0.008868525736033916, 0.018859686329960823, 0.07774554193019867, 0.054550036787986755, 0.27576303482055664, 0.1262982189655304, 0.06817155331373215, 0.06489472836256027, 0.02961319498717785, 0.0905478373169899, 0.17955242097377777, -0.04619118198752403, 0.18362373113632202, -0.06867803633213043, -0.18565845489501953, 0.038355518132448196, -0.013611708767712116, 0.05560639500617981, 0.1625892072916031, -0.010438076220452785, 0.03667459264397621, 0.07544669508934021, 0.1377580612897873, 0.11926194280385971, 0.07736333459615707, -0.053890686482191086, -0.0123245595023036, -0.01958155632019043, -0.06044450402259827, 0.12492989748716354, -0.05211121216416359, -0.054029930382966995, -0.024178169667720795, 0.05096742510795593, 0.04076123610138893, 0.0885038748383522, 0.0038742104079574347, -0.2888331115245819, 0.029169974848628044, 0.046290311962366104, -0.07681751251220703, -0.0919661894440651, 0.09150734543800354, -0.013760070316493511, -0.16257907450199127, 0.015721965581178665, -0.028128916397690773, 0.09203445166349411, -0.0021014101803302765, 0.07018017023801804, -0.09545313566923141, -0.03008081018924713, -0.042241763323545456, 0.14214786887168884, -0.3904902935028076, 0.20279116928577423, -0.013757672160863876, 0.042110588401556015, -0.11073333024978638, 0.007113968953490257, 0.08234450221061707, 0.16876806318759918, 0.10588139295578003, -0.06181146577000618, -0.12684638798236847, -0.10665581375360489, -0.10000576823949814, -0.002122523495927453, 0.02025875821709633, -0.01400430966168642, 0.02576049603521824, -0.11316744238138199, -0.006981417536735535, 0.045942071825265884, 0.00005032867193222046, -0.1327565312385559, -0.1645476371049881, 0.0024751632008701563, 0.05992445722222328, 0.12014522403478622, -0.031115030869841576, -0.08387571573257446, -0.09976387768983841, 0.17837554216384888, 0.05166070535778999, -0.0018781357211992145, -0.12678098678588867, -0.04034697636961937, -0.05093073844909668, -0.028619080781936646, 0.07061373442411423, 0.0115804523229599, 0.11983727663755417, -0.0776883065700531, -0.08326467871665955, 0.10231561213731766, -0.11192917823791504, -0.05676306411623955, -0.10391143709421158, 0.037218786776065826, -0.03583132475614548, -0.00300496444106102, 0.10846041887998581, 0.03224887326359749, -0.06038367375731468, -0.06040918827056885, -0.030626386404037476, 0.002953569171950221, -0.024285854771733284, -0.10091901570558548, -0.12347328662872314, -0.11263954639434814, -0.02959991805255413, -0.12120452523231506, 0.22577744722366333, 0.14880536496639252, -0.08410761505365372, 0.14291362464427948, 0.19940108060836792, -0.11654946208000183, -0.31744757294654846, -0.06689527630805969, -0.0593874417245388, 0.013890923000872135, 0.03878229483962059, -0.12954181432724, 0.1002892553806305, 0.01007110346108675, -0.07845258712768555, -0.02825743705034256, -0.14488260447978973, -0.1637139469385147, 0.2514044940471649, 0.017858991399407387, 0.2469845563173294, -0.09900328516960144, -0.0562664233148098, -0.15211105346679688, 0.03663777932524681, 0.07304413616657257, -0.07589749246835709, 0.07363534718751907, 0.04355068877339363, 0.08316291868686676, 0.03649081289768219, -0.01850447990000248, 0.052836280316114426, -0.054001953452825546, 0.0728193074464798, -0.16864342987537384, -0.03414404019713402, 0.027340056374669075, -0.022704705595970154, 0.10976030677556992, -0.0730476900935173, 0.02634204924106598, -0.030788259580731392, -0.07285180687904358, 0.03122970648109913, 0.06520438939332962, -0.0009521509637124836, -0.1139688566327095, 0.0039772349409759045, -0.012743611820042133, 0.010530095547437668, -0.05437423661351204, 0.0461973138153553, -0.04407922551035881, 0.13509979844093323, 0.1618819236755371, 0.23810654878616333, -0.05280663073062897, 0.09310457110404968, -0.033480700105428696, -0.11406292766332626, 0.08189824223518372, -0.08512503653764725, 0.032127875834703445, 0.07843150943517685, -0.044101208448410034, 0.16950587928295135, 0.05329832062125206, 0.014960072003304958, -0.01104823499917984, 0.15519900619983673, -0.16232933104038239, 0.024055518209934235, -0.08237790316343307, 0.12376146763563156, 0.05088603496551514, 0.0014241976896300912, 0.1342407464981079, -0.09748697280883789, -0.014323112554848194, 0.022257447242736816, 0.007675370201468468, -0.033624421805143356, 0.10129436105489731, 0.04010987654328346, 0.014930672012269497, -0.07691510766744614, 0.03148246556520462, 0.0725991502404213, 0.015066240914165974, 0.04953509941697121, 0.0274610985070467, -0.0900094285607338, -0.1050470843911171, 0.005259819328784943, 0.2579692006111145, -0.19413785636425018, -0.08945966511964798, -0.025965580716729164, -0.12841492891311646, 0.014765934087336063, 0.11006096005439758, 0.07310435175895691, 0.044437263160943985, -0.06245476007461548, -0.02698042243719101, -0.11322736740112305, 0.10147830843925476, 0.009620734490454197, 0.04834596440196037, -0.1538265496492386, 0.08177132904529572, -0.029066724702715874, 0.00937450211495161, -0.09466302394866943, -0.042740534991025925, -0.1231655478477478, 0.03438199684023857, -0.1424497365951538, -0.03872010484337807, -0.03764064237475395, -0.0057784151285886765, 0.060039203613996506, -0.010813248343765736, -0.026481660082936287, -0.03197747841477394, -0.09286012500524521, 0.03200497850775719, -0.00152523850556463, 0.05066218599677086, -0.047269031405448914, -0.034692924469709396, 0.0329509861767292, -0.007188139017671347, 0.0626831129193306, 0.005206009838730097, -0.02624976634979248, 0.05140461027622223, -0.1694229692220688, 0.02440795861184597, 0.082680344581604, 0.006990185473114252, 0.028766458854079247, -0.03343239799141884, -0.007944096811115742, 0.10507944226264954, 0.045716822147369385, 0.04421989992260933, -0.0016878187889233232, -0.09214121103286743, 0.04705626890063286, 0.06632386893033981, -0.12569046020507812, -0.02528555504977703, -0.032815009355545044, -0.0014488169690594077, -0.03754571080207825, 0.21374857425689697, -0.12118947505950928, 0.04649411514401436, -0.040373969823122025, 0.03989865258336067, -0.03320062533020973, -0.133179172873497, -0.09410917013883591, -0.12265769392251968, -0.029428137466311455, -0.006662990897893906, 0.26298657059669495, 0.1437089890241623, -0.026487676426768303, 0.03210904821753502, 0.06094885990023613, 0.06782585382461548, 0.017442740499973297, 0.1971471905708313, 0.11334112286567688, 0.022985318675637245, -0.12867672741413116, 0.071586012840271, 0.04378187656402588, -0.05412180349230766, -0.0045047118328511715, -0.009332980960607529, -0.0896068587899208, 0.06856201589107513, 0.07274024933576584, -0.017855705693364143, -0.07965993136167526, -0.14463277161121368, -0.1213601604104042, 0.037436988204717636, -0.09540953487157822, 0.01893600821495056, 0.1667509526014328, -0.04064670577645302, -0.010678790509700775, -0.05300956964492798, -0.04512229934334755, -0.2152797430753708, -0.15672186017036438, -0.11946940422058105, -0.0855121985077858, 0.02102506160736084, -0.043301086872816086, 0.052764277905225754, 0.029187506064772606, 0.06313791871070862, -0.05562494322657585, 0.09108006209135056, -0.09731566905975342, -0.005315596237778664, 0.015475943684577942, -0.061675529927015305, 0.010126902721822262, -0.21131323277950287, -0.008043201640248299, -0.1311621069908142, 0.03128067031502724, -0.03407350182533264, -0.01647294871509075, 0.011170338839292526, 0.0016343420138582587, -0.04606010392308235, -0.022542638704180717, -0.018754055723547935, 0.030862459912896156, 0.018654366955161095, 0.04895023629069328, 0.011602147482335567, -0.007601312827318907, 0.04120941832661629, 0.2065572589635849, -0.04314391687512398, -0.18267865478992462, -0.13445086777210236, 0.23455701768398285, 0.008398239500820637, 0.12134525924921036, -0.06294604390859604, 0.00014360684144776314, 0.037056922912597656, 0.3312853276729584, 0.2920037806034088, -0.06591223180294037, 0.016222776845097542, -0.03496059402823448, -0.0027518460992723703, -0.010928671807050705, 0.15421736240386963, 0.028664609417319298, 0.17177550494670868, -0.04381604865193367, 0.043717969208955765, -0.016995223239064217, -0.08461270481348038, -0.04839999973773956, 0.1234268918633461, -0.023147767409682274, -0.008460386656224728, -0.025535622611641884, 0.07288745045661926, -0.10863745957612991, 0.12828007340431213, -0.14435164630413055, -0.04277050495147705, -0.06639285385608673, 0.03365843743085861, 0.09799379110336304, -0.002150370506569743, 0.03132012113928795, -0.028657719492912292, -0.01699075475335121, 0.02508445642888546, -0.035447027534246445, -0.09260720759630203, -0.014365178532898426, 0.0811554342508316, -0.0027090583462268114, 0.19214791059494019, -0.017216810956597328, 0.056720808148384094, 0.07222237437963486, -0.00482771685346961, -0.08674640208482742, 0.1075904443860054, -0.0048650032840669155, -0.05891861394047737, 0.12367065995931625, -0.010637915693223476, 0.01167653501033783, 0.01576627418398857, 0.004334752913564444, -0.1429060995578766, 0.12325593084096909, -0.09570106863975525, -0.09618059545755386, -0.048408687114715576, 0.0977308452129364, -0.025742121040821075, 0.14952276647090912, 0.07897619158029556, -0.021133622154593468, 0.017186688259243965, -0.026781605556607246, 0.05889764428138733, -0.02527386136353016, -0.13093510270118713, -0.030025644227862358, -0.19915463030338287, -0.0342443622648716, 0.06796575337648392, -0.01800817996263504, -0.27645865082740784, -0.07106360793113708, -0.08814076334238052, -0.034605976194143295, -0.13013334572315216, 0.0708061084151268, 0.23948776721954346, 0.02763688750565052, -0.012005307711660862, -0.1307791769504547, -0.004188409075140953, 0.03875477984547615, -0.04872049018740654, -0.09937731176614761 ]
null
null
transformers
GGUF version for [Test157t/Hex-Macaroniac-7b](https://huggingface.co/Test157t/Hex-Macaroniac-7b) ![BbNKAgz3sRPbzedRZnFib.jpeg](https://cdn-uploads.huggingface.co/production/uploads/65c6e1fb06180e5f7247fe38/WXIjjA-Z4Icd8JvlDFEp_.jpeg)
{"library_name": "transformers", "pipeline_tag": "text-generation"}
text-generation
konz00/Hex-Macaroniac-7b-GGUF
[ "transformers", "gguf", "text-generation", "endpoints_compatible", "region:us" ]
2024-02-15T02:41:35+00:00
[]
[]
TAGS #transformers #gguf #text-generation #endpoints_compatible #region-us
GGUF version for Test157t/Hex-Macaroniac-7b !URL
[]
[ "TAGS\n#transformers #gguf #text-generation #endpoints_compatible #region-us \n" ]
[ 25 ]
[ "passage: TAGS\n#transformers #gguf #text-generation #endpoints_compatible #region-us \n" ]
[ -0.009579749777913094, 0.016051435843110085, -0.00778944743797183, -0.0606827586889267, 0.17959557473659515, 0.04390456900000572, 0.05772693082690239, 0.08898092061281204, 0.0652620941400528, 0.0004920579376630485, 0.13663704693317413, 0.12208554148674011, 0.004242816939949989, 0.07835976779460907, -0.07947206497192383, -0.2472376823425293, 0.10715211927890778, 0.056565575301647186, -0.08158249408006668, 0.053426869213581085, 0.046937670558691025, -0.01570023223757744, 0.09193092584609985, -0.045532554388046265, -0.20299391448497772, 0.046174556016922, 0.03110901266336441, -0.06774543225765228, 0.07453011721372604, 0.10531035810709, 0.09625300019979477, 0.0060175945982337, -0.15509317815303802, -0.23109441995620728, 0.03242243826389313, 0.017691269516944885, -0.10005950927734375, 0.001726817456074059, 0.032374605536460876, -0.10376358032226562, 0.0906355157494545, 0.04741545394062996, -0.09837349504232407, 0.0898582860827446, -0.1877211183309555, -0.08756867796182632, -0.04876207560300827, -0.03491636738181114, 0.01505382638424635, 0.06442180275917053, -0.013353051617741585, -0.04188106954097748, -0.0708744004368782, 0.08128168433904648, 0.15614952147006989, -0.3134511411190033, -0.0046891276724636555, 0.18256108462810516, 0.0725054144859314, 0.0586109533905983, -0.061780765652656555, 0.12141754478216171, 0.019321439787745476, -0.012183133512735367, -0.06757590919733047, -0.08256939053535461, -0.018916055560112, 0.1169932410120964, -0.06847090274095535, -0.06659001857042313, 0.19365370273590088, -0.03753431886434555, 0.03441791608929634, -0.019127504900097847, -0.07204557210206985, -0.045780349522829056, -0.04394662380218506, 0.049071304500103, -0.03188550844788551, 0.14522109925746918, 0.050162553787231445, -0.09602239727973938, -0.10060478001832962, -0.03710633143782616, -0.20041054487228394, 0.26695820689201355, -0.02112429216504097, 0.08602497726678848, -0.22190633416175842, 0.02547726221382618, -0.15691377222537994, -0.07777521759271622, -0.03994326665997505, -0.08868958801031113, 0.0007165444549173117, 0.03496411815285683, -0.1054995059967041, -0.05400620773434639, 0.1524859070777893, 0.10100986063480377, -0.039991434663534164, 0.06680738180875778, -0.05496573448181152, 0.09457684308290482, 0.007357186172157526, 0.08487115800380707, 0.004251695703715086, -0.02679322101175785, 0.0028159006033092737, -0.21648989617824554, -0.02913138084113598, -0.07440675050020218, -0.13762322068214417, -0.05771903321146965, -0.06727461516857147, 0.08988132327795029, -0.011513961479067802, 0.05322519689798355, -0.014487922191619873, 0.029230542480945587, 0.01647862233221531, -0.04698542132973671, -0.016856428235769272, -0.0006139421020634472, 0.042123325169086456, 0.13298702239990234, -0.05349437892436981, 0.002476717112585902, -0.04667213186621666, 0.045286741107702255, -0.07442079484462738, -0.010620219632983208, -0.03220843896269798, -0.017421729862689972, 0.01238013245165348, -0.14715172350406647, 0.04519771412014961, -0.1117033064365387, -0.17961354553699493, 0.012365331873297691, 0.02541954070329666, -0.027730686590075493, 0.05330321565270424, -0.043738897889852524, -0.03978561982512474, 0.05650516226887703, -0.06196942925453186, -0.05383311212062836, -0.07760827243328094, 0.06954456865787506, 0.006530481390655041, 0.07355255633592606, -0.19058609008789062, 0.06655817478895187, -0.02438404969871044, 0.021579977124929428, -0.137967050075531, 0.10443931818008423, -0.06279003620147705, 0.17324736714363098, 0.0018789154710248113, 0.022720806300640106, -0.14477799832820892, 0.07889829576015472, -0.08049003779888153, 0.18953093886375427, -0.06065136939287186, -0.12099350243806839, 0.33334603905677795, -0.07543940842151642, -0.13493743538856506, 0.0689801499247551, 0.02829323336482048, -0.025411535054445267, 0.09032180160284042, 0.2727244794368744, 0.037348054349422455, 0.007425226736813784, 0.08740752190351486, 0.17363785207271576, -0.11715899407863617, -0.08976823836565018, 0.02972390130162239, -0.07228533923625946, -0.13154412806034088, 0.038542065769433975, 0.017456594854593277, 0.13168790936470032, -0.033883508294820786, -0.006262729875743389, -0.034181054681539536, -0.007485904730856419, 0.02858523279428482, -0.005158690735697746, 0.10653847455978394, -0.06756997108459473, 0.00851160567253828, -0.0791502594947815, -0.04450387507677078, -0.02837834320962429, 0.03613412007689476, -0.02884381264448166, 0.10680101811885834, -0.04231800138950348, 0.08531270921230316, -0.1458498239517212, -0.1421574354171753, -0.025719186291098595, 0.10338063538074493, -0.032638587057590485, 0.11590561270713806, 0.07598087936639786, -0.07653150707483292, -0.010251731611788273, 0.036094751209020615, 0.14052505791187286, -0.01232639979571104, -0.005007986444979906, -0.02063787169754505, 0.10085758566856384, -0.0727856233716011, -0.07437864691019058, -0.08648914098739624, 0.02121381089091301, 0.17789724469184875, 0.07818994671106339, 0.020845280960202217, -0.013459769077599049, -0.0017346810782328248, -0.0031804873142391443, -0.030983738601207733, -0.022077839821577072, 0.07634211331605911, -0.006328213028609753, -0.12348480522632599, 0.22044001519680023, -0.12334394454956055, 0.27089813351631165, 0.1863541156053543, -0.21157202124595642, 0.040252525359392166, -0.07312045991420746, -0.003348080674186349, 0.0510278157889843, 0.08169369399547577, -0.038423482328653336, 0.1006215363740921, -0.011515987105667591, 0.14560577273368835, -0.02439459040760994, -0.015194443054497242, -0.014426624402403831, -0.024794336408376694, -0.04696030542254448, 0.040061965584754944, 0.0040283603593707085, -0.1499948799610138, 0.21265046298503876, 0.16437600553035736, 0.10120455920696259, 0.2301403433084488, -0.052775606513023376, -0.006179410964250565, 0.06962192803621292, 0.04979134351015091, -0.02862548828125, -0.023439612239599228, -0.27340593934059143, -0.04803221672773361, 0.058091115206480026, 0.08838862180709839, 0.1594451367855072, -0.11775633692741394, -0.07357568293809891, 0.011395692825317383, -0.07802775502204895, -0.032421816140413284, 0.10294470191001892, -0.0038269758224487305, 0.07158119976520538, 0.01997542567551136, 0.05874188244342804, 0.11971315741539001, -0.013143796473741531, -0.07740460336208344, 0.19619908928871155, -0.14169634878635406, -0.28958284854888916, -0.1811203956604004, -0.19683189690113068, -0.04825779050588608, 0.08782356977462769, 0.1169954314827919, -0.1564473658800125, -0.03824961185455322, 0.04453830048441887, 0.12497194111347198, -0.13104239106178284, 0.03814001381397247, 0.021633630618453026, 0.0357159860432148, -0.10764139890670776, -0.08001222461462021, -0.05617373436689377, -0.010578549467027187, -0.03604588285088539, 0.08127055317163467, -0.1475372612476349, 0.11227642744779587, 0.13462066650390625, 0.07316968590021133, 0.10140632838010788, -0.012971841730177402, 0.20399945974349976, -0.13861972093582153, -0.08125321567058563, 0.17675228416919708, -0.0056007071398198605, 0.059339489787817, 0.08977903425693512, 0.0006586898816749454, -0.13959509134292603, -0.011487981304526329, -0.027896802872419357, -0.13169744610786438, -0.1854529231786728, -0.0752529427409172, -0.1555994153022766, 0.0391971692442894, -0.035490378737449646, 0.0891667827963829, 0.11595811694860458, 0.07381458580493927, 0.023958778008818626, 0.009837509132921696, 0.03174016997218132, 0.05706017464399338, 0.16474942862987518, -0.012225826270878315, 0.06657849252223969, -0.08221769332885742, -0.06015617400407791, 0.09020262211561203, 0.08751114457845688, 0.17842808365821838, 0.1306467205286026, 0.14687080681324005, 0.040677037090063095, -0.025401374325156212, 0.1577799767255783, 0.0967152863740921, 0.0021733229514211416, -0.06181129068136215, -0.014838159084320068, -0.0008649908122606575, -0.03697579726576805, 0.0287077184766531, 0.01853453554213047, -0.2132999449968338, -0.030649419873952866, -0.18822017312049866, 0.13289867341518402, 0.022944875061511993, 0.04152572527527809, -0.13737933337688446, -0.012605909258127213, 0.11767126619815826, -0.018848104402422905, -0.11539360880851746, 0.08806689828634262, 0.07253143936395645, -0.09908895939588547, 0.10482010990381241, -0.04038228839635849, 0.12548591196537018, -0.012067651376128197, 0.09052639454603195, -0.054524339735507965, -0.09063196182250977, 0.02122526243329048, 0.09754706919193268, -0.2797391414642334, 0.20971086621284485, 0.02209780551493168, -0.037529509514570236, -0.05445846915245056, 0.003275763476267457, 0.02146379090845585, 0.17078383266925812, 0.13981258869171143, 0.024530909955501556, -0.11239399015903473, -0.07275530695915222, 0.0757327750325203, 0.050253208726644516, 0.1992543339729309, -0.08020731806755066, -0.029789097607135773, -0.010982216335833073, 0.0018502890598028898, -0.03299505263566971, 0.0070411707274615765, 0.0035219250712543726, -0.22908763587474823, 0.06321971863508224, 0.06728363037109375, 0.1331171989440918, -0.007174460217356682, 0.0938580259680748, -0.0987543985247612, 0.20914298295974731, -0.10293351858854294, -0.0611550472676754, -0.12668845057487488, -0.07016585767269135, 0.055333029478788376, -0.0459693968296051, 0.06438574194908142, -0.09787745028734207, 0.004825790412724018, -0.09091890603303909, -0.2175038456916809, 0.11621169745922089, -0.08599694073200226, 0.05296287685632706, -0.025739246979355812, 0.14477863907814026, -0.06959626078605652, -0.03430679440498352, 0.02216348610818386, 0.01117691956460476, -0.07169883698225021, -0.14873681962490082, 0.036700405180454254, 0.0059575242921710014, 0.012029055505990982, 0.11613360792398453, -0.012293081730604172, 0.07422421872615814, 0.019829515367746353, -0.03598923981189728, 0.2794061303138733, 0.14528462290763855, -0.023540012538433075, 0.16387425363063812, 0.1182202398777008, -0.06954378634691238, -0.3015587329864502, -0.05622600018978119, -0.17103037238121033, -0.009528614580631256, -0.13196390867233276, -0.22375774383544922, 0.08889705687761307, 0.03337733820080757, -0.002300830790773034, 0.23360541462898254, -0.22890694439411163, -0.03639558330178261, 0.09377763420343399, -0.008005589246749878, 0.4988834857940674, -0.18573682010173798, -0.15002183616161346, -0.0720629021525383, -0.2446787804365158, 0.14136669039726257, -0.07708404213190079, 0.11433251202106476, 0.0017991859931498766, 0.04815256968140602, 0.02332199737429619, -0.06559228152036667, 0.17698130011558533, 0.04580013081431389, 0.03142178803682327, -0.08555825054645538, -0.008231335319578648, 0.07476440072059631, 0.008679834194481373, -0.010487770661711693, -0.06746388971805573, -0.007445100229233503, -0.1339777708053589, -0.055862344801425934, -0.07582037150859833, 0.03865489736199379, 0.0799262747168541, -0.020358197391033173, -0.05047819763422012, -0.05339556559920311, -0.016170836985111237, 0.04169759154319763, 0.25741541385650635, -0.1039213016629219, 0.1687774360179901, 0.025945115834474564, 0.04493308067321777, -0.20477426052093506, -0.004618740640580654, -0.046843063086271286, -0.01980600692331791, 0.0935465469956398, -0.16110028326511383, 0.05819813162088394, 0.07944425195455551, -0.05106005072593689, 0.06884531676769257, 0.1359410583972931, 0.014532159082591534, 0.021173717454075813, 0.1467812955379486, -0.18967050313949585, -0.12311629951000214, -0.07608576118946075, -0.1257346123456955, 0.13505223393440247, 0.10438317060470581, 0.13074788451194763, 0.10108643770217896, 0.01909380964934826, -0.04022539034485817, -0.0070203510113060474, -0.0709037110209465, 0.01386108249425888, 0.00817179773002863, 0.02584969438612461, -0.15160691738128662, 0.0819007083773613, -0.03959338366985321, -0.18340301513671875, -0.008320484310388565, 0.128557488322258, -0.15399695932865143, -0.09042810648679733, -0.11891915649175644, 0.12760032713413239, -0.14738141000270844, -0.033070772886276245, -0.010086797177791595, -0.1264544129371643, 0.06237873435020447, 0.23380206525325775, 0.04952940717339516, 0.1272558718919754, -0.0008513382636010647, 0.004208550788462162, 0.027461236342787743, -0.10672212392091751, -0.05929316207766533, 0.018589265644550323, -0.09067489206790924, -0.040657006204128265, -0.06846732646226883, 0.1384158432483673, -0.08021344989538193, -0.05024665221571922, -0.1991536170244217, 0.023239627480506897, -0.12186726182699203, -0.05760553851723671, -0.12232958525419235, -0.05573355779051781, 0.017527537420392036, -0.032693274319171906, -0.03974439203739166, -0.039665255695581436, -0.13093158602714539, 0.009885036386549473, -0.03841420263051987, 0.00771184591576457, -0.05131201073527336, 0.012187164276838303, 0.09882780909538269, -0.03772640600800514, 0.12072275578975677, 0.17711962759494781, -0.06689894944429398, 0.1500464826822281, -0.26132556796073914, -0.11999466270208359, 0.10191045701503754, -0.045100219547748566, 0.03038318082690239, 0.1263023316860199, 0.03314247727394104, 0.04125086963176727, 0.01655580848455429, 0.0664728581905365, -0.021218232810497284, -0.11617148667573929, -0.023018088191747665, -0.08339186012744904, -0.08072522282600403, -0.05001387000083923, -0.05469566211104393, 0.1577877551317215, 0.04257164150476456, 0.05085008218884468, -0.021031958982348442, 0.08851058781147003, 0.029373785480856895, 0.019602583721280098, 0.031050896272063255, -0.19086246192455292, 0.05620424449443817, -0.07514713704586029, 0.003966974094510078, 0.022788207978010178, 0.34499022364616394, -0.05453638732433319, 0.01073573436588049, 0.02188924327492714, 0.022196220234036446, 0.09242333471775055, 0.02054937183856964, 0.30792322754859924, 0.15773355960845947, -0.047039639204740524, -0.11123811453580856, 0.0990198403596878, 0.02290377765893936, -0.04754915088415146, 0.12262656539678574, 0.050504595041275024, -0.07515083253383636, 0.1626535952091217, -0.051976338028907776, -0.014229681342840195, -0.016774345189332962, -0.061775192618370056, -0.03942084684967995, 0.030239198356866837, -0.003761051222681999, 0.02561202459037304, 0.17811238765716553, -0.06015368551015854, 0.09523767977952957, 0.023459063842892647, -0.047965023666620255, -0.1466878056526184, -0.1085495874285698, -0.04624064639210701, -0.1821233481168747, 0.05373648926615715, -0.11874468624591827, 0.07605502009391785, 0.1288948655128479, 0.05164184421300888, 0.002530656987801194, 0.1609726846218109, 0.04395661503076553, -0.10577259957790375, 0.10061278194189072, -0.06672453135251999, 0.047171708196401596, 0.07377781718969345, -0.03795863687992096, -0.06771766394376755, -0.08864762634038925, -0.00720007810741663, 0.08406072109937668, -0.015572836622595787, 0.02659783698618412, -0.16683253645896912, -0.07534679025411606, -0.05712404102087021, 0.11183422058820724, -0.07889340072870255, 0.11370791494846344, 0.0017686465289443731, -0.05400395765900612, 0.04991501197218895, 0.2105201780796051, -0.059385403990745544, 0.025879031047225, -0.04132382199168205, 0.07631102204322815, 0.05708041042089462, 0.1591852456331253, -0.09176953881978989, -0.017336152493953705, -0.08061710000038147, 0.3502628803253174, 0.23014387488365173, -0.052850354462862015, 0.02067919820547104, 0.05188370496034622, 0.05179297551512718, 0.2027982473373413, 0.10301763564348221, 0.06460312008857727, 0.27206170558929443, -0.04580971226096153, -0.079527847468853, 0.035237859934568405, -0.04374440386891365, -0.12283738702535629, 0.12384460121393204, 0.04579421505331993, -0.06362751126289368, -0.053369324654340744, 0.13975189626216888, -0.2425307184457779, 0.11192592978477478, 0.019153717905282974, -0.15003347396850586, -0.00391812901943922, -0.0630134865641594, 0.08068794012069702, -0.0178346186876297, 0.10188694298267365, -0.0034701921977102757, -0.15066954493522644, 0.009997948072850704, 0.05140922963619232, -0.28469008207321167, -0.026178516447544098, 0.01790185272693634, -0.009450799785554409, -0.03424009680747986, -0.03393350541591644, -0.046783559024333954, 0.07180610299110413, 0.035729311406612396, -0.007510825525969267, 0.05646010860800743, -0.03149822726845741, -0.042454853653907776, -0.05317913368344307, 0.08055443316698074, -0.013036716729402542, -0.12445653975009918, 0.05525088682770729, -0.15915890038013458, 0.056089311838150024, 0.0342070534825325, -0.047501031309366226, 0.014744436368346214, -0.05187036097049713, -0.11269349604845047, 0.03578566014766693, 0.06540670990943909, 0.016533343121409416, 0.015689637511968613, -0.024289604276418686, 0.013355252332985401, 0.013228170573711395, -0.014545965008437634, -0.12462245672941208, -0.045247167348861694, -0.11810245364904404, 0.19476470351219177, -0.026068395003676414, -0.15855787694454193, 0.03035876713693142, -0.05968214571475983, 0.13963234424591064, -0.0934998020529747, 0.08311305940151215, 0.13333196938037872, 0.03356841579079628, -0.028563642874360085, -0.1888405978679657, 0.09133175015449524, 0.07190821319818497, -0.05834801867604256, -0.10380897670984268 ]
null
null
diffusers
# Textual inversion text2image fine-tuning - JiafengMao/textual_inversion_pigg_XL These are textual inversion adaption weights for stabilityai/stable-diffusion-xl-base-1.0. You can find some example images in the following. ![img_0](./image_0.png) ![img_1](./image_1.png) ![img_2](./image_2.png) ![img_3](./image_3.png)
{"license": "creativeml-openrail-m", "tags": ["stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "diffusers", "textual_inversion"], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "inference": true}
text-to-image
JiafengMao/textual_inversion_pigg_XL
[ "diffusers", "tensorboard", "safetensors", "stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "textual_inversion", "base_model:stabilityai/stable-diffusion-xl-base-1.0", "license:creativeml-openrail-m", "endpoints_compatible", "diffusers:StableDiffusionXLPipeline", "region:us" ]
2024-02-15T02:42:45+00:00
[]
[]
TAGS #diffusers #tensorboard #safetensors #stable-diffusion #stable-diffusion-diffusers #text-to-image #textual_inversion #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionXLPipeline #region-us
# Textual inversion text2image fine-tuning - JiafengMao/textual_inversion_pigg_XL These are textual inversion adaption weights for stabilityai/stable-diffusion-xl-base-1.0. You can find some example images in the following. !img_0 !img_1 !img_2 !img_3
[ "# Textual inversion text2image fine-tuning - JiafengMao/textual_inversion_pigg_XL\nThese are textual inversion adaption weights for stabilityai/stable-diffusion-xl-base-1.0. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3" ]
[ "TAGS\n#diffusers #tensorboard #safetensors #stable-diffusion #stable-diffusion-diffusers #text-to-image #textual_inversion #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionXLPipeline #region-us \n", "# Textual inversion text2image fine-tuning - JiafengMao/textual_inversion_pigg_XL\nThese are textual inversion adaption weights for stabilityai/stable-diffusion-xl-base-1.0. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3" ]
[ 104, 86 ]
[ "passage: TAGS\n#diffusers #tensorboard #safetensors #stable-diffusion #stable-diffusion-diffusers #text-to-image #textual_inversion #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionXLPipeline #region-us \n# Textual inversion text2image fine-tuning - JiafengMao/textual_inversion_pigg_XL\nThese are textual inversion adaption weights for stabilityai/stable-diffusion-xl-base-1.0. You can find some example images in the following. \n\n!img_0\n!img_1\n!img_2\n!img_3" ]
[ -0.09320984035730362, -0.09730938822031021, -0.004164781887084246, 0.006134308874607086, 0.10318045318126678, 0.014364457689225674, 0.21738842129707336, 0.10046011209487915, 0.0609545037150383, 0.10391253232955933, 0.024115310981869698, 0.08498818427324295, 0.028549570590257645, 0.11409338563680649, -0.0840671956539154, -0.2322908341884613, 0.0003957197768613696, 0.015879809856414795, -0.06485102325677872, 0.06596716493368149, 0.10263217985630035, -0.04729340225458145, 0.09804161638021469, 0.022156506776809692, -0.08402799814939499, 0.016013501212000847, 0.04001851752400398, -0.07952604442834854, 0.07700522243976593, 0.06938442587852478, 0.04096150025725365, 0.13398469984531403, 0.018017994239926338, -0.17332226037979126, 0.027581507340073586, 0.05024437606334686, -0.014150735922157764, 0.06914926320314407, -0.018823152408003807, -0.03336348012089729, 0.0781933143734932, -0.07895689457654953, -0.03093132935464382, 0.025237038731575012, -0.05528045445680618, -0.015026927925646305, -0.062057819217443466, 0.0459873266518116, 0.06715993583202362, 0.00019514418090693653, 0.013511044904589653, 0.0386422798037529, -0.012748874723911285, 0.03275588899850845, 0.2884983718395233, -0.2890869677066803, -0.0053018853068351746, 0.10161683708429337, 0.10362238436937332, 0.12613821029663086, -0.05886232107877731, 0.10992087423801422, 0.0388132743537426, -0.03686691075563431, 0.06992311030626297, -0.05715782567858696, 0.16362541913986206, -0.03023437038064003, -0.09986836463212967, 0.10585988312959671, 0.15385180711746216, -0.012080720625817776, -0.07390264421701431, -0.22888264060020447, -0.03133413568139076, 0.1353476494550705, -0.071779765188694, -0.03758203238248825, 0.002925469307228923, 0.008223427459597588, 0.07157228887081146, -0.061523765325546265, -0.0914762020111084, -0.05335663631558418, -0.06435194611549377, 0.19829830527305603, -0.0231870599091053, -0.004697795491665602, 0.042336199432611465, 0.10348289459943771, -0.1557951122522354, -0.13881050050258636, 0.02116333320736885, -0.06089642271399498, 0.027831632643938065, 0.0100393733009696, -0.025003239512443542, -0.2487429827451706, 0.021041247993707657, 0.03285090625286102, 0.03510873019695282, 0.010382812470197678, -0.03586117550730705, 0.11391060799360275, -0.02903548628091812, 0.04957630857825279, -0.08288127183914185, -0.01781035028398037, 0.017087111249566078, 0.0773853287100792, 0.09676866233348846, -0.02456866018474102, -0.08773201704025269, -0.05222560092806816, -0.06765280663967133, 0.008689000271260738, -0.03935525193810463, 0.04409705102443695, -0.09180248528718948, -0.04234141483902931, 0.16507543623447418, -0.0479128435254097, 0.03258226439356804, -0.02891652099788189, 0.023485854268074036, 0.20285677909851074, 0.14294734597206116, 0.032089509069919586, -0.01722181774675846, 0.09962176531553268, -0.05308174341917038, 0.06216474995017052, 0.017794262617826462, -0.1137644499540329, -0.003974258434027433, -0.16373610496520996, -0.026394274085760117, -0.1039244681596756, -0.08499671518802643, 0.06640524417161942, 0.024865159764885902, -0.03441118448972702, 0.07075699418783188, 0.03150444105267525, -0.016919566318392754, 0.03565583378076553, 0.017900923267006874, -0.11174565553665161, -0.03023599646985531, 0.032420869916677475, 0.018292807042598724, 0.11223860085010529, -0.005717029329389334, -0.000825899769552052, -0.04208594560623169, 0.015052498318254948, -0.31664881110191345, 0.0836382582783699, -0.07310525327920914, -0.029282480478286743, -0.07621683180332184, -0.031892742961645126, -0.06131071224808693, 0.07196491211652756, 0.037204109132289886, 0.19754981994628906, -0.20671901106834412, -0.09673817455768585, 0.12194990366697311, -0.18098939955234528, -0.08057690411806107, 0.07244905829429626, 0.04202304780483246, 0.0506628155708313, 0.04094864800572395, 0.08392296731472015, 0.0048400782980024815, -0.2595149278640747, 0.03079024702310562, -0.009888898581266403, -0.06128108128905296, 0.0012134863063693047, 0.03785378858447075, -0.006612809840589762, 0.06769900768995285, 0.02920643985271454, -0.12931600213050842, 0.07641090452671051, -0.05253240838646889, 0.0272821094840765, -0.041573695838451385, 0.004089785739779472, 0.10021446645259857, 0.02850491926074028, 0.03313441202044487, -0.007410934194922447, -0.08580019325017929, 0.09718292206525803, 0.006620561238378286, -0.0741753950715065, 0.05731919780373573, -0.029534460976719856, 0.13169483840465546, -0.13547490537166595, 0.0006634433520957828, -0.11088196933269501, 0.016610905528068542, 0.019802462309598923, 0.17270366847515106, 0.01082522515207529, 0.04752035066485405, 0.13358652591705322, 0.04373789206147194, -0.04803742095828056, -0.009032828733325005, 0.08761618286371231, 0.016653066501021385, -0.07440897077322006, -0.17778155207633972, 0.031473975628614426, -0.08305436372756958, 0.09322213381528854, -0.203130304813385, 0.06317125260829926, -0.02544228546321392, 0.15017394721508026, 0.11824405938386917, -0.027614803984761238, 0.02556372433900833, -0.018709227442741394, -0.07864310592412949, -0.029637323692440987, 0.01854194886982441, 0.0014582544099539518, -0.06971975415945053, 0.25557926297187805, -0.16189613938331604, 0.18576599657535553, 0.1228957399725914, 0.011865757405757904, -0.04773601517081261, -0.17357315123081207, 0.013719581067562103, 0.007244417443871498, -0.028217677026987076, -0.0046744514256715775, 0.025833338499069214, 0.034615516662597656, 0.1285453736782074, -0.04224308207631111, 0.02668035961687565, 0.05482103303074837, -0.07592695951461792, -0.04853871092200279, 0.03452691808342934, 0.03436937928199768, -0.015909377485513687, 0.025889957323670387, 0.1291857212781906, -0.007922844961285591, 0.06504590064287186, -0.0295921191573143, -0.11189039796590805, -0.008452075533568859, 0.03868890181183815, 0.0778440311551094, 0.10999684780836105, -0.018042633309960365, -0.011813421733677387, 0.022235756739974022, -0.04297333210706711, 0.024072987958788872, -0.08169388771057129, -0.0030354273039847612, 0.07770349085330963, -0.04387636110186577, 0.09695538878440857, 0.06863340735435486, -0.03661945462226868, 0.1098211407661438, -0.08826123923063278, -0.017220333218574524, -0.022740738466382027, 0.0024814531207084656, -0.10548456013202667, 0.11338775604963303, -0.06917150318622589, -0.14233148097991943, -0.14349141716957092, -0.05757540836930275, 0.014263064600527287, 0.011110490188002586, 0.028342057019472122, -0.018954642117023468, -0.10067601501941681, -0.11568831652402878, 0.08199438452720642, 0.11475791782140732, 0.03951658308506012, 0.023120375350117683, -0.05985451117157936, 0.01242737751454115, -0.0754578560590744, -0.011708092875778675, -0.038999900221824646, 0.08515821397304535, 0.05349740758538246, 0.023233795538544655, 0.0924362763762474, 0.11184511333703995, 0.0015088252257555723, 0.013568077236413956, 0.01824798993766308, 0.024555226787924767, -0.032396622002124786, 0.12704822421073914, 0.17750301957130432, -0.01179790124297142, 0.06262750178575516, 0.06688562035560608, 0.026784196496009827, 0.0011279004393145442, 0.016507737338542938, 0.002431197790428996, -0.06651519238948822, -0.09319686889648438, -0.07920178025960922, -0.0794399157166481, -0.032093364745378494, 0.010149066336452961, 0.011760929599404335, 0.08855610340833664, 0.08098233491182327, 0.028101405128836632, -0.007102621719241142, 0.0972834900021553, 0.08213163167238235, 0.08006424456834793, -0.009416338056325912, 0.08618244528770447, -0.04264279082417488, -0.0636243000626564, 0.10039068758487701, -0.08962433785200119, 0.10750303417444229, -0.06874039024114609, 0.09420127421617508, 0.03311082720756531, -0.02456665225327015, 0.09001625329256058, 0.13729457557201385, -0.073171466588974, -0.06151476502418518, -0.013137544505298138, -0.11058728396892548, 0.0504353791475296, 0.05819953978061676, -0.011764295399188995, -0.04185944050550461, -0.06731449067592621, 0.04237942025065422, 0.05842619761824608, 0.023813193663954735, 0.10120612382888794, -0.2415156215429306, -0.019585225731134415, 0.019894545897841454, 0.03808928653597832, -0.018987538293004036, -0.03168834000825882, 0.15664128959178925, -0.006289947312325239, 0.08418027311563492, -0.10202746838331223, 0.056727487593889236, 0.027530642226338387, -0.02715103141963482, -0.025313910096883774, 0.166452094912529, -0.04706040769815445, -0.045044656842947006, -0.16692772507667542, 0.044493041932582855, 0.021502656862139702, -0.00198355526663363, -0.03422193229198456, 0.023959996178746223, 0.046734973788261414, 0.13684439659118652, 0.09013120830059052, -0.004872580990195274, 0.10378459841012955, -0.03271854668855667, -0.12948334217071533, -0.03900932893157005, 0.06433486193418503, -0.045874349772930145, 0.046238791197538376, 0.02048598602414131, -0.04017127677798271, 0.06461948901414871, -0.012036599218845367, -0.20906758308410645, -0.15957173705101013, 0.06359440088272095, 0.015190327540040016, -0.09473727643489838, -0.08723349124193192, -0.1017434298992157, -0.11443399637937546, 0.26443830132484436, -0.024596577510237694, -0.06940148770809174, -0.13027024269104004, -0.022350678220391273, 0.08277271687984467, -0.03689613193273544, 0.021299831569194794, -0.012465592473745346, 0.10266097635030746, -0.07039941847324371, -0.15402725338935852, 0.11340264230966568, -0.08044232428073883, -0.07129787653684616, -0.07888125628232956, 0.1132015511393547, -0.010254697874188423, -0.02507743425667286, -0.0067682513035833836, -0.012601621448993683, 0.06826332211494446, -0.07312378287315369, 0.019931504502892494, 0.10685966163873672, 0.017471639439463615, 0.06383930891752243, -0.07637131214141846, -0.07006446272134781, -0.017200039699673653, 0.026660693809390068, 0.10299066454172134, 0.12337096035480499, -0.07701951265335083, 0.08629810065031052, 0.07854284346103668, -0.0331457182765007, -0.15700608491897583, -0.021040141582489014, -0.05442796275019646, 0.040121372789144516, 0.0565803088247776, -0.0535576231777668, 0.12864194810390472, 0.046869173645973206, 0.013871688395738602, 0.20958906412124634, -0.3311648368835449, -0.1286727637052536, 0.03058740310370922, 0.10633856803178787, 0.08429969847202301, -0.15527063608169556, -0.09618795663118362, 0.03030819445848465, -0.1497279405593872, 0.04266708344221115, -0.05363373085856438, 0.04243796318769455, -0.04515592008829117, -0.07823798805475235, 0.019423384219408035, -0.058653708547353745, 0.11837676912546158, -0.056405406445264816, 0.0030603271443396807, -0.09167848527431488, 0.028628382831811905, 0.05024869740009308, -0.04010091349482536, 0.07580579072237015, -0.2456858903169632, 0.006355201359838247, -0.1446562111377716, -0.032419707626104355, 0.023897431790828705, 0.04750779643654823, -0.005736987572163343, -0.03370858356356621, -0.03991695120930672, -0.003994243685156107, -0.017027072608470917, -0.023897849023342133, 0.07149407267570496, -0.031006868928670883, 0.06828849762678146, 0.16882270574569702, 0.0726759284734726, -0.052637021988630295, -0.07653241604566574, -0.045089032500982285, -0.0019314740784466267, 0.049177590757608414, -0.09703729301691055, 0.012493488378822803, 0.11759580671787262, 0.05742551013827324, 0.13398192822933197, 0.04638975113630295, -0.04735812917351723, 0.03719177097082138, 0.12809348106384277, -0.10145581513643265, -0.006446319632232189, -0.034604981541633606, 0.014750722795724869, 0.0442400760948658, 0.07158704847097397, 0.14122456312179565, -0.04198594391345978, 0.024692118167877197, 0.004322985652834177, 0.0633404552936554, -0.045245569199323654, 0.11213700473308563, 0.02607758343219757, 0.008730640634894371, -0.058471713215112686, 0.07054506242275238, -0.05997169017791748, -0.09373636543750763, -0.041250549256801605, 0.08266763389110565, -0.088259257376194, -0.03642560914158821, 0.09137343615293503, 0.13235467672348022, -0.08970307558774948, 0.014335761778056622, -0.0854262188076973, -0.16806557774543762, -0.003954913932830095, 0.10159328579902649, 0.032308995723724365, -0.043309107422828674, -0.03685542568564415, -0.015056418254971504, -0.05833881348371506, 0.051903340965509415, 0.138505756855011, 0.0958179235458374, -0.18610364198684692, -0.0595562681555748, -0.025167271494865417, -0.048631295561790466, -0.07859665900468826, 0.01128139067441225, -0.05096398666501045, -0.06386294960975647, -0.07221781462430954, 0.07293552160263062, -0.15976285934448242, -0.06588298082351685, -0.021457279101014137, -0.06431598961353302, 0.0063608004711568356, -0.0026285068597644567, -0.008433548733592033, -0.020793873816728592, -0.03202391043305397, 0.018881136551499367, -0.09057684242725372, -0.04338593780994415, -0.0203746035695076, -0.1380433291196823, 0.09823448210954666, 0.06244361028075218, -0.04768788442015648, -0.00402161804959178, -0.21104860305786133, -0.00003511466275085695, 0.15088817477226257, -0.01201170589774847, -0.0006720692617818713, -0.05211176723241806, 0.06241678074002266, 0.005680326838046312, -0.008861622773110867, -0.038792431354522705, 0.05212501063942909, -0.061707962304353714, 0.04262835159897804, -0.12870638072490692, 0.011819812469184399, -0.06897581368684769, 0.06946000456809998, 0.19240379333496094, 0.08723938465118408, 0.12017699331045151, -0.12593814730644226, 0.049429092556238174, -0.147581085562706, 0.02392795868217945, 0.03670847788453102, -0.04785028100013733, 0.12301553040742874, 0.0044317892752587795, -0.02101818472146988, -0.032731909304857254, 0.13720940053462982, -0.044843561947345734, -0.13967396318912506, 0.0113868722692132, -0.05488135665655136, 0.08013580739498138, 0.052515167742967606, 0.18580473959445953, 0.02746063657104969, 0.027896959334611893, -0.1344609260559082, 0.05865371227264404, 0.1261097490787506, -0.11213243752717972, 0.12999959290027618, 0.11929361522197723, -0.1058594137430191, 0.13200020790100098, 0.023359637707471848, -0.06113499402999878, -0.050630513578653336, 0.08940920233726501, -0.1243993416428566, 0.09804746508598328, 0.009438141249120235, 0.08565321564674377, 0.17903783917427063, -0.05401585251092911, -0.00971929170191288, 0.06450904905796051, -0.041671816259622574, -0.05826007202267647, -0.12980423867702484, -0.03734689578413963, -0.1855921596288681, -0.009934235364198685, -0.06198571249842644, 0.02484758384525776, -0.02530718594789505, 0.04529011994600296, 0.06772207468748093, 0.08902976661920547, 0.055899202823638916, -0.07875066995620728, 0.11468148976564407, -0.005679655820131302, -0.07395471632480621, 0.06468521803617477, 0.005344975274056196, 0.022950896993279457, 0.017100155353546143, 0.0024445548187941313, 0.06727098673582077, 0.059917546808719635, 0.047645203769207, -0.020884841680526733, -0.08705878257751465, 0.0106408866122365, -0.0015530354576185346, 0.0047803656198084354, 0.1299307644367218, 0.055784646421670914, -0.017568642273545265, -0.03814994543790817, 0.12048377096652985, -0.032868485897779465, -0.04198687523603439, -0.12333717197179794, 0.05612088739871979, -0.03545130789279938, 0.048270564526319504, -0.06458351016044617, -0.11632411926984787, -0.006937767378985882, 0.18035149574279785, 0.1706504225730896, -0.07867129147052765, 0.026394980028271675, -0.025800319388508797, -0.017926549538969994, 0.0496971569955349, 0.016325457021594048, 0.035993341356515884, 0.22642481327056885, -0.0049054911360144615, -0.005006146617233753, -0.06722868233919144, -0.06176145747303963, -0.06709929555654526, 0.0035792598500847816, 0.02228519506752491, -0.02741045132279396, -0.04037097841501236, 0.08116940408945084, 0.008612049743533134, -0.11844668537378311, 0.054665010422468185, -0.12217626720666885, -0.0016432511620223522, -0.06562062352895737, 0.05645064264535904, -0.02395292930305004, 0.005940019618719816, -0.028917156159877777, -0.045964960008859634, 0.024317724630236626, 0.02433702163398266, -0.07033118605613708, 0.08708419650793076, -0.06585077196359634, -0.1392504870891571, 0.0293272715061903, -0.01059259008616209, 0.01868286356329918, 0.046518467366695404, 0.052905499935150146, -0.059393178671598434, 0.024140629917383194, -0.053157344460487366, -0.07296457141637802, -0.039631690829992294, 0.10563142597675323, -0.012455802410840988, -0.02612672932446003, 0.04858453571796417, -0.15631455183029175, 0.05019347742199898, -0.021359829232096672, -0.1579841673374176, -0.07233842462301254, -0.027875486761331558, -0.0463116280734539, 0.08801531046628952, 0.037956610321998596, 0.012491029687225819, 0.03511599078774452, -0.0015676316106691957, 0.03467188775539398, 0.04664771258831024, 0.06021028384566307, 0.040597960352897644, -0.11184606701135635, 0.0095134312286973, 0.02521693892776966, 0.04612581431865692, -0.22841963171958923, -0.05411844700574875, -0.1084146499633789, -0.012518703006207943, -0.0514095239341259, 0.0764184221625328, 0.25252702832221985, 0.03269701451063156, -0.01057489588856697, -0.22363527119159698, 0.07237381488084793, 0.1323971450328827, -0.10205182433128357, -0.0030863224528729916 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
aidonuts/enthralling-etchings-132-s600b
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T02:52:27+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 60, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.04654794931411743, 0.16618601977825165, -0.005445904564112425, 0.01853804849088192, 0.0981811136007309, 0.011998992413282394, 0.06433123350143433, 0.11398410052061081, -0.0230073444545269, 0.11406639218330383, 0.03047988750040531, 0.10172267258167267, 0.11317981779575348, 0.14841650426387787, -0.002152352826669812, -0.22403094172477722, 0.050844956189394, -0.12105348706245422, -0.033293843269348145, 0.11749980598688126, 0.1483822613954544, -0.09928343445062637, 0.07274559140205383, -0.029687678441405296, -0.012143402360379696, -0.030057786032557487, -0.05890674889087677, -0.046214159578084946, 0.04651786759495735, 0.06640566885471344, 0.06770290434360504, 0.0071083661168813705, 0.09012923389673233, -0.2696533799171448, 0.018959321081638336, 0.07145345956087112, -0.002759667346253991, 0.06957992166280746, 0.06404146552085876, -0.07107418030500412, 0.10337356477975845, -0.05106033384799957, 0.14650006592273712, 0.08365883678197861, -0.09081148356199265, -0.1895141303539276, -0.08866965025663376, 0.09882009029388428, 0.17572562396526337, 0.04925641790032387, -0.02320658043026924, 0.09761467576026917, -0.08769196271896362, 0.015438909642398357, 0.04981724172830582, -0.07620415836572647, -0.05378096550703049, 0.05986575037240982, 0.07907199114561081, 0.06627275794744492, -0.12434766441583633, -0.02885502204298973, 0.005009706597775221, 0.010980482213199139, 0.0769270583987236, 0.01728810742497444, 0.146672785282135, 0.0338633768260479, -0.12615777552127838, -0.04880760237574577, 0.09869225323200226, 0.03395522013306618, -0.04422314465045929, -0.24749068915843964, -0.03152675926685333, -0.030810698866844177, -0.029386121779680252, -0.03716538846492767, 0.04340358078479767, -0.007673026993870735, 0.08638741075992584, -0.0060646249912679195, -0.07403432577848434, -0.03937075287103653, 0.06169692054390907, 0.0672287791967392, 0.02999979443848133, -0.013745363801717758, 0.010938193649053574, 0.11620724946260452, 0.1095694974064827, -0.12054188549518585, -0.05555335059762001, -0.06393084675073624, -0.08656639605760574, -0.040790557861328125, 0.034162238240242004, 0.03456587344408035, 0.05349370837211609, 0.25305667519569397, 0.015654386952519417, 0.059652652591466904, 0.034477248787879944, 0.007892133668065071, 0.05848940089344978, 0.11044429242610931, -0.06018859148025513, -0.10444226115942001, -0.02648012898862362, 0.08843598514795303, 0.008199662901461124, -0.03287925571203232, -0.05088530853390694, 0.06019928678870201, 0.01946467161178589, 0.11926145106554031, 0.09061790257692337, 0.010536285117268562, -0.07121123373508453, -0.061038948595523834, 0.1891259253025055, -0.16544590890407562, 0.04322727024555206, 0.035097137093544006, -0.03903156518936157, 0.00019933005387429148, 0.013914269395172596, 0.016625655815005302, -0.025983380153775215, 0.09017423540353775, -0.054113563150167465, -0.04145489260554314, -0.11186197400093079, -0.03383193537592888, 0.033762916922569275, 0.008953776210546494, -0.035059962421655655, -0.033713940531015396, -0.08351044356822968, -0.07577689737081528, 0.09320491552352905, -0.07346344739198685, -0.04878907650709152, -0.01804324984550476, -0.07530532777309418, 0.022395428270101547, 0.019394835457205772, 0.07707412540912628, -0.02362251654267311, 0.04399976506829262, -0.05189276114106178, 0.05863580107688904, 0.11207318305969238, 0.03570080175995827, -0.05736649036407471, 0.06062258034944534, -0.23834340274333954, 0.09552820026874542, -0.07409077137708664, 0.05591456592082977, -0.153293639421463, -0.024439791217446327, 0.04788333550095558, 0.008784620091319084, -0.009650949388742447, 0.13416339457035065, -0.21702027320861816, -0.02536402828991413, 0.1717337965965271, -0.10057014971971512, -0.07069246470928192, 0.05619903281331062, -0.04835370555520058, 0.10988964140415192, 0.03825836628675461, -0.025690359994769096, 0.06171267107129097, -0.1267417073249817, 0.003717758459970355, -0.05005312338471413, -0.017048977315425873, 0.1548657864332199, 0.07182947546243668, -0.07217690348625183, 0.07399354875087738, 0.025708531960844994, -0.0246540866792202, -0.04625825211405754, -0.015164627693593502, -0.10536660254001617, 0.014689887873828411, -0.06369215250015259, 0.014470234513282776, -0.020807426422834396, -0.09071163833141327, -0.027962757274508476, -0.17504668235778809, -0.03014434315264225, 0.08651752024888992, -0.008693269453942776, -0.01803150773048401, -0.1178668737411499, 0.009341353550553322, 0.04177580401301384, 0.0061247628182172775, -0.13462838530540466, -0.04812471568584442, 0.02780051715672016, -0.1600649207830429, 0.034652888774871826, -0.05392369255423546, 0.04932025074958801, 0.025790516287088394, -0.028889117762446404, -0.026493212208151817, 0.021633783355355263, 0.005992184858769178, -0.011999987065792084, -0.24343903362751007, -0.028118690475821495, -0.024888472631573677, 0.1682123839855194, -0.20917098224163055, 0.03546025976538658, 0.07867541164159775, 0.15366052091121674, 0.011240328662097454, -0.04177491366863251, 0.005974748637527227, -0.06935794651508331, -0.02736494317650795, -0.05875484645366669, -0.0047869328409433365, -0.03310677409172058, -0.04545191675424576, 0.04568447172641754, -0.16510973870754242, -0.032636504620313644, 0.09776268899440765, 0.06289951503276825, -0.13922683894634247, -0.020621931180357933, -0.03630133345723152, -0.049253206700086594, -0.04911839962005615, -0.0605199858546257, 0.10893940925598145, 0.05891856551170349, 0.04574795812368393, -0.05928509309887886, -0.07568105310201645, -0.001827909960411489, -0.013898161239922047, -0.017864689230918884, 0.09759635478258133, 0.0751434788107872, -0.13251115381717682, 0.09224759042263031, 0.09603385627269745, 0.07919023185968399, 0.09113933145999908, -0.02355697751045227, -0.08261934667825699, -0.045987509191036224, 0.031442027539014816, 0.020124373957514763, 0.13039541244506836, -0.024294709786772728, 0.04352088272571564, 0.042134687304496765, -0.019369594752788544, 0.014752166345715523, -0.08687400817871094, 0.033972494304180145, 0.028472330421209335, -0.016721390187740326, 0.050190530717372894, -0.03876714035868645, 0.02440318465232849, 0.08830609917640686, 0.045322712510824203, 0.03507532551884651, 0.015493292361497879, -0.05206458270549774, -0.1083620935678482, 0.16405931115150452, -0.12714070081710815, -0.22483378648757935, -0.13936103880405426, 0.0037376401014626026, 0.035628627985715866, -0.015835661441087723, 0.002417160663753748, -0.059374887496232986, -0.12220635265111923, -0.08858037739992142, 0.015140829607844353, 0.04942670464515686, -0.09028962254524231, -0.06437795609235764, 0.058117836713790894, 0.03889724239706993, -0.14560972154140472, 0.017612040042877197, 0.04854894429445267, -0.09789852797985077, -0.006774199660867453, 0.08094939589500427, 0.0698540136218071, 0.1770169734954834, 0.017703235149383545, -0.021850809454917908, 0.032354529947042465, 0.20614571869373322, -0.13538233935832977, 0.11083246022462845, 0.13607586920261383, -0.09041404724121094, 0.08072979003190994, 0.19951270520687103, 0.03932560607790947, -0.10153959691524506, 0.031980328261852264, 0.02283124253153801, -0.0284719280898571, -0.24526868760585785, -0.07212468236684799, -0.004402178805321455, -0.058010730892419815, 0.07660572230815887, 0.09286724030971527, 0.08215958625078201, 0.012304253876209259, -0.09310996532440186, -0.08154371380805969, 0.05942574888467789, 0.10367169976234436, 0.024584239348769188, -0.010839897207915783, 0.08998730033636093, -0.034100502729415894, 0.019626356661319733, 0.0853661298751831, 0.005239574704319239, 0.17840281128883362, 0.05159219726920128, 0.18830420076847076, 0.07925192266702652, 0.07219027727842331, 0.009912233799695969, 0.013080619275569916, 0.018877580761909485, 0.03300119563937187, -0.002769160782918334, -0.08440786600112915, -0.02248465269804001, 0.11566436290740967, 0.06668911874294281, 0.010815348476171494, 0.015172341838479042, -0.04104290530085564, 0.07965951412916183, 0.1831512451171875, -0.007656289264559746, -0.1783534437417984, -0.057547420263290405, 0.07553383708000183, -0.09879875183105469, -0.09854305535554886, -0.013454320840537548, 0.03072015568614006, -0.17046253383159637, 0.023390959948301315, -0.02239842526614666, 0.1106182336807251, -0.14194999635219574, -0.020490378141403198, 0.07218493521213531, 0.07199500501155853, 0.004729843698441982, 0.05758659541606903, -0.16417601704597473, 0.10671813786029816, 0.008950476534664631, 0.06779605895280838, -0.09610627591609955, 0.1008887067437172, -0.004196076653897762, -0.02063460275530815, 0.1393408179283142, 0.002700034761801362, -0.06884108483791351, -0.0763031542301178, -0.08754398673772812, -0.009632662869989872, 0.12754282355308533, -0.1419651061296463, 0.08767123520374298, -0.037212442606687546, -0.0424150750041008, -0.0017086371080949903, -0.10206665843725204, -0.11638247221708298, -0.18888559937477112, 0.06001543253660202, -0.13492922484874725, 0.03152317553758621, -0.10799519717693329, -0.032371897250413895, -0.030304040759801865, 0.19337286055088043, -0.23447458446025848, -0.07199826091527939, -0.1475764364004135, -0.10233612358570099, 0.1443224400281906, -0.0501345656812191, 0.08485390990972519, -0.007241467013955116, 0.16846685111522675, 0.019060896709561348, -0.02531743235886097, 0.0971490666270256, -0.09173708409070969, -0.19302815198898315, -0.07869284600019455, 0.15662524104118347, 0.13260218501091003, 0.031680017709732056, -0.002461588243022561, 0.036563750356435776, -0.015421539545059204, -0.11935004591941833, 0.015969349071383476, 0.1787186712026596, 0.06237189099192619, 0.02331034652888775, -0.027346095070242882, -0.11273157596588135, -0.06900003552436829, -0.028530338779091835, 0.03054865077137947, 0.17762407660484314, -0.07057618349790573, 0.18207968771457672, 0.14163152873516083, -0.05922834202647209, -0.20400173962116241, 0.010538800619542599, 0.03055560030043125, 0.0009220078936778009, 0.02591954916715622, -0.20123432576656342, 0.08688826113939285, 0.004683020059019327, -0.05110127478837967, 0.13194532692432404, -0.17217805981636047, -0.14451217651367188, 0.0765485092997551, 0.038384392857551575, -0.19559739530086517, -0.12913893163204193, -0.09174312651157379, -0.045869920402765274, -0.18591414391994476, 0.09569250047206879, 0.0305706188082695, 0.010893458500504494, 0.03030681423842907, 0.029179483652114868, 0.019487828016281128, -0.0418255440890789, 0.18391458690166473, -0.024792250245809555, 0.026594700291752815, -0.08539514988660812, -0.06927408277988434, 0.03743394836783409, -0.052842434495687485, 0.07349982857704163, -0.023486759513616562, 0.007861839607357979, -0.10348054021596909, -0.042148489505052567, -0.03735732287168503, 0.015448716469109058, -0.09657872468233109, -0.08514349907636642, -0.045032672584056854, 0.09675803780555725, 0.09690850973129272, -0.033646680414676666, -0.028050623834133148, -0.07533035427331924, 0.04412057250738144, 0.19926515221595764, 0.1785389482975006, 0.042153384536504745, -0.08034496754407883, -0.004150947090238333, -0.010121207684278488, 0.04310847446322441, -0.20463712513446808, 0.06283636391162872, 0.05450061708688736, 0.01973269321024418, 0.11436162889003754, -0.019565396010875702, -0.15359151363372803, -0.07263088971376419, 0.06303015351295471, -0.060181066393852234, -0.19620554149150848, 0.00867035984992981, 0.060603946447372437, -0.16371412575244904, -0.04535605385899544, 0.04643881320953369, -0.005620351992547512, -0.038163937628269196, 0.021896906197071075, 0.09194854646921158, 0.0026654244866222143, 0.07427921891212463, 0.05387866869568825, 0.0827430784702301, -0.10537070035934448, 0.08090532571077347, 0.08839722722768784, -0.08452684432268143, 0.023530138656497, 0.10478579998016357, -0.059433579444885254, -0.03440561518073082, 0.020135708153247833, 0.08153781294822693, 0.01775863952934742, -0.040019966661930084, 0.013229827396571636, -0.10452935844659805, 0.05954122915863991, 0.08839859813451767, 0.032507482916116714, 0.016702456399798393, 0.03425082191824913, 0.04607953503727913, -0.07238735258579254, 0.12142276018857956, 0.031868141144514084, 0.017129309475421906, -0.036505792289972305, -0.040896978229284286, 0.019542274996638298, -0.03214648738503456, -0.005015232600271702, -0.03023446537554264, -0.07695909589529037, -0.014793801121413708, -0.1626158058643341, -0.011131818406283855, -0.05648450180888176, 0.010329355485737324, 0.03204665705561638, -0.032609567046165466, 0.008124498650431633, 0.009250079281628132, -0.07695289701223373, -0.0663459524512291, -0.020460480824112892, 0.09540658444166183, -0.16213038563728333, 0.022481130436062813, 0.08244425803422928, -0.12187694013118744, 0.09281346201896667, 0.016204802319407463, -0.006236857734620571, 0.025038830935955048, -0.1475188434123993, 0.034843120723962784, -0.03386561945080757, 0.010836300440132618, 0.04373383894562721, -0.21569781005382538, -0.00004886732858722098, -0.033673107624053955, -0.06639216095209122, -0.009451326914131641, -0.03672455996274948, -0.11508306115865707, 0.1058407872915268, 0.007236586883664131, -0.08753558248281479, -0.03186136856675148, 0.029325377196073532, 0.0838974118232727, -0.021959776058793068, 0.15145497024059296, -0.008370938710868359, 0.07429654151201248, -0.16209737956523895, -0.018623165786266327, -0.006028574425727129, 0.022658247500658035, -0.01664556935429573, -0.01111356820911169, 0.044031109660863876, -0.022746501490473747, 0.17925859987735748, -0.030318550765514374, 0.02272745408117771, 0.06815794110298157, 0.019072026014328003, -0.030184008181095123, 0.10406795144081116, 0.04094860330224037, 0.02014910988509655, 0.018591465428471565, 0.003289656015112996, -0.04647882282733917, -0.03173251822590828, -0.19407226145267487, 0.07288651913404465, 0.15608493983745575, 0.09729263186454773, -0.016707008704543114, 0.07954329252243042, -0.10199416428804398, -0.1109243705868721, 0.12477338314056396, -0.04797708988189697, -0.002418199321255088, -0.07150927931070328, 0.13247236609458923, 0.1437523066997528, -0.1859612911939621, 0.07269313186407089, -0.0699717253446579, -0.04708027467131615, -0.10980689525604248, -0.19441905617713928, -0.05561789125204086, -0.049456022679805756, -0.016053348779678345, -0.04698808491230011, 0.07504211366176605, 0.054538097232580185, 0.006766852922737598, -0.0023397188633680344, 0.06506035476922989, -0.031050674617290497, -0.0037882844917476177, 0.032597362995147705, 0.06591679900884628, 0.012734474614262581, -0.030802709981799126, 0.016619903966784477, -0.013545602560043335, 0.045626189559698105, 0.06578011065721512, 0.04976864159107208, -0.02938537672162056, 0.014603170566260815, -0.038539156317710876, -0.10249634087085724, 0.043612558394670486, -0.024421939626336098, -0.0789753645658493, 0.15477414429187775, 0.023680059239268303, 0.007779473438858986, -0.020137663930654526, 0.23901568353176117, -0.0738423764705658, -0.0964353010058403, -0.14737580716609955, 0.10557299107313156, -0.038081806153059006, 0.05800395458936691, 0.04625935107469559, -0.10226529091596603, 0.018044332042336464, 0.1338089406490326, 0.16182038187980652, -0.039008259773254395, 0.020095856860280037, 0.031135575845837593, 0.00566398398950696, -0.03622615709900856, 0.04847532883286476, 0.06906453520059586, 0.16569648683071136, -0.04632584750652313, 0.09100406616926193, 0.0019041687482967973, -0.09579581767320633, -0.038361791521310806, 0.11069868505001068, -0.016052277758717537, 0.019335128366947174, -0.05818064883351326, 0.11742528527975082, -0.06386786699295044, -0.23783175647258759, 0.06453443318605423, -0.0684293657541275, -0.13765870034694672, -0.02378307841718197, 0.08207765966653824, -0.012955902144312859, 0.027587108314037323, 0.0730307325720787, -0.07240920513868332, 0.201939657330513, 0.03798431158065796, -0.05499868467450142, -0.055047210305929184, 0.0805421993136406, -0.10008571296930313, 0.2739645540714264, 0.01557221356779337, 0.04601577669382095, 0.10384146869182587, -0.009341772645711899, -0.13838784396648407, 0.019836371764540672, 0.09581108391284943, -0.10502193123102188, 0.04196618124842644, 0.19815568625926971, -0.0014755994779989123, 0.12389086186885834, 0.07657600939273834, -0.07551808655261993, 0.0478031262755394, -0.08054235577583313, -0.06760486960411072, -0.09260394424200058, 0.09703279286623001, -0.07772123068571091, 0.14251399040222168, 0.13876807689666748, -0.05074559152126312, 0.012724342755973339, -0.031311117112636566, 0.044293127954006195, -0.00010600237874314189, 0.10321761667728424, 0.004272161517292261, -0.1832672357559204, 0.024692710489034653, 0.005650998093187809, 0.10749758034944534, -0.16033467650413513, -0.09566054493188858, 0.042343202978372574, 0.003505636239424348, -0.0672195628285408, 0.1290110945701599, 0.05665452033281326, 0.04342988133430481, -0.03997718170285225, -0.03521440550684929, -0.0060732318088412285, 0.13561366498470306, -0.10713256150484085, 0.0009933578548952937 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlmr-lstm-crf-resume-ner3 This model is a fine-tuned version of [hiendang7613/xlmr-lstm-crf-resume-ner3](https://huggingface.co/hiendang7613/xlmr-lstm-crf-resume-ner3) on the fcv_dataset dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 128 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 100 ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
{"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["fcv_dataset"], "base_model": "hiendang7613/xlmr-lstm-crf-resume-ner3", "model-index": [{"name": "xlmr-lstm-crf-resume-ner3", "results": []}]}
token-classification
hiendang7613/xlmr-lstm-crf-resume-ner3
[ "transformers", "safetensors", "xlm-roberta", "token-classification", "generated_from_trainer", "dataset:fcv_dataset", "base_model:hiendang7613/xlmr-lstm-crf-resume-ner3", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-15T02:53:17+00:00
[]
[]
TAGS #transformers #safetensors #xlm-roberta #token-classification #generated_from_trainer #dataset-fcv_dataset #base_model-hiendang7613/xlmr-lstm-crf-resume-ner3 #license-mit #autotrain_compatible #endpoints_compatible #region-us
# xlmr-lstm-crf-resume-ner3 This model is a fine-tuned version of hiendang7613/xlmr-lstm-crf-resume-ner3 on the fcv_dataset dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 128 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 100 ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
[ "# xlmr-lstm-crf-resume-ner3\n\nThis model is a fine-tuned version of hiendang7613/xlmr-lstm-crf-resume-ner3 on the fcv_dataset dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 32\n- eval_batch_size: 32\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 128\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 100", "### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #safetensors #xlm-roberta #token-classification #generated_from_trainer #dataset-fcv_dataset #base_model-hiendang7613/xlmr-lstm-crf-resume-ner3 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n", "# xlmr-lstm-crf-resume-ner3\n\nThis model is a fine-tuned version of hiendang7613/xlmr-lstm-crf-resume-ner3 on the fcv_dataset dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 32\n- eval_batch_size: 32\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 128\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 100", "### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1" ]
[ 89, 58, 6, 12, 8, 3, 113, 33 ]
[ "passage: TAGS\n#transformers #safetensors #xlm-roberta #token-classification #generated_from_trainer #dataset-fcv_dataset #base_model-hiendang7613/xlmr-lstm-crf-resume-ner3 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# xlmr-lstm-crf-resume-ner3\n\nThis model is a fine-tuned version of hiendang7613/xlmr-lstm-crf-resume-ner3 on the fcv_dataset dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 32\n- eval_batch_size: 32\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 128\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 100### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1" ]
[ -0.11198334395885468, 0.1772160530090332, -0.003713583806529641, 0.06395334750413895, 0.12606748938560486, 0.029043182730674744, 0.09085119515657425, 0.13713791966438293, -0.043740686029195786, 0.12877154350280762, 0.07809972763061523, 0.02292197197675705, 0.06700318306684494, 0.16127552092075348, -0.015053002163767815, -0.20801648497581482, 0.006131189409643412, -0.020306430757045746, -0.07763273268938065, 0.072365403175354, 0.1215689480304718, -0.11275457590818405, 0.08278673142194748, 0.0008235960849560797, -0.12688048183918, 0.0248924121260643, -0.03688618540763855, -0.04934906214475632, 0.07213534414768219, 0.030609985813498497, 0.07037962228059769, -0.002894781995564699, 0.1348026543855667, -0.2274244874715805, -0.0032376006711274385, 0.061295926570892334, 0.026972105726599693, 0.10272027552127838, 0.06108739227056503, 0.006502753123641014, 0.049043938517570496, -0.1732800453901291, 0.069344162940979, 0.03112390823662281, -0.0640374943614006, -0.16005900502204895, -0.1000903844833374, 0.104000024497509, 0.09631989896297455, 0.07135546207427979, 0.015109419822692871, 0.1128506287932396, -0.042382486164569855, 0.0633123442530632, 0.22017304599285126, -0.2269117832183838, -0.050362687557935715, 0.09013479202985764, 0.057116176933050156, 0.042575713247060776, -0.11550600826740265, 0.0004448912513907999, 0.07208481431007385, -0.002807296346873045, 0.1063263788819313, -0.0056677269749343395, -0.03930402919650078, 0.0010950666619464755, -0.10876411199569702, -0.03619825094938278, 0.18469023704528809, 0.0868353471159935, -0.032435134053230286, -0.12501567602157593, -0.034379806369543076, -0.12768028676509857, -0.021909737959504128, -0.02775648795068264, 0.028555890545248985, -0.0705462396144867, -0.03780791535973549, -0.032934579998254776, -0.06452485173940659, -0.05160399153828621, 0.02896670624613762, 0.10957831889390945, 0.03979267552495003, 0.0069714877754449844, 0.0074773808009922504, 0.09841931611299515, -0.03295496106147766, -0.12628807127475739, -0.04146253317594528, -0.0025913254357874393, -0.12111512571573257, -0.06119740754365921, -0.014118568040430546, 0.05063788592815399, 0.028911402449011803, 0.1585143506526947, -0.048637084662914276, 0.04849957674741745, 0.082705058157444, -0.011351747438311577, -0.0004940970684401691, 0.14938591420650482, -0.06801959872245789, -0.09758565574884415, 0.005682338494807482, 0.08432897925376892, 0.012119659222662449, 0.0006896477425470948, -0.0860297903418541, -0.04075087979435921, 0.06534159183502197, 0.059140369296073914, -0.0305472444742918, 0.03084055334329605, -0.029366323724389076, -0.027154995128512383, 0.07300231605768204, -0.12658420205116272, 0.04345839470624924, -0.0008463950362056494, -0.07125263661146164, -0.014389476738870144, 0.009607677347958088, -0.013166313990950584, -0.018756497651338577, 0.038801949471235275, -0.08128662407398224, -0.033558160066604614, -0.05774400383234024, -0.06213188171386719, 0.014570570550858974, -0.05870446935296059, 0.01354395505040884, -0.07371465116739273, -0.1991603672504425, -0.04323922097682953, 0.005031826440244913, -0.07285723835229874, -0.0672745406627655, -0.012396431528031826, -0.0494295209646225, 0.027976423501968384, -0.011777800507843494, 0.06941662728786469, -0.03234831243753433, 0.03743460029363632, 0.03134896233677864, 0.04162056744098663, 0.06318061798810959, 0.025522371754050255, -0.06477641314268112, 0.052461352199316025, -0.11889076232910156, 0.07892701774835587, -0.09781134128570557, 0.03192148357629776, -0.12819817662239075, -0.08143047243356705, -0.01400740072131157, -0.01818472519516945, 0.04169962927699089, 0.11302975565195084, -0.14806240797042847, -0.02225366234779358, 0.1546323299407959, -0.03602683171629906, -0.10796864330768585, 0.07479238510131836, -0.028600040823221207, -0.013871556147933006, 0.05381346493959427, 0.10263185948133469, 0.16418279707431793, -0.1500386893749237, -0.05944940447807312, 0.06855817884206772, 0.08631479740142822, 0.013012287206947803, 0.07422339171171188, -0.013513226062059402, 0.046270307153463364, 0.0189540795981884, -0.10652264952659607, 0.00697632972151041, -0.07763882726430893, -0.05917850509285927, -0.04891538619995117, -0.0780593603849411, 0.036246903240680695, 0.0012154090218245983, 0.01421916764229536, -0.051673367619514465, -0.10274337232112885, 0.0551266074180603, 0.16969959437847137, -0.06381717324256897, -0.001138020190410316, -0.06615269184112549, 0.05688255652785301, -0.043152835220098495, -0.025035066530108452, -0.16759109497070312, -0.14087596535682678, 0.059271812438964844, -0.06792067736387253, -0.000027132378818350844, 0.03605734929442406, 0.049768656492233276, 0.0644988864660263, -0.03506319969892502, -0.041030898690223694, -0.06215810775756836, -0.011605706997215748, -0.08499685674905777, -0.15629935264587402, -0.07400201261043549, -0.035073522478342056, 0.1632806658744812, -0.24227076768875122, 0.013015972450375557, -0.011823632754385471, 0.14363016188144684, 0.022776959463953972, -0.053563524037599564, 0.008851603604853153, -0.0037952500861138105, -0.03049341030418873, -0.11541163921356201, -0.00870704185217619, -0.0030403004493564367, -0.08875429630279541, -0.050405438989400864, -0.1298685520887375, 0.04370416700839996, 0.05809327960014343, 0.12173531949520111, -0.10137789696455002, -0.08280298113822937, -0.04910578951239586, -0.03075924701988697, -0.0797312781214714, -0.05143718421459198, 0.1486310064792633, 0.004660032223910093, 0.11919271945953369, -0.06445741653442383, -0.08501557260751724, 0.004735851660370827, 0.006320740561932325, -0.026256727054715157, 0.07951141893863678, 0.052798278629779816, -0.09156617522239685, 0.09321863204240799, 0.07723414152860641, -0.02023361623287201, 0.09536013752222061, -0.04497189074754715, -0.08942946791648865, -0.04704413190484047, 0.014982718043029308, -0.010047976858913898, 0.10950618982315063, -0.033549483865499496, 0.007962162606418133, 0.05019769445061684, -0.0015829293988645077, 0.015050123445689678, -0.11013638228178024, 0.005131933372467756, 0.050237491726875305, -0.046992283314466476, 0.051085151731967926, -0.018827401101589203, 0.013899268582463264, 0.07547923177480698, 0.004931438714265823, -0.01813860796391964, 0.00033550558146089315, -0.015308991074562073, -0.0771990567445755, 0.18615315854549408, -0.10325587540864944, -0.1471739113330841, -0.13242723047733307, 0.028721027076244354, -0.0398833192884922, -0.016463851556181908, -0.02996106818318367, -0.06502357870340347, -0.07849115133285522, -0.09194427728652954, -0.00976740475744009, -0.03227442502975464, 0.0011280286125838757, 0.09145321696996689, 0.016418198123574257, 0.08136864751577377, -0.12513038516044617, 0.01619068719446659, 0.0042781219817698, -0.05132519453763962, -0.020713631063699722, 0.05477561429142952, 0.10891078412532806, 0.10418981313705444, -0.014396848157048225, 0.023633701726794243, -0.032033637166023254, 0.22274154424667358, -0.08036857843399048, 0.013326616026461124, 0.12916414439678192, 0.03473367914557457, 0.04750261455774307, 0.10692189633846283, 0.02349805273115635, -0.06412813067436218, 0.023538347333669662, 0.04687868803739548, -0.009650669060647488, -0.22288480401039124, -0.04592782258987427, -0.03626413643360138, -0.052334263920784, 0.11499586701393127, 0.04709235578775406, 0.03866871818900108, 0.0626121535897255, -0.053278014063835144, 0.058968450874090195, -0.02124049887061119, 0.08751822263002396, 0.05728847160935402, 0.03620517998933792, 0.07456959784030914, -0.02160808816552162, -0.01506536640226841, 0.06160373240709305, 0.03928626328706741, 0.23200532793998718, -0.0320255309343338, 0.1633499413728714, 0.004928514361381531, 0.12937913835048676, -0.03742204234004021, 0.033458638936281204, 0.02326766960322857, -0.014510678127408028, 0.017695972695946693, -0.0710015818476677, -0.004393541254103184, 0.04345693439245224, -0.0024874908849596977, 0.058898478746414185, -0.08284182846546173, 0.07501675188541412, 0.020429514348506927, 0.19834023714065552, 0.06366482377052307, -0.304411917924881, -0.0626738891005516, 0.03283178061246872, 0.001501750317402184, -0.07538998872041702, 0.0012135939905419946, 0.13670557737350464, -0.12688089907169342, 0.07374969124794006, -0.03160976618528366, 0.07502935826778412, -0.0548609159886837, -0.017972763627767563, 0.06182991713285446, 0.1467607170343399, -0.0013876548036932945, 0.09918469935655594, -0.1224704161286354, 0.17694343626499176, 0.02570868656039238, 0.06644793599843979, -0.051073625683784485, 0.06342410296201706, -0.011637425981462002, 0.07834061980247498, 0.14014136791229248, -0.002392534166574478, -0.09788449108600616, -0.1954805552959442, -0.10847760736942291, 0.005920750088989735, 0.10271847248077393, -0.08649606257677078, 0.10382988303899765, -0.04779819771647453, -0.02464253082871437, 0.03500622138381004, -0.07350265979766846, -0.14342528581619263, -0.15327703952789307, 0.017110390588641167, 0.02494548074901104, -0.030491864308714867, -0.08630592375993729, -0.08178023248910904, -0.0868266150355339, 0.17701534926891327, -0.051572542637586594, -0.0445764996111393, -0.1511680781841278, 0.10156341642141342, 0.15627659857273102, -0.08921521157026291, 0.027058426290750504, 0.015146869234740734, 0.18593409657478333, 0.026866327971220016, -0.06934903562068939, 0.03015572763979435, -0.08839771151542664, -0.1595085710287094, -0.055619578808546066, 0.14614124596118927, 0.028988191857933998, 0.041181303560733795, 0.010049046948552132, -0.0011054797796532512, 0.02789878472685814, -0.09066961705684662, 0.005915206857025623, 0.1039055585861206, 0.08779268711805344, 0.07148011028766632, -0.0661846473813057, 0.023300306871533394, -0.02480808086693287, 0.002007479313760996, 0.11623496562242508, 0.20489460229873657, -0.10295560956001282, 0.03968443349003792, 0.03239540383219719, -0.07758785039186478, -0.15755487978458405, 0.05343307554721832, 0.1143854632973671, 0.04056280851364136, 0.06680164486169815, -0.1485486626625061, 0.1178613156080246, 0.10056518763303757, -0.011945288628339767, 0.03378136456012726, -0.2781219482421875, -0.11832772195339203, 0.08318514376878738, 0.06392759084701538, -0.07364460080862045, -0.1235654428601265, -0.062958724796772, -0.04674724116921425, -0.10201767086982727, 0.07232195883989334, -0.038033969700336456, 0.08000462502241135, -0.015424593351781368, 0.0362401120364666, 0.044381946325302124, -0.03400976583361626, 0.158378466963768, 0.05620802938938141, 0.08547461032867432, -0.03420848771929741, 0.029407020658254623, 0.08868173509836197, -0.08707109093666077, 0.08417857438325882, -0.05614220350980759, 0.07760842144489288, -0.1710413098335266, -0.019278695806860924, -0.04392014443874359, 0.06306183338165283, -0.0545576848089695, -0.04966742917895317, -0.06597372144460678, 0.04682043567299843, 0.0642174780368805, -0.025294486433267593, 0.07727525383234024, 0.04801817238330841, 0.05051835998892784, 0.1198565885424614, 0.07176665961742401, 0.055594250559806824, -0.14725527167320251, -0.0018118879524990916, 0.00006444570317398757, 0.03798043355345726, -0.10652485489845276, 0.0301260594278574, 0.1116066426038742, 0.05261398106813431, 0.13134066760540009, 0.008928913623094559, -0.05309649929404259, -0.024588026106357574, 0.03911278769373894, -0.07365201413631439, -0.14267367124557495, -0.03471909090876579, -0.02776973508298397, -0.1678064614534378, 0.025185197591781616, 0.10186798125505447, -0.045389022678136826, -0.010487817227840424, -0.03568251430988312, 0.03484627604484558, 0.00553538603708148, 0.1819465011358261, 0.06879781186580658, 0.0759827122092247, -0.07501495629549026, 0.09672477096319199, 0.08412586897611618, -0.026110142469406128, 0.03253830969333649, 0.025457259267568588, -0.07544339448213577, -0.007528876420110464, 0.05625840649008751, 0.09891543537378311, -0.043989721685647964, -0.04021947458386421, -0.05730780214071274, -0.09080874919891357, 0.02826330065727234, 0.05872712656855583, 0.04161698743700981, -0.02148256078362465, 0.0037673835176974535, -0.0013672822387889028, -0.13627426326274872, 0.11878912895917892, 0.05352921783924103, 0.09008077532052994, -0.17464210093021393, 0.09042102843523026, -0.010611923411488533, 0.0395081490278244, -0.006839806213974953, 0.025678221136331558, -0.07313869148492813, -0.04539819806814194, -0.09515538066625595, 0.020556161180138588, -0.04124283045530319, -0.0000035837292671203613, -0.009825095534324646, -0.026271836832165718, -0.023292360827326775, 0.04179638624191284, -0.042910993099212646, -0.06667470932006836, 0.008798412047326565, 0.08977450430393219, -0.12141553312540054, -0.01708114892244339, 0.013842473737895489, -0.10587688535451889, 0.08996091783046722, 0.028960859403014183, 0.05848982185125351, -0.007753957528620958, -0.0836394876241684, 0.01238009799271822, 0.05758444964885712, 0.05970514565706253, 0.057590700685977936, -0.11461973190307617, 0.001782607869245112, -0.04078434780240059, 0.006850551348179579, -0.013491982594132423, -0.030601030215620995, -0.12512992322444916, -0.03463806211948395, -0.08367044478654861, -0.009069329127669334, -0.05881185829639435, 0.05454317480325699, 0.09799768775701523, 0.003586424980312586, 0.119393490254879, -0.04963697865605354, 0.03991665691137314, -0.2264838069677353, -0.023647066205739975, -0.005730285309255123, -0.008807656355202198, -0.08288557827472687, -0.008846800774335861, 0.08173990994691849, -0.03248758986592293, 0.1382835954427719, -0.029039278626441956, 0.08419468253850937, 0.019305557012557983, -0.031891804188489914, 0.049268584698438644, 0.018007272854447365, 0.15362794697284698, 0.05180417001247406, -0.013504291884601116, 0.08375316113233566, -0.004271224606782198, 0.09475640952587128, 0.002972931833937764, 0.1172361746430397, 0.13555970788002014, -0.05181808024644852, 0.05409453064203262, 0.07357084006071091, -0.1443987339735031, -0.10359781235456467, 0.08832095563411713, -0.0602489672601223, 0.0935106948018074, -0.03252020105719566, 0.10207343846559525, 0.10976100713014603, -0.18716274201869965, 0.021916817873716354, -0.03451303020119667, -0.11439124494791031, -0.09609164297580719, -0.09207260608673096, -0.0852426290512085, -0.11767812073230743, 0.016077842563390732, -0.14406327903270721, 0.008476858027279377, 0.06519730389118195, 0.0015215349849313498, 0.004312141798436642, 0.13808013498783112, -0.025286421179771423, -0.0019061763305217028, 0.044212471693754196, 0.037807896733284, 0.014352140948176384, -0.0406787283718586, -0.04338623955845833, 0.06394419819116592, 0.00012468855129554868, 0.0822886973619461, -0.05320841073989868, 0.05260652303695679, 0.06259694695472717, 0.01529052946716547, -0.067011259496212, 0.004513142164796591, -0.005362666677683592, 0.0060341134667396545, 0.03583970665931702, 0.0479714460670948, 0.008538518100976944, -0.058267612010240555, 0.2222682684659958, -0.06782901287078857, -0.019265778362751007, -0.12903688848018646, 0.09084877371788025, 0.050187524408102036, 0.01481702458113432, 0.06862819194793701, -0.10610955953598022, -0.018227968364953995, 0.12619701027870178, 0.12111937999725342, -0.03750963509082794, -0.014935972169041634, -0.027429021894931793, -0.01884988509118557, -0.0541626401245594, 0.0855848416686058, 0.08404324948787689, 0.024867627769708633, -0.02300449088215828, -0.006215437315404415, 0.019107980653643608, -0.029720937833189964, -0.07132775336503983, 0.07074330747127533, -0.0016761217266321182, 0.025881120935082436, -0.012989307753741741, 0.058152154088020325, 0.022873099893331528, -0.19966544210910797, 0.04774585738778114, -0.13183008134365082, -0.1670820415019989, -0.02246272563934326, 0.04513392597436905, 0.012861286289989948, 0.042984962463378906, -0.014699799008667469, -0.019464027136564255, 0.1632983237504959, -0.01635882630944252, -0.08353319764137268, -0.08546619862318039, 0.06741796433925629, -0.061951179057359695, 0.22342267632484436, -0.004395348951220512, 0.07497710734605789, 0.09946803003549576, -0.023827098309993744, -0.15347272157669067, 0.0164655614644289, 0.081386499106884, -0.040061671286821365, 0.061006393283605576, 0.1419476568698883, -0.03086220845580101, 0.10337994992733002, 0.05370120704174042, -0.11179040372371674, -0.03314100205898285, -0.012992148287594318, -0.000006533704890898662, -0.10527264326810837, 0.01277585607022047, -0.04757694900035858, 0.17086927592754364, 0.1768409162759781, -0.06138845905661583, -0.018414659425616264, -0.033521875739097595, 0.03623001649975777, 0.05186391621828079, 0.11506666988134384, 0.013938331045210361, -0.17978999018669128, 0.030835499987006187, -0.005038651637732983, 0.038012973964214325, -0.18641112744808197, -0.08483076095581055, 0.034329745918512344, -0.06163223832845688, -0.03718459978699684, 0.12021186947822571, 0.05029578506946564, 0.031785011291503906, -0.03714548796415329, -0.0810949057340622, -0.05387285724282265, 0.12205111980438232, -0.15935111045837402, -0.055565815418958664 ]
null
null
transformers
# NeuralLogic-7B-V NeuralLogic-7B-V is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [Kukedlc/Triunvirato-7b](https://huggingface.co/Kukedlc/Triunvirato-7b) * [mlabonne/NeuralHermes-2.5-Mistral-7B](https://huggingface.co/mlabonne/NeuralHermes-2.5-Mistral-7B) ## 🧩 Configuration ```yaml slices: - sources: - model: Kukedlc/Triunvirato-7b layer_range: [0, 32] - model: mlabonne/NeuralHermes-2.5-Mistral-7B layer_range: [0, 32] merge_method: slerp base_model: Kukedlc/Triunvirato-7b parameters: t: - filter: self_attn value: [0.1, 0.6, 0.3, 0.7, 1] - filter: mlp value: [0.9, 0.4, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ``` ## 💻 Usage ```python !pip install -qU transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "Kukedlc/NeuralLogic-7B-V" messages = [{"role": "user", "content": "What is a large language model?"}] tokenizer = AutoTokenizer.from_pretrained(model) prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ```
{"tags": ["merge", "mergekit", "lazymergekit", "Kukedlc/Triunvirato-7b", "mlabonne/NeuralHermes-2.5-Mistral-7B"], "base_model": ["Kukedlc/Triunvirato-7b", "mlabonne/NeuralHermes-2.5-Mistral-7B"]}
text-generation
Kukedlc/NeuralLogic-7B-V
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "Kukedlc/Triunvirato-7b", "mlabonne/NeuralHermes-2.5-Mistral-7B", "base_model:Kukedlc/Triunvirato-7b", "base_model:mlabonne/NeuralHermes-2.5-Mistral-7B", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T02:53:46+00:00
[]
[]
TAGS #transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #Kukedlc/Triunvirato-7b #mlabonne/NeuralHermes-2.5-Mistral-7B #base_model-Kukedlc/Triunvirato-7b #base_model-mlabonne/NeuralHermes-2.5-Mistral-7B #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# NeuralLogic-7B-V NeuralLogic-7B-V is a merge of the following models using LazyMergekit: * Kukedlc/Triunvirato-7b * mlabonne/NeuralHermes-2.5-Mistral-7B ## Configuration ## Usage
[ "# NeuralLogic-7B-V\n\nNeuralLogic-7B-V is a merge of the following models using LazyMergekit:\n* Kukedlc/Triunvirato-7b\n* mlabonne/NeuralHermes-2.5-Mistral-7B", "## Configuration", "## Usage" ]
[ "TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #Kukedlc/Triunvirato-7b #mlabonne/NeuralHermes-2.5-Mistral-7B #base_model-Kukedlc/Triunvirato-7b #base_model-mlabonne/NeuralHermes-2.5-Mistral-7B #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# NeuralLogic-7B-V\n\nNeuralLogic-7B-V is a merge of the following models using LazyMergekit:\n* Kukedlc/Triunvirato-7b\n* mlabonne/NeuralHermes-2.5-Mistral-7B", "## Configuration", "## Usage" ]
[ 120, 58, 4, 3 ]
[ "passage: TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #Kukedlc/Triunvirato-7b #mlabonne/NeuralHermes-2.5-Mistral-7B #base_model-Kukedlc/Triunvirato-7b #base_model-mlabonne/NeuralHermes-2.5-Mistral-7B #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# NeuralLogic-7B-V\n\nNeuralLogic-7B-V is a merge of the following models using LazyMergekit:\n* Kukedlc/Triunvirato-7b\n* mlabonne/NeuralHermes-2.5-Mistral-7B## Configuration## Usage" ]
[ -0.013844859786331654, -0.011602162383496761, -0.0023680776357650757, 0.00947654154151678, 0.09292692691087723, 0.08292701840400696, 0.17529812455177307, 0.09945034235715866, 0.03433153033256531, 0.05885445699095726, 0.07621511071920395, 0.16097185015678406, 0.007181650027632713, 0.0048358263447880745, -0.011505188420414925, -0.24814248085021973, 0.1250292956829071, 0.03044966794550419, 0.013221279717981815, 0.03414314612746239, 0.07486648112535477, -0.04866805300116539, 0.07227183878421783, -0.007035371381789446, -0.08229311555624008, -0.04171421751379967, 0.04944341629743576, -0.02655794657766819, 0.10517683625221252, 0.07234467566013336, 0.06021350249648094, 0.04612918570637703, 0.042130034416913986, -0.09495369344949722, 0.03696165606379509, -0.012298638001084328, -0.004037014674395323, 0.08319111913442612, 0.05704345181584358, -0.017376096919178963, 0.16581374406814575, -0.022787652909755707, 0.018333153799176216, 0.01081604603677988, -0.09209935367107391, -0.0913756787776947, -0.0795777440071106, 0.10700330883264542, 0.0465170219540596, 0.04912729933857918, -0.034024037420749664, 0.16850803792476654, -0.0030524232424795628, 0.07333196699619293, 0.21660301089286804, -0.3135565221309662, -0.05217886343598366, 0.14938391745090485, 0.040911830961704254, -0.05901027470827103, 0.024931607767939568, 0.10222002863883972, 0.05332496389746666, -0.011007693596184254, 0.06060726195573807, -0.05458451434969902, 0.10945754498243332, -0.07288529723882675, -0.14715895056724548, 0.032724060118198395, 0.03929678350687027, 0.005621961317956448, 0.01531671267002821, -0.0458107627928257, -0.07012355327606201, 0.1195613369345665, -0.06978819519281387, -0.05762379243969917, -0.02239123359322548, -0.04186065495014191, 0.04593745246529579, -0.06887796521186829, -0.024359045550227165, -0.034038152545690536, -0.0687045231461525, 0.18515048921108246, 0.03955858573317528, -0.022234365344047546, 0.004339860286563635, 0.05572390556335449, -0.010099802166223526, -0.10338465869426727, 0.032861120998859406, -0.056095488369464874, -0.01523752510547638, -0.02097325026988983, -0.05882271006703377, -0.13851170241832733, 0.09023857861757278, 0.09177688509225845, 0.034880731254816055, -0.0021758279763162136, 0.0035505203995853662, 0.021910903975367546, -0.022177455946803093, -0.005270432215183973, -0.21002204716205597, -0.04444152116775513, 0.06139856576919556, 0.12878580391407013, 0.090861976146698, 0.008873610757291317, -0.10173451900482178, -0.039899446070194244, 0.09298187494277954, 0.02558768168091774, 0.026588566601276398, 0.11601433157920837, -0.07057894766330719, -0.1017136499285698, 0.10573528707027435, -0.03349820151925087, -0.007249929942190647, 0.009785483591258526, -0.07056604325771332, 0.07200565189123154, 0.009361851960420609, 0.02373976819217205, -0.0137679073959589, 0.05294881761074066, -0.1027308851480484, -0.012541001662611961, -0.04967827349901199, -0.14462342858314514, 0.01464444026350975, -0.014942063950002193, -0.025838661938905716, -0.12914565205574036, -0.17743130028247833, 0.012819795869290829, 0.06489374488592148, -0.07480426877737045, 0.033609505742788315, -0.0759502649307251, -0.043936584144830704, 0.014350770972669125, 0.005860683973878622, -0.03965111821889877, 0.001827975851483643, -0.009270135313272476, -0.010468512773513794, 0.08835618942975998, -0.1213926449418068, 0.010777256451547146, -0.07269380241632462, 0.07472612708806992, -0.13496993482112885, 0.04697798192501068, -0.0669291615486145, 0.05281440541148186, -0.09489845484495163, -0.02667759545147419, -0.004434880334883928, -0.015028932131826878, 0.09706047922372818, 0.10751651972532272, -0.12710487842559814, -0.06727886945009232, 0.0662723183631897, -0.10619013011455536, -0.13601192831993103, 0.09569264948368073, 0.0025131350848823786, 0.12092041224241257, 0.09526513516902924, 0.20095594227313995, 0.06863777339458466, -0.017294924706220627, -0.06290677189826965, -0.10200006514787674, 0.006505351513624191, 0.04181818664073944, 0.05489245057106018, 0.08469992130994797, -0.14509451389312744, 0.05960153415799141, -0.0028403704054653645, 0.09516210854053497, -0.0547795407474041, -0.058425698429346085, -0.02702004462480545, -0.10655172169208527, 0.17215506732463837, -0.011173916980624199, 0.07057303190231323, -0.04848780855536461, -0.05194500833749771, 0.11825241148471832, 0.11903537809848785, -0.08251605927944183, -0.0025862371549010277, -0.1054241955280304, 0.11541111022233963, -0.06422639638185501, 0.065439872443676, -0.08873485028743744, -0.15993423759937286, 0.03587133064866066, -0.05392573028802872, 0.04494967311620712, -0.013003658503293991, 0.03230537101626396, 0.04217341169714928, -0.09067024290561676, -0.031246667727828026, 0.05976644158363342, 0.04076322168111801, -0.015246903523802757, -0.2066187560558319, -0.018429111689329147, -0.04108113795518875, 0.1998189091682434, -0.11252523213624954, 0.027644654735922813, -0.04155990108847618, 0.21996256709098816, -0.003538279328495264, 0.02316637709736824, 0.06191130355000496, -0.0003704771806951612, -0.03369751572608948, 0.003290161956101656, 0.117437943816185, -0.07807225733995438, -0.12379740923643112, 0.05572138726711273, -0.09399669617414474, 0.052236247807741165, 0.0830138772726059, 0.09685845673084259, -0.01775357872247696, 0.05562157556414604, 0.025279240682721138, -0.04268866777420044, 0.09712401032447815, -0.052774492651224136, -0.0034187277778983116, -0.001271128305234015, 0.05958956480026245, -0.08027333766222, -0.005660271737724543, 0.028085952624678612, -0.013634677976369858, -0.05403144285082817, 0.048294126987457275, -0.010008386336266994, -0.2977651059627533, 0.04651079699397087, 0.07646184414625168, -0.05828956142067909, 0.09211567044258118, 0.041467007249593735, -0.0030523473396897316, -0.09522690623998642, -0.04071551561355591, 0.06224743276834488, 0.0465041920542717, -0.04516427963972092, 0.042541053146123886, 0.0452006533741951, -0.023522302508354187, 0.03568962961435318, -0.05643949285149574, 0.005787841975688934, -0.036858703941106796, -0.002443572971969843, 0.1311108022928238, 0.11251871287822723, -0.02445392683148384, 0.09461217373609543, 0.007202116772532463, -0.068605437874794, 0.009783036075532436, -0.007451018318533897, -0.0776180550456047, 0.15372876822948456, -0.14805565774440765, -0.25957250595092773, -0.10353522002696991, -0.009879506193101406, -0.12529084086418152, -0.011309151537716389, 0.019095486029982567, -0.017485007643699646, -0.03597467392683029, -0.09543207287788391, 0.12211564183235168, 0.09867145866155624, 0.022939566522836685, 0.01988225243985653, 0.0018077254062518477, 0.0072012548334896564, -0.0882406011223793, 0.003236709861084819, -0.04347569867968559, -0.05166596174240112, 0.04828744754195213, -0.07085517048835754, 0.09403977543115616, 0.14528986811637878, 0.029284115880727768, -0.04931386932730675, -0.03400929272174835, 0.22633780539035797, -0.028347045183181763, 0.07357581704854965, 0.12879012525081635, -0.010784988291561604, 0.006723672151565552, 0.1860949546098709, 0.03183763846755028, -0.05713089928030968, 0.04104616120457649, 0.01183420978486538, -0.003426965093240142, -0.1742994636297226, -0.11975327134132385, -0.00824365671724081, 0.025417378172278404, 0.0429077111184597, 0.016094902530312538, 0.08711301535367966, 0.10158799588680267, -0.051695775240659714, -0.018049685284495354, 0.00731766177341342, 0.09035269916057587, 0.2480459213256836, 0.01265338622033596, 0.11206495761871338, 0.00030927290208637714, -0.04033621400594711, 0.02102256938815117, 0.004193651955574751, 0.1265663504600525, -0.003644368378445506, 0.11538306623697281, 0.03218982368707657, 0.06445026397705078, 0.04097745195031166, 0.07481896877288818, -0.0587727427482605, -0.045698750764131546, -0.02951745316386223, -0.078739695250988, -0.009046347811818123, 0.014851170592010021, -0.03273392841219902, 0.06982830166816711, -0.03120400942862034, 0.014971457421779633, 0.07746973633766174, 0.08110956847667694, 0.1349387764930725, -0.24982041120529175, -0.10336894541978836, 0.03975026309490204, 0.018034735694527626, -0.008900870569050312, -0.020387519150972366, 0.04142044857144356, -0.05478044971823692, 0.09242025762796402, -0.04443934187293053, 0.05661764368414879, -0.04570074379444122, 0.01755988784134388, -0.0820147767663002, 0.0699571818113327, -0.0013423117343336344, 0.08140894770622253, -0.15009073913097382, 0.129098579287529, 0.022376317530870438, 0.04594426229596138, 0.012090097181499004, -0.0236460380256176, 0.06735449284315109, 0.1582302302122116, 0.013880942016839981, 0.003432187717407942, 0.014146094210445881, -0.00578252924606204, -0.0772014781832695, -0.022447161376476288, 0.031436629593372345, -0.030797356739640236, 0.1332108974456787, -0.030790576711297035, -0.05883706733584404, 0.011517285369336605, 0.06451742351055145, -0.24273717403411865, -0.08324802666902542, 0.09222394227981567, 0.07305515557527542, 0.08238795399665833, -0.09133336693048477, -0.09674899280071259, -0.055067360401153564, 0.10544996708631516, 0.005160720553249121, -0.03472667559981346, -0.0783342570066452, 0.02822965383529663, 0.1679197996854782, -0.08705494552850723, 0.027545131742954254, -0.04205271229147911, 0.07692789286375046, -0.03860016167163849, -0.15219582617282867, 0.062491197139024734, -0.09441094100475311, -0.12742742896080017, -0.026509635150432587, 0.09951150417327881, -0.020583363249897957, 0.05051548406481743, 0.029288364574313164, 0.03784966841340065, 0.03206029161810875, -0.02848321571946144, 0.0010378849692642689, 0.14578042924404144, -0.0881311371922493, 0.060473501682281494, -0.12849418818950653, -0.12797506153583527, -0.09953445196151733, 0.005161929409950972, 0.1436290740966797, 0.274407297372818, -0.03268394619226456, 0.041339099407196045, 0.16592396795749664, -0.06751649081707001, -0.2602400779724121, -0.00800877995789051, 0.05895400419831276, -0.008110175840556622, 0.024064188823103905, -0.14261431992053986, 0.0644921287894249, 0.11760072410106659, -0.017853690311312675, 0.018952617421746254, -0.3083667457103729, -0.13945528864860535, 0.059703484177589417, 0.13656911253929138, 0.17743535339832306, -0.1211777999997139, -0.03904895484447479, -0.0848018005490303, -0.0886177271604538, 0.17452006042003632, -0.07228083908557892, 0.12505057454109192, -0.031386371701955795, -0.011802325956523418, 0.02038801833987236, -0.004324796609580517, 0.11423252522945404, -0.08050846308469772, 0.016227811574935913, -0.04353967308998108, -0.06095785275101662, 0.18185681104660034, -0.025949008762836456, 0.07671812176704407, -0.021102825179696083, 0.029730450361967087, 0.050063107162714005, -0.0433247834444046, -0.08541291952133179, 0.04315569996833801, -0.016951214522123337, -0.07598377764225006, -0.06476692110300064, 0.10521436482667923, 0.06100558862090111, 0.02464110404253006, 0.10284024477005005, -0.004374385345727205, 0.05284585431218147, 0.11279845982789993, 0.07709522545337677, -0.07376047968864441, 0.09089449793100357, 0.01314656063914299, -0.06077531725168228, 0.058728404343128204, 0.023601321503520012, -0.03696389123797417, 0.09426459670066833, -0.03890259563922882, 0.10008595138788223, 0.02708556316792965, -0.048268623650074005, -0.016325466334819794, 0.10219231247901917, -0.12264415621757507, -0.2658468782901764, -0.023880116641521454, 0.030738556757569313, -0.04176557436585426, 0.08675816655158997, 0.24792057275772095, -0.07769337296485901, -0.0244830921292305, 0.01609058864414692, -0.020449575036764145, -0.04105859249830246, 0.13789613544940948, 0.009373881854116917, 0.04232186824083328, -0.07108979672193527, 0.017010299488902092, 0.06299137324094772, -0.07913342863321304, -0.002947902772575617, 0.08809329569339752, -0.12726904451847076, -0.09974189847707748, -0.10403278470039368, 0.20549660921096802, -0.09319115430116653, -0.017218273133039474, -0.15396395325660706, -0.07814349234104156, -0.0055602234788239, 0.1825384795665741, 0.07979671657085419, 0.013571358285844326, -0.013101313263177872, -0.04482511058449745, -0.06551611423492432, 0.13201527297496796, 0.008296623826026917, 0.11526520550251007, -0.12977305054664612, -0.012798386625945568, -0.02915574423968792, -0.00882677361369133, -0.07799076288938522, -0.0010805202182382345, -0.17636436223983765, -0.05141286551952362, -0.22547994554042816, -0.03344399109482765, -0.12947258353233337, -0.040796294808387756, 0.008522857911884785, -0.019421664997935295, -0.014826606959104538, -0.004461628384888172, -0.03584156930446625, -0.026940777897834778, 0.0033438459504395723, 0.07468636333942413, -0.06187132000923157, -0.04459395632147789, -0.013279582373797894, -0.02863300032913685, 0.07866567373275757, 0.017836634069681168, 0.027196839451789856, 0.06362415105104446, -0.08407124131917953, -0.038289133459329605, 0.03437875211238861, 0.01920534297823906, 0.06332544982433319, -0.17826911807060242, -0.03188811615109444, -0.007878211326897144, 0.008143593557178974, 0.022524628788232803, 0.09123877435922623, -0.05045605078339577, -0.017484286800026894, -0.046266116201877594, 0.00047696783440187573, -0.019414523616433144, -0.010752053931355476, -0.03560337796807289, 0.00631632748991251, 0.16744443774223328, -0.08331169188022614, 0.006613054778426886, -0.10802344232797623, -0.0017875183839350939, -0.042713914066553116, -0.15244802832603455, -0.05891609564423561, -0.03212793543934822, 0.017307642847299576, -0.026290925219655037, 0.021207690238952637, -0.023661905899643898, -0.192291259765625, 0.013808335177600384, -0.02664492465555668, 0.011484699323773384, 0.04904929921030998, 0.136605903506279, 0.0930766761302948, -0.016803031787276268, -0.1079363226890564, 0.07654254138469696, 0.0238749198615551, 0.10539703816175461, 0.08483240753412247, 0.034353721886873245, 0.036053504794836044, 0.08298777788877487, 0.09140025824308395, 0.05813754349946976, -0.07255467772483826, 0.064808189868927, -0.036073386669158936, 0.05188128724694252, -0.015330959111452103, 0.13582982122898102, 0.16595065593719482, -0.08494318276643753, 0.020989498123526573, -0.013641871511936188, -0.053929638117551804, -0.05093549191951752, -0.17419008910655975, -0.13271759450435638, -0.17111583054065704, -0.0314759835600853, -0.1512242704629898, -0.09306246787309647, 0.006721455603837967, 0.02723391354084015, -0.002591146621853113, 0.19630520045757294, 0.03166602551937103, 0.007462365552783012, 0.034828558564186096, -0.021441834047436714, -0.08649960160255432, 0.017239410430192947, -0.023297114297747612, -0.036158088594675064, 0.008722767233848572, -0.02152559719979763, 0.006124500650912523, -0.06660237908363342, 0.023733843117952347, -0.0014858758077025414, -0.11794756352901459, 0.027109291404485703, 0.028633471578359604, 0.025716742500662804, 0.059386737644672394, 0.05611490458250046, -0.0944591611623764, -0.027069753035902977, 0.13906222581863403, -0.02130044996738434, -0.11254908889532089, -0.08684022724628448, 0.24197092652320862, -0.05776600539684296, 0.06239372864365578, -0.006933481898158789, -0.036626409739255905, -0.03172024339437485, 0.18330858647823334, 0.25664031505584717, -0.07121334969997406, -0.028355224058032036, -0.015855595469474792, -0.0016601416282355785, -0.02973749302327633, 0.09890192747116089, 0.010458774864673615, 0.1506146490573883, -0.06098824739456177, 0.025241365656256676, -0.04210473969578743, -0.05732794478535652, -0.09122739732265472, -0.08564162254333496, 0.06008214130997658, 0.010261829942464828, -0.01867796666920185, 0.09294070303440094, -0.13123719394207, 0.011906458996236324, -0.041530974209308624, -0.11729368567466736, -0.12558065354824066, -0.11284466832876205, 0.05400465056300163, -0.0381777323782444, 0.12236958742141724, -0.06682558357715607, 0.009444733150303364, 0.004751079715788364, -0.0313127227127552, -0.04451735317707062, -0.0722576379776001, 0.0664338767528534, -0.002873851917684078, 0.09723704308271408, -0.042679935693740845, 0.04333481937646866, 0.13367491960525513, -0.04024568200111389, -0.12531867623329163, 0.05275607109069824, 0.039509233087301254, -0.05725373327732086, 0.09084713459014893, 0.13551479578018188, -0.009848522022366524, 0.15611456334590912, 0.05210421606898308, -0.1687183678150177, 0.021021196618676186, 0.09015487879514694, 0.019994085654616356, -0.04315454140305519, 0.08857329934835434, -0.06145979464054108, 0.1334560364484787, 0.17373377084732056, -0.06067503243684769, -0.024100378155708313, -0.018831288442015648, 0.015118957497179508, 0.09380420297384262, 0.11963612586259842, -0.029149532318115234, -0.24780438840389252, -0.05054357647895813, 0.02962009608745575, 0.013832923956215382, -0.2716841995716095, -0.09993649274110794, -0.12476522475481033, -0.013913113623857498, -0.09014406055212021, 0.07674449682235718, 0.1531023532152176, -0.0005246862419880927, -0.01391368918120861, -0.15269023180007935, -0.06986676156520844, 0.06769411265850067, -0.11563429236412048, -0.14040975272655487 ]
null
null
peft
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.8.2
{"library_name": "peft", "base_model": "mistralai/mistral-7b-v0.1"}
null
pawan2411/ESGcombinedData-LoRA
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:mistralai/mistral-7b-v0.1", "region:us" ]
2024-02-15T02:54:32+00:00
[ "1910.09700" ]
[]
TAGS #peft #safetensors #arxiv-1910.09700 #base_model-mistralai/mistral-7b-v0.1 #region-us
# Model Card for Model ID ## Model Details ### Model Description - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact ### Framework versions - PEFT 0.8.2
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ "TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-mistralai/mistral-7b-v0.1 #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ 39, 6, 3, 54, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4, 11 ]
[ "passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-mistralai/mistral-7b-v0.1 #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2" ]
[ -0.11899120360612869, 0.2104109823703766, -0.0027906037867069244, 0.03118092752993107, 0.08750686794519424, 0.01975245587527752, 0.05095892399549484, 0.12541480362415314, -0.010217074304819107, 0.10841858386993408, 0.07001832872629166, 0.10695755481719971, 0.1116432398557663, 0.21738633513450623, 0.0034132348373532295, -0.1874093860387802, 0.027393320575356483, -0.09075714647769928, 0.001194633892737329, 0.12612509727478027, 0.14381824433803558, -0.1023091971874237, 0.08312153071165085, -0.014105898328125477, -0.00324136926792562, -0.03772240877151489, -0.06921181827783585, -0.022615013644099236, 0.04079878330230713, 0.03793589398264885, 0.0607740543782711, -0.007416970562189817, 0.0948612317442894, -0.26137417554855347, 0.018804360181093216, 0.045681312680244446, -0.001243007485754788, 0.09123539924621582, 0.10082155466079712, -0.04216529428958893, 0.11987194418907166, -0.030301248654723167, 0.1391197293996811, 0.08784705400466919, -0.08632715791463852, -0.22086915373802185, -0.0670289620757103, 0.08970329910516739, 0.1917114555835724, 0.07608018815517426, -0.0395476408302784, 0.12784133851528168, -0.07498826086521149, 0.020953109487891197, 0.03453224524855614, -0.08773895353078842, -0.06842830777168274, 0.06039523333311081, 0.1154782623052597, 0.06072290614247322, -0.12916715443134308, -0.03324943780899048, 0.027636824175715446, 0.0360838957130909, 0.07855849713087082, 0.010215988382697105, 0.16562971472740173, 0.0292774997651577, -0.1423611342906952, -0.0449204184114933, 0.14038434624671936, 0.023653971031308174, -0.03839389979839325, -0.22771398723125458, -0.006226068362593651, -0.09090570360422134, -0.02805851399898529, -0.051995646208524704, 0.03531576693058014, 0.00938830804079771, 0.11613599210977554, -0.03334273770451546, -0.09409711509943008, -0.019048072397708893, 0.09529106318950653, 0.0460384376347065, 0.023521756753325462, -0.018422352150082588, 0.004062324296683073, 0.1254531294107437, 0.0788412019610405, -0.13312995433807373, -0.05889269337058067, -0.07672346383333206, -0.04302334785461426, -0.036276597529649734, 0.046698108315467834, 0.04188153147697449, 0.05743258446455002, 0.2494467794895172, -0.024082006886601448, 0.06012117117643356, 0.07300996780395508, 0.01736772246658802, 0.04929600656032562, 0.0966423824429512, -0.05298350006341934, -0.1652916967868805, -0.01097804680466652, 0.09561049193143845, -0.0024318716023117304, -0.027305593714118004, -0.05136279761791229, 0.04288328438997269, 0.03416229411959648, 0.11071319878101349, 0.10742215812206268, -0.013462834060192108, -0.07816566526889801, -0.06269408017396927, 0.21524156630039215, -0.1540936529636383, 0.04594319313764572, 0.0227572750300169, -0.008074215613305569, -0.0430738665163517, 0.009797831065952778, 0.01713946834206581, -0.029479963704943657, 0.08153048902750015, -0.06970912218093872, -0.04280419275164604, -0.12419557571411133, -0.02671237476170063, 0.02813059277832508, 0.0069215320982038975, -0.03680720552802086, -0.03637115657329559, -0.07490842789411545, -0.09751129150390625, 0.10658825933933258, -0.05975869670510292, -0.05761215090751648, -0.029488956555724144, -0.09243814647197723, 0.026475999504327774, 0.02690735086798668, 0.07841771095991135, -0.0283634215593338, 0.04294651746749878, -0.014280539937317371, 0.06343408674001694, 0.07684655487537384, 0.03162851557135582, -0.0701599270105362, 0.0626249760389328, -0.19254496693611145, 0.07934091985225677, -0.07944734394550323, 0.03436725214123726, -0.1621738076210022, -0.010008329525589943, 0.016450313851237297, 0.022174149751663208, 0.03150167316198349, 0.162149116396904, -0.2168390452861786, -0.028007516637444496, 0.15547558665275574, -0.1031072735786438, -0.12041089683771133, 0.034475840628147125, -0.04259898141026497, 0.16461659967899323, 0.025213448330760002, -0.006136021576821804, 0.10210162401199341, -0.15874549746513367, -0.027002932503819466, -0.018608734011650085, -0.002704378217458725, 0.08258302509784698, 0.08742301166057587, -0.08641579747200012, 0.023058362305164337, 0.0146971819922328, -0.05407104268670082, -0.02110249735414982, -0.04342940077185631, -0.10645578801631927, 0.008305773138999939, -0.08344396203756332, 0.020540902391076088, -0.005029383581131697, -0.08987408876419067, -0.004362586885690689, -0.15881355106830597, -0.04557062312960625, 0.08658547699451447, 0.004673588555306196, -0.024090982973575592, -0.10393372923135757, 0.04802441596984863, -0.03700239956378937, -0.023048674687743187, -0.13933387398719788, -0.021634133532643318, 0.017478477209806442, -0.13603205978870392, -0.009776646271348, -0.11717197299003601, 0.06696562469005585, 0.013203580863773823, -0.054635390639305115, -0.04384777322411537, -0.001310712774284184, 0.006946849171072245, -0.05403148755431175, -0.24189414083957672, -0.03113928996026516, -0.05030995234847069, 0.1510135680437088, -0.22088944911956787, 0.04134318232536316, 0.03748974949121475, 0.12860433757305145, 0.002885582624003291, -0.06691576540470123, 0.026448113843798637, -0.07280705869197845, -0.027068400755524635, -0.07414884865283966, -0.004095721058547497, -0.004159152042120695, -0.0370136983692646, 0.018883368000388145, -0.11956165730953217, -0.04157957434654236, 0.10083581507205963, 0.06484252214431763, -0.15107908844947815, -0.00035795787698589265, -0.040047090500593185, -0.06116493418812752, -0.07491880655288696, -0.06662304699420929, 0.09442504495382309, 0.0546271875500679, 0.04000966250896454, -0.07317358255386353, -0.07457275688648224, 0.010716375894844532, -0.02254772186279297, -0.017548859119415283, 0.11212087422609329, 0.07500045746564865, -0.10377515852451324, 0.09629248827695847, 0.0788206085562706, 0.0361144132912159, 0.08473557233810425, -0.025471409782767296, -0.10529723763465881, -0.031559593975543976, 0.04582323133945465, 0.010942596010863781, 0.1705540418624878, -0.06342114508152008, 0.05431486293673515, 0.04662635922431946, -0.03894418478012085, 0.047194186598062515, -0.09163188934326172, 0.011139070615172386, 0.00856103003025055, -0.013093167915940285, 0.02231733314692974, -0.02610756829380989, 0.007508622016757727, 0.0788954496383667, 0.053542375564575195, 0.031003881245851517, 0.02448081038892269, -0.03389669582247734, -0.1369483917951584, 0.18215389549732208, -0.09885990619659424, -0.2377728968858719, -0.16241411864757538, 0.06152905523777008, 0.05220453068614006, -0.01511597540229559, 0.01769239455461502, -0.05573000758886337, -0.10734280198812485, -0.0847989022731781, 0.00046723702689632773, 0.030200952664017677, -0.05645965784788132, -0.06711659580469131, 0.04522988945245743, 0.04284689947962761, -0.12306949496269226, 0.03284240886569023, 0.06167784705758095, -0.016469528898596764, -0.0024301158264279366, 0.06098673492670059, 0.09022294729948044, 0.1855623722076416, -0.005363462958484888, 0.00008336466271430254, 0.05975780263543129, 0.2758288085460663, -0.1553190052509308, 0.1200755387544632, 0.13222041726112366, -0.06896428763866425, 0.07671049982309341, 0.18976257741451263, 0.03277582302689552, -0.09641318023204803, 0.02280794270336628, 0.025671953335404396, -0.020486317574977875, -0.2634943723678589, -0.055133309215307236, -0.015147299505770206, -0.0872906967997551, 0.0739329606294632, 0.08931269496679306, 0.07867603749036789, 0.037940382957458496, -0.06808945536613464, -0.09906767308712006, 0.03016929142177105, 0.10526300966739655, -0.030869536101818085, 0.0061525809578597546, 0.08358584344387054, -0.0394597090780735, 0.011515969410538673, 0.09563713520765305, -0.012425171211361885, 0.15721388161182404, 0.056327346712350845, 0.10829852521419525, 0.08148375153541565, 0.09563787281513214, -0.003909681923687458, 0.03556659072637558, 0.016407016664743423, 0.025135071948170662, 0.016537167131900787, -0.08634054660797119, 0.02168934978544712, 0.11328601092100143, 0.03605898097157478, 0.032747525721788406, 0.020769743248820305, -0.04227204993367195, 0.046472806483507156, 0.1904388964176178, 0.01574656553566456, -0.21045322716236115, -0.08163321018218994, 0.06131252646446228, -0.0797181949019432, -0.14668919146060944, -0.011666201055049896, 0.03460833802819252, -0.16206876933574677, 0.020196013152599335, -0.03935380280017853, 0.10276901721954346, -0.09335257858037949, -0.041819557547569275, 0.10713896155357361, 0.057994525879621506, -0.019242338836193085, 0.048500217497348785, -0.17638587951660156, 0.12012625485658646, 0.030222132802009583, 0.0765254944562912, -0.08575019985437393, 0.1028914600610733, 0.00008579371205996722, -0.00756382429972291, 0.1660338193178177, 0.004995278548449278, -0.05058227479457855, -0.08208775520324707, -0.10200401395559311, -0.01031569018959999, 0.09008632600307465, -0.13855160772800446, 0.07414360344409943, -0.026600223034620285, -0.03145773708820343, -0.004074910655617714, -0.09121152013540268, -0.13191068172454834, -0.16607145965099335, 0.055129095911979675, -0.09648937731981277, 0.02554628625512123, -0.08837458491325378, -0.056441351771354675, 0.008477063849568367, 0.17824631929397583, -0.23039597272872925, -0.1053895503282547, -0.15075267851352692, -0.1112571731209755, 0.16023197770118713, -0.04188292473554611, 0.08615583926439285, 0.0012770009925588965, 0.16243714094161987, 0.013505544513463974, -0.014202363789081573, 0.09846492856740952, -0.09434974938631058, -0.19023121893405914, -0.05550258606672287, 0.16379012167453766, 0.1432182490825653, 0.03071731887757778, -0.013496004045009613, 0.0307376179844141, -0.05934348702430725, -0.12109185755252838, 0.02490292862057686, 0.16549956798553467, 0.07581592351198196, -0.019114432856440544, -0.022096702829003334, -0.11402136087417603, -0.05272067338228226, -0.0418752059340477, -0.009879505261778831, 0.19278131425380707, -0.07334691286087036, 0.15842516720294952, 0.11453307420015335, -0.05686240643262863, -0.2083483636379242, 0.036364179104566574, 0.04623817279934883, 0.020556772127747536, 0.041772302240133286, -0.18514220416545868, 0.08956161141395569, -0.010792738758027554, -0.0777331218123436, 0.1655220240354538, -0.16835978627204895, -0.1391320824623108, 0.10392885655164719, 0.03230835869908333, -0.219212144613266, -0.13583151996135712, -0.10011351853609085, -0.0220856461673975, -0.1366918534040451, 0.05043289065361023, 0.0038158383686095476, 0.006202526390552521, 0.02082069404423237, 0.0094809764996171, 0.0269132349640131, -0.05105128884315491, 0.20862632989883423, -0.028510915115475655, 0.005549795459955931, -0.04871257394552231, -0.08579101413488388, 0.029847655445337296, -0.05017323046922684, 0.10737714916467667, -0.0049947588704526424, 0.029168613255023956, -0.16571524739265442, -0.039589956402778625, -0.05140593275427818, 0.027232803404331207, -0.09209448099136353, -0.08746706694364548, -0.041040968149900436, 0.09178666770458221, 0.09658870846033096, -0.027361160144209862, 0.00045020479592494667, -0.08927948027849197, 0.06264150887727737, 0.20813333988189697, 0.19560374319553375, 0.06387996673583984, -0.05983371287584305, 0.01795409806072712, -0.034251920878887177, 0.04724997654557228, -0.223591148853302, 0.04190773516893387, 0.05601511523127556, 0.01956031285226345, 0.07329723238945007, -0.011246003210544586, -0.15119317173957825, -0.0727582573890686, 0.0852903351187706, -0.05616509169340134, -0.16914114356040955, -0.028858166188001633, 0.02165849879384041, -0.20600873231887817, -0.03540804609656334, 0.02838703989982605, -0.01915847510099411, -0.03853688761591911, 0.020697718486189842, 0.0823936015367508, -0.023388970643281937, 0.10135126858949661, 0.08541107177734375, 0.09329536557197571, -0.10278649628162384, 0.0669015422463417, 0.07240721583366394, -0.03671887889504433, 0.02878006361424923, 0.11342639476060867, -0.04931100085377693, -0.036981772631406784, 0.08106772601604462, 0.10270316898822784, 0.021037597209215164, -0.05635508522391319, 0.009420645423233509, -0.049376364797353745, 0.058464664965867996, 0.10032541304826736, 0.027604063972830772, 0.0036632090341299772, 0.05964985862374306, 0.031105278059840202, -0.0881335437297821, 0.11222723126411438, 0.06078741326928139, 0.016999710351228714, -0.05536302179098129, -0.039029255509376526, -0.011816776357591152, -0.018256165087223053, -0.020726386457681656, -0.004097717814147472, -0.08436880260705948, -0.007573968265205622, -0.10702309012413025, 0.022529296576976776, -0.08071092516183853, 0.007537417113780975, 0.02828921750187874, -0.049284279346466064, 0.00033262601937167346, 0.002159812254831195, -0.07109886407852173, -0.05224272981286049, -0.01369281206279993, 0.08059820532798767, -0.13290037214756012, 0.04153348505496979, 0.07174966484308243, -0.1041741594672203, 0.07168220728635788, -0.007336880546063185, 0.009418381378054619, 0.0019502349896356463, -0.15107163786888123, 0.05669509246945381, -0.02520727552473545, -0.011557059362530708, 0.017567388713359833, -0.19832535088062286, -0.007824988104403019, -0.03900846093893051, -0.06478390842676163, 0.008458448573946953, -0.010714758187532425, -0.1209070086479187, 0.09941264986991882, 0.0016967884730547667, -0.060168445110321045, -0.025457488372921944, 0.03961818665266037, 0.1051812618970871, -0.021033890545368195, 0.13395929336547852, -0.022546125575900078, 0.07223593443632126, -0.17463983595371246, -0.006616184953600168, -0.013467080891132355, 0.04694100096821785, -0.029865462332963943, -0.029412945732474327, 0.058880507946014404, -0.02295377478003502, 0.173000305891037, -0.012983821332454681, 0.06646580249071121, 0.0533200278878212, 0.014708022587001324, 0.022111468017101288, 0.08146478980779648, 0.06383035331964493, -0.0073390258476138115, -0.0006018347921781242, 0.0356985442340374, -0.0037003627512604, -0.04633275046944618, -0.16286350786685944, 0.05826741084456444, 0.15707318484783173, 0.05463059991598129, 0.0268808975815773, 0.02067190408706665, -0.1162528544664383, -0.08538912981748581, 0.11824384331703186, -0.02366606518626213, -0.03377113863825798, -0.06890562921762466, 0.17843325436115265, 0.14185191690921783, -0.20020009577274323, 0.06809397041797638, -0.05254495516419411, -0.04699171707034111, -0.1380850076675415, -0.17562271654605865, -0.05755474418401718, -0.05008191242814064, -0.02682449482381344, -0.060899972915649414, 0.04967869445681572, 0.046304624527692795, 0.001968495547771454, -0.01798810251057148, 0.1018579825758934, 0.0172536913305521, -0.025425124913454056, 0.04544275254011154, 0.065976582467556, 0.03887335583567619, -0.09320461750030518, 0.008296609856188297, -0.00037831871304661036, 0.01730966940522194, 0.0677262619137764, 0.018432751297950745, -0.06181145831942558, 0.0236885454505682, -0.01950075849890709, -0.12290865182876587, 0.03987628594040871, -0.01355690136551857, -0.03947757929563522, 0.15086233615875244, 0.036326438188552856, 0.007413273211568594, -0.02049829810857773, 0.22611406445503235, -0.08354852348566055, -0.07188935577869415, -0.14664708077907562, 0.06304332613945007, -0.06742242723703384, 0.03209489956498146, 0.03237859159708023, -0.12204160541296005, 0.01070651039481163, 0.1633777916431427, 0.1291772723197937, -0.011536736972630024, 0.01035639550536871, 0.04495443031191826, 0.004723056219518185, -0.03933286294341087, 0.02222437597811222, 0.04765155538916588, 0.17141854763031006, -0.07124623656272888, 0.06253436952829361, -0.01087181456387043, -0.08464547246694565, -0.016942167654633522, 0.0997907742857933, -0.007830440998077393, 0.000256451457971707, -0.06457310914993286, 0.14534370601177216, -0.07701791077852249, -0.21098949015140533, 0.06295657902956009, -0.06316857784986496, -0.14034275710582733, -0.04186270013451576, 0.03536170721054077, -0.018833987414836884, 0.004339688457548618, 0.07316447794437408, -0.04546234756708145, 0.19132816791534424, 0.037816885858774185, -0.05347967520356178, -0.08586016297340393, 0.05529412999749184, -0.15502649545669556, 0.2805453836917877, 0.020817916840314865, 0.04546976834535599, 0.10791340470314026, -0.018484557047486305, -0.14902734756469727, 0.010769309476017952, 0.10542773455381393, -0.06915343552827835, 0.06083641201257706, 0.17130382359027863, 0.0024306944105774164, 0.1273282915353775, 0.05708400532603264, -0.051982104778289795, 0.03411924093961716, -0.09642670303583145, -0.04751727730035782, -0.11139443516731262, 0.08134733140468597, -0.08297251909971237, 0.16113770008087158, 0.11421211808919907, -0.07081950455904007, -0.0008512743515893817, -0.01784730888903141, 0.0876304879784584, 0.010066651739180088, 0.11150365322828293, 0.01405317336320877, -0.18902921676635742, 0.033097609877586365, 0.009436176158487797, 0.1056588813662529, -0.18665140867233276, -0.054596561938524246, 0.04207892343401909, -0.020130937919020653, -0.08136475831270218, 0.11764061450958252, 0.03680787980556488, 0.03269727900624275, -0.03870633989572525, -0.037169020622968674, 0.009012813679873943, 0.1432064026594162, -0.1120939776301384, -0.01275254599750042 ]
null
null
null
# **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="arnabmukherjee/q-taxi_v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
{"tags": ["Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "q-taxi_v3", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "Taxi-v3", "type": "Taxi-v3"}, "metrics": [{"type": "mean_reward", "value": "7.50 +/- 2.77", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
arnabmukherjee/q-taxi_v3
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
2024-02-15T02:55:24+00:00
[]
[]
TAGS #Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
# Q-Learning Agent playing1 Taxi-v3 This is a trained model of a Q-Learning agent playing Taxi-v3 . ## Usage
[ "# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ "TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n", "# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ 32, 33 ]
[ "passage: TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ 0.048862796276807785, -0.16549694538116455, -0.005485367961227894, 0.02960980497300625, 0.1345081776380539, -0.01784728653728962, 0.11895976960659027, 0.07759871333837509, -0.07461097836494446, -0.055395450443029404, 0.1418241262435913, 0.09088201075792313, 0.055222880095243454, 0.05699880048632622, 0.09511256217956543, -0.27440664172172546, 0.048217080533504486, -0.02918700873851776, 0.05621987581253052, 0.11878681182861328, 0.0670095682144165, -0.040441032499074936, 0.061956584453582764, 0.11818158626556396, -0.1018151044845581, -0.007344264071434736, 0.035402704030275345, -0.09440053254365921, 0.17413531243801117, 0.07204403728246689, 0.12337774783372879, 0.05132639780640602, 0.179361954331398, -0.12762396037578583, 0.024310702458024025, -0.0010275895474478602, -0.10138072073459625, -0.03909514099359512, -0.012415820732712746, -0.08349097520112991, 0.03230205550789833, 0.23522862792015076, 0.07199250161647797, 0.06632792949676514, -0.17707863450050354, -0.06584878265857697, -0.04375573247671127, 0.069611094892025, 0.14951466023921967, 0.03758616745471954, -0.033800311386585236, 0.1684885323047638, -0.2564343810081482, 0.05066783353686333, 0.037275806069374084, -0.42313119769096375, 0.017119819298386574, 0.1507398933172226, 0.15090937912464142, 0.06909667700529099, -0.10573802888393402, 0.013512322679162025, 0.051325585693120956, -0.0005318621988408267, 0.024325110018253326, 0.006554204970598221, 0.15601307153701782, 0.08537693321704865, -0.1487821787595749, -0.058576688170433044, 0.17441977560520172, -0.03788546845316887, -0.02613203600049019, -0.039745692163705826, 0.0067160045728087425, -0.06427708268165588, -0.004067842848598957, -0.1777995079755783, 0.00734262028709054, 0.06666424125432968, -0.014348524622619152, 0.014901017770171165, -0.035522811114788055, -0.0966939702630043, -0.023098144680261612, -0.08592145889997482, 0.01677769608795643, -0.006319406442344189, -0.10187895596027374, 0.05002119392156601, -0.061138734221458435, 0.0014382408699020743, -0.05123179033398628, -0.15047866106033325, -0.049055423587560654, -0.03481535613536835, 0.1474713832139969, -0.0044205985032022, -0.01873963139951229, -0.03164304047822952, 0.15474793314933777, 0.049551334232091904, -0.05370146036148071, 0.05625450983643532, 0.07605006545782089, 0.23867930471897125, 0.10401605814695358, 0.10196955502033234, -0.06798075139522552, 0.10180158913135529, -0.12330973148345947, -0.08915644884109497, -0.17508824169635773, 0.11820860952138901, 0.00015364694991149008, 0.1317785084247589, -0.12023144960403442, 0.07898581773042679, -0.067511186003685, 0.013453764840960503, 0.01636839471757412, 0.0820009782910347, -0.012399360537528992, 0.10676060616970062, -0.005061192903667688, -0.06941985338926315, 0.014177112840116024, 0.05935845896601677, 0.03754841163754463, -0.038601722568273544, -0.03192409873008728, -0.05762290954589844, -0.05065649375319481, -0.10128600150346756, -0.06447898596525192, 0.018573462963104248, -0.007677143905311823, -0.1833900660276413, -0.06407523155212402, 0.00897200871258974, 0.015712225809693336, -0.03988850116729736, -0.05148044601082802, -0.15265507996082306, -0.042461175471544266, -0.015450406819581985, -0.03500641882419586, -0.06214277446269989, -0.0383245050907135, 0.046435944736003876, -0.07560601085424423, 0.013364278711378574, 0.023342855274677277, 0.05405820533633232, -0.025881100445985794, 0.06068144738674164, -0.08357544988393784, 0.09493788331747055, -0.1540430635213852, -0.03271956741809845, -0.025445878505706787, -0.041183918714523315, 0.1752462536096573, 0.06099751964211464, -0.015994304791092873, 0.15260063111782074, -0.17141541838645935, -0.058121129870414734, 0.15596486628055573, 0.008629098534584045, -0.09967197477817535, -0.003560945624485612, -0.09397093951702118, 0.1428760588169098, 0.08571921288967133, 0.2478504776954651, 0.12005335837602615, -0.22748184204101562, 0.055358242243528366, 0.12515293061733246, -0.14365963637828827, 0.10365243256092072, 0.07344598323106766, 0.005470725707709789, -0.18886831402778625, -0.06843198090791702, -0.06121627986431122, 0.1053021252155304, -0.08522345870733261, -0.0776243582367897, 0.09323626756668091, -0.05086790770292282, 0.24641476571559906, -0.028281206265091896, 0.06174173951148987, -0.026681531220674515, -0.1389324963092804, -0.01723906397819519, 0.060955192893743515, 0.05258452147245407, -0.024835573509335518, -0.25895482301712036, 0.13646544516086578, 0.048650871962308884, 0.025074828416109085, 0.004106190986931324, -0.05691491439938545, 0.016934165731072426, 0.1511998474597931, 0.020012924447655678, 0.13717477023601532, 0.027723990380764008, 0.0706823319196701, -0.006239562761038542, -0.10560829937458038, -0.04169593006372452, 0.061916545033454895, -0.08518962562084198, -0.06641357392072678, 0.011197872459888458, -0.06935211271047592, -0.11783787608146667, -0.12166737765073776, -0.026334572583436966, -0.02980303019285202, -0.07444227486848831, 0.02368103712797165, 0.06536602973937988, -0.06702698022127151, -0.0023908785078674555, 0.007125476840883493, -0.011537045240402222, 0.16434046626091003, 0.011393417604267597, -0.007796820718795061, 0.1328643560409546, -0.11533161997795105, 0.12461213022470474, 0.049438029527664185, -0.024806302040815353, -0.04662557691335678, 0.0014137453399598598, -0.057529181241989136, 0.029044216498732567, -0.04390640929341316, 0.02774495631456375, 0.20111067593097687, 0.02772962674498558, 0.11389166116714478, -0.0656520202755928, 0.04385066404938698, -0.007961965166032314, -0.009693224914371967, 0.018563594669103622, 0.07608018070459366, 0.07813210040330887, -0.1324140727519989, 0.02262016013264656, 0.22455167770385742, 0.1385764330625534, 0.18313980102539062, -0.010877152904868126, 0.06325667351484299, -0.04875868931412697, 0.027505528181791306, 0.024100203067064285, 0.10314226150512695, -0.10732068121433258, -0.0322517491877079, -0.025407759472727776, 0.023599207401275635, -0.08197105675935745, -0.1055799350142479, -0.090115025639534, 0.01222382951527834, -0.03125503659248352, -0.15570329129695892, 0.13300658762454987, -0.10451057553291321, 0.01802753657102585, 0.04692702740430832, -0.22163605690002441, 0.11530312895774841, 0.014291439205408096, -0.10303618758916855, 0.11281087249517441, -0.12051989883184433, -0.08699832111597061, -0.05777236074209213, -0.18658851087093353, 0.05280197039246559, 0.04673841595649719, 0.05166793242096901, -0.18521739542484283, 0.024835903197526932, 0.05545609071850777, 0.13426995277404785, -0.09743253141641617, -0.07142634689807892, -0.15038461983203888, 0.016068490222096443, -0.033661190420389175, -0.16029728949069977, -0.005609163548797369, -0.032781440764665604, -0.18849676847457886, -0.04539939761161804, -0.15086813271045685, -0.034627582877874374, 0.20464378595352173, 0.026907702907919884, 0.09480511397123337, -0.07926445454359055, 0.3802889585494995, -0.042039383202791214, -0.06146497279405594, -0.01321389526128769, -0.07072482258081436, 0.02512686513364315, 0.13271741569042206, 0.0036099457647651434, -0.017886579036712646, -0.0037857077550143003, 0.0024592927657067776, -0.06234965845942497, -0.13400450348854065, 0.0028710351325571537, 0.03905198723077774, 0.1874423623085022, 0.004639793653041124, 0.06659388542175293, 0.03133883699774742, 0.057546284049749374, 0.07748064398765564, 0.030926106497645378, 0.0011591583024710417, -0.01591806672513485, 0.06604493409395218, -0.11684755235910416, 0.042466625571250916, -0.030429253354668617, -0.10143838077783585, -0.013183288276195526, 0.07950251549482346, 0.12755028903484344, 0.17849206924438477, -0.04790908098220825, 0.17489230632781982, 0.13580141961574554, 0.16576050221920013, 0.049315933138132095, -0.020801831036806107, -0.08773037046194077, -0.06118565797805786, 0.004774159751832485, -0.031952597200870514, 0.04869702458381653, 0.3231290578842163, 0.037619613111019135, -0.09036035090684891, 0.11149907857179642, 0.009480619803071022, 0.05359881371259689, 0.022797370329499245, -0.11162138730287552, 0.11170321702957153, 0.07968773692846298, -0.06341761350631714, -0.07602835446596146, 0.16758501529693604, -0.1109386757016182, -0.26646625995635986, -0.11410990357398987, -0.012305386364459991, 0.07903840392827988, 0.005651174578815699, 0.05498376116156578, -0.11829282343387604, -0.16034497320652008, -0.034191906452178955, 0.1335442066192627, -0.3077351450920105, 0.2065143585205078, -0.0198091771453619, 0.06707923114299774, -0.039657969027757645, -0.07026876509189606, 0.09694647043943405, 0.13174086809158325, 0.29124146699905396, 0.01396956667304039, 0.04841272905468941, -0.15176129341125488, -0.0976925864815712, 0.0018439020495861769, 0.015482662245631218, -0.02563396655023098, 0.028520405292510986, -0.0540912002325058, 0.008404579944908619, -0.018086453899741173, 0.2102297693490982, -0.11316607892513275, 0.004344627261161804, -0.06968966871500015, -0.11707738786935806, 0.19409789144992828, -0.07178345322608948, -0.04543264955282211, -0.14959357678890228, -0.15512511134147644, -0.004174166824668646, -0.02413962036371231, -0.019664527848362923, -0.17603960633277893, -0.18804074823856354, -0.05204557999968529, -0.005645004566758871, -0.003464865731075406, 0.05867868289351463, -0.07517234236001968, -0.04805335775017738, 0.1009904220700264, -0.07743175327777863, -0.056063808500766754, -0.1103200614452362, 0.1391381323337555, 0.06248528137803078, 0.16743235290050507, 0.05907081440091133, 0.0006117874872870743, 0.11471151560544968, -0.02913086675107479, 0.11103474348783493, -0.11291708797216415, -0.17145049571990967, -0.08334989100694656, -0.018775060772895813, 0.09519003331661224, -0.04789286106824875, 0.0028788831550627947, 0.2550160884857178, 0.14880181849002838, -0.0897710770368576, 0.27680760622024536, 0.04414956644177437, -0.09375058114528656, -0.18432219326496124, -0.15961645543575287, 0.03759992495179176, 0.060025621205568314, 0.13095876574516296, -0.057205069810152054, -0.08483537286520004, -0.08492398262023926, -0.07478608191013336, -0.13140805065631866, -0.24232175946235657, -0.030598774552345276, 0.22874866425991058, 0.08656918257474899, 0.08219650387763977, -0.012482990510761738, -0.01186054851859808, 0.00526038184762001, 0.02680150233209133, 0.12018456310033798, -0.13341329991817474, 0.11107480525970459, 0.022198403254151344, 0.044267985969781876, 0.009712530300021172, 0.07929777354001999, 0.03375575691461563, -0.003218587953597307, -0.0006439819699153304, -0.0988350659608841, -0.2596651017665863, 0.0816885456442833, -0.01623627357184887, -0.09960969537496567, 0.014988959766924381, 0.02061903104186058, -0.2089255303144455, 0.011128270998597145, -0.019883770495653152, -0.03150356933474541, -0.06483490765094757, -0.10664787143468857, -0.056551624089479446, 0.04928823933005333, 0.10853826254606247, 0.011660109274089336, 0.05354316532611847, -0.0404130220413208, 0.07917837053537369, 0.0826287642121315, 0.15132710337638855, 0.06795957684516907, -0.190711110830307, -0.10953907668590546, -0.0414445661008358, 0.12121522426605225, -0.12505418062210083, 0.036917757242918015, 0.053161121904850006, -0.016534561291337013, 0.14621229469776154, 0.1070784479379654, -0.07452095299959183, 0.11915595084428787, 0.08904775977134705, -0.04094788804650307, -0.23367151618003845, -0.07120766490697861, 0.11133213341236115, 0.07195597887039185, -0.03961895406246185, 0.018120890483260155, -0.04960581287741661, -0.013980977237224579, 0.048759616911411285, -0.0538676381111145, -0.07230538129806519, 0.004421027842909098, 0.1247575581073761, 0.1029362753033638, -0.04655474051833153, 0.01296416949480772, 0.037371400743722916, 0.003788623260334134, 0.04730486497282982, 0.0407949760556221, -0.08269952982664108, -0.04124005511403084, 0.02782733179628849, 0.37552911043167114, -0.010165480896830559, -0.020456433296203613, 0.018555615097284317, -0.19949445128440857, 0.09135842323303223, 0.13205479085445404, 0.04697350412607193, 0.004247748292982578, -0.08139242231845856, 0.026877427473664284, -0.010625290684401989, 0.09936143457889557, -0.07806670665740967, -0.05493134260177612, -0.21631066501140594, -0.025010565295815468, 0.017490221187472343, 0.24077683687210083, -0.08458559215068817, -0.12801732122898102, -0.20628872513771057, 0.13128381967544556, -0.11333390325307846, -0.03695881739258766, -0.024473199620842934, 0.03926658630371094, -0.01989821158349514, 0.06291737407445908, -0.0710630789399147, 0.006373001262545586, -0.11024709790945053, 0.055267609655857086, 0.04204455390572548, 0.1229788213968277, 0.014207782223820686, 0.02016810141503811, 0.05822525918483734, -0.01837925612926483, 0.07173580676317215, -0.06203491613268852, -0.04550490900874138, 0.14224006235599518, -0.020255116745829582, -0.04152837023139, -0.0483345128595829, -0.036874305456876755, 0.11981741338968277, -0.05059147998690605, -0.007141099311411381, -0.054929375648498535, -0.06906463205814362, 0.03462086617946625, -0.009175732731819153, -0.008798843249678612, 0.06801853328943253, 0.04024988040328026, -0.026994358748197556, 0.005263668950647116, 0.03447828069329262, -0.10330043733119965, -0.04955084249377251, 0.16955432295799255, -0.0749620869755745, 0.10274054110050201, -0.031069839373230934, 0.018015999346971512, 0.005847334861755371, -0.022399673238396645, -0.015360680408775806, -0.1457086056470871, -0.06137600541114807, -0.09489979594945908, 0.11565322428941727, 0.08146517723798752, 0.03358805552124977, 0.04274565726518631, 0.019532648846507072, -0.04414922371506691, -0.038583990186452866, 0.12961317598819733, 0.08133101463317871, 0.012996876612305641, 0.01137041300535202, 0.01941833831369877, -0.020302120596170425, 0.0028480992186814547, -0.01250747125595808, -0.07239153981208801, -0.05874783173203468, 0.09400010108947754, 0.1600283533334732, -0.06127211079001427, -0.13325586915016174, -0.020593497902154922, 0.04988488554954529, 0.0014717020094394684, -0.08777432143688202, 0.04833676666021347, 0.15805292129516602, -0.05623878911137581, 0.03216489031910896, -0.09984751045703888, -0.07263360917568207, -0.16060975193977356, -0.10029061883687973, -0.06092562898993492, -0.28350353240966797, 0.09752398729324341, 0.006392303854227066, -0.014731393195688725, 0.059529416263103485, 0.051305368542671204, -0.052508849650621414, 0.07068239152431488, -0.18146829307079315, -0.007054794579744339, 0.03497592359781265, -0.13212306797504425, 0.02475893869996071, -0.2378365397453308, 0.10198072344064713, -0.04623803123831749, -0.1519704908132553, -0.04004510119557381, 0.0641569048166275, -0.09540136158466339, -0.01822364516556263, -0.0475153923034668, -0.01922670193016529, 0.01624443754553795, -0.009348669089376926, -0.031147832050919533, 0.13716529309749603, 0.02827494591474533, -0.03268734738230705, 0.005254602525383234, 0.0223685409873724, 0.03955082967877388, -0.0969657450914383, -0.05986930429935455, 0.08311155438423157, -0.031056145206093788, 0.14728976786136627, 0.000341245875461027, 0.04181376099586487, -0.06758682429790497, 0.2593761384487152, 0.2023983597755432, -0.12479214370250702, 0.008118697442114353, -0.021801479160785675, 0.012670028023421764, -0.041751839220523834, 0.13110700249671936, 0.013386172242462635, 0.12186761200428009, -0.17513342201709747, -0.01036517322063446, -0.0818324014544487, -0.04501292482018471, 0.06702108681201935, 0.14714950323104858, 0.15742522478103638, 0.03436789661645889, -0.07328428328037262, 0.06722653657197952, -0.30119743943214417, 0.20540550351142883, -0.1346001923084259, -0.01498429011553526, -0.040251150727272034, -0.058389630168676376, 0.061147745698690414, 0.11309876292943954, 0.10832664370536804, -0.021150551736354828, -0.0905047357082367, -0.04486766457557678, -0.039378076791763306, -0.13019338250160217, -0.02718670479953289, 0.1654091775417328, 0.06799814850091934, 0.31520840525627136, -0.017577875405550003, 0.07702425122261047, 0.034410297870635986, 0.06451138854026794, 0.004519328009337187, 0.09537279605865479, 0.07960964739322662, -0.06345855444669724, -0.07373003661632538, -0.001637450186535716, 0.05033271387219429, 0.14567798376083374, -0.03826142102479935, -0.18691548705101013, 0.15858715772628784, 0.07192251086235046, -0.13762691617012024, -0.05777517706155777, 0.08409425616264343, -0.0739973932504654, 0.0550808347761631, 0.08115427941083908, 0.015876613557338715, -0.017793258652091026, -0.004664506763219833, 0.06074233725667, 0.024694660678505898, -0.02343848906457424, 0.003570882137864828, -0.08337053656578064, -0.04151543974876404, 0.07267895340919495, -0.0844460055232048, -0.20546193420886993, -0.0957019031047821, -0.07551700621843338, 0.030557552352547646, -0.0649830624461174, 0.12575586140155792, 0.1717868149280548, 0.0593598335981369, -0.03307248651981354, -0.10721943527460098, -0.035562749952077866, 0.07602505385875702, -0.044773899018764496, -0.09409699589014053 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
kyone/another_model
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T02:57:55+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 57, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.05622259899973869, 0.16002345085144043, -0.004987028427422047, 0.023115945979952812, 0.0962471067905426, 0.011845538392663002, 0.06785304099321365, 0.11496778577566147, -0.020396295934915543, 0.11142492294311523, 0.03292480856180191, 0.0972127765417099, 0.11474913358688354, 0.16215258836746216, 0.004439093638211489, -0.23455148935317993, 0.04782992601394653, -0.12695099413394928, -0.033447545021772385, 0.11785799264907837, 0.14491069316864014, -0.10402194410562515, 0.07766910642385483, -0.030544815585017204, -0.009361269883811474, -0.03290390968322754, -0.06365230679512024, -0.05152205005288124, 0.05037128925323486, 0.06932847946882248, 0.06591591984033585, 0.007509593386203051, 0.09122733771800995, -0.2655104100704193, 0.02280162274837494, 0.07630051672458649, -0.0015554219717159867, 0.07497020810842514, 0.048351652920246124, -0.08209776133298874, 0.0788840726017952, -0.05696587264537811, 0.14718368649482727, 0.08216129243373871, -0.08924587815999985, -0.1965435892343521, -0.08464295417070389, 0.10284840315580368, 0.18357418477535248, 0.05158785358071327, -0.024141347035765648, 0.10476154088973999, -0.08419200032949448, 0.008797040209174156, 0.06024181470274925, -0.06443428993225098, -0.05412506312131882, 0.06934051215648651, 0.07975570857524872, 0.07967228442430496, -0.13025140762329102, -0.014651902951300144, 0.011243549175560474, 0.007594773545861244, 0.08504551649093628, 0.022028017789125443, 0.14595499634742737, 0.04393624886870384, -0.13030564785003662, -0.044304780662059784, 0.09771761298179626, 0.04345165938138962, -0.053857799619436264, -0.2537047266960144, -0.024983759969472885, -0.03927002474665642, -0.03094942681491375, -0.038562554866075516, 0.04431856796145439, -0.011080716736614704, 0.08032315224409103, -0.01118796318769455, -0.08149448037147522, -0.041395120322704315, 0.06544242054224014, 0.062143467366695404, 0.026896316558122635, -0.01158317644149065, 0.00973866879940033, 0.1224486380815506, 0.10907839238643646, -0.12763150036334991, -0.05768941715359688, -0.06755511462688446, -0.08307720720767975, -0.04300352931022644, 0.03337155282497406, 0.044020529836416245, 0.04436098039150238, 0.2466370165348053, 0.01108562108129263, 0.05453123152256012, 0.045806169509887695, 0.010608446784317493, 0.06787561625242233, 0.11606968939304352, -0.062306761741638184, -0.09178462624549866, -0.029058339074254036, 0.09215214103460312, 0.006741520017385483, -0.042814407497644424, -0.060904473066329956, 0.06479041278362274, 0.012608112767338753, 0.12110785394906998, 0.08444269746541977, 0.0026690615341067314, -0.07305197417736053, -0.06963318586349487, 0.18848419189453125, -0.1598394364118576, 0.047875016927719116, 0.031182926148176193, -0.038971830159425735, -0.0014042917173355818, 0.008752269670367241, 0.02394084818661213, -0.020246321335434914, 0.08923295140266418, -0.05574449151754379, -0.03784004598855972, -0.11079790443181992, -0.03252100944519043, 0.030985163524746895, 0.0051483530551195145, -0.027043871581554413, -0.033837489783763885, -0.09040277451276779, -0.059588029980659485, 0.0922931432723999, -0.07471107691526413, -0.04984431713819504, -0.013726521283388138, -0.07691634446382523, 0.023329194635152817, 0.016799474135041237, 0.08357251435518265, -0.02157396264374256, 0.0384126678109169, -0.0560205839574337, 0.0631464347243309, 0.11269522458314896, 0.029363946989178658, -0.053069718182086945, 0.05750001594424248, -0.24315528571605682, 0.10326608270406723, -0.07320205867290497, 0.050549428910017014, -0.15059062838554382, -0.026000602170825005, 0.044471126049757004, 0.00805877335369587, -0.013138634152710438, 0.14088952541351318, -0.21621745824813843, -0.0323486253619194, 0.16741067171096802, -0.0939871072769165, -0.07602590322494507, 0.059108685702085495, -0.05233629792928696, 0.10869261622428894, 0.04351044446229935, -0.02232111617922783, 0.060673557221889496, -0.14475463330745697, -0.01067100279033184, -0.04139741137623787, -0.02402937039732933, 0.16397778689861298, 0.07567544281482697, -0.06286642700433731, 0.08052356541156769, 0.024165838956832886, -0.017831770703196526, -0.04484899342060089, -0.023361295461654663, -0.10819391161203384, 0.009856974706053734, -0.06032416597008705, 0.02424289658665657, -0.025761527940630913, -0.09367526322603226, -0.02868773601949215, -0.1802000105381012, -0.009223134256899357, 0.0881323292851448, -0.011722641065716743, -0.021903391927480698, -0.12039245665073395, 0.011948852799832821, 0.031212422996759415, 0.002984174294397235, -0.13029038906097412, -0.05838731303811073, 0.027675874531269073, -0.16422230005264282, 0.03272955119609833, -0.05597274377942085, 0.05056252330541611, 0.03445037454366684, -0.03187771514058113, -0.033117350190877914, 0.009550533257424831, 0.006354342680424452, -0.010578392073512077, -0.2502359449863434, -0.02440580166876316, -0.0219739843159914, 0.17386503517627716, -0.21793730556964874, 0.04213962331414223, 0.07686693966388702, 0.14929872751235962, 0.006240781396627426, -0.038500864058732986, 0.010139784775674343, -0.08222103863954544, -0.030560437589883804, -0.0643099993467331, -0.012082485482096672, -0.03717579320073128, -0.05608142167329788, 0.05165567249059677, -0.16133594512939453, -0.028727244585752487, 0.1057019829750061, 0.06860516220331192, -0.14001330733299255, -0.019125886261463165, -0.04171464592218399, -0.043496038764715195, -0.05877087265253067, -0.0552728995680809, 0.1185101792216301, 0.05596614256501198, 0.04696191847324371, -0.06956122815608978, -0.07775315642356873, 0.007865429855883121, -0.017090093344449997, -0.017978519201278687, 0.08920905739068985, 0.07311701774597168, -0.12023317068815231, 0.09247473627328873, 0.10194233059883118, 0.09365488588809967, 0.108615942299366, -0.017981963232159615, -0.08929306268692017, -0.04584396257996559, 0.02045595459640026, 0.013332244008779526, 0.14797501266002655, -0.01403066236525774, 0.056954506784677505, 0.03922648727893829, -0.01123172789812088, 0.012020308524370193, -0.09384570270776749, 0.027314940467476845, 0.034342724829912186, -0.020308034494519234, 0.03796098753809929, -0.04001156985759735, 0.019826533272862434, 0.08712323755025864, 0.04676510766148567, 0.04415108636021614, 0.011758276261389256, -0.04233846068382263, -0.10904491692781448, 0.173858180642128, -0.12615609169006348, -0.24583272635936737, -0.14115718007087708, 0.0015609683468937874, 0.04152948409318924, -0.009671499952673912, 0.003867273684591055, -0.07054664939641953, -0.11710625886917114, -0.0934595838189125, 0.018713686615228653, 0.04491026699542999, -0.07426843047142029, -0.0596279613673687, 0.059872306883335114, 0.03894329443573952, -0.14430272579193115, 0.022237464785575867, 0.047419775277376175, -0.09032250195741653, -0.006925572175532579, 0.08398029953241348, 0.06729988008737564, 0.17764869332313538, 0.009659109637141228, -0.021044570952653885, 0.03080335259437561, 0.21258224546909332, -0.14283664524555206, 0.11252175271511078, 0.14021345973014832, -0.09024007618427277, 0.08099348843097687, 0.1948828399181366, 0.039186809211969376, -0.10478170961141586, 0.03259138762950897, 0.02489176020026207, -0.028939135372638702, -0.25018003582954407, -0.0680207833647728, 0.002590036718174815, -0.04892077296972275, 0.07092583924531937, 0.0918794497847557, 0.09946957975625992, 0.015428726561367512, -0.09732488542795181, -0.08017807453870773, 0.0468163788318634, 0.10640767961740494, 0.0070237633772194386, -0.01532268337905407, 0.08905128389596939, -0.03260866180062294, 0.018378758803009987, 0.0954233929514885, 0.00412675691768527, 0.17459604144096375, 0.05586163327097893, 0.17767499387264252, 0.07751350849866867, 0.06634163856506348, 0.019167855381965637, 0.0069374511949718, 0.02067388966679573, 0.017508454620838165, -0.004214957356452942, -0.08522020280361176, -0.00457410141825676, 0.12029227614402771, 0.06321834027767181, 0.024303704500198364, 0.0137604009360075, -0.03941800817847252, 0.08438141644001007, 0.17332784831523895, 0.0020201504230499268, -0.18486954271793365, -0.07240456342697144, 0.07921045273542404, -0.0910051167011261, -0.10552998632192612, -0.03353073075413704, 0.03346012532711029, -0.1747758537530899, 0.02097497321665287, -0.017018353566527367, 0.10809773951768875, -0.13855572044849396, -0.018670624122023582, 0.06328251957893372, 0.07232730835676193, -0.0028869258239865303, 0.06308864802122116, -0.153975248336792, 0.1050168052315712, 0.016289174556732178, 0.06754438579082489, -0.09747608006000519, 0.10138221830129623, -0.006303760688751936, -0.007241528946906328, 0.13875643908977509, 0.010596190579235554, -0.05694379657506943, -0.08987913280725479, -0.10555228590965271, -0.008462639525532722, 0.12933635711669922, -0.15157614648342133, 0.0847775787115097, -0.028662750497460365, -0.043171048164367676, 0.0024383023846894503, -0.1199452206492424, -0.1302652359008789, -0.1875755488872528, 0.058235347270965576, -0.1366453617811203, 0.039557021111249924, -0.10582595318555832, -0.04340389743447304, -0.028466427698731422, 0.2041483372449875, -0.2317875325679779, -0.0682469978928566, -0.1541893482208252, -0.08429346233606339, 0.14446710050106049, -0.04730919376015663, 0.08914490789175034, -0.0013825427740812302, 0.19013537466526031, 0.024473950266838074, -0.02387205697596073, 0.10308998823165894, -0.09543927758932114, -0.19450686872005463, -0.08603953570127487, 0.15582145750522614, 0.13931062817573547, 0.03702725097537041, -0.004593946039676666, 0.029260434210300446, -0.020000332966446877, -0.12535293400287628, 0.025526588782668114, 0.1793687790632248, 0.07859015464782715, 0.023437971249222755, -0.025896867737174034, -0.10993997752666473, -0.06524094194173813, -0.0335373692214489, 0.02718053013086319, 0.18264614045619965, -0.07421271502971649, 0.1900695115327835, 0.13626199960708618, -0.05445687845349312, -0.1955246478319168, 0.018216576427221298, 0.040417760610580444, 0.010847307741641998, 0.03138056397438049, -0.2078717201948166, 0.09027513861656189, 0.0014845491386950016, -0.05172133818268776, 0.141556978225708, -0.174949511885643, -0.1512570083141327, 0.06491631269454956, 0.0364508256316185, -0.19348180294036865, -0.117862768471241, -0.08817066252231598, -0.046907443553209305, -0.17498233914375305, 0.10519181191921234, 0.016932250931859016, 0.009516867808997631, 0.03492651879787445, 0.02640140987932682, 0.011080757714807987, -0.03873949125409126, 0.19461296498775482, -0.02505207620561123, 0.029532426968216896, -0.08079101145267487, -0.06136554479598999, 0.0607450045645237, -0.05577658861875534, 0.07896649837493896, -0.020188091322779655, 0.012835816480219364, -0.1100873053073883, -0.0468425452709198, -0.027396185323596, 0.017321845516562462, -0.09195652604103088, -0.09473495930433273, -0.05146971344947815, 0.09373841434717178, 0.08845265954732895, -0.036603908985853195, -0.04043547809123993, -0.07348548620939255, 0.0325477197766304, 0.17183002829551697, 0.17659065127372742, 0.038550034165382385, -0.08084331452846527, -0.005880105309188366, -0.01188716571778059, 0.04436201974749565, -0.22519725561141968, 0.06208868324756622, 0.04557957127690315, 0.015879612416028976, 0.11362850666046143, -0.018783990293741226, -0.16298477351665497, -0.06594224274158478, 0.06143777072429657, -0.06664001196622849, -0.18599680066108704, 0.0032026967965066433, 0.058006007224321365, -0.1646854728460312, -0.037671029567718506, 0.042260222136974335, -0.0045668939128518105, -0.04300284758210182, 0.01627597212791443, 0.08071378618478775, 0.005054219625890255, 0.07112491130828857, 0.05733523517847061, 0.0842885971069336, -0.10417009145021439, 0.07519911974668503, 0.08007751405239105, -0.08229218423366547, 0.031453702598810196, 0.08910130709409714, -0.061817802488803864, -0.03069761022925377, 0.032593827694654465, 0.07753410935401917, 0.019773589447140694, -0.041717879474163055, 0.008655321784317493, -0.09745000302791595, 0.06339588761329651, 0.09504765272140503, 0.03549657016992569, 0.014742289669811726, 0.034356739372015, 0.04988397657871246, -0.07460241764783859, 0.11766603589057922, 0.022336218506097794, 0.01780087500810623, -0.044981084764003754, -0.05459042266011238, 0.032110098749399185, -0.022974027320742607, -0.010163158178329468, -0.03885438293218613, -0.07015778869390488, -0.018130742013454437, -0.15929651260375977, -0.014899281784892082, -0.04085385054349899, 0.007158880587667227, 0.02551902085542679, -0.03834335505962372, 0.007963370531797409, 0.012195355258882046, -0.07085035741329193, -0.061454467475414276, -0.022903166711330414, 0.09224231541156769, -0.16436699032783508, 0.025155464187264442, 0.08285263180732727, -0.12099926173686981, 0.09775067120790482, 0.021939631551504135, 0.0031351554207503796, 0.028338242322206497, -0.1542527824640274, 0.04096807911992073, -0.024365095421671867, 0.01272035762667656, 0.04409142583608627, -0.22033950686454773, 0.001463581225834787, -0.03818526118993759, -0.05954346805810928, -0.010227864608168602, -0.033079732209444046, -0.11291328817605972, 0.09883669763803482, 0.008058897219598293, -0.08219768106937408, -0.030809206888079643, 0.03451729565858841, 0.08243680745363235, -0.02608415111899376, 0.15152283012866974, 0.0016822130419313908, 0.07172226905822754, -0.17519205808639526, -0.021702464669942856, -0.011611736379563808, 0.02207101881504059, -0.014536668546497822, -0.015496513806283474, 0.042471300810575485, -0.02421419881284237, 0.19108575582504272, -0.026401294395327568, 0.038726791739463806, 0.06405707448720932, 0.01593620702624321, -0.014801506884396076, 0.10957890748977661, 0.05975057929754257, 0.02399693801999092, 0.022115202620625496, 0.007329683285206556, -0.039842452853918076, -0.014149460941553116, -0.19538825750350952, 0.06474217027425766, 0.1377464383840561, 0.08781574666500092, -0.01322576031088829, 0.07683692127466202, -0.10024392604827881, -0.12397097796201706, 0.11215250939130783, -0.06283260136842728, -0.007701667957007885, -0.06531554460525513, 0.13346771895885468, 0.14944057166576385, -0.18992236256599426, 0.06835456937551498, -0.06228158622980118, -0.05332518368959427, -0.11744599789381027, -0.1957325041294098, -0.055616896599531174, -0.056456826627254486, -0.014700124971568584, -0.048795297741889954, 0.07307228446006775, 0.05693497136235237, 0.012962869368493557, 0.003600025549530983, 0.0766802653670311, -0.015357231721282005, 0.0008028073934838176, 0.03077360987663269, 0.06600049883127213, 0.013312965631484985, -0.02929985709488392, 0.020537450909614563, -0.007275243755429983, 0.04005419462919235, 0.06378308683633804, 0.038119763135910034, -0.02801438421010971, 0.01591232419013977, -0.03770609200000763, -0.10940317064523697, 0.0409080907702446, -0.028551526367664337, -0.08112191408872604, 0.13721226155757904, 0.02428387477993965, 0.005870606284588575, -0.02180131897330284, 0.24582624435424805, -0.07231455296278, -0.09001907706260681, -0.1473579704761505, 0.10211005061864853, -0.04095151647925377, 0.06560079753398895, 0.04110138490796089, -0.10732010751962662, 0.013498948886990547, 0.12688814103603363, 0.15896959602832794, -0.044884394854307175, 0.020156091079115868, 0.03252736106514931, 0.003683826420456171, -0.04006262496113777, 0.05253688618540764, 0.0694650411605835, 0.14883354306221008, -0.04907030612230301, 0.08928520232439041, 0.005485867150127888, -0.10256236046552658, -0.03822692111134529, 0.11808354407548904, -0.017866896465420723, 0.018703164532780647, -0.057248231023550034, 0.11889533698558807, -0.059861693531274796, -0.23005777597427368, 0.06317704170942307, -0.0720362737774849, -0.14286935329437256, -0.021647587418556213, 0.07456772774457932, -0.017636949196457863, 0.02658887766301632, 0.07326807081699371, -0.07681973278522491, 0.19899281859397888, 0.038975972682237625, -0.05729197710752487, -0.05658522993326187, 0.0789351835846901, -0.114089734852314, 0.2792985737323761, 0.01164181251078844, 0.04984506592154503, 0.10365619510412216, -0.016686614602804184, -0.13768579065799713, 0.015234606340527534, 0.09244892746210098, -0.09004336595535278, 0.03869183734059334, 0.2132277488708496, -0.002569539239630103, 0.1152428612112999, 0.07714667171239853, -0.07265080511569977, 0.04592108353972435, -0.1130065843462944, -0.0718315914273262, -0.086885966360569, 0.09441597014665604, -0.07240451127290726, 0.14123490452766418, 0.12318195402622223, -0.053516924381256104, 0.010368985123932362, -0.031209774315357208, 0.04651070013642311, 0.007842876948416233, 0.10365527868270874, 0.010769560933113098, -0.18099099397659302, 0.022656621411442757, 0.018202748149633408, 0.10856854915618896, -0.17241089046001434, -0.09672945737838745, 0.04725200682878494, 0.001958663808181882, -0.059874359518289566, 0.1282012164592743, 0.057909298688173294, 0.04923510178923607, -0.043742597103118896, -0.017267800867557526, -0.009560109116137028, 0.13584671914577484, -0.10737434774637222, -0.0021453071385622025 ]
null
null
null
# Lora of trento/トレント/特伦托 (Azur Lane) ## What Is This? This is the LoRA model of waifu trento/トレント/特伦托 (Azur Lane). ## How Is It Trained? * This model is trained with [HCP-Diffusion](https://github.com/7eu7d7/HCP-Diffusion). * The [auto-training framework](https://github.com/deepghs/cyberharem) is maintained by [DeepGHS Team](https://huggingface.co/deepghs). * The base model used for training is [deepghs/animefull-latest](https://huggingface.co/deepghs/animefull-latest). * Dataset used for training is the `stage3-p480-800` in [CyberHarem/trento_azurlane](https://huggingface.co/datasets/CyberHarem/trento_azurlane), which contains 145 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 1480 steps, 40 checkpoints were saved and evaluated. * **Trigger word is `trento_azurlane`.** * Pruned core tags for this waifu are `long_hair, breasts, hair_over_one_eye, large_breasts, purple_hair, red_eyes, bangs, very_long_hair, eyewear_on_head, sunglasses, blue_hair`. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. ## How to Use It? ### If You Are Using A1111 WebUI v1.7+ **Just use it like the classic LoRA**. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 1184, you need to download [`1184/trento_azurlane.pt`](https://huggingface.co/CyberHarem/trento_azurlane/resolve/main/1184/trento_azurlane.pt) as the embedding and [`1184/trento_azurlane.safetensors`](https://huggingface.co/CyberHarem/trento_azurlane/resolve/main/1184/trento_azurlane.safetensors) for loading Lora. By using both files together, you can generate images for the desired characters. ## Which Step Should I Use? We selected 5 good steps for you to choose. The best one is step 1184. 1600 images (1.69 GiB) were generated for auto-testing. ![Metrics Plot](metrics_plot.png) The base model used for generating preview images is [Meina/MeinaMix_V11](https://huggingface.co/Meina/MeinaMix_V11). Here are the preview of the recommended steps: | Step | Epoch | CCIP | AI Corrupt | Bikini Plus | Score | Download | pattern_0_0 | pattern_0_1 | pattern_1 | pattern_2_0 | pattern_2_1 | portrait_0 | portrait_1 | portrait_2 | full_body_0 | full_body_1 | profile_0 | profile_1 | free_0 | free_1 | shorts | maid_0 | maid_1 | miko | yukata | suit | china | bikini_0 | bikini_1 | bikini_2 | sit | squat | kneel | jump | crossed_arms | angry | smile | cry | grin | n_lie_0 | n_lie_1 | n_stand_0 | n_stand_1 | n_stand_2 | n_sex_0 | n_sex_1 | |-------:|--------:|:----------|:-------------|:--------------|:----------|:----------------------------------------------------------------------------------------------------|:----------------------------------------------|:----------------------------------------------|:------------------------------------------|:----------------------------------------------|:----------------------------------------------|:--------------------------------------------|:--------------------------------------------|:--------------------------------------------|:----------------------------------------------|:----------------------------------------------|:------------------------------------------|:------------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:--------------------------------|:------------------------------------|:--------------------------------|:----------------------------------|:----------------------------------------|:----------------------------------------|:----------------------------------------|:------------------------------|:----------------------------------|:----------------------------------|:--------------------------------|:------------------------------------------------|:----------------------------------|:----------------------------------|:------------------------------|:--------------------------------|:--------------------------------------|:--------------------------------------|:------------------------------------------|:------------------------------------------|:------------------------------------------|:--------------------------------------|:--------------------------------------| | 1184 | 33 | **0.969** | 0.954 | 0.833 | **0.743** | [Download](https://huggingface.co/CyberHarem/trento_azurlane/resolve/main/1184/trento_azurlane.zip) | ![pattern_0_0](1184/previews/pattern_0_0.png) | ![pattern_0_1](1184/previews/pattern_0_1.png) | ![pattern_1](1184/previews/pattern_1.png) | ![pattern_2_0](1184/previews/pattern_2_0.png) | ![pattern_2_1](1184/previews/pattern_2_1.png) | ![portrait_0](1184/previews/portrait_0.png) | ![portrait_1](1184/previews/portrait_1.png) | ![portrait_2](1184/previews/portrait_2.png) | ![full_body_0](1184/previews/full_body_0.png) | ![full_body_1](1184/previews/full_body_1.png) | ![profile_0](1184/previews/profile_0.png) | ![profile_1](1184/previews/profile_1.png) | ![free_0](1184/previews/free_0.png) | ![free_1](1184/previews/free_1.png) | ![shorts](1184/previews/shorts.png) | ![maid_0](1184/previews/maid_0.png) | ![maid_1](1184/previews/maid_1.png) | ![miko](1184/previews/miko.png) | ![yukata](1184/previews/yukata.png) | ![suit](1184/previews/suit.png) | ![china](1184/previews/china.png) | ![bikini_0](1184/previews/bikini_0.png) | ![bikini_1](1184/previews/bikini_1.png) | ![bikini_2](1184/previews/bikini_2.png) | ![sit](1184/previews/sit.png) | ![squat](1184/previews/squat.png) | ![kneel](1184/previews/kneel.png) | ![jump](1184/previews/jump.png) | ![crossed_arms](1184/previews/crossed_arms.png) | ![angry](1184/previews/angry.png) | ![smile](1184/previews/smile.png) | ![cry](1184/previews/cry.png) | ![grin](1184/previews/grin.png) | ![n_lie_0](1184/previews/n_lie_0.png) | ![n_lie_1](1184/previews/n_lie_1.png) | ![n_stand_0](1184/previews/n_stand_0.png) | ![n_stand_1](1184/previews/n_stand_1.png) | ![n_stand_2](1184/previews/n_stand_2.png) | ![n_sex_0](1184/previews/n_sex_0.png) | ![n_sex_1](1184/previews/n_sex_1.png) | | 1036 | 29 | 0.954 | 0.955 | 0.838 | 0.733 | [Download](https://huggingface.co/CyberHarem/trento_azurlane/resolve/main/1036/trento_azurlane.zip) | ![pattern_0_0](1036/previews/pattern_0_0.png) | ![pattern_0_1](1036/previews/pattern_0_1.png) | ![pattern_1](1036/previews/pattern_1.png) | ![pattern_2_0](1036/previews/pattern_2_0.png) | ![pattern_2_1](1036/previews/pattern_2_1.png) | ![portrait_0](1036/previews/portrait_0.png) | ![portrait_1](1036/previews/portrait_1.png) | ![portrait_2](1036/previews/portrait_2.png) | ![full_body_0](1036/previews/full_body_0.png) | ![full_body_1](1036/previews/full_body_1.png) | ![profile_0](1036/previews/profile_0.png) | ![profile_1](1036/previews/profile_1.png) | ![free_0](1036/previews/free_0.png) | ![free_1](1036/previews/free_1.png) | ![shorts](1036/previews/shorts.png) | ![maid_0](1036/previews/maid_0.png) | ![maid_1](1036/previews/maid_1.png) | ![miko](1036/previews/miko.png) | ![yukata](1036/previews/yukata.png) | ![suit](1036/previews/suit.png) | ![china](1036/previews/china.png) | ![bikini_0](1036/previews/bikini_0.png) | ![bikini_1](1036/previews/bikini_1.png) | ![bikini_2](1036/previews/bikini_2.png) | ![sit](1036/previews/sit.png) | ![squat](1036/previews/squat.png) | ![kneel](1036/previews/kneel.png) | ![jump](1036/previews/jump.png) | ![crossed_arms](1036/previews/crossed_arms.png) | ![angry](1036/previews/angry.png) | ![smile](1036/previews/smile.png) | ![cry](1036/previews/cry.png) | ![grin](1036/previews/grin.png) | ![n_lie_0](1036/previews/n_lie_0.png) | ![n_lie_1](1036/previews/n_lie_1.png) | ![n_stand_0](1036/previews/n_stand_0.png) | ![n_stand_1](1036/previews/n_stand_1.png) | ![n_stand_2](1036/previews/n_stand_2.png) | ![n_sex_0](1036/previews/n_sex_0.png) | ![n_sex_1](1036/previews/n_sex_1.png) | | 666 | 19 | 0.953 | 0.975 | 0.833 | 0.724 | [Download](https://huggingface.co/CyberHarem/trento_azurlane/resolve/main/666/trento_azurlane.zip) | ![pattern_0_0](666/previews/pattern_0_0.png) | ![pattern_0_1](666/previews/pattern_0_1.png) | ![pattern_1](666/previews/pattern_1.png) | ![pattern_2_0](666/previews/pattern_2_0.png) | ![pattern_2_1](666/previews/pattern_2_1.png) | ![portrait_0](666/previews/portrait_0.png) | ![portrait_1](666/previews/portrait_1.png) | ![portrait_2](666/previews/portrait_2.png) | ![full_body_0](666/previews/full_body_0.png) | ![full_body_1](666/previews/full_body_1.png) | ![profile_0](666/previews/profile_0.png) | ![profile_1](666/previews/profile_1.png) | ![free_0](666/previews/free_0.png) | ![free_1](666/previews/free_1.png) | ![shorts](666/previews/shorts.png) | ![maid_0](666/previews/maid_0.png) | ![maid_1](666/previews/maid_1.png) | ![miko](666/previews/miko.png) | ![yukata](666/previews/yukata.png) | ![suit](666/previews/suit.png) | ![china](666/previews/china.png) | ![bikini_0](666/previews/bikini_0.png) | ![bikini_1](666/previews/bikini_1.png) | ![bikini_2](666/previews/bikini_2.png) | ![sit](666/previews/sit.png) | ![squat](666/previews/squat.png) | ![kneel](666/previews/kneel.png) | ![jump](666/previews/jump.png) | ![crossed_arms](666/previews/crossed_arms.png) | ![angry](666/previews/angry.png) | ![smile](666/previews/smile.png) | ![cry](666/previews/cry.png) | ![grin](666/previews/grin.png) | ![n_lie_0](666/previews/n_lie_0.png) | ![n_lie_1](666/previews/n_lie_1.png) | ![n_stand_0](666/previews/n_stand_0.png) | ![n_stand_1](666/previews/n_stand_1.png) | ![n_stand_2](666/previews/n_stand_2.png) | ![n_sex_0](666/previews/n_sex_0.png) | ![n_sex_1](666/previews/n_sex_1.png) | | 925 | 26 | 0.940 | 0.959 | 0.838 | 0.717 | [Download](https://huggingface.co/CyberHarem/trento_azurlane/resolve/main/925/trento_azurlane.zip) | ![pattern_0_0](925/previews/pattern_0_0.png) | ![pattern_0_1](925/previews/pattern_0_1.png) | ![pattern_1](925/previews/pattern_1.png) | ![pattern_2_0](925/previews/pattern_2_0.png) | ![pattern_2_1](925/previews/pattern_2_1.png) | ![portrait_0](925/previews/portrait_0.png) | ![portrait_1](925/previews/portrait_1.png) | ![portrait_2](925/previews/portrait_2.png) | ![full_body_0](925/previews/full_body_0.png) | ![full_body_1](925/previews/full_body_1.png) | ![profile_0](925/previews/profile_0.png) | ![profile_1](925/previews/profile_1.png) | ![free_0](925/previews/free_0.png) | ![free_1](925/previews/free_1.png) | ![shorts](925/previews/shorts.png) | ![maid_0](925/previews/maid_0.png) | ![maid_1](925/previews/maid_1.png) | ![miko](925/previews/miko.png) | ![yukata](925/previews/yukata.png) | ![suit](925/previews/suit.png) | ![china](925/previews/china.png) | ![bikini_0](925/previews/bikini_0.png) | ![bikini_1](925/previews/bikini_1.png) | ![bikini_2](925/previews/bikini_2.png) | ![sit](925/previews/sit.png) | ![squat](925/previews/squat.png) | ![kneel](925/previews/kneel.png) | ![jump](925/previews/jump.png) | ![crossed_arms](925/previews/crossed_arms.png) | ![angry](925/previews/angry.png) | ![smile](925/previews/smile.png) | ![cry](925/previews/cry.png) | ![grin](925/previews/grin.png) | ![n_lie_0](925/previews/n_lie_0.png) | ![n_lie_1](925/previews/n_lie_1.png) | ![n_stand_0](925/previews/n_stand_0.png) | ![n_stand_1](925/previews/n_stand_1.png) | ![n_stand_2](925/previews/n_stand_2.png) | ![n_sex_0](925/previews/n_sex_0.png) | ![n_sex_1](925/previews/n_sex_1.png) | | 444 | 13 | 0.933 | **0.985** | **0.845** | 0.716 | [Download](https://huggingface.co/CyberHarem/trento_azurlane/resolve/main/444/trento_azurlane.zip) | ![pattern_0_0](444/previews/pattern_0_0.png) | ![pattern_0_1](444/previews/pattern_0_1.png) | ![pattern_1](444/previews/pattern_1.png) | ![pattern_2_0](444/previews/pattern_2_0.png) | ![pattern_2_1](444/previews/pattern_2_1.png) | ![portrait_0](444/previews/portrait_0.png) | ![portrait_1](444/previews/portrait_1.png) | ![portrait_2](444/previews/portrait_2.png) | ![full_body_0](444/previews/full_body_0.png) | ![full_body_1](444/previews/full_body_1.png) | ![profile_0](444/previews/profile_0.png) | ![profile_1](444/previews/profile_1.png) | ![free_0](444/previews/free_0.png) | ![free_1](444/previews/free_1.png) | ![shorts](444/previews/shorts.png) | ![maid_0](444/previews/maid_0.png) | ![maid_1](444/previews/maid_1.png) | ![miko](444/previews/miko.png) | ![yukata](444/previews/yukata.png) | ![suit](444/previews/suit.png) | ![china](444/previews/china.png) | ![bikini_0](444/previews/bikini_0.png) | ![bikini_1](444/previews/bikini_1.png) | ![bikini_2](444/previews/bikini_2.png) | ![sit](444/previews/sit.png) | ![squat](444/previews/squat.png) | ![kneel](444/previews/kneel.png) | ![jump](444/previews/jump.png) | ![crossed_arms](444/previews/crossed_arms.png) | ![angry](444/previews/angry.png) | ![smile](444/previews/smile.png) | ![cry](444/previews/cry.png) | ![grin](444/previews/grin.png) | ![n_lie_0](444/previews/n_lie_0.png) | ![n_lie_1](444/previews/n_lie_1.png) | ![n_stand_0](444/previews/n_stand_0.png) | ![n_stand_1](444/previews/n_stand_1.png) | ![n_stand_2](444/previews/n_stand_2.png) | ![n_sex_0](444/previews/n_sex_0.png) | ![n_sex_1](444/previews/n_sex_1.png) | ## Anything Else? Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. ## All Steps We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * [Steps From 1147 to 1480](all/0.md) * [Steps From 777 to 1110](all/1.md) * [Steps From 407 to 740](all/2.md) * [Steps From 37 to 370](all/3.md)
{"license": "mit", "tags": ["art", "not-for-all-audiences"], "datasets": ["CyberHarem/trento_azurlane"], "pipeline_tag": "text-to-image"}
text-to-image
CyberHarem/trento_azurlane
[ "art", "not-for-all-audiences", "text-to-image", "dataset:CyberHarem/trento_azurlane", "license:mit", "region:us" ]
2024-02-15T02:59:22+00:00
[]
[]
TAGS #art #not-for-all-audiences #text-to-image #dataset-CyberHarem/trento_azurlane #license-mit #region-us
Lora of trento/トレント/特伦托 (Azur Lane) =================================== What Is This? ------------- This is the LoRA model of waifu trento/トレント/特伦托 (Azur Lane). How Is It Trained? ------------------ * This model is trained with HCP-Diffusion. * The auto-training framework is maintained by DeepGHS Team. * The base model used for training is deepghs/animefull-latest. * Dataset used for training is the 'stage3-p480-800' in CyberHarem/trento\_azurlane, which contains 145 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 1480 steps, 40 checkpoints were saved and evaluated. * Trigger word is 'trento\_azurlane'. * Pruned core tags for this waifu are 'long\_hair, breasts, hair\_over\_one\_eye, large\_breasts, purple\_hair, red\_eyes, bangs, very\_long\_hair, eyewear\_on\_head, sunglasses, blue\_hair'. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. How to Use It? -------------- ### If You Are Using A1111 WebUI v1.7+ Just use it like the classic LoRA. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 1184, you need to download '1184/trento\_azurlane.pt' as the embedding and '1184/trento\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters. Which Step Should I Use? ------------------------ We selected 5 good steps for you to choose. The best one is step 1184. 1600 images (1.69 GiB) were generated for auto-testing. !Metrics Plot The base model used for generating preview images is Meina/MeinaMix\_V11. Here are the preview of the recommended steps: Anything Else? -------------- Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. All Steps --------- We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * Steps From 1147 to 1480 * Steps From 777 to 1110 * Steps From 407 to 740 * Steps From 37 to 370
[ "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 1184, you need to download '1184/trento\\_azurlane.pt' as the embedding and '1184/trento\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 1184.\n\n\n1600 images (1.69 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 1147 to 1480\n* Steps From 777 to 1110\n* Steps From 407 to 740\n* Steps From 37 to 370" ]
[ "TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/trento_azurlane #license-mit #region-us \n", "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 1184, you need to download '1184/trento\\_azurlane.pt' as the embedding and '1184/trento\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 1184.\n\n\n1600 images (1.69 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 1147 to 1480\n* Steps From 777 to 1110\n* Steps From 407 to 740\n* Steps From 37 to 370" ]
[ 44, 38, 470 ]
[ "passage: TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/trento_azurlane #license-mit #region-us \n### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file." ]
[ 0.008282329887151718, -0.010813342407345772, -0.004074349533766508, 0.08539935946464539, 0.08425532281398773, 0.08042465895414352, 0.22647781670093536, 0.07790467143058777, 0.10676363855600357, -0.0700766071677208, 0.07396978884935379, 0.07299508154392242, -0.006240255665034056, 0.025329170748591423, -0.03649500012397766, -0.15687617659568787, -0.06543301790952682, -0.03148656338453293, -0.010370979085564613, 0.019909244030714035, 0.08352497965097427, 0.0061955321580171585, 0.10663237422704697, -0.05436847358942032, -0.04093346372246742, 0.055826153606176376, -0.025097806006669998, -0.04665796831250191, 0.034060459583997726, 0.08629161864519119, 0.12184813618659973, 0.014943474903702736, 0.07090160250663757, -0.15960951149463654, 0.06758925318717957, -0.00043760257540270686, -0.11019965261220932, -0.005722942296415567, 0.027205707505345345, -0.03855954483151436, 0.11782544106245041, 0.03865690901875496, -0.10896294564008713, 0.040901169180870056, -0.13707749545574188, -0.00007192894554464146, -0.06121373549103737, 0.032826848328113556, 0.13215431571006775, 0.058375973254442215, 0.030140932649374008, 0.06062062457203865, -0.05663517490029335, 0.0809461697936058, 0.12442934513092041, -0.1326274275779724, -0.07233463227748871, 0.1000954806804657, 0.03202112764120102, 0.13176438212394714, -0.09077596664428711, 0.09795486181974411, 0.07601366937160492, -0.04549431428313255, -0.14508414268493652, -0.09481175988912582, -0.2063887119293213, -0.013651732355356216, 0.010551892220973969, 0.02012687176465988, 0.4020841717720032, 0.05731542780995369, 0.03533409908413887, 0.07533325999975204, -0.06924085319042206, 0.02679830603301525, -0.09773319959640503, 0.13292832672595978, 0.04507775604724884, 0.0954304188489914, -0.0478585809469223, -0.10751506686210632, -0.11773879081010818, -0.06786275655031204, -0.08627261966466904, -0.011874516494572163, 0.02533727139234543, 0.1077716052532196, -0.1937468945980072, 0.0035574401263147593, -0.044977638870477676, -0.12492809444665909, 0.0251364316791296, -0.10240250825881958, 0.17301319539546967, 0.06414522230625153, -0.018081439658999443, -0.01041432749480009, 0.24924996495246887, 0.1322508454322815, 0.18650662899017334, 0.043472565710544586, -0.1132921427488327, 0.128951296210289, 0.033446408808231354, -0.08006887882947922, -0.021035827696323395, -0.09900471568107605, 0.14369933307170868, -0.05235183611512184, 0.10962948948144913, -0.057374730706214905, -0.12090787291526794, 0.023666420951485634, -0.10513497143983841, 0.06715977191925049, 0.04195716232061386, 0.005420645698904991, -0.05129195749759674, 0.03747595474123955, 0.03204168751835823, -0.03553204610943794, -0.004661572631448507, -0.010691546835005283, -0.055320046842098236, 0.05347089469432831, 0.0971202403306961, 0.035075344145298004, 0.05537336319684982, -0.0029992389027029276, -0.0194384902715683, -0.005464870948344469, -0.052430104464292526, -0.00952139962464571, 0.04785977303981781, 0.03169257938861847, 0.08751177042722702, -0.15986180305480957, -0.08018863946199417, -0.018526870757341385, 0.06257569789886475, -0.0006314760539680719, 0.1010202169418335, -0.008470374159514904, 0.059924736618995667, 0.0018041738076135516, -0.01684623956680298, 0.025845197960734367, -0.09996909648180008, 0.08771630376577377, -0.017897581681609154, 0.0874844640493393, -0.20351150631904602, -0.004067422356456518, -0.04503501206636429, 0.012504429556429386, 0.0659249946475029, -0.007325366605073214, -0.10321648418903351, 0.12527233362197876, -0.016409287229180336, 0.06849177926778793, -0.11006209254264832, 0.04767181724309921, 0.031114287674427032, 0.07873784750699997, -0.10193626582622528, 0.016479527577757835, 0.12158041447401047, -0.1338706612586975, -0.15960119664669037, 0.09549065679311752, -0.023995263502001762, 0.039851780980825424, 0.04339044168591499, 0.16023612022399902, 0.17169998586177826, -0.1983289122581482, -0.003034234279766679, 0.05588124692440033, -0.024280661717057228, -0.08259642124176025, -0.014868009835481644, 0.10890352725982666, 0.01824786514043808, 0.03317156806588173, -0.029797837138175964, 0.11758898943662643, -0.034727998077869415, -0.0836668536067009, -0.028252294287085533, -0.08033901453018188, -0.07588053494691849, 0.047221872955560684, -0.0077016931027174, -0.05055078864097595, 0.014058544300496578, -0.14244472980499268, 0.16145148873329163, 0.013284537941217422, 0.01936211623251438, -0.07703632861375809, 0.12789760529994965, 0.006900557782500982, -0.00357690523378551, 0.009308338165283203, -0.048418883234262466, -0.10468889027833939, 0.24354608356952667, 0.08693308383226395, 0.09011869877576828, 0.060541801154613495, -0.0547943152487278, -0.06861642003059387, 0.02333882823586464, 0.01363224908709526, -0.038245752453804016, 0.02468820847570896, -0.10466381907463074, 0.05136343464255333, -0.016549127176404, 0.026860862970352173, -0.009906958788633347, -0.0265278872102499, 0.06920379400253296, 0.013759948313236237, -0.021141724660992622, 0.09519041329622269, 0.04380428418517113, -0.019361350685358047, -0.07481934130191803, -0.002927987137809396, 0.07450024038553238, -0.0027160674799233675, -0.07247327268123627, 0.02409626729786396, -0.008897383697330952, 0.045799337327480316, 0.19817107915878296, -0.22987370193004608, 0.04042237997055054, -0.003093279432505369, 0.04705164581537247, 0.04142823442816734, -0.007614717818796635, -0.034031592309474945, 0.021473510190844536, -0.0187489353120327, 0.07229593396186829, -0.024920349940657616, 0.0696261003613472, -0.023185184225440025, -0.14202307164669037, -0.016355453059077263, -0.0306364968419075, 0.16771630942821503, -0.1765207201242447, 0.060667071491479874, 0.18073487281799316, -0.12262150645256042, 0.12240287661552429, 0.0016650217585265636, -0.0072533320635557175, 0.008674788288772106, 0.04381673038005829, 0.008214942179620266, 0.0982264056801796, -0.06702104955911636, -0.018879758194088936, 0.022178776562213898, -0.0877949669957161, 0.03697758913040161, -0.124397873878479, -0.1114298552274704, -0.06630093604326248, -0.033238861709833145, -0.038637012243270874, 0.024496283382177353, -0.051227785646915436, 0.07824406772851944, -0.08899204432964325, -0.08928871899843216, -0.02236027456820011, -0.08222305774688721, 0.02886631339788437, 0.013155392371118069, -0.06328742206096649, -0.1503150463104248, -0.12399573624134064, -0.09351621568202972, -0.13751858472824097, 0.0011159320129081607, 0.06992671638727188, -0.11040053516626358, -0.03848977014422417, 0.01808357797563076, -0.06495580822229385, 0.09383523464202881, -0.08236445486545563, 0.010198683477938175, 0.06008844077587128, -0.03673027828335762, -0.16999603807926178, 0.001603154232725501, -0.0668797567486763, -0.06685683131217957, 0.15437588095664978, -0.16929307579994202, 0.19011878967285156, -0.018016181886196136, 0.048109881579875946, 0.06342428922653198, 0.028362590819597244, 0.12176361680030823, -0.11412513256072998, 0.07776015251874924, 0.1977071762084961, 0.04869340360164642, 0.08170712739229202, 0.12269801646471024, 0.08196268230676651, -0.11131227761507034, 0.03758753836154938, 0.07711627334356308, -0.10073084384202957, -0.10076382011175156, -0.0500476099550724, -0.11592653393745422, -0.06812942773103714, 0.058655235916376114, 0.059391092509031296, 0.06967144459486008, 0.12589722871780396, -0.054868184030056, -0.0017256314167752862, 0.09905458986759186, 0.04804457724094391, 0.08953096717596054, 0.01683102734386921, 0.05628281459212303, -0.13994239270687103, -0.04944632202386856, 0.16652587056159973, 0.2315073162317276, 0.22512611746788025, 0.01962064765393734, 0.07298325002193451, 0.11808299273252487, 0.08429175615310669, 0.08771807700395584, 0.05228864401578903, 0.00023861810041125864, 0.021056968718767166, -0.07885155081748962, -0.050924867391586304, 0.015541283413767815, -0.0034568828996270895, -0.04667365923523903, -0.1475001871585846, 0.09923001378774643, 0.009007398970425129, 0.08571012318134308, 0.13895487785339355, 0.03472043201327324, -0.11271388083696365, 0.16269664466381073, 0.09361519664525986, 0.10084450244903564, -0.0646032840013504, 0.1290625035762787, 0.05449122190475464, -0.0017845388501882553, 0.16057944297790527, 0.02358253113925457, 0.15534687042236328, -0.030475694686174393, -0.0737319365143776, -0.06107965111732483, -0.0430646650493145, 0.0063719418831169605, 0.036253608763217926, -0.2254086583852768, 0.10276028513908386, 0.056738466024398804, 0.01969154365360737, -0.007298523560166359, -0.05700656399130821, 0.17491494119167328, 0.1561840921640396, 0.08053689450025558, 0.026332896202802658, -0.02502117119729519, -0.010458644479513168, -0.08495860546827316, 0.04703335464000702, 0.023672211915254593, 0.076255202293396, -0.03815580531954765, -0.10344894230365753, -0.024292241781949997, 0.0005030549946241081, 0.01821139268577099, -0.08166766911745071, -0.11558423936367035, -0.042093098163604736, 0.24328485131263733, -0.06631390005350113, 0.04872102662920952, 0.054761696606874466, 0.012791960500180721, -0.025002457201480865, 0.05101504549384117, -0.020833760499954224, -0.023694556206464767, -0.03616500645875931, 0.004828114993870258, 0.010725312866270542, -0.04735611751675606, -0.05794346705079079, -0.030992012470960617, -0.09874158352613449, -0.09814251214265823, 0.014736657962203026, -0.04957880079746246, 0.016566026955842972, -0.029971206560730934, 0.01829170063138008, -0.10447345674037933, -0.030612008646130562, 0.03577180951833725, 0.024762259796261787, -0.071014404296875, -0.12813657522201538, -0.01224446576088667, -0.021332694217562675, -0.062030430883169174, 0.02564748004078865, -0.12487475574016571, -0.07716001570224762, -0.05327502638101578, -0.026671120896935463, 0.1258077621459961, 0.21835005283355713, -0.021679963916540146, -0.001931511564180255, 0.1453677862882614, -0.10922257602214813, -0.3158858120441437, -0.15663912892341614, -0.16139505803585052, -0.1012473776936531, 0.03198167309165001, -0.06657449901103973, 0.013788237236440182, 0.08517415076494217, -0.03529374673962593, 0.1872539073228836, -0.21980080008506775, -0.09199072420597076, 0.07433941960334778, 0.09523287415504456, 0.311598002910614, -0.24274620413780212, 0.010884993709623814, -0.11865094304084778, -0.03170520067214966, 0.010256299749016762, -0.097236767411232, 0.12291891872882843, 0.04259512573480606, 0.06496585160493851, -0.003449220908805728, -0.007799386512488127, 0.13594377040863037, -0.07768324762582779, 0.13582588732242584, -0.11611241102218628, -0.10489313304424286, 0.22075606882572174, -0.03234528750181198, -0.003250677138566971, -0.19785763323307037, -0.03203533589839935, -0.023314282298088074, 0.03524085879325867, -0.003110745223239064, 0.0602274052798748, -0.010070440359413624, -0.012673539109528065, -0.12396930158138275, -0.016036909073591232, -0.02329271100461483, 0.06379352509975433, 0.21959012746810913, -0.06581304967403412, -0.05858122929930687, 0.052337296307086945, -0.0017596986144781113, 0.11469529569149017, 0.017049040645360947, -0.055554233491420746, -0.041331708431243896, 0.08914345502853394, -0.18456166982650757, 0.05997580662369728, 0.01265899557620287, -0.006231851410120726, 0.014781540259718895, 0.014157948084175587, 0.02633136697113514, 0.11894959956407547, 0.17933149635791779, 0.008452641777694225, -0.034571513533592224, -0.022000879049301147, 0.03709634393453598, 0.12339653074741364, -0.010473179630935192, 0.11632666736841202, 0.02590874210000038, 0.03374149277806282, 0.009637615643441677, 0.05500314384698868, -0.07731061428785324, -0.07423298060894012, 0.1030057966709137, -0.05530332401394844, -0.08312983810901642, 0.09137063473463058, 0.05299244821071625, 0.06543762236833572, -0.000489863392431289, 0.04218664765357971, 0.01862739957869053, -0.1251554638147354, 0.036815907806158066, 0.19205355644226074, -0.08393417298793793, -0.0660633072257042, -0.07672525942325592, 0.012873529456555843, -0.11675265431404114, 0.08288386464118958, 0.03614602982997894, -0.0301668643951416, 0.1108819916844368, -0.04206503555178642, -0.037554781883955, 0.008016821928322315, -0.049737341701984406, 0.041313741356134415, -0.1469770222902298, -0.19383984804153442, 0.04718223214149475, -0.009142467752099037, -0.06726271659135818, -0.09483914077281952, -0.09299442917108536, 0.06278831511735916, -0.1670549362897873, 0.14427681267261505, -0.058776579797267914, 0.05846452713012695, -0.04135176166892052, -0.04885268583893776, -0.11214787513017654, -0.018919453024864197, -0.055735789239406586, -0.01975416950881481, 0.059845928102731705, 0.011880212463438511, -0.11924335360527039, -0.11491818726062775, 0.0563240721821785, -0.0028067228849977255, -0.008558911271393299, 0.007124139927327633, -0.07055697590112686, 0.008817676454782486, -0.24156984686851501, -0.0655660554766655, 0.08495832979679108, 0.04395012930035591, -0.08835533261299133, 0.12597329914569855, 0.05126292631030083, -0.015497235581278801, 0.033960144966840744, -0.0011329733533784747, 0.16689974069595337, -0.0797010064125061, 0.04159247875213623, -0.1262105405330658, -0.1731049120426178, -0.02389020472764969, 0.03589846193790436, 0.22303318977355957, 0.07912511378526688, 0.11672980338335037, -0.052598655223846436, 0.021664995700120926, -0.01778273656964302, 0.0715470090508461, 0.01280985027551651, -0.10162119567394257, -0.07091189920902252, -0.16508078575134277, -0.06298089772462845, -0.05426793172955513, 0.15426214039325714, 0.05036142095923424, -0.1249975860118866, -0.0051876697689294815, 0.11407212167978287, -0.18308088183403015, -0.01597343012690544, 0.17056508362293243, -0.04376618564128876, 0.025858253240585327, -0.15629389882087708, 0.019595321267843246, 0.07591155171394348, -0.03232826665043831, -0.0026199223939329386, 0.13214965164661407, 0.0026176662649959326, 0.0007820423925295472, 0.04264834150671959, -0.02793293260037899, 0.07264292240142822, -0.06878387928009033, 0.06788726896047592, 0.003358921268954873, -0.04995625466108322, -0.11324694752693176, 0.1852465122938156, -0.016159532591700554, 0.01072340551763773, -0.0586254820227623, -0.003384943585842848, -0.10301882773637772, -0.11151672154664993, -0.07265331596136093, -0.13264641165733337, 0.06854915618896484, -0.06584469228982925, 0.02443259209394455, -0.008723746985197067, 0.017740529030561447, -0.07514446973800659, 0.006306178402155638, -0.19250893592834473, -0.05326400324702263, 0.01495970506221056, -0.014337600208818913, -0.011178134940564632, -0.04625440388917923, -0.03890366479754448, 0.011694192886352539, -0.05164412409067154, -0.06003273278474808, 0.06364379078149796, 0.10639195889234543, 0.054585617035627365, -0.1623663753271103, -0.10112671554088593, -0.06857531517744064, 0.03490159288048744, 0.07479880005121231, 0.17290617525577545, 0.03337572515010834, -0.011132853105664253, 0.050581399351358414, 0.151534304022789, 0.009729825891554356, -0.0869985818862915, -0.06487001478672028, -0.12361104041337967, -0.12492654472589493, -0.006881828419864178, -0.06833257526159286, -0.026779094710946083, 0.026897422969341278, 0.2548786401748657, 0.20219403505325317, -0.15274816751480103, 0.032977279275655746, -0.07114651054143906, 0.03964370861649513, -0.024736188352108, 0.1609419584274292, 0.04555198922753334, 0.14881309866905212, -0.03511627018451691, -0.03532186523079872, -0.05836447328329086, 0.023668447509407997, -0.10249540209770203, 0.039833493530750275, -0.012034764513373375, -0.06987304985523224, -0.06263766437768936, 0.11242013424634933, -0.1120753213763237, 0.06432294100522995, 0.17286133766174316, -0.14126835763454437, -0.020742956548929214, -0.04081796109676361, 0.05051301419734955, 0.10977227985858917, 0.02455800585448742, -0.0776948407292366, -0.031207459047436714, 0.013563400134444237, 0.027669237926602364, -0.1758531928062439, -0.11001501232385635, 0.003011969616636634, -0.11818236857652664, 0.13543878495693207, -0.009678139351308346, -0.0004592125478666276, 0.0407450906932354, -0.061238110065460205, 0.0015723780961707234, 0.1600077599287033, 0.021761320531368256, -0.039338789880275726, -0.01365713495761156, -0.07665890455245972, -0.10612933337688446, 0.0696246325969696, 0.08500825613737106, 0.044314123690128326, 0.002849658252671361, 0.1620713472366333, -0.015161372721195221, -0.041400935500860214, 0.1328699141740799, -0.17339855432510376, 0.09267029911279678, 0.0008569446508772671, -0.017984483391046524, -0.06585881859064102, -0.04805313050746918, 0.0356052927672863, 0.08710567653179169, -0.16091357171535492, -0.04442844167351723, 0.053282227367162704, -0.10502125322818756, 0.059783875942230225, 0.042521797120571136, -0.09108574688434601, 0.013886374421417713, -0.12058424204587936, -0.0016499054618179798, -0.10400617867708206, 0.051281098276376724, 0.20200037956237793, -0.02688520960509777, 0.013284526765346527, -0.1212056428194046, 0.046520017087459564, -0.030889146029949188, -0.040804922580718994, -0.06705444306135178 ]
null
null
transformers
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64c14f6b02e1f8f67c73bd05/pf4d6FA7DriRtVq5HCkxd.png) ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64c14f6b02e1f8f67c73bd05/e4u8VYfDBh11u60rFYJHF.png) This model is a finetune of jondurbin's excellent [bagel](https://huggingface.co/jondurbin/bagel-34b-v0.2) model. It has been trained with new datasets and a new technique, which we will share to the community soon. This model has not utilised any form of merging. ### Evaluation Results | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K | | --- | --- | --- | --- | --- | --- | --- | | 77.29 | 74.23 | 86.76 | 76.66 | 70.22 | 83.66 | 72.18 | ### Contamination Results With reference model jondurbin/bagel-34b-v0.2: | ARC | TruthfulQA | GSM8K | | --- | --- | --- | | 0.08| 0.38| 0.88|
{"license": "other", "license_name": "yi-license", "license_link": "https://huggingface.co/01-ai/Yi-34B-200K/blob/main/LICENSE", "base_model": "jondurbin/bagel-34b-v0.2"}
text-generation
LoneStriker/Smaug-34B-v0.1-2.4bpw-h6-exl2
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "base_model:jondurbin/bagel-34b-v0.2", "license:other", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T02:59:31+00:00
[]
[]
TAGS #transformers #safetensors #llama #text-generation #conversational #base_model-jondurbin/bagel-34b-v0.2 #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
!image/png !image/png This model is a finetune of jondurbin's excellent bagel model. It has been trained with new datasets and a new technique, which we will share to the community soon. This model has not utilised any form of merging. ### Evaluation Results ### Contamination Results With reference model jondurbin/bagel-34b-v0.2: ARC: 0.08, TruthfulQA: 0.38, GSM8K: 0.88
[ "### Evaluation Results", "### Contamination Results\n\n\nWith reference model jondurbin/bagel-34b-v0.2:\n\n\nARC: 0.08, TruthfulQA: 0.38, GSM8K: 0.88" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #conversational #base_model-jondurbin/bagel-34b-v0.2 #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Evaluation Results", "### Contamination Results\n\n\nWith reference model jondurbin/bagel-34b-v0.2:\n\n\nARC: 0.08, TruthfulQA: 0.38, GSM8K: 0.88" ]
[ 72, 5, 40 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #base_model-jondurbin/bagel-34b-v0.2 #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Evaluation Results### Contamination Results\n\n\nWith reference model jondurbin/bagel-34b-v0.2:\n\n\nARC: 0.08, TruthfulQA: 0.38, GSM8K: 0.88" ]
[ -0.04525256156921387, 0.11974851042032242, -0.0023604375310242176, 0.012170374393463135, 0.03191312029957771, -0.05886117368936539, 0.17323105037212372, 0.05917754024267197, -0.04918491840362549, 0.019062822684645653, 0.16053636372089386, 0.18094675242900848, -0.008187057450413704, 0.1386977881193161, -0.12849180400371552, -0.028414348140358925, 0.10181622952222824, -0.001769914524629712, 0.06670404970645905, 0.11774345487356186, 0.1055540069937706, -0.04559142142534256, 0.11616022884845734, -0.000478314672363922, -0.047079816460609436, 0.057999301701784134, 0.07410825043916702, -0.11545513570308685, 0.10300738364458084, 0.07263334095478058, 0.022486131638288498, 0.08843635022640228, -0.020663965493440628, -0.12247072160243988, 0.0387629009783268, -0.0251071285456419, -0.06758589297533035, 0.06031336635351181, 0.01187992561608553, -0.06556987017393112, 0.07975111901760101, 0.038842957466840744, -0.01809644140303135, 0.08711165934801102, -0.10160006582736969, -0.013214037753641605, -0.07276489585638046, 0.07464108616113663, 0.12269559502601624, 0.07799924165010452, -0.033364187926054, 0.17100784182548523, -0.03681858628988266, 0.10829465091228485, 0.12859967350959778, -0.2861159145832062, 0.012810084968805313, 0.14649921655654907, 0.015632623806595802, -0.014787672087550163, -0.02818293496966362, 0.127168670296669, 0.11412127315998077, -0.038246817886829376, -0.05368415266275406, -0.05218542739748955, -0.05121970549225807, 0.04191635921597481, -0.028297731652855873, -0.031205181032419205, 0.24603690207004547, 0.047117650508880615, -0.09339313209056854, -0.03837243467569351, -0.08512642234563828, -0.04135356843471527, 0.005270859692245722, 0.008432362228631973, -0.026538344100117683, 0.04785020276904106, -0.07052703201770782, 0.038619764149188995, -0.09188854694366455, -0.07066746056079865, -0.10828105360269547, 0.18378926813602448, -0.010209337808191776, 0.03299311175942421, -0.08632496744394302, 0.07517071068286896, -0.06521965563297272, -0.14724700152873993, -0.08512477576732635, -0.06403377652168274, 0.11848216503858566, 0.01366229448467493, 0.016455238685011864, -0.0035586729645729065, 0.1555667370557785, 0.1479959487915039, -0.028327224776148796, 0.0006063813925720751, -0.09035154432058334, 0.0205267071723938, -0.011179042980074883, 0.017841985449194908, -0.046613987535238266, -0.00803209189325571, 0.09749525040388107, -0.025481615215539932, 0.13511274755001068, -0.022282762452960014, -0.08684057742357254, -0.002710809698328376, 0.042806316167116165, 0.10156551003456116, -0.0140672093257308, 0.08425431698560715, -0.052823781967163086, 0.040074653923511505, 0.07843570411205292, -0.05424026772379875, -0.05436987802386284, 0.028668098151683807, 0.015002930536866188, -0.01679180935025215, 0.09796661883592606, 0.052758943289518356, 0.01722336933016777, 0.02221459522843361, -0.09820661693811417, -0.05040527135133743, -0.007780633866786957, -0.07567847520112991, 0.044102612882852554, 0.026247646659612656, 0.03990452364087105, -0.20050622522830963, -0.2024843990802765, 0.017882604151964188, -0.021485023200511932, -0.028603633865714073, -0.023840170353651047, -0.01781654916703701, -0.03690173849463463, 0.020091397687792778, -0.05038277804851532, -0.0453876256942749, -0.09905695915222168, 0.07193168252706528, 0.015691842883825302, 0.05465209111571312, -0.11070174723863602, 0.010288162156939507, -0.12189900130033493, 0.055784162133932114, -0.0008082325221039355, -0.024562334641814232, -0.06984174251556396, 0.12232629209756851, -0.09646580368280411, 0.004350407049059868, -0.06777644157409668, -0.009058627299964428, 0.031188344582915306, 0.2439412772655487, -0.056106749922037125, -0.0483710803091526, 0.14184477925300598, -0.10290747135877609, -0.146474689245224, 0.09316866844892502, -0.02648594044148922, 0.09551043808460236, 0.13569827377796173, 0.10372859984636307, -0.08362486213445663, -0.12071352452039719, -0.04245147854089737, 0.031019972637295723, -0.021407349035143852, 0.032973140478134155, 0.056698575615882874, -0.04596508666872978, -0.15885303914546967, 0.040428876876831055, 0.10893809050321579, 0.05181001126766205, -0.03325556591153145, -0.06067856773734093, -0.06387163698673248, -0.09264726936817169, 0.0875277891755104, -0.056905005127191544, 0.03776533529162407, -0.11693596839904785, -0.06886328756809235, -0.09717018157243729, 0.07803769409656525, -0.028727812692523003, -0.011137898080050945, -0.128780797123909, 0.16040876507759094, -0.12373936921358109, 0.010161287151277065, -0.036752041429281235, -0.006272847764194012, -0.0440339632332325, 0.017466934397816658, 0.021833565086126328, -0.047452982515096664, 0.04236302152276039, 0.050389427691698074, -0.047711387276649475, -0.0021017601247876883, 0.07062821090221405, 0.023353898897767067, -0.023983022198081017, -0.12796476483345032, 0.09416372328996658, -0.028211945667862892, 0.1304721236228943, -0.09916528314352036, 0.01916375569999218, 0.0667835995554924, 0.06526730209589005, 0.0018856950337067246, 0.009114283137023449, 0.08137739449739456, -0.04938162863254547, -0.08771296590566635, -0.01176049280911684, 0.03573398292064667, -0.0019473363645374775, -0.12689007818698883, 0.13121262192726135, -0.1341925412416458, 0.24663500487804413, 0.15531542897224426, -0.049411214888095856, -0.026963306590914726, 0.01810063049197197, -0.0019616030622273684, 0.026632843539118767, -0.029577618464827538, -0.007773878984153271, 0.05558868125081062, -0.03407000005245209, 0.10185427963733673, -0.08900423347949982, 0.014143328182399273, 0.01616692915558815, -0.10224787890911102, -0.049550969153642654, 0.11395908892154694, 0.022674305364489555, -0.21802012622356415, 0.1282704472541809, 0.15238741040229797, -0.012105832807719707, 0.12698812782764435, 0.024525528773665428, -0.031773846596479416, 0.0069692861288785934, 0.01455776859074831, -0.008089111186563969, 0.07266543805599213, -0.14309173822402954, 0.05488405376672745, 0.07744058221578598, -0.02179626002907753, 0.04645966365933418, -0.0939142256975174, -0.04648003354668617, 0.011251467280089855, -0.05396752431988716, -0.07915814965963364, 0.10041439533233643, -0.03633032739162445, 0.14521215856075287, -0.03023960255086422, -0.05886389687657356, 0.0597975067794323, 0.00873294286429882, -0.14194972813129425, 0.20696020126342773, -0.05018157884478569, -0.2539389729499817, -0.15807662904262543, -0.1259155571460724, -0.12086830288171768, 0.07073884457349777, 0.10442011058330536, -0.12064488232135773, -0.10227535665035248, -0.09715291857719421, 0.005357338115572929, 0.05081578716635704, 0.025437192991375923, 0.06566254049539566, 0.02001352049410343, 0.03598915413022041, -0.1263357549905777, -0.03752841427922249, 0.009426312521100044, -0.044434282928705215, 0.06842487305402756, -0.10476984828710556, 0.136917844414711, 0.13368754088878632, -0.005514014046639204, 0.010711542330682278, -0.010348239913582802, 0.2177879810333252, -0.03046317957341671, -0.011799454689025879, 0.2519260346889496, -0.05130639299750328, 0.023998664692044258, 0.1930372565984726, 0.00034387398045510054, -0.09298261255025864, 0.060426268726587296, -0.03400789573788643, -0.07959618419408798, -0.2291390299797058, -0.08585230261087418, -0.005240465048700571, 0.09024275094270706, -0.01967555843293667, 0.06269354373216629, 0.17713089287281036, 0.11363354325294495, -0.0586823895573616, -0.042059000581502914, 0.07949195057153702, 0.10106640309095383, 0.1497822254896164, -0.033522699028253555, 0.13658195734024048, -0.0753566175699234, -0.07923572510480881, 0.0819254070520401, -0.011328653432428837, 0.0074868290685117245, 0.10603635758161545, 0.022833775728940964, 0.0897650346159935, 0.04280072823166847, 0.11630502343177795, 0.10067429393529892, 0.018635738641023636, -0.06781565397977829, -0.03464825078845024, -0.08098717033863068, -0.012124256230890751, 0.0859992504119873, -0.1179543063044548, 0.029849160462617874, 0.025358248502016068, -0.043092723935842514, 0.07799109816551208, -0.04413076490163803, 0.16482216119766235, -0.2662866413593292, -0.08682014048099518, 0.09255517274141312, 0.018726011738181114, -0.08015884459018707, 0.05407383292913437, -0.0405050627887249, -0.04566274955868721, 0.09152435511350632, 0.010035226121544838, 0.07530494779348373, 0.057576797902584076, 0.06461914628744125, -0.12900123000144958, -0.14413250982761383, -0.06746280193328857, 0.0970132127404213, -0.2956734597682953, 0.20926906168460846, 0.05372067540884018, -0.003084513358771801, -0.05033077299594879, -0.02428804524242878, 0.030122704803943634, 0.1391783356666565, 0.1509656459093094, -0.01981334201991558, -0.06564483791589737, -0.06485388427972794, -0.12255921214818954, 0.05480664595961571, -0.00819749291986227, 0.03382544592022896, 0.034050218760967255, 0.010311155579984188, -0.004561418201774359, -0.0026016244664788246, 0.09061220288276672, -0.21963302791118622, -0.046655286103487015, 0.05248333513736725, 0.10832510888576508, 0.12074069678783417, -0.06447115540504456, -0.07513546943664551, -0.08508055657148361, 0.23702093958854675, -0.0625227615237236, -0.10194088518619537, -0.09014973789453506, -0.021780971437692642, 0.01762191392481327, -0.0663316398859024, 0.02208435907959938, -0.08036525547504425, 0.024505993351340294, -0.010240960866212845, -0.14934593439102173, 0.10480043292045593, -0.049296315759420395, -0.12117796391248703, -0.025198617950081825, 0.15699182450771332, -0.05874878540635109, 0.03404039889574051, 0.05122647061944008, 0.014375274069607258, -0.030037662014365196, -0.10779748857021332, 0.009121141396462917, 0.0329253114759922, 0.030129294842481613, 0.04136652126908302, -0.048795122653245926, -0.17945589125156403, -0.05609149858355522, -0.06241508945822716, 0.1563376933336258, 0.3060047924518585, -0.0401444248855114, 0.09593530744314194, 0.1683492660522461, -0.04123794659972191, -0.2577572762966156, -0.07495813071727753, -0.15076082944869995, 0.03437121585011482, -0.036793097853660583, -0.016836166381835938, 0.060471221804618835, 0.055813200771808624, -0.08341040462255478, 0.12228112667798996, -0.1719401627779007, -0.1605696976184845, 0.22616682946681976, 0.05346270650625229, 0.41733214259147644, -0.12373313307762146, -0.06255658715963364, -0.1327088475227356, -0.2468731850385666, 0.046127799898386, -0.13185331225395203, 0.07590366154909134, 0.013204272836446762, 0.09867428243160248, 0.0189632847905159, -0.058187421411275864, 0.1694808453321457, -0.01832689717411995, 0.07232312113046646, -0.08308630436658859, -0.07397674769163132, 0.023600811138749123, 0.007810712326318026, 0.11084768921136856, -0.08314908295869827, 0.07256767153739929, -0.02639179117977619, -0.059907618910074234, -0.022033901885151863, 0.07684612274169922, -0.02306857518851757, -0.08536715060472488, -0.08441435545682907, -0.04730981960892677, -0.017372172325849533, -0.022709108889102936, 0.19115155935287476, -0.11369631439447403, 0.17018923163414001, 0.19033950567245483, 0.07518313825130463, -0.13623355329036713, 0.09502947330474854, 0.027228476479649544, -0.10156794637441635, 0.05541050061583519, -0.20051199197769165, 0.07639018446207047, 0.061430923640728, -0.0012111717369407415, 0.10185527056455612, 0.043502096086740494, -0.030594022944569588, 0.061301153153181076, 0.09786342829465866, -0.22464708983898163, 0.10059235244989395, -0.05757242441177368, -0.0007416980806738138, -0.008538971655070782, 0.10571233928203583, 0.1879286766052246, -0.028340907767415047, 0.0011921939440071583, -0.008149935863912106, 0.04514331743121147, -0.057173922657966614, 0.13196025788784027, 0.04099012538790703, 0.01646367833018303, -0.11045177280902863, 0.07405988872051239, -0.015266580507159233, -0.0029268215876072645, 0.002587819704785943, -0.068906769156456, -0.11984097957611084, -0.07245661318302155, -0.0835028663277626, 0.14852671325206757, -0.0926840528845787, -0.06982482969760895, -0.13130107522010803, -0.12495394051074982, 0.005092257168143988, 0.2045643925666809, 0.07511714845895767, 0.1291884481906891, -0.006392376031726599, -0.11234872788190842, 0.0015836305683478713, 0.09020483493804932, 0.04129631817340851, 0.013021300546824932, -0.11610196530818939, 0.029401004314422607, -0.027276942506432533, 0.04721243306994438, -0.08050666749477386, 0.0022118939086794853, -0.13819047808647156, 0.003819393692538142, -0.20548875629901886, 0.0315348319709301, -0.06913459300994873, -0.028310168534517288, -0.02152065746486187, -0.05790617689490318, -0.08014465868473053, 0.023150626569986343, -0.06844570487737656, 0.027660731226205826, 0.015278948470950127, 0.0657072514295578, -0.09135891497135162, -0.019728990271687508, 0.07895719259977341, -0.008307642303407192, 0.07862573117017746, 0.0676976889371872, -0.07427100837230682, 0.05139647051692009, -0.24940131604671478, 0.0018649037228897214, 0.11401069164276123, -0.037043459713459015, -0.013385551981627941, -0.0676761120557785, 0.011837771162390709, 0.08617475628852844, 0.0029287610668689013, 0.06422247737646103, 0.06184908375144005, -0.059269651770591736, -0.08718816190958023, -0.05930265039205551, -0.07730169594287872, 0.0080960588529706, -0.04978354275226593, 0.06966052949428558, -0.01240470726042986, 0.13221296668052673, -0.08040022104978561, -0.023575492203235626, -0.0887303352355957, 0.04240598529577255, -0.012022001668810844, -0.13256709277629852, -0.21663416922092438, -0.05064918100833893, 0.0026295320130884647, -0.014342094771564007, 0.21714670956134796, -0.02622874826192856, -0.08874227851629257, 0.04999305680394173, 0.06934843957424164, 0.16823916137218475, -0.0008650150266475976, 0.25804299116134644, 0.035310011357069016, 0.03383481875061989, -0.06693612784147263, 0.023270899429917336, 0.045881520956754684, -0.10601496696472168, 0.05687053129076958, 0.09904928505420685, -0.02789405547082424, 0.09040209650993347, -0.024938447400927544, -0.04645003005862236, 0.05246574431657791, -0.09910298138856888, -0.05317483842372894, 0.023233721032738686, 0.009847880341112614, 0.036207474768161774, 0.23055145144462585, -0.06486792117357254, -0.03515565022826195, -0.02273271046578884, -0.021145790815353394, -0.19307371973991394, -0.14153456687927246, -0.14047746360301971, -0.10968784242868423, 0.06616965681314468, -0.07175037264823914, -0.01185464859008789, 0.09433726966381073, 0.03147280588746071, -0.05254008248448372, 0.11112251877784729, -0.07656412571668625, -0.0848955363035202, 0.024617794901132584, -0.02266838401556015, 0.0006497151916846633, -0.020755551755428314, -0.017286498099565506, 0.03757702186703682, 0.00966620072722435, 0.013766100630164146, 0.021621860563755035, 0.06936486810445786, 0.0357549786567688, -0.10845433920621872, -0.05757184326648712, -0.06448255479335785, 0.07902612537145615, 0.04891064390540123, 0.13602788746356964, 0.0036216711159795523, 0.003788225119933486, 0.05403226241469383, 0.17218798398971558, -0.03813575953245163, -0.052631642669439316, -0.07094701379537582, 0.13891363143920898, -0.019248971715569496, 0.11031370609998703, -0.013745217584073544, -0.04466104134917259, 0.03597381338477135, 0.23662206530570984, 0.18739384412765503, -0.03022376075387001, 0.01773729734122753, -0.09219540655612946, 0.03275680169463158, 0.04832259193062782, -0.022437097504734993, 0.04993830993771553, 0.1407332569360733, -0.06442586332559586, -0.03680576756596565, 0.014846207574009895, 0.00917789712548256, -0.09620551764965057, 0.07313837110996246, -0.03960747644305229, -0.07642187923192978, -0.08454439043998718, 0.07770311087369919, -0.10824597626924515, 0.09002173691987991, -0.07597295939922333, -0.13074138760566711, -0.0626433864235878, 0.012174737639725208, 0.056368015706539154, 0.0513683557510376, 0.03065972402691841, -0.07269394397735596, -0.0027795112691819668, 0.007419312372803688, 0.005608659703284502, -0.1359582543373108, -0.03462883085012436, 0.05317911133170128, 0.04610811918973923, 0.12341704219579697, 0.025619041174650192, 0.053931187838315964, 0.11613762378692627, -0.05143020674586296, -0.07409734278917313, 0.14772026240825653, 0.01969115622341633, -0.10262318700551987, 0.040930528193712234, -0.03211837634444237, 0.03614179790019989, 0.07360602170228958, 0.054300498217344284, -0.08762802183628082, 0.04163828119635582, -0.007832719944417477, -0.09112028777599335, -0.06365055590867996, 0.048358459025621414, -0.09754876792430878, 0.08241371810436249, 0.03510749340057373, -0.10332797467708588, 0.0019209374440833926, -0.019443746656179428, 0.08583100140094757, 0.030733279883861542, -0.1817665547132492, 0.04598311707377434, -0.17440858483314514, 0.041772615164518356, 0.014958573505282402, -0.014534308575093746, -0.32099953293800354, -0.012852340936660767, -0.09132730960845947, 0.034026045352220535, -0.1058378517627716, 0.04856368154287338, 0.1373571902513504, 0.0041321394965052605, -0.019345324486494064, -0.21664179861545563, 0.0576627179980278, 0.0628793016076088, -0.050371285527944565, -0.14817924797534943 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
null
MaggieZhang/test
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
2024-02-15T03:06:34+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 31, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.06646376848220825, 0.2168014943599701, -0.00225935154594481, 0.023818302899599075, 0.1271018385887146, -0.001635765191167593, 0.04218708351254463, 0.13324736058712006, -0.020175931975245476, 0.11144465953111649, 0.046588581055402756, 0.09377603232860565, 0.09928803145885468, 0.18404334783554077, 0.04859916493296623, -0.2059975117444992, 0.007056170143187046, -0.09090408682823181, 0.014076028019189835, 0.1116579994559288, 0.13719257712364197, -0.10291384905576706, 0.08272874355316162, -0.04045208916068077, -0.02019004337489605, 0.00012576708104461432, -0.09259183704853058, -0.07032395154237747, 0.06885425746440887, 0.06264153122901917, 0.051234472543001175, 0.001456156256608665, 0.09140396863222122, -0.2864592671394348, 0.017265573143959045, 0.08406311273574829, 0.0027674848679453135, 0.06290827691555023, 0.07236549258232117, -0.07389893382787704, 0.11328595131635666, -0.08021481335163116, 0.13019037246704102, 0.08625296503305435, -0.062064990401268005, -0.23071379959583282, -0.07525765895843506, 0.0963398814201355, 0.12251301854848862, 0.06215599179267883, -0.022921854630112648, 0.15455181896686554, -0.06248689442873001, 0.012971068732440472, 0.1294165402650833, -0.11526761949062347, -0.05572471022605896, 0.061741601675748825, 0.11775490641593933, 0.10740239918231964, -0.14110268652439117, -0.0017287094378843904, 0.04900608956813812, 0.029121357947587967, 0.08589313924312592, 0.022661056369543076, 0.12003941088914871, 0.04652795568108559, -0.13695219159126282, -0.04037507623434067, 0.12011898308992386, 0.038862764835357666, -0.06446044892072678, -0.2168138176202774, -0.006778308190405369, -0.0601806715130806, -0.014732478186488152, -0.07019448280334473, 0.039128515869379044, -0.02470310963690281, 0.07317749410867691, -0.04465159401297569, -0.1063927412033081, -0.0421026237308979, 0.0892222449183464, 0.07748593389987946, 0.011527054943144321, -0.02519804798066616, 0.04627908393740654, 0.13455867767333984, 0.05402068421244621, -0.10399353504180908, -0.07017925381660461, -0.06942764669656754, -0.09420394152402878, -0.04035796597599983, 0.056760527193546295, 0.031942449510097504, 0.02665667235851288, 0.22703726589679718, 0.016653569415211678, 0.04155244305729866, 0.0224777739495039, 0.01032855175435543, 0.043662428855895996, 0.0955500528216362, -0.05303520709276199, -0.15660029649734497, -0.04072032496333122, 0.09077946096658707, -0.0027527001220732927, -0.036689214408397675, -0.03966725245118141, 0.03849169611930847, 0.06843466311693192, 0.13122352957725525, 0.07552056759595871, -0.017929591238498688, -0.04813180863857269, -0.030096933245658875, 0.23523783683776855, -0.1493375599384308, 0.04426715523004532, -0.02271856553852558, -0.01804111897945404, -0.03908449783921242, 0.03597262129187584, 0.022118929773569107, -0.000004518366949923802, 0.09706240892410278, -0.058981191366910934, -0.05378659814596176, -0.10168042778968811, -0.03272576630115509, 0.04088849574327469, -0.013975566253066063, -0.010589460842311382, -0.09025166928768158, -0.09490354359149933, -0.04766594246029854, 0.05537205561995506, -0.05123869329690933, -0.03770573064684868, 0.009465423412621021, -0.08151785284280777, -0.005444355774670839, -0.005417742300778627, 0.10699385404586792, -0.03222226724028587, 0.04445803165435791, -0.027600755915045738, 0.05225523188710213, 0.09919606149196625, 0.031576547771692276, -0.0773419588804245, 0.0561848059296608, -0.22559374570846558, 0.07503069192171097, -0.11481974273920059, 0.04335082694888115, -0.1704932004213333, -0.042439818382263184, 0.005444696638733149, 0.0139949731528759, 0.013206101022660732, 0.12720820307731628, -0.19255615770816803, -0.01654396951198578, 0.13260798156261444, -0.09212633967399597, -0.118110790848732, 0.07884611934423447, -0.029701577499508858, 0.1624738723039627, 0.04682036489248276, -0.027025915682315826, 0.09224298596382141, -0.16434773802757263, -0.07092688232660294, -0.00949116237461567, -0.01727987825870514, 0.12109188735485077, 0.07512219995260239, -0.05991523340344429, 0.046571120619773865, 0.02832140028476715, -0.038078423589468, -0.04424772411584854, -0.050857074558734894, -0.10884185880422592, -0.01070026308298111, -0.08987759798765182, 0.04065500199794769, -0.01250192429870367, -0.07916021347045898, -0.029885273426771164, -0.18612512946128845, -0.0030564051121473312, 0.10038342326879501, 0.0035033065360039473, -0.005652366206049919, -0.08666291832923889, 0.026358824223279953, -0.03112892620265484, -0.008404186926782131, -0.16764774918556213, -0.04399421438574791, 0.046902090311050415, -0.16094985604286194, 0.020117372274398804, -0.06413903087377548, 0.06334125250577927, 0.03641495108604431, -0.05590536445379257, -0.0248766727745533, -0.01730942726135254, 0.011945613659918308, -0.05083848536014557, -0.18994836509227753, -0.056277405470609665, -0.037882111966609955, 0.149809330701828, -0.25956398248672485, 0.032966937869787216, 0.051140617579221725, 0.14649195969104767, 0.00406361510977149, -0.05115427449345589, 0.01429014839231968, -0.05360214412212372, -0.054652128368616104, -0.06746816635131836, -0.006135428790003061, -0.027576493099331856, -0.05147203803062439, 0.019243421033024788, -0.1755700707435608, -0.021410830318927765, 0.09424154460430145, 0.12876708805561066, -0.1486445665359497, -0.018640631809830666, -0.048725154250860214, -0.06339836865663528, -0.0715010017156601, -0.07038594037294388, 0.10712739825248718, 0.0513901449739933, 0.04796046018600464, -0.07435787469148636, -0.07092321664094925, 0.02726263552904129, 0.006906150374561548, -0.03382374346256256, 0.08727246522903442, 0.05199531093239784, -0.09209315478801727, 0.0756213590502739, 0.1092359870672226, 0.07177663594484329, 0.09363535046577454, 0.01574566215276718, -0.11756632477045059, -0.028492970392107964, 0.036266472190618515, 0.02740776725113392, 0.1465986967086792, -0.05952361226081848, 0.04016614332795143, 0.04494241625070572, -0.04170418903231621, 0.022319864481687546, -0.08787637203931808, 0.024075502529740334, 0.025203049182891846, -0.0034381982404738665, 0.06284574419260025, -0.02525499276816845, -0.0050758360885083675, 0.07016654312610626, 0.047779910266399384, 0.04621000960469246, 0.009655474685132504, -0.01720241829752922, -0.1047825813293457, 0.16950392723083496, -0.0951867327094078, -0.269941508769989, -0.17632324993610382, 0.026197833940386772, 0.04035249724984169, -0.022378476336598396, 0.031619444489479065, -0.07056326419115067, -0.10630585998296738, -0.1060405746102333, -0.002429972169920802, 0.01714223250746727, -0.06364088505506516, -0.0741225928068161, 0.07348573952913284, 0.04382912442088127, -0.14902326464653015, 0.038552410900592804, 0.055694397538900375, -0.057955220341682434, -0.0233661737293005, 0.09118817001581192, 0.12397737801074982, 0.14583967626094818, -0.021366750821471214, -0.028626007959246635, 0.029004426673054695, 0.19620531797409058, -0.13469526171684265, 0.10371150821447372, 0.13814030587673187, -0.04545360431075096, 0.08360563963651657, 0.1560150384902954, 0.029186224564909935, -0.08317049592733383, 0.05044832453131676, 0.04082648828625679, -0.043159641325473785, -0.2666129767894745, -0.0534592866897583, 0.012832709588110447, -0.06255637854337692, 0.09786593168973923, 0.10183793306350708, 0.11542957276105881, 0.034910861402750015, -0.07166364789009094, -0.043925940990448, -0.0058974819257855415, 0.11737963557243347, -0.05490213260054588, -0.012639665976166725, 0.07686592638492584, -0.05086168646812439, 0.005355054512619972, 0.10266812145709991, 0.02973790094256401, 0.17442677915096283, 0.020399179309606552, 0.11231429129838943, 0.06195578724145889, 0.08633565157651901, 0.0007386076031252742, 0.02951662428677082, 0.05147615820169449, 0.017203815281391144, -0.002300140680745244, -0.10421168059110641, -0.006156572140753269, 0.1449710875749588, 0.028103826567530632, 0.029669636860489845, -0.0018948549404740334, -0.005003341939300299, 0.05121048167347908, 0.1746254414319992, -0.011592294089496136, -0.22072425484657288, -0.0845772922039032, 0.06936841458082199, -0.06218599155545235, -0.12968985736370087, -0.026130788028240204, 0.045467354357242584, -0.17519839107990265, 0.026703642681241035, -0.027433741837739944, 0.0919293761253357, -0.09345759451389313, -0.02221956104040146, 0.03687324374914169, 0.084866963326931, -0.014529162086546421, 0.08703910559415817, -0.14498743414878845, 0.11886418610811234, 0.02978132851421833, 0.09024628251791, -0.11081171780824661, 0.07909037172794342, -0.007550720125436783, 0.009180475026369095, 0.19379350543022156, -0.011335089802742004, -0.03514958545565605, -0.08774717897176743, -0.11210042238235474, -0.013537433929741383, 0.12687496840953827, -0.1243172138929367, 0.08773399889469147, -0.015198243781924248, -0.044079482555389404, 0.00937260314822197, -0.12100647389888763, -0.17273177206516266, -0.19628387689590454, 0.05585884302854538, -0.09575839340686798, 0.025643249973654747, -0.11914430558681488, -0.07089093327522278, -0.02952558360993862, 0.241120383143425, -0.1745356321334839, -0.06510113179683685, -0.1468164622783661, -0.046294767409563065, 0.1662203073501587, -0.04437198117375374, 0.0718095526099205, -0.0208172257989645, 0.20345525443553925, 0.005988610442727804, -0.004939318168908358, 0.06724198162555695, -0.08892562240362167, -0.16873881220817566, -0.06771010160446167, 0.1510489284992218, 0.11680185794830322, 0.04907919466495514, -0.002248800592496991, 0.0011772146681323647, -0.016943959519267082, -0.1137804463505745, -0.0033210667315870523, 0.16037839651107788, 0.03878779336810112, 0.025986969470977783, -0.05243593826889992, -0.08797456324100494, -0.06899320334196091, -0.06853509694337845, 0.06221301481127739, 0.19590823352336884, -0.10376439243555069, 0.1700313836336136, 0.147536963224411, -0.07305635511875153, -0.23175598680973053, 0.035342130810022354, 0.04983805492520332, 0.0014306638622656465, 0.04886869341135025, -0.18252557516098022, 0.10521943867206573, 0.019543392583727837, -0.05505957826972008, 0.13485197722911835, -0.1557481735944748, -0.1552847921848297, 0.0722852572798729, 0.03904085233807564, -0.22423844039440155, -0.1354004591703415, -0.09622503817081451, -0.05825018882751465, -0.14065024256706238, 0.06054598465561867, -0.002136280992999673, 0.015948504209518433, 0.03500790148973465, -0.0015643214574083686, 0.027123261243104935, -0.058935679495334625, 0.18609118461608887, -0.004065449349582195, 0.020676052197813988, -0.060264769941568375, -0.0478842556476593, 0.09839435666799545, -0.06130504235625267, 0.12208222597837448, 0.004057085141539574, 0.01594383642077446, -0.10362856835126877, -0.048314861953258514, -0.04328322783112526, 0.05154227837920189, -0.07548051327466965, -0.10070807486772537, -0.043625857681035995, 0.08841723203659058, 0.07005169242620468, -0.03383097052574158, 0.00549331633374095, -0.07189501076936722, 0.10019614547491074, 0.17795267701148987, 0.17573626339435577, 0.009926567785441875, -0.07241068035364151, 0.01677953451871872, -0.04142116755247116, 0.044231921434402466, -0.2513144314289093, 0.03756171092391014, 0.06098250672221184, 0.029438555240631104, 0.09217222779989243, -0.020435843616724014, -0.1820858269929886, -0.04050002992153168, 0.08094815909862518, -0.05452597141265869, -0.22617179155349731, -0.019085140898823738, 0.0954197570681572, -0.2020406424999237, -0.007372708059847355, 0.03995226323604584, -0.048725228756666183, -0.023169852793216705, 0.00010950004070764408, 0.06317184865474701, 0.002471912419423461, 0.09773622453212738, 0.0735151618719101, 0.09715340286493301, -0.08337292820215225, 0.10562895983457565, 0.10150538384914398, -0.09572599828243256, 0.03605884686112404, 0.06754924356937408, -0.05300498008728027, -0.043293699622154236, 0.03665391728281975, 0.033023297786712646, 0.005234600510448217, -0.060321882367134094, 0.013913018628954887, -0.036497246474027634, 0.044923391193151474, 0.08326134830713272, 0.03754979372024536, -0.013354414142668247, 0.06462216377258301, 0.03401726484298706, -0.10898099094629288, 0.10366570204496384, 0.01731540448963642, 0.04105307161808014, -0.08384523540735245, -0.019968897104263306, 0.035425446927547455, 0.030576206743717194, -0.01765924133360386, -0.02306121215224266, -0.02860277332365513, -0.01614218018949032, -0.14299540221691132, -0.023106401786208153, -0.07243485748767853, 0.006181265693157911, 0.014656842686235905, -0.031884219497442245, -0.011233693920075893, 0.02475680410861969, -0.06979699432849884, -0.07426341623067856, -0.006949664559215307, 0.09833318740129471, -0.15115703642368317, 0.008848577737808228, 0.06907843053340912, -0.11088496446609497, 0.08190931379795074, -0.008411259390413761, 0.016245156526565552, 0.022527478635311127, -0.15448406338691711, 0.05601610988378525, 0.0008648968650959432, 0.01916889287531376, 0.025886621326208115, -0.16471809148788452, 0.004104440100491047, -0.04661374166607857, -0.02149827405810356, -0.00004464812809601426, -0.02647159807384014, -0.12325995415449142, 0.06858719140291214, -0.015622655861079693, -0.035931166261434555, -0.02701525390148163, 0.0539589487016201, 0.07888586074113846, -0.027474910020828247, 0.10445091128349304, -0.008690856397151947, 0.04941811040043831, -0.16801609098911285, -0.02470702864229679, -0.04982255399227142, 0.019377702847123146, 0.009884213097393513, -0.007693959400057793, 0.04183054715394974, -0.00976533442735672, 0.21883612871170044, -0.05075952783226967, 0.1607085019350052, 0.05847611650824547, -0.017352959141135216, -0.0007513365126214921, 0.06180921941995621, 0.05997028574347496, 0.04658793285489082, 0.009480604901909828, 0.023740366101264954, -0.022450892254710197, -0.006695089396089315, -0.15932634472846985, 0.01890849508345127, 0.14999441802501678, 0.06301083415746689, 0.024745315313339233, 0.05866100639104843, -0.12775006890296936, -0.12135478109121323, 0.09311001747846603, -0.026755332946777344, 0.00928465835750103, -0.08245618641376495, 0.1358020007610321, 0.14980104565620422, -0.14000412821769714, 0.05256148427724838, -0.06134212389588356, -0.05217423290014267, -0.10388828068971634, -0.12032219022512436, -0.05887215584516525, -0.053666237741708755, 0.002330566756427288, -0.03760887682437897, 0.054546963423490524, 0.03344334661960602, -0.009351172484457493, -0.00022941511997487396, 0.13597318530082703, -0.019751882180571556, -0.0028988157864660025, 0.048313532024621964, 0.03693558648228645, 0.02373051457107067, -0.05275435373187065, 0.02940409444272518, 0.02539868652820587, 0.032232340425252914, 0.06546790152788162, 0.033412106335163116, -0.047448933124542236, 0.03804153576493263, -0.0025254099164158106, -0.11207924783229828, 0.019641218706965446, -0.00460948096588254, -0.0742158442735672, 0.1268945336341858, 0.0407399944961071, 0.010224059224128723, -0.03741471841931343, 0.24361543357372284, -0.06653323769569397, -0.06378097087144852, -0.13251738250255585, 0.10491154342889786, -0.0027236645109951496, 0.06476365029811859, 0.023412218317389488, -0.1284150779247284, 0.005243356805294752, 0.13858191668987274, 0.12181595712900162, 0.0045748427510261536, 0.009228081442415714, 0.0518609918653965, 0.0025186820421367884, -0.06998204439878464, 0.054019294679164886, 0.06992026418447495, 0.12919506430625916, -0.07847554981708527, 0.07680778950452805, 0.0006860480643808842, -0.08370215445756912, -0.02947772853076458, 0.11312682181596756, -0.0409729965031147, 0.03491825982928276, -0.047444481402635574, 0.10916327685117722, -0.05787910893559456, -0.29412412643432617, 0.02350960113108158, -0.09588567912578583, -0.15202060341835022, -0.018367812037467957, 0.05944539234042168, -0.02624768204987049, 0.018029648810625076, 0.06971040368080139, -0.06011629104614258, 0.20098382234573364, 0.0335683599114418, -0.07864278554916382, -0.0664360448718071, 0.04837050288915634, -0.06564252078533173, 0.2949807047843933, 0.008418165147304535, 0.02863333560526371, 0.10770907253026962, -0.03253700211644173, -0.18271861970424652, 0.010723991319537163, 0.1133992001414299, -0.08056149631738663, 0.08200647681951523, 0.19000613689422607, -0.012578671798110008, 0.1209007054567337, 0.05294662341475487, -0.047376248985528946, 0.04217283055186272, -0.03389401361346245, -0.051268599927425385, -0.10752558708190918, 0.058453381061553955, -0.05909625440835953, 0.15447644889354706, 0.10152646154165268, -0.05671518296003342, -0.004550917539745569, -0.05555408447980881, 0.04875178262591362, 0.01804669201374054, 0.12263146042823792, 0.02951994352042675, -0.1865430772304535, 0.032826557755470276, -0.01144319772720337, 0.10186848044395447, -0.25588861107826233, -0.08421015739440918, 0.08833149075508118, -0.011924264021217823, -0.05105875805020332, 0.10560628771781921, 0.057650718837976456, 0.04243382066488266, -0.043439045548439026, -0.10480839014053345, -0.02186836116015911, 0.14663739502429962, -0.1469624787569046, -0.025013303384184837 ]
null
null
diffusers
# plum <Gallery /> ## Download model [Download](/lylosn/plum/tree/main) them in the Files & versions tab.
{"tags": ["text-to-image", "stable-diffusion", "lora", "diffusers", "template:sd-lora"], "widget": [{"text": "plum", "output": {"url": "images/profile_plum (17).png"}}], "base_model": "InstantX/InstantID"}
text-to-image
lylosn/plum
[ "diffusers", "text-to-image", "stable-diffusion", "lora", "template:sd-lora", "base_model:InstantX/InstantID", "region:us" ]
2024-02-15T03:07:14+00:00
[]
[]
TAGS #diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-InstantX/InstantID #region-us
# plum <Gallery /> ## Download model Download them in the Files & versions tab.
[ "# plum\n\n<Gallery />", "## Download model\n\n\nDownload them in the Files & versions tab." ]
[ "TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-InstantX/InstantID #region-us \n", "# plum\n\n<Gallery />", "## Download model\n\n\nDownload them in the Files & versions tab." ]
[ 47, 8, 14 ]
[ "passage: TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-InstantX/InstantID #region-us \n# plum\n\n<Gallery />## Download model\n\n\nDownload them in the Files & versions tab." ]
[ -0.07747066766023636, 0.007895318791270256, -0.0037919834721833467, 0.01482907310128212, 0.1215486004948616, 0.048235174268484116, 0.13110210001468658, 0.07494860887527466, 0.0862073227763176, 0.0778968557715416, 0.039883702993392944, 0.12544988095760345, 0.012029966339468956, 0.19545978307724, -0.03512014448642731, -0.24596811830997467, 0.04464305564761162, 0.012501967139542103, -0.035124361515045166, 0.0420561209321022, 0.028840655460953712, -0.04317428916692734, 0.13047637045383453, -0.07558687031269073, -0.0334915928542614, 0.011908410117030144, -0.020044509321451187, -0.028920885175466537, -0.0405663326382637, 0.052899330854415894, 0.02240690030157566, 0.09903533011674881, 0.09613079577684402, -0.15205493569374084, 0.04699110984802246, 0.03639855235815048, -0.06626633554697037, 0.03181920945644379, 0.0022401383612304926, -0.07625158131122589, 0.14088620245456696, -0.017718961462378502, -0.008449177257716656, 0.054878246039152145, -0.08430637419223785, 0.016843192279338837, -0.010634716600179672, -0.021986309438943863, 0.09653918445110321, -0.006343060173094273, 0.03816407546401024, 0.015542080625891685, -0.0246000774204731, 0.05044999346137047, 0.2232123762369156, -0.18017931282520294, -0.02803575061261654, 0.24244184792041779, 0.10941754281520844, 0.23004871606826782, -0.011931732296943665, 0.10043223202228546, 0.08710908144712448, -0.03538668155670166, -0.028579512611031532, -0.04298321157693863, 0.05593306943774223, -0.01667986623942852, -0.07223779708147049, 0.04779956862330437, 0.31434667110443115, 0.04082747548818588, -0.0091609051451087, -0.0626991018652916, -0.052505649626255035, 0.07661686837673187, -0.15132033824920654, 0.07545848190784454, 0.05670253187417984, 0.03438105434179306, 0.0037888376973569393, -0.12729620933532715, -0.047156453132629395, -0.06858666241168976, -0.06681104749441147, 0.11852776259183884, -0.01838778518140316, 0.0785323828458786, -0.07151556760072708, 0.03964546322822571, -0.06884271651506424, -0.14549744129180908, 0.02563987858593464, -0.11276647448539734, 0.09475278109312057, 0.0867556631565094, -0.03125675767660141, -0.026787754148244858, 0.11761824786663055, 0.07584038376808167, 0.15346872806549072, 0.04333490878343582, -0.07191068679094315, 0.12618066370487213, 0.03656500577926636, -0.0030177750159054995, 0.005907372105866671, -0.08986584842205048, 0.05765717849135399, 0.03182331472635269, 0.10374418646097183, -0.057214729487895966, -0.1797119379043579, 0.0422758050262928, -0.1710984706878662, 0.0062665510922670364, 0.06910654902458191, -0.0038712217938154936, -0.09477546066045761, -0.05147358030080795, 0.18938283622264862, 0.004764984827488661, -0.05240211263298988, -0.03493994101881981, -0.061186790466308594, 0.19452986121177673, 0.09148452430963516, -0.012646944262087345, 0.07288791984319687, 0.03530266135931015, -0.11162706464529037, -0.01008972804993391, -0.026989378035068512, -0.025941867381334305, 0.016447383910417557, -0.1703607589006424, 0.04041004180908203, -0.13846814632415771, -0.23917832970619202, 0.025534695014357567, 0.044297538697719574, -0.06558302044868469, 0.018990477547049522, -0.011914416216313839, 0.003686488838866353, 0.03919383883476257, -0.023012250661849976, -0.0807909294962883, -0.08365324139595032, 0.0828230232000351, -0.024549046531319618, 0.15625935792922974, -0.0954752042889595, 0.03202477842569351, -0.06489332020282745, 0.03728748857975006, -0.26637887954711914, 0.046236552298069, -0.1360986828804016, 0.08411331474781036, -0.07344429939985275, -0.052196986973285675, -0.09027235954999924, 0.049776069819927216, -0.025767629966139793, 0.20901846885681152, -0.15038038790225983, -0.06891129165887833, 0.10326164215803146, -0.14998313784599304, -0.0810491293668747, -0.008904014714062214, 0.040519725531339645, 0.018525581806898117, 0.047648243606090546, 0.19315186142921448, 0.012992335483431816, -0.23637539148330688, 0.0888051763176918, 0.12147228419780731, -0.026460682973265648, -0.04979673773050308, 0.04812794178724289, 0.034843411296606064, 0.08294112980365753, 0.05659982189536095, -0.08170916140079498, 0.10070988535881042, -0.044857680797576904, 0.011847639456391335, -0.017485735937952995, -0.09925561398267746, 0.039196304976940155, 0.060539014637470245, 0.02751949615776539, 0.01029235403984785, -0.004538984037935734, 0.08444485068321228, 0.08815911412239075, 0.012285343371331692, -0.004582736175507307, -0.010984248481690884, 0.1697544902563095, -0.09624388068914413, -0.010234925895929337, -0.11188151687383652, -0.09458183497190475, -0.006490881089121103, 0.08445344865322113, 0.01678619720041752, 0.031388457864522934, 0.06032981723546982, 0.005527209490537643, -0.039363134652376175, 0.013669265434145927, 0.0348837710916996, -0.013446886092424393, 0.021542832255363464, -0.10082679986953735, 0.04014516994357109, -0.05967524275183678, 0.020614510402083397, -0.16146087646484375, 0.01973799057304859, 0.010082519613206387, 0.12209005653858185, 0.07107345759868622, -0.03239976242184639, 0.045798514038324356, -0.04668429493904114, -0.07299592345952988, -0.01394885778427124, 0.06845249980688095, 0.009439594112336636, -0.05300867557525635, 0.15624919533729553, -0.051113616675138474, 0.11147419363260269, 0.1955237239599228, -0.07255881279706955, 0.009839232079684734, -0.11527778953313828, 0.02104979008436203, 0.02198224700987339, 0.0117731224745512, -0.08435505628585815, -0.12506423890590668, -0.001932552084326744, 0.10771320760250092, -0.06170962378382683, 0.06901181489229202, 0.08566364645957947, -0.06169704347848892, -0.04526171833276749, 0.03568454831838608, 0.22231784462928772, -0.027851440012454987, 0.09219522774219513, 0.1750272661447525, 0.015230193734169006, 0.1871485412120819, -0.05494610592722893, -0.0893697738647461, 0.07911902666091919, -0.01650199666619301, 0.01337041798979044, 0.16539806127548218, -0.011259187012910843, -0.02145097777247429, 0.03461292386054993, -0.012888017110526562, 0.020571764558553696, -0.06409282237291336, -0.07217610627412796, 0.007702709175646305, -0.03822716698050499, 0.0429784432053566, 0.06593835353851318, -0.09194348007440567, 0.05253325402736664, -0.06478885561227798, -0.00376391364261508, -0.02654441073536873, -0.01621866598725319, -0.03489520773291588, 0.061532195657491684, -0.12050708383321762, -0.12218423932790756, -0.134987935423851, -0.11105555295944214, -0.05392337962985039, 0.029528746381402016, 0.039113134145736694, -0.10305516421794891, -0.04509124159812927, -0.037434205412864685, 0.02531939372420311, 0.03493673354387283, -0.06989060342311859, -0.027740314602851868, 0.02862595207989216, -0.06152227148413658, -0.11080720275640488, -0.008165977895259857, -0.052731119096279144, 0.07462594658136368, 0.04515717178583145, -0.16735681891441345, 0.1333492398262024, 0.08683222532272339, 0.09341264516115189, 0.026508325710892677, 0.045967064797878265, 0.1911633014678955, -0.06911448389291763, 0.07590463757514954, 0.2513492703437805, 0.08314254879951477, 0.03894041106104851, 0.06845642626285553, 0.05552569031715393, -0.10588552802801132, 0.033896055072546005, -0.02383142150938511, -0.10955178737640381, -0.10898788273334503, -0.17081421613693237, -0.1159515529870987, 0.03441663831472397, 0.05418797582387924, 0.013441432267427444, 0.000253685429925099, 0.11539655178785324, -0.016387825831770897, -0.009103868156671524, 0.05283674597740173, 0.06302767992019653, 0.06886383146047592, -0.04415682703256607, 0.06923441588878632, -0.048093486577272415, -0.017416514456272125, 0.17348439991474152, -0.02028004266321659, 0.1733698844909668, 0.040409937500953674, 0.06431207060813904, 0.011881192214787006, -0.019963210448622704, 0.15828709304332733, 0.08881156146526337, 0.012305868789553642, -0.032849133014678955, -0.029962711036205292, -0.09918064624071121, 0.023277195170521736, 0.03957890719175339, -0.009217255748808384, -0.126348614692688, 0.015943733975291252, -0.060907911509275436, -0.00806319061666727, -0.030759546905755997, 0.06521360576152802, -0.24919883906841278, 0.07687809318304062, 0.09897670149803162, 0.13235308229923248, -0.06843249499797821, 0.048225000500679016, 0.05457261949777603, -0.045418016612529755, 0.08193294703960419, 0.019220955669879913, 0.12361856549978256, -0.0061605554074049, -0.06381867080926895, -0.013945895247161388, 0.027852362021803856, -0.0008903858833946288, -0.03192087635397911, -0.017510294914245605, 0.1169431284070015, 0.02080567739903927, -0.048223674297332764, 0.04042995721101761, -0.03206489607691765, 0.09004195779561996, 0.19278794527053833, 0.10637735575437546, 0.03721423074603081, -0.008242514915764332, -0.06395694613456726, -0.11301092058420181, 0.0059487177059054375, 0.08251969516277313, -0.0377468541264534, -0.049994099885225296, 0.015050627291202545, -0.026080574840307236, -0.015404965728521347, -0.015781596302986145, -0.11716397106647491, -0.13395482301712036, -0.025882164016366005, 0.021971862763166428, 0.0295867882668972, -0.005742640700191259, -0.04917542636394501, -0.04176534712314606, 0.0679829865694046, 0.03958243876695633, -0.10463608801364899, -0.07524333149194717, -0.04628204181790352, 0.058857448399066925, 0.016216440126299858, 0.0848589614033699, -0.03923501819372177, 0.037857815623283386, -0.0372755229473114, -0.1558939665555954, 0.025270866230130196, -0.0971628949046135, -0.04878339543938637, -0.11316345632076263, 0.08045272529125214, -0.04677353799343109, -0.01674065925180912, -0.02124079316854477, 0.04895668104290962, 0.025260863825678825, -0.08193571865558624, 0.0033833570778369904, 0.11049281060695648, 0.03354001045227051, 0.12771156430244446, -0.11184480041265488, -0.026040028780698776, 0.011313681490719318, 0.002752151805907488, 0.018620602786540985, 0.19671525061130524, -0.04369155690073967, 0.05399515479803085, 0.11249399185180664, -0.03999588266015053, -0.2127912938594818, -0.04022660106420517, -0.03483198955655098, -0.046824824064970016, 0.090249203145504, -0.15339383482933044, 0.12327559292316437, 0.08570370823144913, -0.04732638597488403, 0.26114147901535034, -0.2820014953613281, -0.0705152228474617, -0.024300402030348778, 0.10520325601100922, 0.2614355981349945, -0.19823777675628662, -0.08429215103387833, -0.09491683542728424, -0.24930866062641144, 0.0517856739461422, 0.014156493358314037, 0.045490000396966934, -0.006883600261062384, -0.03646217659115791, 0.010823211632668972, -0.004115464631468058, 0.19497647881507874, -0.07506445050239563, 0.014345094561576843, -0.10452531278133392, 0.03354421257972717, 0.12351582199335098, -0.017013275995850563, 0.0002514977823011577, -0.21280178427696228, 0.0591600127518177, -0.12025907635688782, 0.010697493329644203, 0.004248961340636015, 0.022864997386932373, -0.014483273029327393, -0.022046292200684547, -0.04452279210090637, 0.013097718358039856, -0.013292447663843632, 0.03316018730401993, 0.1174677237868309, -0.06444557011127472, 0.01184472069144249, 0.14126254618167877, -0.07562515884637833, 0.038666579872369766, -0.17247861623764038, -0.09726185351610184, -0.04193788394331932, 0.07215429097414017, -0.24646516144275665, -0.04111120104789734, 0.09125674515962601, 0.04573466256260872, 0.06788737326860428, 0.01681116409599781, 0.01760661043226719, 0.10250914096832275, 0.154429093003273, -0.07846815139055252, -0.010287214070558548, -0.04114234820008278, -0.016419636085629463, 0.07967393100261688, -0.012045790441334248, 0.09089895337820053, 0.0017909664893522859, 0.030978228896856308, 0.010577422566711903, 0.034518785774707794, -0.03489772975444794, 0.06180718168616295, 0.08940663188695908, -0.004729407373815775, -0.08185632526874542, 0.0870605856180191, -0.006751921493560076, -0.04899674654006958, -0.10084552317857742, 0.09238038212060928, -0.11566215008497238, -0.04982507973909378, 0.04384443536400795, 0.06807015091180801, -0.09682627767324448, 0.03355680778622627, -0.0763433650135994, -0.02883658930659294, -0.03018779680132866, 0.13120850920677185, 0.0851811096072197, -0.03921528905630112, 0.031100669875741005, 0.006849496625363827, 0.05662093311548233, 0.01473903190344572, 0.0626203790307045, 0.0874079242348671, -0.16424241662025452, -0.2534202039241791, 0.02159983478486538, 0.028653757646679878, -0.11055838316679001, -0.05085701122879982, -0.12966494262218475, -0.011393032036721706, -0.14142298698425293, 0.10053279995918274, -0.08393598347902298, -0.043266989290714264, -0.07399244606494904, -0.046996548771858215, -0.05745447799563408, -0.02785543166100979, -0.0416184663772583, -0.015367528423666954, 0.008142230100929737, 0.06915800273418427, -0.05449866130948067, -0.03747550770640373, 0.06851634383201599, -0.08710463345050812, 0.03000178188085556, 0.06041351705789566, -0.02835911326110363, 0.027924492955207825, -0.10544408857822418, -0.027120929211378098, 0.09442895650863647, 0.08772192150354385, -0.00996014941483736, 0.16613039374351501, 0.03651926666498184, -0.025734370574355125, 0.012524254620075226, -0.03554021194577217, 0.003604358993470669, -0.1254129707813263, 0.04568548500537872, -0.08649592846632004, -0.03055201657116413, -0.03937900811433792, -0.004006486386060715, 0.1790512502193451, 0.09805108606815338, 0.08706975728273392, -0.04717203229665756, 0.03697619587182999, -0.07803525030612946, 0.007150989957153797, 0.032635532319545746, -0.10751237720251083, 0.057153597474098206, -0.04741106554865837, -0.006929879076778889, -0.014457128942012787, 0.21280761063098907, 0.04700940474867821, -0.056831520050764084, 0.0075998203828930855, 0.12344218790531158, 0.10689491778612137, 0.0043146004900336266, 0.26263856887817383, 0.07591526955366135, 0.075718455016613, -0.15203548967838287, 0.0912812128663063, 0.0800558403134346, -0.06261930614709854, -0.08279850333929062, 0.14382469654083252, -0.09791207313537598, 0.04387625306844711, 0.06786755472421646, 0.03647070750594139, -0.014342758804559708, 0.01805540919303894, 0.03937868773937225, 0.07342136651277542, -0.012432971969246864, 0.05089525133371353, 0.11123868823051453, -0.04673052951693535, -0.0068677132949233055, 0.04771384596824646, -0.010966256260871887, -0.07991892844438553, -0.17326593399047852, -0.06132267042994499, -0.28178396821022034, 0.047113459557294846, -0.052575018256902695, 0.03168826177716255, 0.12378913164138794, 0.03592299669981003, 0.013944605365395546, 0.00206189788877964, -0.0790642723441124, -0.1195482686161995, 0.08265072852373123, -0.015540235675871372, -0.08350085467100143, -0.03584076464176178, -0.03449990227818489, 0.06586834788322449, -0.039532553404569626, -0.0471029095351696, 0.059515632688999176, 0.018340032547712326, 0.04951569065451622, -0.03521379828453064, -0.06967928260564804, -0.03940251097083092, 0.01725992187857628, -0.07647453248500824, 0.18026834726333618, 0.03856950253248215, 0.021579653024673462, -0.017624113708734512, 0.1961071789264679, -0.01149971503764391, -0.03825727105140686, -0.04565609246492386, 0.0015700090443715453, -0.06749897450208664, 0.09410864859819412, -0.07356010377407074, -0.09806301444768906, -0.05938243865966797, 0.2635190784931183, 0.15415988862514496, -0.10851992666721344, 0.0053609684109687805, 0.0031624429393559694, -0.006783525459468365, 0.049026768654584885, 0.06256166100502014, 0.044541869312524796, 0.2134939432144165, -0.0030863825231790543, -0.07068741321563721, -0.09889781475067139, -0.039394523948431015, -0.09642935544252396, -0.09206278622150421, 0.048187483102083206, -0.09822526574134827, -0.07269486784934998, 0.1402360498905182, -0.17128124833106995, 0.04124468192458153, 0.045747168362140656, -0.15508751571178436, -0.006847194861620665, -0.11108605563640594, 0.07366001605987549, 0.08350226283073425, 0.01725018583238125, -0.10055916756391525, -0.053469497710466385, -0.07606091350317001, -0.01839527301490307, -0.14006097614765167, -0.09783133864402771, -0.026595719158649445, -0.199581578373909, 0.08904103934764862, -0.05202663317322731, -0.017674101516604424, -0.01664683036506176, 0.033898212015628815, -0.007226203568279743, 0.03692178428173065, -0.0030552824027836323, -0.0984998494386673, -0.060270000249147415, 0.07791879028081894, -0.05600378289818764, -0.018777601420879364, 0.03787074610590935, -0.09550672024488449, 0.01604594476521015, 0.15860676765441895, -0.08432074636220932, -0.058845508843660355, 0.011700322851538658, -0.12918247282505035, 0.09131760895252228, 0.004415182862430811, 0.013855275698006153, -0.05664954334497452, -0.00869309064000845, 0.056095194071531296, 0.09184736758470535, -0.09425138682126999, 0.04560406878590584, -0.08094069361686707, -0.13360966742038727, 0.06545902043581009, 0.007768378127366304, -0.10310792177915573, -0.004864047281444073, -0.14910003542900085, 0.046542562544345856, -0.01316075213253498, 0.06267806142568588, 0.2177879512310028, -0.010202725417912006, -0.02742050401866436, -0.2054244428873062, 0.07761581242084503, 0.0829853042960167, -0.07565160095691681, -0.07813931256532669 ]
null
null
null
import gradio openai.api_key = "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" messages = [{"role": "system", "content": "You are a financial experts that specializes in real estate investment and negotiation"}] def CustomChatGPT(user_input): messages.append({"role": "user", "content": user_input}) response = openai.ChatCompletion.create( model = "gpt-3.5-turbo", messages = messages ) ChatGPT_reply = response["choices"][0]["message"]["content"] messages.append({"role": "assistant", "content": ChatGPT_reply}) return ChatGPT_reply demo = gradio.Interface(fn=CustomChatGPT, inputs = "text", outputs = "text", title = "Real Estate Pro") demo.launch(share=True)
{}
null
meeeeeeeeeeeeeeeeeee/lol
[ "region:us" ]
2024-02-15T03:12:59+00:00
[]
[]
TAGS #region-us
import gradio openai.api_key = "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" messages = [{"role": "system", "content": "You are a financial experts that specializes in real estate investment and negotiation"}] def CustomChatGPT(user_input): URL({"role": "user", "content": user_input}) response = URL( model = "gpt-3.5-turbo", messages = messages ) ChatGPT_reply = response["choices"][0]["message"]["content"] URL({"role": "assistant", "content": ChatGPT_reply}) return ChatGPT_reply demo = gradio.Interface(fn=CustomChatGPT, inputs = "text", outputs = "text", title = "Real Estate Pro") URL(share=True)
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
null
# Lora of juno/ジュノー/天后 (Azur Lane) ## What Is This? This is the LoRA model of waifu juno/ジュノー/天后 (Azur Lane). ## How Is It Trained? * This model is trained with [HCP-Diffusion](https://github.com/7eu7d7/HCP-Diffusion). * The [auto-training framework](https://github.com/deepghs/cyberharem) is maintained by [DeepGHS Team](https://huggingface.co/deepghs). * The base model used for training is [deepghs/animefull-latest](https://huggingface.co/deepghs/animefull-latest). * Dataset used for training is the `stage3-p480-800` in [CyberHarem/juno_azurlane](https://huggingface.co/datasets/CyberHarem/juno_azurlane), which contains 59 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 800 steps, 40 checkpoints were saved and evaluated. * **Trigger word is `juno_azurlane`.** * Pruned core tags for this waifu are `pink_hair, long_hair, crown, bangs, mini_crown, ribbon, twintails, pink_eyes, bow, purple_eyes`. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. ## How to Use It? ### If You Are Using A1111 WebUI v1.7+ **Just use it like the classic LoRA**. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 480, you need to download [`480/juno_azurlane.pt`](https://huggingface.co/CyberHarem/juno_azurlane/resolve/main/480/juno_azurlane.pt) as the embedding and [`480/juno_azurlane.safetensors`](https://huggingface.co/CyberHarem/juno_azurlane/resolve/main/480/juno_azurlane.safetensors) for loading Lora. By using both files together, you can generate images for the desired characters. ## Which Step Should I Use? We selected 5 good steps for you to choose. The best one is step 480. 1520 images (1.59 GiB) were generated for auto-testing. ![Metrics Plot](metrics_plot.png) The base model used for generating preview images is [Meina/MeinaMix_V11](https://huggingface.co/Meina/MeinaMix_V11). Here are the preview of the recommended steps: | Step | Epoch | CCIP | AI Corrupt | Bikini Plus | Score | Download | pattern_0_0 | pattern_0_1 | pattern_0_2 | portrait_0 | portrait_1 | portrait_2 | full_body_0 | full_body_1 | profile_0 | profile_1 | free_0 | free_1 | shorts | maid_0 | maid_1 | miko | yukata | suit | china | bikini_0 | bikini_1 | bikini_2 | sit | squat | kneel | jump | crossed_arms | angry | smile | cry | grin | n_lie_0 | n_lie_1 | n_stand_0 | n_stand_1 | n_stand_2 | n_sex_0 | n_sex_1 | |-------:|--------:|:----------|:-------------|:--------------|:----------|:-----------------------------------------------------------------------------------------------|:---------------------------------------------|:---------------------------------------------|:---------------------------------------------|:-------------------------------------------|:-------------------------------------------|:-------------------------------------------|:---------------------------------------------|:---------------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-------------------------------|:-----------------------------------|:-------------------------------|:---------------------------------|:---------------------------------------|:---------------------------------------|:---------------------------------------|:-----------------------------|:---------------------------------|:---------------------------------|:-------------------------------|:-----------------------------------------------|:---------------------------------|:---------------------------------|:-----------------------------|:-------------------------------|:-------------------------------------|:-------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------------|:-------------------------------------|:-------------------------------------| | 480 | 33 | **0.864** | 0.964 | 0.863 | **0.778** | [Download](https://huggingface.co/CyberHarem/juno_azurlane/resolve/main/480/juno_azurlane.zip) | ![pattern_0_0](480/previews/pattern_0_0.png) | ![pattern_0_1](480/previews/pattern_0_1.png) | ![pattern_0_2](480/previews/pattern_0_2.png) | ![portrait_0](480/previews/portrait_0.png) | ![portrait_1](480/previews/portrait_1.png) | ![portrait_2](480/previews/portrait_2.png) | ![full_body_0](480/previews/full_body_0.png) | ![full_body_1](480/previews/full_body_1.png) | ![profile_0](480/previews/profile_0.png) | ![profile_1](480/previews/profile_1.png) | ![free_0](480/previews/free_0.png) | ![free_1](480/previews/free_1.png) | ![shorts](480/previews/shorts.png) | ![maid_0](480/previews/maid_0.png) | ![maid_1](480/previews/maid_1.png) | ![miko](480/previews/miko.png) | ![yukata](480/previews/yukata.png) | ![suit](480/previews/suit.png) | ![china](480/previews/china.png) | ![bikini_0](480/previews/bikini_0.png) | ![bikini_1](480/previews/bikini_1.png) | ![bikini_2](480/previews/bikini_2.png) | ![sit](480/previews/sit.png) | ![squat](480/previews/squat.png) | ![kneel](480/previews/kneel.png) | ![jump](480/previews/jump.png) | ![crossed_arms](480/previews/crossed_arms.png) | ![angry](480/previews/angry.png) | ![smile](480/previews/smile.png) | ![cry](480/previews/cry.png) | ![grin](480/previews/grin.png) | ![n_lie_0](480/previews/n_lie_0.png) | ![n_lie_1](480/previews/n_lie_1.png) | ![n_stand_0](480/previews/n_stand_0.png) | ![n_stand_1](480/previews/n_stand_1.png) | ![n_stand_2](480/previews/n_stand_2.png) | ![n_sex_0](480/previews/n_sex_0.png) | ![n_sex_1](480/previews/n_sex_1.png) | | 520 | 36 | 0.816 | **0.976** | 0.862 | 0.742 | [Download](https://huggingface.co/CyberHarem/juno_azurlane/resolve/main/520/juno_azurlane.zip) | ![pattern_0_0](520/previews/pattern_0_0.png) | ![pattern_0_1](520/previews/pattern_0_1.png) | ![pattern_0_2](520/previews/pattern_0_2.png) | ![portrait_0](520/previews/portrait_0.png) | ![portrait_1](520/previews/portrait_1.png) | ![portrait_2](520/previews/portrait_2.png) | ![full_body_0](520/previews/full_body_0.png) | ![full_body_1](520/previews/full_body_1.png) | ![profile_0](520/previews/profile_0.png) | ![profile_1](520/previews/profile_1.png) | ![free_0](520/previews/free_0.png) | ![free_1](520/previews/free_1.png) | ![shorts](520/previews/shorts.png) | ![maid_0](520/previews/maid_0.png) | ![maid_1](520/previews/maid_1.png) | ![miko](520/previews/miko.png) | ![yukata](520/previews/yukata.png) | ![suit](520/previews/suit.png) | ![china](520/previews/china.png) | ![bikini_0](520/previews/bikini_0.png) | ![bikini_1](520/previews/bikini_1.png) | ![bikini_2](520/previews/bikini_2.png) | ![sit](520/previews/sit.png) | ![squat](520/previews/squat.png) | ![kneel](520/previews/kneel.png) | ![jump](520/previews/jump.png) | ![crossed_arms](520/previews/crossed_arms.png) | ![angry](520/previews/angry.png) | ![smile](520/previews/smile.png) | ![cry](520/previews/cry.png) | ![grin](520/previews/grin.png) | ![n_lie_0](520/previews/n_lie_0.png) | ![n_lie_1](520/previews/n_lie_1.png) | ![n_stand_0](520/previews/n_stand_0.png) | ![n_stand_1](520/previews/n_stand_1.png) | ![n_stand_2](520/previews/n_stand_2.png) | ![n_sex_0](520/previews/n_sex_0.png) | ![n_sex_1](520/previews/n_sex_1.png) | | 540 | 37 | 0.815 | 0.972 | 0.857 | 0.735 | [Download](https://huggingface.co/CyberHarem/juno_azurlane/resolve/main/540/juno_azurlane.zip) | ![pattern_0_0](540/previews/pattern_0_0.png) | ![pattern_0_1](540/previews/pattern_0_1.png) | ![pattern_0_2](540/previews/pattern_0_2.png) | ![portrait_0](540/previews/portrait_0.png) | ![portrait_1](540/previews/portrait_1.png) | ![portrait_2](540/previews/portrait_2.png) | ![full_body_0](540/previews/full_body_0.png) | ![full_body_1](540/previews/full_body_1.png) | ![profile_0](540/previews/profile_0.png) | ![profile_1](540/previews/profile_1.png) | ![free_0](540/previews/free_0.png) | ![free_1](540/previews/free_1.png) | ![shorts](540/previews/shorts.png) | ![maid_0](540/previews/maid_0.png) | ![maid_1](540/previews/maid_1.png) | ![miko](540/previews/miko.png) | ![yukata](540/previews/yukata.png) | ![suit](540/previews/suit.png) | ![china](540/previews/china.png) | ![bikini_0](540/previews/bikini_0.png) | ![bikini_1](540/previews/bikini_1.png) | ![bikini_2](540/previews/bikini_2.png) | ![sit](540/previews/sit.png) | ![squat](540/previews/squat.png) | ![kneel](540/previews/kneel.png) | ![jump](540/previews/jump.png) | ![crossed_arms](540/previews/crossed_arms.png) | ![angry](540/previews/angry.png) | ![smile](540/previews/smile.png) | ![cry](540/previews/cry.png) | ![grin](540/previews/grin.png) | ![n_lie_0](540/previews/n_lie_0.png) | ![n_lie_1](540/previews/n_lie_1.png) | ![n_stand_0](540/previews/n_stand_0.png) | ![n_stand_1](540/previews/n_stand_1.png) | ![n_stand_2](540/previews/n_stand_2.png) | ![n_sex_0](540/previews/n_sex_0.png) | ![n_sex_1](540/previews/n_sex_1.png) | | 420 | 29 | 0.780 | 0.958 | **0.864** | 0.718 | [Download](https://huggingface.co/CyberHarem/juno_azurlane/resolve/main/420/juno_azurlane.zip) | ![pattern_0_0](420/previews/pattern_0_0.png) | ![pattern_0_1](420/previews/pattern_0_1.png) | ![pattern_0_2](420/previews/pattern_0_2.png) | ![portrait_0](420/previews/portrait_0.png) | ![portrait_1](420/previews/portrait_1.png) | ![portrait_2](420/previews/portrait_2.png) | ![full_body_0](420/previews/full_body_0.png) | ![full_body_1](420/previews/full_body_1.png) | ![profile_0](420/previews/profile_0.png) | ![profile_1](420/previews/profile_1.png) | ![free_0](420/previews/free_0.png) | ![free_1](420/previews/free_1.png) | ![shorts](420/previews/shorts.png) | ![maid_0](420/previews/maid_0.png) | ![maid_1](420/previews/maid_1.png) | ![miko](420/previews/miko.png) | ![yukata](420/previews/yukata.png) | ![suit](420/previews/suit.png) | ![china](420/previews/china.png) | ![bikini_0](420/previews/bikini_0.png) | ![bikini_1](420/previews/bikini_1.png) | ![bikini_2](420/previews/bikini_2.png) | ![sit](420/previews/sit.png) | ![squat](420/previews/squat.png) | ![kneel](420/previews/kneel.png) | ![jump](420/previews/jump.png) | ![crossed_arms](420/previews/crossed_arms.png) | ![angry](420/previews/angry.png) | ![smile](420/previews/smile.png) | ![cry](420/previews/cry.png) | ![grin](420/previews/grin.png) | ![n_lie_0](420/previews/n_lie_0.png) | ![n_lie_1](420/previews/n_lie_1.png) | ![n_stand_0](420/previews/n_stand_0.png) | ![n_stand_1](420/previews/n_stand_1.png) | ![n_stand_2](420/previews/n_stand_2.png) | ![n_sex_0](420/previews/n_sex_0.png) | ![n_sex_1](420/previews/n_sex_1.png) | | 700 | 48 | 0.783 | 0.961 | 0.856 | 0.710 | [Download](https://huggingface.co/CyberHarem/juno_azurlane/resolve/main/700/juno_azurlane.zip) | ![pattern_0_0](700/previews/pattern_0_0.png) | ![pattern_0_1](700/previews/pattern_0_1.png) | ![pattern_0_2](700/previews/pattern_0_2.png) | ![portrait_0](700/previews/portrait_0.png) | ![portrait_1](700/previews/portrait_1.png) | ![portrait_2](700/previews/portrait_2.png) | ![full_body_0](700/previews/full_body_0.png) | ![full_body_1](700/previews/full_body_1.png) | ![profile_0](700/previews/profile_0.png) | ![profile_1](700/previews/profile_1.png) | ![free_0](700/previews/free_0.png) | ![free_1](700/previews/free_1.png) | ![shorts](700/previews/shorts.png) | ![maid_0](700/previews/maid_0.png) | ![maid_1](700/previews/maid_1.png) | ![miko](700/previews/miko.png) | ![yukata](700/previews/yukata.png) | ![suit](700/previews/suit.png) | ![china](700/previews/china.png) | ![bikini_0](700/previews/bikini_0.png) | ![bikini_1](700/previews/bikini_1.png) | ![bikini_2](700/previews/bikini_2.png) | ![sit](700/previews/sit.png) | ![squat](700/previews/squat.png) | ![kneel](700/previews/kneel.png) | ![jump](700/previews/jump.png) | ![crossed_arms](700/previews/crossed_arms.png) | ![angry](700/previews/angry.png) | ![smile](700/previews/smile.png) | ![cry](700/previews/cry.png) | ![grin](700/previews/grin.png) | ![n_lie_0](700/previews/n_lie_0.png) | ![n_lie_1](700/previews/n_lie_1.png) | ![n_stand_0](700/previews/n_stand_0.png) | ![n_stand_1](700/previews/n_stand_1.png) | ![n_stand_2](700/previews/n_stand_2.png) | ![n_sex_0](700/previews/n_sex_0.png) | ![n_sex_1](700/previews/n_sex_1.png) | ## Anything Else? Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. ## All Steps We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * [Steps From 620 to 800](all/0.md) * [Steps From 420 to 600](all/1.md) * [Steps From 220 to 400](all/2.md) * [Steps From 20 to 200](all/3.md)
{"license": "mit", "tags": ["art", "not-for-all-audiences"], "datasets": ["CyberHarem/juno_azurlane"], "pipeline_tag": "text-to-image"}
text-to-image
CyberHarem/juno_azurlane
[ "art", "not-for-all-audiences", "text-to-image", "dataset:CyberHarem/juno_azurlane", "license:mit", "region:us" ]
2024-02-15T03:13:08+00:00
[]
[]
TAGS #art #not-for-all-audiences #text-to-image #dataset-CyberHarem/juno_azurlane #license-mit #region-us
Lora of juno/ジュノー/天后 (Azur Lane) ================================ What Is This? ------------- This is the LoRA model of waifu juno/ジュノー/天后 (Azur Lane). How Is It Trained? ------------------ * This model is trained with HCP-Diffusion. * The auto-training framework is maintained by DeepGHS Team. * The base model used for training is deepghs/animefull-latest. * Dataset used for training is the 'stage3-p480-800' in CyberHarem/juno\_azurlane, which contains 59 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 800 steps, 40 checkpoints were saved and evaluated. * Trigger word is 'juno\_azurlane'. * Pruned core tags for this waifu are 'pink\_hair, long\_hair, crown, bangs, mini\_crown, ribbon, twintails, pink\_eyes, bow, purple\_eyes'. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. How to Use It? -------------- ### If You Are Using A1111 WebUI v1.7+ Just use it like the classic LoRA. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 480, you need to download '480/juno\_azurlane.pt' as the embedding and '480/juno\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters. Which Step Should I Use? ------------------------ We selected 5 good steps for you to choose. The best one is step 480. 1520 images (1.59 GiB) were generated for auto-testing. !Metrics Plot The base model used for generating preview images is Meina/MeinaMix\_V11. Here are the preview of the recommended steps: Anything Else? -------------- Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. All Steps --------- We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * Steps From 620 to 800 * Steps From 420 to 600 * Steps From 220 to 400 * Steps From 20 to 200
[ "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 480, you need to download '480/juno\\_azurlane.pt' as the embedding and '480/juno\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 480.\n\n\n1520 images (1.59 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 620 to 800\n* Steps From 420 to 600\n* Steps From 220 to 400\n* Steps From 20 to 200" ]
[ "TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/juno_azurlane #license-mit #region-us \n", "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 480, you need to download '480/juno\\_azurlane.pt' as the embedding and '480/juno\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 480.\n\n\n1520 images (1.59 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 620 to 800\n* Steps From 420 to 600\n* Steps From 220 to 400\n* Steps From 20 to 200" ]
[ 44, 38, 465 ]
[ "passage: TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/juno_azurlane #license-mit #region-us \n### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file." ]
[ 0.010026790201663971, 0.013531080447137356, -0.0043538822792470455, 0.08454758673906326, 0.07959701865911484, 0.07892483472824097, 0.22509224712848663, 0.07900720089673996, 0.12188925594091415, -0.06773224472999573, 0.0890742763876915, 0.07040310651063919, -0.007279727142304182, 0.03276883065700531, -0.03429216146469116, -0.1476564258337021, -0.06647754460573196, -0.029525212943553925, -0.001554134301841259, 0.012416910380125046, 0.07481800019741058, 0.002788530196994543, 0.11049043387174606, -0.049691326916217804, -0.03528183698654175, 0.0598069466650486, -0.027787476778030396, -0.04879675433039665, 0.030614659190177917, 0.08306505531072617, 0.11867057532072067, 0.018644345924258232, 0.0642220601439476, -0.15421314537525177, 0.0680076852440834, -0.005700091831386089, -0.10827980190515518, -0.0008981709834188223, 0.026970000937581062, -0.04266854748129845, 0.11924438923597336, 0.02850973792374134, -0.1094331294298172, 0.04384881258010864, -0.1433405578136444, -0.01735999993979931, -0.05347299575805664, 0.03129515424370766, 0.14425650238990784, 0.04863365367054939, 0.025149362161755562, 0.055666517466306686, -0.04904117062687874, 0.07363047450780869, 0.11650466173887253, -0.13443073630332947, -0.06797204911708832, 0.09347433596849442, 0.014685267582535744, 0.13490505516529083, -0.08247190713882446, 0.09742391854524612, 0.0798969492316246, -0.05042298510670662, -0.14423108100891113, -0.0967773050069809, -0.2182650864124298, -0.00780947832390666, 0.018488101661205292, 0.01733970455825329, 0.4139375388622284, 0.06097254529595375, 0.04048171266913414, 0.06839203089475632, -0.07183048129081726, 0.028796756640076637, -0.09391450136899948, 0.13818779587745667, 0.049677636474370956, 0.09510461986064911, -0.039504967629909515, -0.10058122128248215, -0.11493045091629028, -0.0686783641576767, -0.0778748095035553, -0.02581154741346836, 0.021756114438176155, 0.11471707373857498, -0.19459423422813416, 0.00633764686062932, -0.04224354028701782, -0.13046835362911224, 0.030894579365849495, -0.10038162767887115, 0.18231892585754395, 0.06604573130607605, -0.013013049960136414, 0.004027980379760265, 0.24784651398658752, 0.11538992077112198, 0.18443092703819275, 0.04397092014551163, -0.0972549319267273, 0.1298188865184784, 0.04454624652862549, -0.09393933415412903, -0.01162861380726099, -0.09823617339134216, 0.1466289609670639, -0.05358599126338959, 0.1103256344795227, -0.05952898785471916, -0.11721347272396088, 0.023916274309158325, -0.10648804903030396, 0.06716351956129074, 0.041145000606775284, 0.004317757207900286, -0.05185356363654137, 0.041086774319410324, 0.03546391800045967, -0.033110007643699646, -0.006656778510659933, -0.00645798584446311, -0.0597408264875412, 0.04236467927694321, 0.1063031479716301, 0.03208812326192856, 0.0610135942697525, -0.005962630268186331, -0.026943085715174675, -0.00911224540323019, -0.05084679275751114, -0.0020405673421919346, 0.04854243993759155, 0.03256954252719879, 0.09014945477247238, -0.1553889811038971, -0.07225833088159561, -0.02064385451376438, 0.05593125894665718, 0.0016096344916149974, 0.08718930929899216, -0.013429876416921616, 0.06391361355781555, 0.0009000935242511332, -0.0245197881013155, 0.03090388886630535, -0.10563931614160538, 0.08721311390399933, -0.006878865882754326, 0.09625419229269028, -0.19612039625644684, -0.006397523917257786, -0.05044575035572052, 0.015057172626256943, 0.0652189552783966, -0.0016235826769843698, -0.10739900916814804, 0.12342239916324615, -0.01885630376636982, 0.0751161128282547, -0.10088243335485458, 0.05046459659934044, 0.018282007426023483, 0.08832613378763199, -0.0943698137998581, 0.0041581615805625916, 0.11980708688497543, -0.12982383370399475, -0.16067510843276978, 0.08168523013591766, -0.029190704226493835, 0.03633757680654526, 0.04796764627099037, 0.1494063436985016, 0.15861083567142487, -0.19298920035362244, -0.006070789415389299, 0.0510438047349453, -0.008782988414168358, -0.08126462250947952, -0.012364628724753857, 0.10377377271652222, 0.01716930791735649, 0.03150276094675064, -0.04354367032647133, 0.12453751266002655, -0.03261636942625046, -0.08214982599020004, -0.024393755942583084, -0.08039506524801254, -0.08049912750720978, 0.05193231254816055, -0.003995245788246393, -0.049740858376026154, 0.019566740840673447, -0.1514633148908615, 0.15513616800308228, 0.020279498770833015, 0.013535017147660255, -0.07473887503147125, 0.11683153361082077, -0.0009783790446817875, -0.0015166718512773514, 0.009953542612493038, -0.05226613208651543, -0.10713820904493332, 0.24497859179973602, 0.08750081807374954, 0.0911557599902153, 0.058914169669151306, -0.05353964865207672, -0.06465822458267212, 0.01647247187793255, 0.016460193321108818, -0.03307108208537102, 0.029198991134762764, -0.10796952992677689, 0.047716379165649414, -0.016421735286712646, 0.0353107713162899, -0.0025341014843434095, -0.030212940648198128, 0.06328044831752777, 0.014727907255291939, -0.016973329707980156, 0.09092125296592712, 0.05368401110172272, -0.01748639903962612, -0.06906875967979431, -0.0001619392423890531, 0.08105053752660751, -0.01018737442791462, -0.07337896525859833, 0.027002401649951935, -0.001011599670164287, 0.040527768433094025, 0.20097647607326508, -0.22509880363941193, 0.04290476068854332, 0.017983535304665565, 0.050093092024326324, 0.04070252180099487, -0.006453781854361296, -0.028239209204912186, 0.037784282118082047, -0.023303739726543427, 0.0774770975112915, -0.02247917465865612, 0.07249470055103302, -0.025365686044096947, -0.14166665077209473, -0.013397552073001862, -0.028051218017935753, 0.17543630301952362, -0.16595715284347534, 0.07054857164621353, 0.1742934137582779, -0.12454710155725479, 0.1360272467136383, 0.001584159443154931, -0.011951157823204994, 0.014330304227769375, 0.040752850472927094, 0.004209449980407953, 0.09833180904388428, -0.07309333980083466, -0.027049291878938675, 0.023450816050171852, -0.09402347356081009, 0.03646537661552429, -0.12425798922777176, -0.11076670140028, -0.06459712237119675, -0.03232020139694214, -0.03858952596783638, 0.020552555099129677, -0.05235997959971428, 0.0776660367846489, -0.08301776647567749, -0.08327989280223846, -0.024818897247314453, -0.08268840610980988, 0.021404171362519264, 0.010102522559463978, -0.05638326704502106, -0.1409795880317688, -0.12545934319496155, -0.08742093294858932, -0.14242297410964966, 0.007918985560536385, 0.06501949578523636, -0.11870569735765457, -0.039440594613552094, 0.02507016249001026, -0.048698507249355316, 0.10317052900791168, -0.0864049643278122, 0.01738450862467289, 0.05482076480984688, -0.03460245206952095, -0.16427376866340637, 0.00023392259026877582, -0.06349792331457138, -0.057433705776929855, 0.148419588804245, -0.1617821753025055, 0.1869257390499115, -0.025465477257966995, 0.055911317467689514, 0.0628192275762558, 0.03632283955812454, 0.13153395056724548, -0.10748550295829773, 0.07041347026824951, 0.1953832060098648, 0.04805642366409302, 0.0700136199593544, 0.11251478642225266, 0.0837385356426239, -0.11238422989845276, 0.03435971960425377, 0.07688696682453156, -0.09730416536331177, -0.08689136058092117, -0.05424318090081215, -0.11092496663331985, -0.06440785527229309, 0.0684283897280693, 0.05465379357337952, 0.05461302399635315, 0.1300162822008133, -0.055516041815280914, 0.005046885926276445, 0.09805014729499817, 0.047281984239816666, 0.08481255918741226, 0.014656233601272106, 0.052350860089063644, -0.1433085948228836, -0.059534698724746704, 0.16300798952579498, 0.22805504500865936, 0.2089245319366455, 0.027806168422102928, 0.07499413192272186, 0.11821408569812775, 0.07427435368299484, 0.08848800510168076, 0.05005848780274391, 0.005934832617640495, 0.02090349793434143, -0.07970084995031357, -0.050365936011075974, 0.0011595458490774035, -0.0013689746847376227, -0.04424886032938957, -0.14282655715942383, 0.09416777640581131, 0.0021327210124582052, 0.07888376712799072, 0.14404912292957306, 0.0407068096101284, -0.12001818418502808, 0.15751004219055176, 0.10314439982175827, 0.09449277073144913, -0.06488978862762451, 0.1296989470720291, 0.055408261716365814, -0.004225041717290878, 0.16472768783569336, 0.029226522892713547, 0.15411895513534546, -0.04014676809310913, -0.07573817670345306, -0.07258859276771545, -0.0537782721221447, 0.01319102756679058, 0.03148067370057106, -0.2105511575937271, 0.09851840138435364, 0.05963342636823654, 0.007914894260466099, -0.00568134430795908, -0.053546372801065445, 0.17471355199813843, 0.1572289764881134, 0.08472010493278503, 0.021796461194753647, -0.026519764214754105, -0.004591580014675856, -0.08733803033828735, 0.052534319460392, 0.004858703352510929, 0.059378091245889664, -0.03635517880320549, -0.09575144201517105, -0.02074512466788292, -0.0039005286525934935, 0.026727410033345222, -0.07858893275260925, -0.11162659525871277, -0.04866955056786537, 0.2491271048784256, -0.06196075305342674, 0.05380335450172424, 0.056101731956005096, 0.023860454559326172, -0.03205356374382973, 0.03948808088898659, -0.0276974905282259, -0.016717972233891487, -0.03176077455282211, -0.000006914485311426688, 0.011125706136226654, -0.05738624185323715, -0.05790203809738159, -0.0278446227312088, -0.10163857042789459, -0.10078230500221252, 0.01381196640431881, -0.049565043300390244, 0.01770196482539177, -0.03057118132710457, 0.013733063824474812, -0.10401063412427902, -0.033990710973739624, 0.027408946305513382, 0.02948223426938057, -0.07560823857784271, -0.12578028440475464, -0.0020553499925881624, -0.015252274461090565, -0.06281805783510208, 0.022286465391516685, -0.11760800331830978, -0.09388063102960587, -0.051196709275245667, -0.023385673761367798, 0.12862618267536163, 0.22161541879177094, -0.020989149808883667, 0.005750422365963459, 0.14845387637615204, -0.10566186159849167, -0.3091539442539215, -0.16330091655254364, -0.16460934281349182, -0.10349802672863007, 0.034204404801130295, -0.08121373504400253, 0.01744757406413555, 0.08088313043117523, -0.04149215295910835, 0.19891034066677094, -0.20275209844112396, -0.0911971777677536, 0.08162833750247955, 0.09432649612426758, 0.33259403705596924, -0.24692745506763458, 0.004514758475124836, -0.12114740908145905, -0.05418600142002106, 0.013348892331123352, -0.08341720700263977, 0.11660982668399811, 0.039993803948163986, 0.07623650878667831, -0.004174055065959692, -0.0023923262488096952, 0.14977915585041046, -0.08158310502767563, 0.13083156943321228, -0.11135643720626831, -0.1026635691523552, 0.19969631731510162, -0.034718435257673264, -0.003037534886971116, -0.20083586871623993, -0.0363846980035305, -0.04201946035027504, 0.035820718854665756, -0.009429063647985458, 0.05669715628027916, -0.0077074188739061356, -0.00652057770639658, -0.11876630038022995, -0.020991338416934013, -0.026947148144245148, 0.06174254044890404, 0.2338305115699768, -0.06882573664188385, -0.05292687937617302, 0.04998837411403656, -0.00254370691254735, 0.10615713894367218, 0.01602587103843689, -0.054578766226768494, -0.04196383059024811, 0.10028553009033203, -0.2090526670217514, 0.062301792204380035, 0.0045770034193992615, -0.006501107942312956, 0.010938138701021671, 0.010010323487222195, 0.020574573427438736, 0.12224839627742767, 0.18140624463558197, -0.011305641382932663, -0.017648618668317795, -0.021017590537667274, 0.024389296770095825, 0.13698118925094604, -0.02115660160779953, 0.11198479682207108, 0.022583087906241417, 0.028478847816586494, 0.01216037292033434, 0.05333033204078674, -0.07951702177524567, -0.08364089578390121, 0.09763399511575699, -0.052186910063028336, -0.08825526386499405, 0.0936364158987999, 0.05474637821316719, 0.0628877580165863, -0.0002936365781351924, 0.03801184892654419, 0.01994982920587063, -0.12198997288942337, 0.025408107787370682, 0.21416588127613068, -0.08136038482189178, -0.0647696927189827, -0.0810520350933075, 0.00991086196154356, -0.11871422082185745, 0.07394909858703613, 0.03303125128149986, -0.030913425609469414, 0.11893486231565475, -0.044095881283283234, -0.029327193275094032, 0.0010766304330900311, -0.057149488478899, 0.03612315654754639, -0.14539220929145813, -0.19895191490650177, 0.046987589448690414, -0.011744147166609764, -0.06621746718883514, -0.08645711839199066, -0.0939839705824852, 0.06556093692779541, -0.15706415474414825, 0.14348430931568146, -0.0652717724442482, 0.057201217859983444, -0.03881126642227173, -0.045870713889598846, -0.11237795650959015, -0.01735318824648857, -0.05749170854687691, -0.019788045436143875, 0.057520389556884766, 0.018903590738773346, -0.1169886365532875, -0.11112428456544876, 0.0573110468685627, -0.006434372626245022, -0.00953908171504736, 0.009361089207231998, -0.07061561942100525, 0.012915005907416344, -0.23499006032943726, -0.06917677819728851, 0.09307587146759033, 0.03742650896310806, -0.08243276923894882, 0.12367593497037888, 0.05342244729399681, -0.020215392112731934, 0.034241244196891785, -0.0007490265998058021, 0.17632152140140533, -0.07318571209907532, 0.027775723487138748, -0.11137915402650833, -0.16814333200454712, -0.024306826293468475, 0.036501988768577576, 0.24063195288181305, 0.07626350224018097, 0.11810390651226044, -0.05594533681869507, 0.023484259843826294, -0.01579345390200615, 0.07317563891410828, 0.011296149343252182, -0.1014108881354332, -0.07263538241386414, -0.16426962614059448, -0.0628989040851593, -0.06004444882273674, 0.16543325781822205, 0.04516726732254028, -0.12922045588493347, -0.004238903988152742, 0.11975903064012527, -0.17815172672271729, -0.010381218045949936, 0.1669190376996994, -0.045551080256700516, 0.029089339077472687, -0.1549077332019806, 0.034505631774663925, 0.08201102912425995, -0.02694765105843544, -0.004323452245444059, 0.12112501263618469, 0.008717822842299938, 0.0065869842655956745, 0.041883938014507294, -0.028457151725888252, 0.09477900713682175, -0.05434450879693031, 0.0675015076994896, 0.0005182363674975932, -0.04305556043982506, -0.12733951210975647, 0.18782569468021393, -0.013569561764597893, 0.009796311147511005, -0.06301875412464142, 0.008303241804242134, -0.10608360171318054, -0.09678197652101517, -0.07334375381469727, -0.14031416177749634, 0.07223954051733017, -0.06351637840270996, 0.012111101299524307, -0.015007412992417812, 0.012657162733376026, -0.07896869629621506, 0.0010760213481262326, -0.18407681584358215, -0.05931498855352402, 0.0156713780015707, -0.009842797182500362, -0.015346731059253216, -0.05649673566222191, -0.03305312618613243, 0.023874538019299507, -0.049892690032720566, -0.06647515296936035, 0.058129459619522095, 0.09906956553459167, 0.06009102985262871, -0.16191931068897247, -0.10492872446775436, -0.0705106109380722, 0.0330769419670105, 0.07077374309301376, 0.1885053664445877, 0.03203992918133736, -0.0055852592922747135, 0.04776448383927345, 0.13845902681350708, 0.013189554214477539, -0.0661550983786583, -0.07449419796466827, -0.12633264064788818, -0.13782505691051483, -0.00848128367215395, -0.060225702822208405, -0.023890001699328423, 0.021200008690357208, 0.23731648921966553, 0.19248834252357483, -0.15151871740818024, 0.03388307988643646, -0.07727275043725967, 0.04160230979323387, -0.02806459739804268, 0.15481173992156982, 0.043174490332603455, 0.14227396249771118, -0.033922094851732254, -0.035542502999305725, -0.06061072275042534, 0.021643565967679024, -0.096873939037323, 0.039013609290122986, -0.014046789146959782, -0.07003960758447647, -0.05887087062001228, 0.10724359005689621, -0.11745094507932663, 0.05782672390341759, 0.19635418057441711, -0.15403473377227783, -0.014933940954506397, -0.04489314556121826, 0.0546228289604187, 0.11285917460918427, 0.02243824489414692, -0.07481906563043594, -0.029938261955976486, 0.01338513195514679, 0.0310162752866745, -0.17578552663326263, -0.11830295622348785, -0.0017665487248450518, -0.13360871374607086, 0.13928677141666412, -0.00836989376693964, -0.0048921480774879456, 0.03389358147978783, -0.0649072527885437, -0.006791213992983103, 0.16464205086231232, 0.021016821265220642, -0.03478477895259857, -0.023562060669064522, -0.07711420953273773, -0.10387305170297623, 0.08068873733282089, 0.08729903399944305, 0.04200766235589981, -0.002032314194366336, 0.16429321467876434, -0.01970687136054039, -0.04249131307005882, 0.1388901174068451, -0.1740514189004898, 0.086015984416008, -0.0018585597863420844, -0.01963929831981659, -0.0715639591217041, -0.04422454163432121, 0.04663538560271263, 0.08553251624107361, -0.16058039665222168, -0.0507589727640152, 0.06249060854315758, -0.0976971909403801, 0.0562962107360363, 0.04299362376332283, -0.09651054441928864, 0.023791132494807243, -0.12226428836584091, -0.007632218301296234, -0.09881968051195145, 0.04562987759709358, 0.19875331223011017, -0.02633798122406006, 0.012535477988421917, -0.14109575748443604, 0.060957420617341995, -0.03143559396266937, -0.03703022003173828, -0.07221098244190216 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
automatic-speech-recognition
spsither/wav2vec2_run9.620
[ "transformers", "safetensors", "wav2vec2", "automatic-speech-recognition", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
2024-02-15T03:22:27+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #wav2vec2 #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 47, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.06877388060092926, 0.1546701192855835, -0.0037609888240695, 0.013798683881759644, 0.11170210689306259, 0.0049477447755634785, 0.07622946053743362, 0.1076156347990036, -0.024175573140382767, 0.12644733488559723, 0.04164152219891548, 0.09870775043964386, 0.11074616760015488, 0.18980292975902557, 0.0015578214079141617, -0.20271944999694824, 0.06667982041835785, -0.11557482928037643, 0.02210802026093006, 0.12125445902347565, 0.14131462574005127, -0.10717527568340302, 0.06805222481489182, -0.03453851491212845, -0.022604284808039665, -0.03256304934620857, -0.06200181692838669, -0.0628168061375618, 0.06936536729335785, 0.060818396508693695, 0.06474827229976654, 0.023958178237080574, 0.07868874818086624, -0.2985154092311859, 0.020363550633192062, 0.07747753709554672, 0.005190075840801001, 0.0596587099134922, 0.07716850191354752, -0.06847380846738815, 0.11357854306697845, -0.0553223080933094, 0.15529125928878784, 0.07729580253362656, -0.09200245141983032, -0.18732582032680511, -0.08171983063220978, 0.09086527675390244, 0.16344711184501648, 0.05807739868760109, -0.035454582422971725, 0.14257195591926575, -0.08119463175535202, 0.015228749252855778, 0.06432900577783585, -0.07448869198560715, -0.04995284602046013, 0.044303327798843384, 0.07393822818994522, 0.09027253836393356, -0.12936420738697052, -0.005840824451297522, 0.04285894334316254, 0.01751609519124031, 0.1045890524983406, 0.0271924901753664, 0.10937820374965668, 0.030452799052000046, -0.13982591032981873, -0.06308452039957047, 0.12294159829616547, 0.03608649969100952, -0.05978325754404068, -0.24299637973308563, -0.007494248915463686, -0.030862024053931236, -0.022421855479478836, -0.0449565127491951, 0.040200937539339066, -0.03043903410434723, 0.0803007185459137, 0.005218773614615202, -0.07346875220537186, -0.0566013865172863, 0.08528164029121399, 0.0660456046462059, 0.024965541437268257, -0.02511134371161461, 0.022877119481563568, 0.11602471768856049, 0.09200266003608704, -0.11191211640834808, -0.07020656764507294, -0.06118712201714516, -0.09110330045223236, -0.04440220445394516, 0.03338851034641266, 0.07138838618993759, 0.04954010248184204, 0.19076436758041382, 0.006971653085201979, 0.05134076997637749, 0.026316070929169655, 0.018496420234441757, 0.061533693224191666, 0.06859898567199707, -0.05315755307674408, -0.12085959315299988, -0.043275654315948486, 0.1195915937423706, 0.008576745167374611, -0.03422791138291359, -0.034871865063905716, 0.05920550227165222, 0.05124519392848015, 0.11922229826450348, 0.06299308687448502, 0.015805674716830254, -0.06944610923528671, -0.041848812252283096, 0.17807698249816895, -0.15696440637111664, 0.01886504516005516, 0.019594965502619743, -0.05179493874311447, -0.028022583574056625, 0.01927095092833042, 0.011918062344193459, -0.028684133663773537, 0.09848573058843613, -0.06384129822254181, -0.037289999425411224, -0.10494036227464676, -0.051826175302267075, 0.03436095267534256, -0.01885044015944004, -0.030469300225377083, -0.04276524484157562, -0.11668366193771362, -0.07342278957366943, 0.06446365267038345, -0.06070359796285629, -0.06312011927366257, -0.04004829749464989, -0.05974921956658363, 0.01184001937508583, -0.0018999426392838359, 0.12804386019706726, -0.03126852586865425, 0.04724927991628647, -0.05154479295015335, 0.07010733336210251, 0.13001501560211182, 0.0328618623316288, -0.06312436610460281, 0.06317896395921707, -0.20583610236644745, 0.10645388811826706, -0.0948607325553894, 0.026716187596321106, -0.16420963406562805, -0.024270139634609222, 0.02872021123766899, 0.03977278992533684, -0.014035328291356564, 0.13902691006660461, -0.1889396458864212, -0.037479519844055176, 0.1823769360780716, -0.1340419203042984, -0.09025664627552032, 0.06442771852016449, -0.056058306246995926, 0.1311984360218048, 0.051679398864507675, -0.016549112275242805, 0.050827931612730026, -0.14181455969810486, -0.021199021488428116, -0.05750836804509163, -0.01345672644674778, 0.14918801188468933, 0.06591099500656128, -0.060217004269361496, 0.03262941166758537, 0.02008114755153656, -0.02076314203441143, -0.052245598286390305, -0.03416990861296654, -0.09862805157899857, 0.003799794940277934, -0.08055862784385681, 0.018423959612846375, -0.026528598740696907, -0.08738208562135696, -0.0410190187394619, -0.1575777381658554, -0.001173238386400044, 0.1026405617594719, 0.0026203012093901634, -0.02646641992032528, -0.10305316001176834, 0.001408840762451291, 0.015838710591197014, -0.010245922021567822, -0.14677146077156067, -0.04217318072915077, 0.026863576844334602, -0.16719304025173187, 0.031281016767024994, -0.045817263424396515, 0.03617605194449425, 0.042714666575193405, -0.04341552406549454, -0.026187991723418236, 0.011214246973395348, 0.01926763355731964, -0.01759723760187626, -0.24584431946277618, -0.01623428985476494, -0.05088721215724945, 0.17665798962116241, -0.2476477026939392, 0.04387471452355385, 0.07402390241622925, 0.1185368224978447, 0.006659833248704672, -0.0473252609372139, 0.03859061002731323, -0.04956425726413727, -0.039547327905893326, -0.06162410229444504, -0.002731422893702984, -0.034249331802129745, -0.04925791174173355, 0.04766050726175308, -0.19274261593818665, -0.0254798773676157, 0.1145588755607605, 0.07196282595396042, -0.16417020559310913, -0.0721944123506546, -0.03388380631804466, -0.060263555496931076, -0.0855790227651596, -0.05511211231350899, 0.10627889633178711, 0.042532145977020264, 0.053568705916404724, -0.07193132489919662, -0.0538090355694294, 0.014475145377218723, -0.008023109287023544, -0.03674730286002159, 0.08616615831851959, 0.07892905920743942, -0.111492820084095, 0.0967666357755661, 0.06781410425901413, 0.06170906499028206, 0.10836543887853622, 0.0035758649464696646, -0.09838994592428207, -0.013410377316176891, 0.028753211721777916, 0.013008177280426025, 0.1445195972919464, -0.08268706500530243, 0.02993486076593399, 0.04475158452987671, -0.029572229832410812, 0.014260980300605297, -0.10948343575000763, 0.020612964406609535, 0.03188888356089592, -0.01410164125263691, 0.016051514074206352, -0.05129382014274597, 0.013738108798861504, 0.10363461822271347, 0.031123731285333633, 0.025897923856973648, 0.016665659844875336, -0.04273077845573425, -0.12888197600841522, 0.17441782355308533, -0.09573886543512344, -0.24906472861766815, -0.13649064302444458, 0.0033230632543563843, 0.04450872540473938, -0.01420661062002182, 0.019941311329603195, -0.06085766479372978, -0.10865217447280884, -0.10793688893318176, 0.02346382476389408, 0.04952440410852432, -0.08567548543214798, -0.05095811188220978, 0.05441328510642052, 0.03898037597537041, -0.12600500881671906, 0.024548007175326347, 0.04095667228102684, -0.07147589325904846, 0.005656755063682795, 0.061115942895412445, 0.08382482826709747, 0.1812773495912552, 0.012779363431036472, -0.015533777885138988, 0.01035984791815281, 0.21022020280361176, -0.14754468202590942, 0.08923394232988358, 0.142924964427948, -0.06379926204681396, 0.07994367927312851, 0.20067699253559113, 0.030222468078136444, -0.0959763154387474, 0.0354040265083313, 0.03157598897814751, -0.03929230570793152, -0.24485765397548676, -0.07799134403467178, 0.004727535881102085, -0.06941798329353333, 0.0999692752957344, 0.08970286697149277, 0.11357339471578598, 0.04878859966993332, -0.10688808560371399, -0.07536104321479797, 0.04997042194008827, 0.11770502477884293, -0.025654911994934082, 0.0004288276832085103, 0.09490229189395905, -0.032173965126276016, 0.024045821279287338, 0.09091470390558243, 0.01785297878086567, 0.1891387403011322, 0.045389045029878616, 0.13416282832622528, 0.08966030925512314, 0.05892613157629967, 0.02283613197505474, 0.020396918058395386, 0.022836502641439438, 0.028627371415495872, -0.02071341499686241, -0.08800762891769409, -0.01406664215028286, 0.1445012241601944, 0.03501417487859726, 0.03224355727434158, 0.005818283185362816, -0.03822546452283859, 0.07026989012956619, 0.16923215985298157, 0.01291902456432581, -0.22557523846626282, -0.06553208827972412, 0.07285686582326889, -0.07819344103336334, -0.10939628630876541, -0.00628721434623003, 0.039236925542354584, -0.1781243532896042, 0.0453440323472023, -0.016895415261387825, 0.09935811161994934, -0.11019659787416458, -0.022818224504590034, 0.03339223191142082, 0.06351818144321442, -0.033710017800331116, 0.07605454325675964, -0.20844414830207825, 0.14833855628967285, 0.007355031557381153, 0.06984888762235641, -0.10627210140228271, 0.07959222793579102, 0.018262188881635666, 0.0005360859213396907, 0.16532482206821442, -0.0075689139775931835, -0.07650822401046753, -0.08155251294374466, -0.07923656702041626, -0.010918287560343742, 0.10160883516073227, -0.10205793380737305, 0.08789419382810593, -0.006757213734090328, -0.030893130227923393, -0.00026032759342342615, -0.11519953608512878, -0.1342930644750595, -0.18055365979671478, 0.04992220178246498, -0.10558607429265976, 0.04552379995584488, -0.11181014776229858, -0.062069665640592575, -0.04111560434103012, 0.18840233981609344, -0.20550832152366638, -0.07671810686588287, -0.14316488802433014, -0.08166468888521194, 0.11773297190666199, -0.036535169929265976, 0.08007847517728806, 0.008441719226539135, 0.20702308416366577, -0.00666013965383172, 0.002528243465349078, 0.08686443418264389, -0.09668374806642532, -0.2072489857673645, -0.09340810775756836, 0.14340825378894806, 0.12398830056190491, 0.045563604682683945, -0.0001787850633263588, 0.021285003051161766, -0.004406071733683348, -0.11160994321107864, 0.036765191704034805, 0.1599014699459076, 0.08414851129055023, 0.041826896369457245, -0.023910723626613617, -0.15188267827033997, -0.1039518192410469, -0.06143968924880028, 0.022748636081814766, 0.18740743398666382, -0.06844107806682587, 0.17012163996696472, 0.157639279961586, -0.061386726796627045, -0.20854754745960236, 0.031976643949747086, 0.03363525867462158, -0.008795025758445263, 0.0332365483045578, -0.20113597810268402, 0.06802120804786682, 0.01531505398452282, -0.057996444404125214, 0.1332528293132782, -0.16826434433460236, -0.15160627663135529, 0.08843177556991577, 0.07692008465528488, -0.20126505196094513, -0.12921905517578125, -0.09711465984582901, -0.05218008533120155, -0.10807206481695175, 0.08772927522659302, -0.006655422504991293, 0.007214459590613842, 0.037578340619802475, 0.02635364979505539, 0.015357093885540962, -0.05328182876110077, 0.19721722602844238, 0.0011987579055130482, 0.044046565890312195, -0.07511261850595474, -0.077226422727108, 0.034381043165922165, -0.06312628090381622, 0.07982822507619858, -0.020660031586885452, 0.0017429457511752844, -0.11481664329767227, -0.06663372367620468, -0.05009456351399422, 0.029989875853061676, -0.08466581255197525, -0.09467059373855591, -0.051657307893037796, 0.09798348695039749, 0.09048279374837875, -0.03396918624639511, -0.06807554513216019, -0.10042613744735718, 0.06601390987634659, 0.22872091829776764, 0.18910692632198334, 0.06991440057754517, -0.06895517557859421, -0.0038870053831487894, -0.026509825140237808, 0.05879383906722069, -0.20851773023605347, 0.044600993394851685, 0.036500073969364166, 0.032537586987018585, 0.13215065002441406, -0.02442602440714836, -0.16357013583183289, -0.043075863271951675, 0.056227099150419235, -0.06633396446704865, -0.16863006353378296, 0.005107434932142496, 0.09075167030096054, -0.15091724693775177, -0.04752274975180626, 0.030901111662387848, -0.03220430761575699, -0.02397167682647705, 0.00030637482996098697, 0.08078145235776901, 0.020850084722042084, 0.1107739508152008, 0.06640642136335373, 0.11335843801498413, -0.10278842598199844, 0.08162284642457962, 0.08386309444904327, -0.11347422748804092, 0.04244251549243927, 0.05978094041347504, -0.06325716525316238, -0.03386267274618149, 0.016484335064888, 0.0787876546382904, 0.03214597329497337, -0.08122093230485916, 0.0026990212500095367, -0.11556044965982437, 0.06788678467273712, 0.14209748804569244, 0.03322440758347511, 0.007564007304608822, 0.04558844491839409, 0.031089849770069122, -0.09967122226953506, 0.10952559113502502, 0.0327114500105381, 0.03264835476875305, -0.052766215056180954, 0.007493352517485619, 0.044093240052461624, -0.012370331212878227, -0.01659340038895607, -0.04159332811832428, -0.062125492841005325, -0.004501889459788799, -0.15752804279327393, 0.029296958819031715, -0.06990371644496918, 0.009181820787489414, 0.0195058211684227, -0.03118128329515457, 0.001035416848026216, 0.014971627853810787, -0.0777391716837883, -0.03601877763867378, -0.00462498189881444, 0.10573451966047287, -0.15904870629310608, 0.012398114427924156, 0.0838126391172409, -0.12594857811927795, 0.0813586562871933, -0.0006106876535341144, -0.01206875778734684, 0.022131776437163353, -0.14767099916934967, 0.06096983700990677, -0.00651735020801425, 0.005330943502485752, 0.022080490365624428, -0.20231451094150543, 0.0010611782781779766, -0.046166326850652695, -0.0580565482378006, -0.006821162533015013, -0.034208331257104874, -0.10881488770246506, 0.10119375586509705, 0.01840946450829506, -0.0807829275727272, -0.019118202850222588, 0.049314580857753754, 0.10984907299280167, -0.05423201248049736, 0.13843025267124176, -0.022093484178185463, 0.05561875179409981, -0.17508383095264435, -0.015010466799139977, -0.01884511485695839, 0.01675039529800415, -0.032699406147003174, -0.0063448576256632805, 0.053761400282382965, -0.021795762702822685, 0.23006084561347961, -0.03329315781593323, 0.022746775299310684, 0.0662616565823555, -0.007395898457616568, -0.02466614730656147, 0.09141410142183304, 0.05831921473145485, 0.019823938608169556, 0.023462723940610886, 0.009678727947175503, -0.051977336406707764, -0.011846045032143593, -0.1287335902452469, 0.08032830059528351, 0.17006289958953857, 0.0832807645201683, -0.0011417492059990764, 0.05661620944738388, -0.11824764311313629, -0.08884397894144058, 0.10315068811178207, -0.03696487843990326, -0.008325101807713509, -0.05479050800204277, 0.14003127813339233, 0.16284166276454926, -0.1792466789484024, 0.06529472023248672, -0.06703231483697891, -0.054111137986183167, -0.1079135313630104, -0.1702733039855957, -0.06385406106710434, -0.04134172946214676, -0.003200325183570385, -0.056672241538763046, 0.07026970386505127, 0.10425727069377899, 0.015394158661365509, 0.007145122159272432, 0.08924684673547745, -0.034410521388053894, 0.003967431839555502, 0.04615078866481781, 0.05031316727399826, 0.015370454639196396, -0.06289559602737427, 0.003805057378485799, 0.012086667120456696, 0.03619912639260292, 0.05767577514052391, 0.03358588367700577, -0.015441972762346268, 0.00826429296284914, -0.019517268985509872, -0.0962890237569809, 0.0407244898378849, -0.028659315779805183, -0.04762914776802063, 0.14599058032035828, 0.023316938430070877, -0.005744231399148703, -0.019850272685289383, 0.22833019495010376, -0.06841307878494263, -0.08293036371469498, -0.13890130817890167, 0.1406106948852539, -0.04129096865653992, 0.054532211273908615, 0.048289187252521515, -0.10287833213806152, 0.031274814158678055, 0.14709845185279846, 0.14302049577236176, -0.028337303549051285, 0.01196619775146246, 0.009999874047935009, 0.005250520538538694, -0.026724260300397873, 0.052909236401319504, 0.049603480845689774, 0.12155342847108841, -0.06124946475028992, 0.09144628793001175, -0.0038096080534160137, -0.08695073425769806, -0.01940424181520939, 0.13583695888519287, -0.001434069243259728, 0.020704632624983788, -0.08129720389842987, 0.11675985902547836, -0.06527755409479141, -0.2561015188694, 0.060353249311447144, -0.06762448698282242, -0.14944049715995789, -0.018578823655843735, 0.027211744338274002, 0.0003355915832798928, 0.021279368549585342, 0.06146527826786041, -0.06275594234466553, 0.15064457058906555, 0.03758588433265686, -0.07729688286781311, -0.07095571607351303, 0.07545747607946396, -0.0798204317688942, 0.2952599823474884, 0.007051850203424692, 0.05692324787378311, 0.09223286807537079, -0.033274851739406586, -0.1323377937078476, 0.049896061420440674, 0.09064158797264099, -0.06194010376930237, 0.06410481035709381, 0.20840007066726685, -0.011975160799920559, 0.12260035425424576, 0.07416624575853348, -0.08735647797584534, 0.05223854258656502, -0.07405798882246017, -0.09430453926324844, -0.08655916899442673, 0.08934324234724045, -0.06278510391712189, 0.15317323803901672, 0.12562185525894165, -0.04725475609302521, 0.0027636797167360783, -0.025733815506100655, 0.054841578006744385, -0.0038393251597881317, 0.11300427466630936, 0.026762498542666435, -0.19724777340888977, 0.03347480297088623, -0.01826278306543827, 0.10099007189273834, -0.2592698633670807, -0.08135145157575607, 0.039587851613759995, -0.009570525959134102, -0.05378785356879234, 0.11855222284793854, 0.06144152209162712, 0.04968099668622017, -0.0558135025203228, -0.05388732627034187, 0.0009833982912823558, 0.1646765172481537, -0.10682281851768494, -0.0031281758565455675 ]
null
null
peft
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.8.2
{"library_name": "peft", "base_model": "microsoft/Orca-2-7b"}
null
nicejames/orca-2-7B-v01-fine-tuned-using-ludwig-4bit
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:microsoft/Orca-2-7b", "region:us" ]
2024-02-15T03:22:56+00:00
[ "1910.09700" ]
[]
TAGS #peft #safetensors #arxiv-1910.09700 #base_model-microsoft/Orca-2-7b #region-us
# Model Card for Model ID ## Model Details ### Model Description - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact ### Framework versions - PEFT 0.8.2
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ "TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-microsoft/Orca-2-7b #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ 35, 6, 3, 54, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4, 11 ]
[ "passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-microsoft/Orca-2-7b #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2" ]
[ -0.10687855631113052, 0.1982012689113617, -0.0038277041167020798, 0.037026334553956985, 0.0878944993019104, 0.018719671294093132, 0.055328574031591415, 0.12798389792442322, -0.038458164781332016, 0.1106455847620964, 0.07358640432357788, 0.10911573469638824, 0.1047355979681015, 0.20201736688613892, -0.0012977122096344829, -0.19787117838859558, 0.0218181349337101, -0.09631700813770294, -0.0019713628571480513, 0.1251724362373352, 0.14370658993721008, -0.09507370740175247, 0.07201100885868073, -0.023878522217273712, -0.01014705840498209, -0.03828539326786995, -0.07100825756788254, -0.03315627947449684, 0.034338727593421936, 0.05058370903134346, 0.06214354187250137, 0.00008773159788688645, 0.0867680236697197, -0.2657345235347748, 0.017779596149921417, 0.03728719428181648, -0.0018732281168922782, 0.08286840468645096, 0.09712915867567062, -0.046539127826690674, 0.12224840372800827, -0.0389997772872448, 0.13750259578227997, 0.07946313172578812, -0.10226511210203171, -0.22182707488536835, -0.06950899213552475, 0.07886537164449692, 0.16667938232421875, 0.07997101545333862, -0.04381309822201729, 0.14129428565502167, -0.10247156023979187, 0.0170722808688879, 0.037182364612817764, -0.07430697232484818, -0.07602965086698532, 0.06468889117240906, 0.11612676084041595, 0.0672411099076271, -0.14481611549854279, -0.03919844329357147, 0.02197086624801159, 0.03714854270219803, 0.07630181312561035, 0.022646985948085785, 0.14712248742580414, 0.03253014013171196, -0.14529123902320862, -0.045280590653419495, 0.12072110176086426, 0.041110262274742126, -0.035029247403144836, -0.2198582887649536, 0.010002875700592995, -0.09748525172472, -0.017697086557745934, -0.0472298301756382, 0.04064169153571129, -0.0021193516440689564, 0.09096667915582657, -0.026321446523070335, -0.1009623259305954, -0.008636095561087132, 0.08319967240095139, 0.05215226486325264, 0.014690390788018703, -0.026229845359921455, 0.007808137219399214, 0.11319270730018616, 0.06240789219737053, -0.12412475049495697, -0.06954283267259598, -0.0668473169207573, -0.03542141243815422, -0.04755961522459984, 0.029661819338798523, 0.039465103298425674, 0.057959090918302536, 0.23753789067268372, -0.016918392851948738, 0.049930304288864136, 0.057919736951589584, 0.026986850425601006, 0.03915201500058174, 0.09589415043592453, -0.0474887415766716, -0.13498547673225403, -0.016810936853289604, 0.09620869904756546, -0.011447805911302567, -0.021781394258141518, -0.050786688923835754, 0.035229239612817764, 0.027858564630150795, 0.11572803556919098, 0.09974590688943863, -0.014473726972937584, -0.08236981183290482, -0.0493394136428833, 0.21526913344860077, -0.1433643251657486, 0.049200672656297684, 0.025389716029167175, -0.014677456580102444, -0.04630810394883156, 0.006667213048785925, 0.020662706345319748, -0.023029709234833717, 0.09389685094356537, -0.060075875371694565, -0.03293721005320549, -0.11266091465950012, -0.011248263530433178, 0.034951549023389816, 0.019597336649894714, -0.022973114624619484, -0.04146966338157654, -0.058546457439661026, -0.09529439359903336, 0.1052720695734024, -0.07286257296800613, -0.06009887158870697, -0.03222799673676491, -0.10978727787733078, 0.016675056889653206, 0.025776894763112068, 0.10555189847946167, -0.024173451587557793, 0.0434199795126915, -0.011914066970348358, 0.0614079087972641, 0.07030993700027466, 0.036198340356349945, -0.07532789558172226, 0.062480468302965164, -0.18767450749874115, 0.09565505385398865, -0.07579248398542404, 0.031135329976677895, -0.15042825043201447, -0.01134167704731226, 0.0024753431789577007, 0.01569943316280842, 0.033345356583595276, 0.1555810123682022, -0.196992889046669, -0.029536720365285873, 0.15525338053703308, -0.09637086093425751, -0.11616922169923782, 0.037844546139240265, -0.0489269383251667, 0.1652156114578247, 0.010320560075342655, -0.0047587258741259575, 0.08439058810472488, -0.14843007922172546, -0.020463721826672554, -0.03491990268230438, -0.0005775797180831432, 0.10083235800266266, 0.09356941282749176, -0.0733669251203537, 0.039875805377960205, 0.014451298862695694, -0.03981127589941025, -0.032782625406980515, -0.054608989506959915, -0.11636707931756973, -0.002986728912219405, -0.08334885537624359, 0.025352245196700096, -0.015028567053377628, -0.08064892143011093, -0.005699039902538061, -0.16361688077449799, -0.031787414103746414, 0.07533062249422073, 0.01368411909788847, -0.016778437420725822, -0.09571249037981033, 0.029027560725808144, -0.038316112011671066, -0.029020072892308235, -0.1601903736591339, -0.017412006855010986, 0.018055781722068787, -0.1476333737373352, 0.01802913472056389, -0.10407460480928421, 0.07474282383918762, 0.014467090368270874, -0.06681551039218903, -0.03667111694812775, -0.012186716310679913, 0.017174087464809418, -0.05442817881703377, -0.2340492606163025, -0.01856367476284504, -0.05891209468245506, 0.1620403230190277, -0.23445309698581696, 0.03435438498854637, 0.04121793061494827, 0.12960128486156464, 0.01363993901759386, -0.059074461460113525, 0.023046176880598068, -0.06478417664766312, -0.024662740528583527, -0.07244608551263809, -0.004635788034647703, -0.0032158768735826015, -0.04291297122836113, 0.017405064776539803, -0.11658445745706558, -0.04775961861014366, 0.1063615083694458, 0.08896128833293915, -0.1677372008562088, -0.020882995799183846, -0.04560055956244469, -0.06872144341468811, -0.0724637508392334, -0.0586540549993515, 0.0975876897573471, 0.0526837557554245, 0.027212083339691162, -0.07351625710725784, -0.06089251860976219, 0.009259404614567757, -0.028571421280503273, -0.027641961351037025, 0.11431065201759338, 0.06993219256401062, -0.11019627004861832, 0.09658503532409668, 0.08590199798345566, 0.020662225782871246, 0.07490333914756775, -0.021306008100509644, -0.1096348762512207, -0.035759810358285904, 0.04052812606096268, 0.01994824968278408, 0.15058055520057678, -0.0776485800743103, 0.06401316076517105, 0.040735647082328796, -0.029119141399860382, 0.046763285994529724, -0.09578729420900345, 0.007324716076254845, 0.007560563739389181, -0.017086761072278023, 0.0146686602383852, -0.02835634909570217, 0.008230018429458141, 0.08366572111845016, 0.05051590874791145, 0.03245481848716736, 0.04433904215693474, -0.024814974516630173, -0.128061905503273, 0.174525648355484, -0.09249167144298553, -0.24483390152454376, -0.1687208116054535, 0.05446386709809303, 0.05355731397867203, -0.022206630557775497, 0.03343435376882553, -0.04412432387471199, -0.09971024096012115, -0.083454929292202, 0.012708136811852455, 0.03737049177289009, -0.06629430502653122, -0.0748123973608017, 0.06963611394166946, 0.04253761097788811, -0.1077754944562912, 0.03526423126459122, 0.06006213650107384, -0.018434016034007072, 0.016616739332675934, 0.045895643532276154, 0.08737726509571075, 0.176386296749115, -0.010704546235501766, -0.009481181390583515, 0.0588473379611969, 0.26883047819137573, -0.1531955450773239, 0.11137170344591141, 0.11936749517917633, -0.07167188823223114, 0.07623914629220963, 0.19538399577140808, 0.031949546188116074, -0.101502425968647, 0.030483348295092583, 0.033989254385232925, -0.018575655296444893, -0.2784092426300049, -0.043095748871564865, -0.01590108871459961, -0.09942476451396942, 0.07584627717733383, 0.08889169245958328, 0.09038715809583664, 0.037056948989629745, -0.054756004363298416, -0.11800802499055862, 0.03147420287132263, 0.10293892025947571, -0.03585941344499588, 0.006103770341724157, 0.0815812200307846, -0.02625410631299019, 0.015206865034997463, 0.09810247272253036, -0.017946848645806313, 0.1709204912185669, 0.07092803716659546, 0.10494667291641235, 0.07411918044090271, 0.08130185306072235, -0.0002411148598184809, 0.017286909744143486, 0.03412409871816635, 0.016238827258348465, 0.016395872458815575, -0.08328196406364441, 0.03032505512237549, 0.1150730550289154, 0.04692104086279869, 0.03319405019283295, 0.01209926512092352, -0.04681574925780296, 0.05396311357617378, 0.19191035628318787, 0.019209614023566246, -0.20965899527072906, -0.07841207832098007, 0.05271046236157417, -0.08278649300336838, -0.1364440619945526, -0.020120752975344658, 0.02075284905731678, -0.16634146869182587, 0.0070562600158154964, -0.03974159434437752, 0.10356786847114563, -0.08129218220710754, -0.0433017797768116, 0.07948260754346848, 0.07046211510896683, -0.023290392011404037, 0.06500200182199478, -0.20895510911941528, 0.1351221352815628, 0.012863919138908386, 0.07596515119075775, -0.08620038628578186, 0.0920979231595993, -0.005015313159674406, -0.005275916773825884, 0.1742461621761322, -0.0004170782631263137, -0.07611178606748581, -0.056245654821395874, -0.0935971736907959, -0.008577565662562847, 0.09796394407749176, -0.13084003329277039, 0.06926705688238144, -0.01833062246441841, -0.036832746118307114, 0.0021563288755714893, -0.07296339422464371, -0.1126292422413826, -0.16454492509365082, 0.0486929826438427, -0.09750043600797653, 0.03523535281419754, -0.090256467461586, -0.05909505486488342, 0.01208637934178114, 0.17837080359458923, -0.18054284155368805, -0.08250012993812561, -0.13690082728862762, -0.0901012197136879, 0.16985757648944855, -0.042849887162446976, 0.07814905792474747, -0.0016773788956925273, 0.17477703094482422, 0.024363115429878235, 0.0006035502883605659, 0.09731742739677429, -0.08712022751569748, -0.19079549610614777, -0.05893193185329437, 0.155519500374794, 0.12855882942676544, 0.03713332861661911, -0.009027170017361641, 0.02839798666536808, -0.05885403975844383, -0.10793828964233398, 0.031201789155602455, 0.1251198649406433, 0.07631176710128784, -0.014902212657034397, -0.03737245872616768, -0.11141790449619293, -0.06941290944814682, -0.0625859946012497, 0.0037012258544564247, 0.20623891055583954, -0.07309786975383759, 0.1596554070711136, 0.12682048976421356, -0.057366326451301575, -0.20139813423156738, 0.04301499202847481, 0.057113248854875565, 0.019144967198371887, 0.038629624992609024, -0.18297402560710907, 0.09416980296373367, 0.012581863440573215, -0.06730658560991287, 0.1412292718887329, -0.16572219133377075, -0.14519765973091125, 0.09124842286109924, 0.04130068048834801, -0.23704025149345398, -0.14104324579238892, -0.09665405750274658, -0.023479215800762177, -0.11122787743806839, 0.07110898196697235, 0.019713321700692177, 0.022351432591676712, 0.030462022870779037, 0.020863203331828117, 0.02713906392455101, -0.048105914145708084, 0.2245209813117981, -0.03803945705294609, 0.0060460749082267284, -0.05343335494399071, -0.10280017554759979, 0.03818415477871895, -0.04816779866814613, 0.09046132117509842, 0.002508975565433502, 0.026706453412771225, -0.13687027990818024, -0.04835830628871918, -0.05830143764615059, 0.022264918312430382, -0.0926186740398407, -0.0854908898472786, -0.033047616481781006, 0.10693074017763138, 0.08932284265756607, -0.02866450324654579, 0.002819797722622752, -0.10257783532142639, 0.06773577630519867, 0.2047501802444458, 0.18752846121788025, 0.05547831580042839, -0.06354036927223206, 0.011546345427632332, -0.036988113075494766, 0.03575422987341881, -0.21607784926891327, 0.04090657830238342, 0.06107569858431816, 0.02354058250784874, 0.08422750979661942, -0.012783624231815338, -0.13903653621673584, -0.06905630975961685, 0.07704705744981766, -0.03622017428278923, -0.15885043144226074, -0.014436904340982437, 0.04779858887195587, -0.21719036996364594, -0.052144523710012436, 0.01913273148238659, -0.019170334562659264, -0.040272437036037445, 0.03069329634308815, 0.07593740522861481, -0.03519800305366516, 0.10124369710683823, 0.08826375752687454, 0.09359744936227798, -0.09743709117174149, 0.07364672422409058, 0.08741246163845062, -0.04279070720076561, 0.013809796422719955, 0.10519561171531677, -0.04319550096988678, -0.033253129571676254, 0.0873754695057869, 0.09659513831138611, 0.019760170951485634, -0.05578383803367615, 0.015712495893239975, -0.04998503997921944, 0.0626085177063942, 0.10506176203489304, 0.03442595526576042, -0.00717223109677434, 0.051657821983098984, 0.033593278378248215, -0.10488200187683105, 0.10789581388235092, 0.05922168120741844, 0.022009460255503654, -0.04808250814676285, -0.03488989546895027, -0.013845025561749935, -0.01112434733659029, -0.017359765246510506, -0.0033669318072497845, -0.09296518564224243, -0.004891127813607454, -0.09724052250385284, 0.040476419031620026, -0.06939581781625748, 0.015653548762202263, 0.03413194790482521, -0.06178456172347069, 0.003768151393160224, 0.0003026011399924755, -0.07370556890964508, -0.051456019282341, -0.01822877675294876, 0.07888293266296387, -0.13191448152065277, 0.021895745769143105, 0.07591346651315689, -0.10823783278465271, 0.07187820971012115, 0.015600395388901234, 0.0026982519775629044, 0.004072374198585749, -0.1737859547138214, 0.061134278774261475, -0.018267905339598656, -0.006444934289902449, 0.011439919471740723, -0.2101336419582367, -0.015407958999276161, -0.0371420681476593, -0.052147507667541504, 0.012997074984014034, -0.030583731830120087, -0.1280163675546646, 0.08893552422523499, -0.009944038465619087, -0.08770205080509186, -0.009924951009452343, 0.041970498859882355, 0.09422961622476578, -0.02449570596218109, 0.12902791798114777, -0.018262775614857674, 0.08166035264730453, -0.1681193858385086, -0.005393511150032282, -0.01741906814277172, 0.034826405346393585, -0.015552517957985401, -0.027861367911100388, 0.06020686402916908, -0.011155433021485806, 0.17581988871097565, -0.016375429928302765, 0.07614202052354813, 0.05552144721150398, -0.004941405262798071, 0.020591458305716515, 0.08399660885334015, 0.06076737865805626, -0.001946604112163186, -0.0035083522088825703, 0.04447783902287483, -0.009573666378855705, -0.050924211740493774, -0.1558661162853241, 0.08873379975557327, 0.15938962996006012, 0.055821504443883896, 0.0195966474711895, 0.03721106797456741, -0.10392233729362488, -0.0711476281285286, 0.13523845374584198, -0.003995783627033234, -0.040242888033390045, -0.07504373043775558, 0.17023102939128876, 0.13382194936275482, -0.1995449662208557, 0.07646548748016357, -0.06628043204545975, -0.051886845380067825, -0.12384674698114395, -0.1537577211856842, -0.07164572924375534, -0.04075117036700249, -0.03024279698729515, -0.05589529871940613, 0.04211259260773659, 0.05365516245365143, 0.0074208322912454605, -0.019279133528470993, 0.10912252962589264, 0.028156571090221405, -0.01571846567094326, 0.0458945706486702, 0.05214013159275055, 0.0345955453813076, -0.1067180335521698, 0.011086443439126015, 0.004104801919311285, 0.02522752247750759, 0.05800772085785866, 0.030190015211701393, -0.055355727672576904, 0.009119541384279728, -0.025866175070405006, -0.121585913002491, 0.035545703023672104, -0.015572297386825085, -0.043816253542900085, 0.14871302247047424, 0.03601183369755745, 0.01761755906045437, -0.014631764031946659, 0.23607689142227173, -0.07017694413661957, -0.0811753049492836, -0.1577238291501999, 0.06170769780874252, -0.08545635640621185, 0.03337876498699188, 0.033214021474123, -0.11469510197639465, 0.01661819964647293, 0.15619690716266632, 0.11576171219348907, -0.02002456784248352, 0.011039803735911846, 0.06755843013525009, 0.0014067701995372772, -0.04056350141763687, 0.011765316128730774, 0.050716180354356766, 0.14378613233566284, -0.0812976211309433, 0.06497219204902649, -0.006863172631710768, -0.07393161207437515, -0.011535213328897953, 0.09561772644519806, -0.006328902207314968, 0.001296713831834495, -0.07016493380069733, 0.13992656767368317, -0.09890378266572952, -0.22899021208286285, 0.06886713951826096, -0.06700205057859421, -0.148993581533432, -0.05161943659186363, 0.03467195853590965, -0.018935875967144966, 0.015544258058071136, 0.0681605190038681, -0.05476875603199005, 0.16760022938251495, 0.04107232019305229, -0.03509023040533066, -0.07962212711572647, 0.0572139248251915, -0.12356412410736084, 0.27410927414894104, 0.01566120609641075, 0.06076684594154358, 0.10600827634334564, -0.015056195668876171, -0.1492050588130951, 0.007876323536038399, 0.11031296849250793, -0.06961613148450851, 0.07868678867816925, 0.16736643016338348, -0.005776694510132074, 0.13293123245239258, 0.06213664263486862, -0.05220787972211838, 0.03710320591926575, -0.06454377621412277, -0.05101386085152626, -0.10145208984613419, 0.08615441620349884, -0.0822802409529686, 0.15433424711227417, 0.12603887915611267, -0.0668829157948494, -0.008985340595245361, -0.01536030974239111, 0.0939909815788269, 0.010890315286815166, 0.10275671631097794, 0.00925779901444912, -0.1919061690568924, 0.032800618559122086, 0.022009072825312614, 0.11708221584558487, -0.21411506831645966, -0.06759718805551529, 0.05691089108586311, -0.027092965319752693, -0.0741300880908966, 0.109661765396595, 0.04356330633163452, 0.030220909044146538, -0.04273420572280884, -0.043754447251558304, 0.00521125877276063, 0.15530161559581757, -0.12153259664773941, -0.010699034668505192 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
aidonuts/enthralling-etchings-132-s800
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T03:23:45+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 60, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.04654794931411743, 0.16618601977825165, -0.005445904564112425, 0.01853804849088192, 0.0981811136007309, 0.011998992413282394, 0.06433123350143433, 0.11398410052061081, -0.0230073444545269, 0.11406639218330383, 0.03047988750040531, 0.10172267258167267, 0.11317981779575348, 0.14841650426387787, -0.002152352826669812, -0.22403094172477722, 0.050844956189394, -0.12105348706245422, -0.033293843269348145, 0.11749980598688126, 0.1483822613954544, -0.09928343445062637, 0.07274559140205383, -0.029687678441405296, -0.012143402360379696, -0.030057786032557487, -0.05890674889087677, -0.046214159578084946, 0.04651786759495735, 0.06640566885471344, 0.06770290434360504, 0.0071083661168813705, 0.09012923389673233, -0.2696533799171448, 0.018959321081638336, 0.07145345956087112, -0.002759667346253991, 0.06957992166280746, 0.06404146552085876, -0.07107418030500412, 0.10337356477975845, -0.05106033384799957, 0.14650006592273712, 0.08365883678197861, -0.09081148356199265, -0.1895141303539276, -0.08866965025663376, 0.09882009029388428, 0.17572562396526337, 0.04925641790032387, -0.02320658043026924, 0.09761467576026917, -0.08769196271896362, 0.015438909642398357, 0.04981724172830582, -0.07620415836572647, -0.05378096550703049, 0.05986575037240982, 0.07907199114561081, 0.06627275794744492, -0.12434766441583633, -0.02885502204298973, 0.005009706597775221, 0.010980482213199139, 0.0769270583987236, 0.01728810742497444, 0.146672785282135, 0.0338633768260479, -0.12615777552127838, -0.04880760237574577, 0.09869225323200226, 0.03395522013306618, -0.04422314465045929, -0.24749068915843964, -0.03152675926685333, -0.030810698866844177, -0.029386121779680252, -0.03716538846492767, 0.04340358078479767, -0.007673026993870735, 0.08638741075992584, -0.0060646249912679195, -0.07403432577848434, -0.03937075287103653, 0.06169692054390907, 0.0672287791967392, 0.02999979443848133, -0.013745363801717758, 0.010938193649053574, 0.11620724946260452, 0.1095694974064827, -0.12054188549518585, -0.05555335059762001, -0.06393084675073624, -0.08656639605760574, -0.040790557861328125, 0.034162238240242004, 0.03456587344408035, 0.05349370837211609, 0.25305667519569397, 0.015654386952519417, 0.059652652591466904, 0.034477248787879944, 0.007892133668065071, 0.05848940089344978, 0.11044429242610931, -0.06018859148025513, -0.10444226115942001, -0.02648012898862362, 0.08843598514795303, 0.008199662901461124, -0.03287925571203232, -0.05088530853390694, 0.06019928678870201, 0.01946467161178589, 0.11926145106554031, 0.09061790257692337, 0.010536285117268562, -0.07121123373508453, -0.061038948595523834, 0.1891259253025055, -0.16544590890407562, 0.04322727024555206, 0.035097137093544006, -0.03903156518936157, 0.00019933005387429148, 0.013914269395172596, 0.016625655815005302, -0.025983380153775215, 0.09017423540353775, -0.054113563150167465, -0.04145489260554314, -0.11186197400093079, -0.03383193537592888, 0.033762916922569275, 0.008953776210546494, -0.035059962421655655, -0.033713940531015396, -0.08351044356822968, -0.07577689737081528, 0.09320491552352905, -0.07346344739198685, -0.04878907650709152, -0.01804324984550476, -0.07530532777309418, 0.022395428270101547, 0.019394835457205772, 0.07707412540912628, -0.02362251654267311, 0.04399976506829262, -0.05189276114106178, 0.05863580107688904, 0.11207318305969238, 0.03570080175995827, -0.05736649036407471, 0.06062258034944534, -0.23834340274333954, 0.09552820026874542, -0.07409077137708664, 0.05591456592082977, -0.153293639421463, -0.024439791217446327, 0.04788333550095558, 0.008784620091319084, -0.009650949388742447, 0.13416339457035065, -0.21702027320861816, -0.02536402828991413, 0.1717337965965271, -0.10057014971971512, -0.07069246470928192, 0.05619903281331062, -0.04835370555520058, 0.10988964140415192, 0.03825836628675461, -0.025690359994769096, 0.06171267107129097, -0.1267417073249817, 0.003717758459970355, -0.05005312338471413, -0.017048977315425873, 0.1548657864332199, 0.07182947546243668, -0.07217690348625183, 0.07399354875087738, 0.025708531960844994, -0.0246540866792202, -0.04625825211405754, -0.015164627693593502, -0.10536660254001617, 0.014689887873828411, -0.06369215250015259, 0.014470234513282776, -0.020807426422834396, -0.09071163833141327, -0.027962757274508476, -0.17504668235778809, -0.03014434315264225, 0.08651752024888992, -0.008693269453942776, -0.01803150773048401, -0.1178668737411499, 0.009341353550553322, 0.04177580401301384, 0.0061247628182172775, -0.13462838530540466, -0.04812471568584442, 0.02780051715672016, -0.1600649207830429, 0.034652888774871826, -0.05392369255423546, 0.04932025074958801, 0.025790516287088394, -0.028889117762446404, -0.026493212208151817, 0.021633783355355263, 0.005992184858769178, -0.011999987065792084, -0.24343903362751007, -0.028118690475821495, -0.024888472631573677, 0.1682123839855194, -0.20917098224163055, 0.03546025976538658, 0.07867541164159775, 0.15366052091121674, 0.011240328662097454, -0.04177491366863251, 0.005974748637527227, -0.06935794651508331, -0.02736494317650795, -0.05875484645366669, -0.0047869328409433365, -0.03310677409172058, -0.04545191675424576, 0.04568447172641754, -0.16510973870754242, -0.032636504620313644, 0.09776268899440765, 0.06289951503276825, -0.13922683894634247, -0.020621931180357933, -0.03630133345723152, -0.049253206700086594, -0.04911839962005615, -0.0605199858546257, 0.10893940925598145, 0.05891856551170349, 0.04574795812368393, -0.05928509309887886, -0.07568105310201645, -0.001827909960411489, -0.013898161239922047, -0.017864689230918884, 0.09759635478258133, 0.0751434788107872, -0.13251115381717682, 0.09224759042263031, 0.09603385627269745, 0.07919023185968399, 0.09113933145999908, -0.02355697751045227, -0.08261934667825699, -0.045987509191036224, 0.031442027539014816, 0.020124373957514763, 0.13039541244506836, -0.024294709786772728, 0.04352088272571564, 0.042134687304496765, -0.019369594752788544, 0.014752166345715523, -0.08687400817871094, 0.033972494304180145, 0.028472330421209335, -0.016721390187740326, 0.050190530717372894, -0.03876714035868645, 0.02440318465232849, 0.08830609917640686, 0.045322712510824203, 0.03507532551884651, 0.015493292361497879, -0.05206458270549774, -0.1083620935678482, 0.16405931115150452, -0.12714070081710815, -0.22483378648757935, -0.13936103880405426, 0.0037376401014626026, 0.035628627985715866, -0.015835661441087723, 0.002417160663753748, -0.059374887496232986, -0.12220635265111923, -0.08858037739992142, 0.015140829607844353, 0.04942670464515686, -0.09028962254524231, -0.06437795609235764, 0.058117836713790894, 0.03889724239706993, -0.14560972154140472, 0.017612040042877197, 0.04854894429445267, -0.09789852797985077, -0.006774199660867453, 0.08094939589500427, 0.0698540136218071, 0.1770169734954834, 0.017703235149383545, -0.021850809454917908, 0.032354529947042465, 0.20614571869373322, -0.13538233935832977, 0.11083246022462845, 0.13607586920261383, -0.09041404724121094, 0.08072979003190994, 0.19951270520687103, 0.03932560607790947, -0.10153959691524506, 0.031980328261852264, 0.02283124253153801, -0.0284719280898571, -0.24526868760585785, -0.07212468236684799, -0.004402178805321455, -0.058010730892419815, 0.07660572230815887, 0.09286724030971527, 0.08215958625078201, 0.012304253876209259, -0.09310996532440186, -0.08154371380805969, 0.05942574888467789, 0.10367169976234436, 0.024584239348769188, -0.010839897207915783, 0.08998730033636093, -0.034100502729415894, 0.019626356661319733, 0.0853661298751831, 0.005239574704319239, 0.17840281128883362, 0.05159219726920128, 0.18830420076847076, 0.07925192266702652, 0.07219027727842331, 0.009912233799695969, 0.013080619275569916, 0.018877580761909485, 0.03300119563937187, -0.002769160782918334, -0.08440786600112915, -0.02248465269804001, 0.11566436290740967, 0.06668911874294281, 0.010815348476171494, 0.015172341838479042, -0.04104290530085564, 0.07965951412916183, 0.1831512451171875, -0.007656289264559746, -0.1783534437417984, -0.057547420263290405, 0.07553383708000183, -0.09879875183105469, -0.09854305535554886, -0.013454320840537548, 0.03072015568614006, -0.17046253383159637, 0.023390959948301315, -0.02239842526614666, 0.1106182336807251, -0.14194999635219574, -0.020490378141403198, 0.07218493521213531, 0.07199500501155853, 0.004729843698441982, 0.05758659541606903, -0.16417601704597473, 0.10671813786029816, 0.008950476534664631, 0.06779605895280838, -0.09610627591609955, 0.1008887067437172, -0.004196076653897762, -0.02063460275530815, 0.1393408179283142, 0.002700034761801362, -0.06884108483791351, -0.0763031542301178, -0.08754398673772812, -0.009632662869989872, 0.12754282355308533, -0.1419651061296463, 0.08767123520374298, -0.037212442606687546, -0.0424150750041008, -0.0017086371080949903, -0.10206665843725204, -0.11638247221708298, -0.18888559937477112, 0.06001543253660202, -0.13492922484874725, 0.03152317553758621, -0.10799519717693329, -0.032371897250413895, -0.030304040759801865, 0.19337286055088043, -0.23447458446025848, -0.07199826091527939, -0.1475764364004135, -0.10233612358570099, 0.1443224400281906, -0.0501345656812191, 0.08485390990972519, -0.007241467013955116, 0.16846685111522675, 0.019060896709561348, -0.02531743235886097, 0.0971490666270256, -0.09173708409070969, -0.19302815198898315, -0.07869284600019455, 0.15662524104118347, 0.13260218501091003, 0.031680017709732056, -0.002461588243022561, 0.036563750356435776, -0.015421539545059204, -0.11935004591941833, 0.015969349071383476, 0.1787186712026596, 0.06237189099192619, 0.02331034652888775, -0.027346095070242882, -0.11273157596588135, -0.06900003552436829, -0.028530338779091835, 0.03054865077137947, 0.17762407660484314, -0.07057618349790573, 0.18207968771457672, 0.14163152873516083, -0.05922834202647209, -0.20400173962116241, 0.010538800619542599, 0.03055560030043125, 0.0009220078936778009, 0.02591954916715622, -0.20123432576656342, 0.08688826113939285, 0.004683020059019327, -0.05110127478837967, 0.13194532692432404, -0.17217805981636047, -0.14451217651367188, 0.0765485092997551, 0.038384392857551575, -0.19559739530086517, -0.12913893163204193, -0.09174312651157379, -0.045869920402765274, -0.18591414391994476, 0.09569250047206879, 0.0305706188082695, 0.010893458500504494, 0.03030681423842907, 0.029179483652114868, 0.019487828016281128, -0.0418255440890789, 0.18391458690166473, -0.024792250245809555, 0.026594700291752815, -0.08539514988660812, -0.06927408277988434, 0.03743394836783409, -0.052842434495687485, 0.07349982857704163, -0.023486759513616562, 0.007861839607357979, -0.10348054021596909, -0.042148489505052567, -0.03735732287168503, 0.015448716469109058, -0.09657872468233109, -0.08514349907636642, -0.045032672584056854, 0.09675803780555725, 0.09690850973129272, -0.033646680414676666, -0.028050623834133148, -0.07533035427331924, 0.04412057250738144, 0.19926515221595764, 0.1785389482975006, 0.042153384536504745, -0.08034496754407883, -0.004150947090238333, -0.010121207684278488, 0.04310847446322441, -0.20463712513446808, 0.06283636391162872, 0.05450061708688736, 0.01973269321024418, 0.11436162889003754, -0.019565396010875702, -0.15359151363372803, -0.07263088971376419, 0.06303015351295471, -0.060181066393852234, -0.19620554149150848, 0.00867035984992981, 0.060603946447372437, -0.16371412575244904, -0.04535605385899544, 0.04643881320953369, -0.005620351992547512, -0.038163937628269196, 0.021896906197071075, 0.09194854646921158, 0.0026654244866222143, 0.07427921891212463, 0.05387866869568825, 0.0827430784702301, -0.10537070035934448, 0.08090532571077347, 0.08839722722768784, -0.08452684432268143, 0.023530138656497, 0.10478579998016357, -0.059433579444885254, -0.03440561518073082, 0.020135708153247833, 0.08153781294822693, 0.01775863952934742, -0.040019966661930084, 0.013229827396571636, -0.10452935844659805, 0.05954122915863991, 0.08839859813451767, 0.032507482916116714, 0.016702456399798393, 0.03425082191824913, 0.04607953503727913, -0.07238735258579254, 0.12142276018857956, 0.031868141144514084, 0.017129309475421906, -0.036505792289972305, -0.040896978229284286, 0.019542274996638298, -0.03214648738503456, -0.005015232600271702, -0.03023446537554264, -0.07695909589529037, -0.014793801121413708, -0.1626158058643341, -0.011131818406283855, -0.05648450180888176, 0.010329355485737324, 0.03204665705561638, -0.032609567046165466, 0.008124498650431633, 0.009250079281628132, -0.07695289701223373, -0.0663459524512291, -0.020460480824112892, 0.09540658444166183, -0.16213038563728333, 0.022481130436062813, 0.08244425803422928, -0.12187694013118744, 0.09281346201896667, 0.016204802319407463, -0.006236857734620571, 0.025038830935955048, -0.1475188434123993, 0.034843120723962784, -0.03386561945080757, 0.010836300440132618, 0.04373383894562721, -0.21569781005382538, -0.00004886732858722098, -0.033673107624053955, -0.06639216095209122, -0.009451326914131641, -0.03672455996274948, -0.11508306115865707, 0.1058407872915268, 0.007236586883664131, -0.08753558248281479, -0.03186136856675148, 0.029325377196073532, 0.0838974118232727, -0.021959776058793068, 0.15145497024059296, -0.008370938710868359, 0.07429654151201248, -0.16209737956523895, -0.018623165786266327, -0.006028574425727129, 0.022658247500658035, -0.01664556935429573, -0.01111356820911169, 0.044031109660863876, -0.022746501490473747, 0.17925859987735748, -0.030318550765514374, 0.02272745408117771, 0.06815794110298157, 0.019072026014328003, -0.030184008181095123, 0.10406795144081116, 0.04094860330224037, 0.02014910988509655, 0.018591465428471565, 0.003289656015112996, -0.04647882282733917, -0.03173251822590828, -0.19407226145267487, 0.07288651913404465, 0.15608493983745575, 0.09729263186454773, -0.016707008704543114, 0.07954329252243042, -0.10199416428804398, -0.1109243705868721, 0.12477338314056396, -0.04797708988189697, -0.002418199321255088, -0.07150927931070328, 0.13247236609458923, 0.1437523066997528, -0.1859612911939621, 0.07269313186407089, -0.0699717253446579, -0.04708027467131615, -0.10980689525604248, -0.19441905617713928, -0.05561789125204086, -0.049456022679805756, -0.016053348779678345, -0.04698808491230011, 0.07504211366176605, 0.054538097232580185, 0.006766852922737598, -0.0023397188633680344, 0.06506035476922989, -0.031050674617290497, -0.0037882844917476177, 0.032597362995147705, 0.06591679900884628, 0.012734474614262581, -0.030802709981799126, 0.016619903966784477, -0.013545602560043335, 0.045626189559698105, 0.06578011065721512, 0.04976864159107208, -0.02938537672162056, 0.014603170566260815, -0.038539156317710876, -0.10249634087085724, 0.043612558394670486, -0.024421939626336098, -0.0789753645658493, 0.15477414429187775, 0.023680059239268303, 0.007779473438858986, -0.020137663930654526, 0.23901568353176117, -0.0738423764705658, -0.0964353010058403, -0.14737580716609955, 0.10557299107313156, -0.038081806153059006, 0.05800395458936691, 0.04625935107469559, -0.10226529091596603, 0.018044332042336464, 0.1338089406490326, 0.16182038187980652, -0.039008259773254395, 0.020095856860280037, 0.031135575845837593, 0.00566398398950696, -0.03622615709900856, 0.04847532883286476, 0.06906453520059586, 0.16569648683071136, -0.04632584750652313, 0.09100406616926193, 0.0019041687482967973, -0.09579581767320633, -0.038361791521310806, 0.11069868505001068, -0.016052277758717537, 0.019335128366947174, -0.05818064883351326, 0.11742528527975082, -0.06386786699295044, -0.23783175647258759, 0.06453443318605423, -0.0684293657541275, -0.13765870034694672, -0.02378307841718197, 0.08207765966653824, -0.012955902144312859, 0.027587108314037323, 0.0730307325720787, -0.07240920513868332, 0.201939657330513, 0.03798431158065796, -0.05499868467450142, -0.055047210305929184, 0.0805421993136406, -0.10008571296930313, 0.2739645540714264, 0.01557221356779337, 0.04601577669382095, 0.10384146869182587, -0.009341772645711899, -0.13838784396648407, 0.019836371764540672, 0.09581108391284943, -0.10502193123102188, 0.04196618124842644, 0.19815568625926971, -0.0014755994779989123, 0.12389086186885834, 0.07657600939273834, -0.07551808655261993, 0.0478031262755394, -0.08054235577583313, -0.06760486960411072, -0.09260394424200058, 0.09703279286623001, -0.07772123068571091, 0.14251399040222168, 0.13876807689666748, -0.05074559152126312, 0.012724342755973339, -0.031311117112636566, 0.044293127954006195, -0.00010600237874314189, 0.10321761667728424, 0.004272161517292261, -0.1832672357559204, 0.024692710489034653, 0.005650998093187809, 0.10749758034944534, -0.16033467650413513, -0.09566054493188858, 0.042343202978372574, 0.003505636239424348, -0.0672195628285408, 0.1290110945701599, 0.05665452033281326, 0.04342988133430481, -0.03997718170285225, -0.03521440550684929, -0.0060732318088412285, 0.13561366498470306, -0.10713256150484085, 0.0009933578548952937 ]
null
null
transformers
#### GPU ```python import torch from transformers import AutoModelForCausalLM, AutoTokenizer def generate_prompt(instruction, input=""): instruction = instruction.strip().replace('\r\n','\n').replace('\n\n','\n') input = input.strip().replace('\r\n','\n').replace('\n\n','\n') if input: return f"""Instruction: {instruction} Input: {input} Response:""" else: return f"""User: hi Assistant: Hi. I am your assistant and I will provide expert full response in full details. Please feel free to ask any question and I will always answer it. User: {instruction} Assistant:""" model = AutoModelForCausalLM.from_pretrained("jetaudio/rwkv-5-v2-3b-16k", trust_remote_code=True, torch_dtype=torch.bfloat16).to(0) tokenizer = AutoTokenizer.from_pretrained("jetaudio/rwkv-5-v2-3b-16k", trust_remote_code=True) text = "介绍一下大熊猫" prompt = generate_prompt(text) inputs = tokenizer(prompt, return_tensors="pt").to(0) output = model.generate(inputs["input_ids"], max_new_tokens=128, do_sample=True, temperature=1.0, top_p=0.3, top_k=0, ) print(tokenizer.decode(output[0].tolist(), skip_special_tokens=True)) ```
{}
text-generation
jetaudio/rwkv-5-v2-3b-16k
[ "transformers", "pytorch", "rwkv5", "text-generation", "custom_code", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-15T03:23:54+00:00
[]
[]
TAGS #transformers #pytorch #rwkv5 #text-generation #custom_code #autotrain_compatible #endpoints_compatible #region-us
#### GPU
[ "#### GPU" ]
[ "TAGS\n#transformers #pytorch #rwkv5 #text-generation #custom_code #autotrain_compatible #endpoints_compatible #region-us \n", "#### GPU" ]
[ 44, 3 ]
[ "passage: TAGS\n#transformers #pytorch #rwkv5 #text-generation #custom_code #autotrain_compatible #endpoints_compatible #region-us \n#### GPU" ]
[ -0.06348849087953568, 0.020802421495318413, -0.0029810641426593065, 0.0804247185587883, 0.1915181428194046, 0.0690026581287384, 0.054393235594034195, 0.15855811536312103, 0.03722719848155975, 0.013889843598008156, 0.16893400251865387, 0.22950111329555511, 0.04601375758647919, 0.13684040307998657, -0.003345726989209652, -0.23110225796699524, 0.07554500550031662, 0.06439270079135895, 0.027814596891403198, 0.1010260209441185, 0.041082724928855896, -0.08131694048643112, 0.1509389728307724, -0.03935874626040459, -0.26542410254478455, -0.03501239791512489, 0.041645217686891556, -0.0731431096792221, 0.1297023743391037, 0.08287856727838516, 0.03849560767412186, 0.02027134783565998, 0.00117978872731328, -0.10482826083898544, 0.027049092575907707, 0.05647151172161102, -0.10157380253076553, 0.08173198252916336, 0.10621488094329834, -0.04130171611905098, 0.13881506025791168, 0.0560644157230854, -0.035791363567113876, 0.05726321414113045, -0.12123918533325195, -0.12823393940925598, -0.03175753355026245, 0.011058079078793526, 0.020773395895957947, 0.07377704232931137, 0.01352107897400856, 0.15789169073104858, -0.06929611414670944, 0.11255249381065369, 0.18230044841766357, -0.28291815519332886, -0.019617756828665733, 0.1512036770582199, 0.03830752149224281, -0.019789330661296844, -0.045118387788534164, 0.051625605672597885, 0.012047521770000458, 0.030631817877292633, -0.010068723000586033, -0.03595559298992157, -0.08759889751672745, 0.022673888131976128, -0.06101302057504654, -0.08470281213521957, 0.15500293672084808, -0.048896439373493195, 0.10938293486833572, -0.026597827672958374, -0.15237607061862946, -0.17455235123634338, -0.010441266000270844, 0.0810210183262825, -0.0903361514210701, 0.030163122341036797, 0.00877110380679369, 0.0010643212590366602, -0.08953127264976501, -0.10024110227823257, -0.14405615627765656, 0.2244642674922943, 0.04955335333943367, 0.08986854553222656, -0.11934322118759155, 0.15561708807945251, -0.0013110877480357885, -0.10890927165746689, -0.008362271822988987, -0.11600992828607559, -0.009105760604143143, 0.08972654491662979, -0.05771046131849289, 0.06519545614719391, 0.08956656605005264, 0.16755251586437225, 0.048252709209918976, 0.036073360592126846, 0.07200983166694641, 0.058287762105464935, -0.03884813189506531, 0.04698265343904495, -0.03797237202525139, -0.07568693906068802, 0.06275227665901184, -0.03461679816246033, 0.017423005774617195, -0.06389392912387848, -0.12977321445941925, -0.07673243433237076, 0.03691283240914345, 0.052215464413166046, 0.05498380586504936, 0.1010521873831749, -0.05571915954351425, -0.014628462493419647, 0.15152756869792938, -0.0638706162571907, -0.014014525339007378, 0.01885904185473919, 0.021272769197821617, 0.04760479927062988, 0.02601003460586071, -0.07956184446811676, -0.07369889318943024, -0.014187092892825603, -0.05673343688249588, -0.04795994982123375, -0.12861034274101257, -0.04446862265467644, 0.016598783433437347, -0.13161872327327728, 0.03701648861169815, -0.1483827382326126, -0.13769805431365967, 0.01346585899591446, 0.025956766679883003, -0.036566101014614105, -0.01709315925836563, 0.004201128147542477, -0.06677249819040298, 0.07254379987716675, -0.027761781588196754, 0.0632140040397644, -0.03583645820617676, 0.12541021406650543, 0.07142522186040878, 0.1158893033862114, -0.09701406955718994, 0.060870151966810226, -0.0771980956196785, -0.020267663523554802, -0.04756373167037964, 0.04477835074067116, -0.06728474795818329, 0.056218478828668594, 0.02469121478497982, -0.0361492745578289, -0.07816143333911896, 0.020395124331116676, 0.018234098330140114, 0.1473688930273056, -0.09752167016267776, -0.08309478312730789, 0.29377439618110657, -0.09729228168725967, -0.15976840257644653, 0.10688190907239914, 0.009500918909907341, -0.08336354047060013, 0.025210833176970482, 0.08055583387613297, 0.03560439869761467, -0.03825875744223595, 0.00038095959462225437, 0.10554821044206619, -0.1014055386185646, -0.11286920309066772, 0.04558367654681206, 0.052854374051094055, -0.07161396741867065, 0.09939533472061157, 0.00930145662277937, 0.10084472596645355, -0.05317353084683418, -0.0435040108859539, -0.08978968113660812, -0.015485589392483234, 0.07179871946573257, 0.07409248501062393, 0.047813206911087036, -0.05887990444898605, -0.08956415206193924, -0.1294514536857605, 0.09703192114830017, -0.03343738988041878, 0.004304847680032253, -0.09289958328008652, 0.1633422076702118, -0.0707794576883316, 0.08605227619409561, -0.15762712061405182, -0.08556319028139114, 0.002537704538553953, 0.06622803956270218, -0.008153989911079407, 0.07895398885011673, 0.0768255740404129, 0.029271777719259262, 0.008202856406569481, -0.03705848380923271, 0.05872363969683647, -0.009154043160378933, -0.02630983106791973, -0.08891652524471283, 0.008828678168356419, -0.05974356457591057, 0.011442465707659721, -0.07898794114589691, 0.01954219676554203, 0.133762925863266, 0.12137728929519653, -0.013231308199465275, 0.027214616537094116, -0.046829793602228165, 0.003081502625718713, -0.11301686614751816, -0.0030328836292028427, 0.10358462482690811, -0.02065254934132099, 0.01273300126194954, 0.13590474426746368, -0.18241477012634277, 0.23966020345687866, 0.23703530430793762, -0.2149249017238617, 0.015118016861379147, 0.04494570195674896, 0.001094894134439528, 0.0007910528802312911, 0.03999706730246544, -0.016820212826132774, 0.06623711436986923, 0.012726757675409317, 0.15751536190509796, -0.008959910832345486, -0.00436873733997345, 0.0664842426776886, -0.012178018689155579, 0.011204276233911514, 0.07358261942863464, 0.1877654641866684, -0.051017045974731445, 0.17780430614948273, 0.10622230917215347, 0.008022884838283062, 0.16475345194339752, 0.03168405219912529, -0.031558744609355927, 0.06674906611442566, -0.036036089062690735, -0.01881754957139492, 0.04040994495153427, -0.19451528787612915, -0.05168624967336655, 0.0959283858537674, -0.04459991678595543, 0.07605300098657608, -0.17464211583137512, 0.00700629036873579, -0.018244953826069832, 0.002840410452336073, 0.03868534415960312, 0.10787271708250046, 0.010240335017442703, 0.0969257727265358, -0.03484441712498665, -0.03511219471693039, 0.10849872976541519, 0.004224576987326145, -0.043559759855270386, 0.1963636726140976, -0.10917184501886368, -0.30187079310417175, -0.18408894538879395, -0.10848742723464966, -0.08284367620944977, 0.062202680855989456, 0.08054070174694061, -0.11291251331567764, -0.05484013631939888, 0.0673886314034462, 0.058117687702178955, 0.001981506822630763, 0.010723374783992767, -0.024386748671531677, 0.03883841633796692, -0.08785197138786316, -0.0922732800245285, -0.0361819826066494, 0.026381516829133034, -0.11449087411165237, 0.19868923723697662, -0.06259840726852417, 0.12623436748981476, 0.15221595764160156, 0.010586869902908802, 0.04366317763924599, -0.001207818859256804, 0.1621994823217392, -0.09440036118030548, 0.015644697472453117, 0.20861060917377472, 0.004263082053512335, 0.05216166377067566, 0.05510670691728592, 0.0011141797294840217, -0.09898500889539719, 0.015314641408622265, -0.07351953536272049, -0.12306969612836838, -0.17129011452198029, -0.08947791904211044, -0.1336301565170288, 0.07391534745693207, 0.06568372249603271, 0.04667133465409279, 0.05477849766612053, 0.1306849867105484, -0.0175804290920496, 0.06962205469608307, -0.019275767728686333, 0.07399458438158035, 0.19955195486545563, -0.0012521796161308885, 0.13809612393379211, -0.12065725028514862, -0.043663982301950455, 0.10904474556446075, 0.10790813714265823, 0.14175482094287872, -0.016512827947735786, 0.12769362330436707, 0.009045789949595928, 0.14518031477928162, 0.14433088898658752, 0.12545065581798553, 0.006427886430174112, -0.0015591845149174333, -0.026588384062051773, 0.007826309651136398, -0.14417335391044617, 0.05089118331670761, -0.026615289971232414, -0.15599524974822998, -0.0345531590282917, -0.0063660587184131145, 0.08463475108146667, 0.08489956706762314, 0.006264079827815294, -0.30496111512184143, -0.047092508524656296, 0.03571180999279022, 0.034883491694927216, -0.1334676742553711, 0.09010223299264908, -0.002284703776240349, -0.13879279792308807, 0.034133121371269226, -0.08074861764907837, 0.09641890227794647, -0.053455956280231476, 0.03151079639792442, 0.012240784242749214, 0.0312592014670372, 0.06586047261953354, 0.1416202187538147, -0.2913568913936615, 0.17925138771533966, -0.01851576194167137, -0.022935455664992332, -0.13310301303863525, 0.0017349038971588016, 0.05430633947253227, 0.12199387699365616, 0.09165683388710022, -0.018352970480918884, -0.08162079006433487, -0.14853553473949432, -0.019544484093785286, 0.03634628653526306, 0.11417461186647415, -0.010366198606789112, -0.0001965691044460982, -0.05338452011346817, -0.027965828776359558, -0.037610094994306564, -0.04765810817480087, 0.055309511721134186, -0.17811129987239838, 0.0926186740398407, 0.01069488562643528, 0.09011568129062653, -0.02540360949933529, -0.06119483709335327, -0.13232539594173431, 0.1563672423362732, -0.12211789935827255, -0.06144725903868675, -0.10605097562074661, -0.044978879392147064, 0.05731469392776489, -0.12422606348991394, 0.08589614927768707, -0.11780473589897156, 0.04216325655579567, -0.048650436103343964, -0.21393494307994843, 0.13172978162765503, -0.12238136678934097, -0.13645023107528687, -0.02103884518146515, 0.08812342584133148, -0.09649248421192169, 0.013235168531537056, -0.016452699899673462, 0.06246371194720268, -0.1473342627286911, -0.11722798645496368, 0.040445223450660706, 0.054213251918554306, 0.0016727036563679576, -0.0033652244601398706, -0.03239521011710167, -0.0573524571955204, 0.016056999564170837, -0.00818627793341875, 0.21990914642810822, 0.20244553685188293, -0.1001833900809288, 0.10578078031539917, 0.1257680207490921, -0.03034919872879982, -0.38162970542907715, -0.07431861013174057, -0.1008039116859436, -0.015035989694297314, -0.08976816385984421, -0.20885725319385529, 0.09208742529153824, -0.019594961777329445, -0.03800075128674507, 0.1787177324295044, -0.1686631441116333, -0.07862883061170578, 0.10624908655881882, 0.06377089768648148, 0.25290820002555847, -0.13070036470890045, -0.037171367555856705, -0.05003344267606735, -0.2968359589576721, 0.13832786679267883, -0.006093269679695368, 0.13260945677757263, -0.07279631495475769, 0.0694291964173317, -0.0005436165374703705, -0.09792860597372055, 0.07968854904174805, -0.021369660273194313, 0.03303016722202301, -0.10009971261024475, 0.0025621652603149414, 0.11031553149223328, 0.01060881931334734, 0.017936712130904198, -0.037153471261262894, 0.06534071266651154, -0.1526513248682022, -0.04485596716403961, -0.11754199862480164, 0.029895661398768425, 0.03050563856959343, -0.05790993198752403, 0.008216774091124535, -0.038968753069639206, -0.02161070890724659, 0.005475493147969246, 0.09353025257587433, 0.013076565228402615, 0.052365709096193314, 0.11637255549430847, 0.05412096902728081, -0.12896135449409485, -0.028554299846291542, -0.07204924523830414, -0.04075879231095314, 0.08786533027887344, -0.09650576114654541, 0.034172963351011276, 0.11358194053173065, -0.03825991973280907, 0.010790808126330376, 0.11783690750598907, -0.01940760388970375, 0.02560836635529995, 0.14034779369831085, -0.23813976347446442, -0.00297153415158391, -0.048886287957429886, -0.0012455509277060628, 0.10031882673501968, 0.0864669606089592, 0.08270714432001114, 0.028883283957839012, -0.042900923639535904, -0.03568131849169731, 0.02263290248811245, -0.04397344961762428, 0.08185082674026489, 0.06372641026973724, 0.05700010806322098, -0.17387236654758453, 0.0573970228433609, -0.03809688985347748, -0.20725250244140625, 0.010159936733543873, 0.17339925467967987, -0.171514630317688, -0.1367061287164688, -0.03184356912970543, 0.06761425733566284, -0.18528755009174347, -0.06087583303451538, -0.08125703781843185, -0.07550682127475739, 0.05725901201367378, 0.06734072417020798, 0.11911220103502274, 0.06140519678592682, -0.0032608702313154936, -0.014125389978289604, -0.06851900368928909, 0.02729359269142151, -0.031626369804143906, 0.06606360524892807, -0.09167598932981491, 0.029435930773615837, 0.008691401220858097, 0.15110522508621216, -0.09743056446313858, -0.019937260076403618, -0.11820823699235916, 0.025399640202522278, -0.06433787196874619, -0.04313555732369423, -0.06267129629850388, -0.017049303278326988, -0.020405787974596024, -0.02984471619129181, -0.037685416638851166, 0.0037399614229798317, -0.1389472335577011, -0.014950024895370007, -0.03974023833870888, -0.0002948916517198086, -0.07435610890388489, 0.028949813917279243, 0.09907127916812897, -0.06326809525489807, 0.10412956029176712, 0.09415322542190552, -0.044488996267318726, 0.09640199691057205, -0.09842275083065033, -0.09704149514436722, 0.055390097200870514, 0.05859449878334999, 0.046426255255937576, 0.0769021287560463, 0.08926157653331757, 0.03893487527966499, 0.0322500579059124, 0.010454759933054447, 0.10761556774377823, -0.10912676155567169, 0.03866369277238846, -0.07381509989500046, -0.09871453791856766, -0.04288771003484726, 0.013530648313462734, 0.04662717878818512, 0.023873409256339073, 0.11963015794754028, -0.016955716535449028, 0.06638216227293015, -0.09828643500804901, 0.020943162962794304, -0.021596405655145645, -0.19001859426498413, -0.07506600022315979, -0.05162481591105461, 0.05888081714510918, 0.009160755202174187, 0.2610054910182953, 0.03520628437399864, -0.06241481378674507, 0.0029263568576425314, 0.06501862406730652, -0.043086934834718704, -0.016019973903894424, 0.17676256597042084, 0.1069493293762207, 0.017430981621146202, -0.08157573640346527, 0.09641439467668533, 0.036932848393917084, 0.008656718768179417, 0.13441318273544312, 0.030444107949733734, 0.0295408982783556, 0.09467480331659317, 0.016120964661240578, -0.062388014048337936, -0.14882057905197144, -0.12373539805412292, -0.1339186131954193, 0.087545245885849, -0.06050347164273262, 0.09178982675075531, 0.13669298589229584, -0.017041850835084915, 0.01744726672768593, -0.06881649047136307, -0.03885218873620033, -0.14929288625717163, -0.10812908411026001, -0.10139729082584381, -0.15911675989627838, 0.059197135269641876, -0.04365623742341995, 0.019142698496580124, 0.13168823719024658, 0.023003915324807167, -0.06437951326370239, 0.18037530779838562, 0.06493167579174042, -0.009728861972689629, 0.00961193535476923, 0.0008756647002883255, -0.03495411202311516, 0.04738270118832588, -0.05963018536567688, -0.08040300756692886, -0.050590116530656815, 0.034764595329761505, 0.0182229895144701, -0.10331561416387558, 0.0721786692738533, -0.06857087463140488, -0.07873699814081192, -0.04824699088931084, 0.05327576398849487, -0.04393789917230606, 0.1714550107717514, -0.024399712681770325, 0.022831935435533524, 0.03766239061951637, 0.15541426837444305, -0.07228261232376099, -0.08601655811071396, -0.05452906712889671, 0.1590331345796585, 0.034915994852781296, 0.07026399672031403, -0.019934460520744324, -0.02786610648036003, -0.059363748878240585, 0.3365967273712158, 0.28732573986053467, -0.043166790157556534, -0.0011051606852561235, 0.05259215459227562, 0.033467989414930344, 0.04915718361735344, 0.21097972989082336, 0.08525042235851288, 0.22680138051509857, -0.11049368232488632, -0.06871871650218964, -0.047329772263765335, -0.029253922402858734, -0.043133001774549484, -0.008450144901871681, 0.0004387795925140381, -0.06424622982740402, -0.05035874992609024, -0.006157173775136471, -0.14314982295036316, 0.1059446707367897, 0.015151637606322765, -0.17602236568927765, -0.03149665892124176, -0.02562171407043934, 0.10872569680213928, -0.0007615323411300778, 0.06535417586565018, -0.023281151428818703, -0.009452123194932938, 0.09220903366804123, 0.01344878040254116, -0.2563682198524475, 0.00042329641291871667, 0.03876722976565361, -0.08308044821023941, 0.07779785990715027, -0.07040213793516159, 0.06210482493042946, 0.10672859102487564, 0.027529796585440636, -0.10079995542764664, 0.08540195226669312, -0.025778697803616524, -0.05247890576720238, -0.005624841433018446, 0.07313285768032074, 0.011286408640444279, -0.1136583611369133, 0.013828392140567303, -0.061411578208208084, 0.057766709476709366, -0.027447190135717392, 0.01840857043862343, -0.023660356178879738, 0.03775031492114067, -0.04557788744568825, 0.11981578916311264, 0.10363515466451645, 0.024831237271428108, -0.03626270219683647, -0.1040838435292244, 0.03298167511820793, 0.00044569899910129607, -0.05554987117648125, -0.11893492937088013, -0.14448001980781555, -0.06018797680735588, 0.08495187759399414, 0.006059000734239817, -0.15859700739383698, -0.031626466661691666, -0.12761488556861877, 0.022996073588728905, -0.17440859973430634, 0.08443991094827652, 0.0780307948589325, 0.02939019910991192, -0.0154415313154459, -0.07151155173778534, -0.007401433773338795, 0.054063908755779266, -0.16961760818958282, -0.1545335054397583 ]
null
null
peft
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.8.2
{"library_name": "peft", "base_model": "google-t5/t5-small"}
null
Queriamin/t5_xsum_summarization_1000steps
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:google-t5/t5-small", "region:us" ]
2024-02-15T03:24:00+00:00
[ "1910.09700" ]
[]
TAGS #peft #safetensors #arxiv-1910.09700 #base_model-google-t5/t5-small #region-us
# Model Card for Model ID ## Model Details ### Model Description - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact ### Framework versions - PEFT 0.8.2
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ "TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-google-t5/t5-small #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ 37, 6, 3, 54, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4, 11 ]
[ "passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-google-t5/t5-small #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2" ]
[ -0.1077658161520958, 0.19646243751049042, -0.0030439540278166533, 0.03554042801260948, 0.0946621522307396, 0.01885688118636608, 0.05619802325963974, 0.12496209144592285, -0.02706761844456196, 0.10974712669849396, 0.06915885955095291, 0.09789934009313583, 0.10723601281642914, 0.222670778632164, 0.00646469509229064, -0.20199885964393616, 0.0255292896181345, -0.0941162109375, -0.0054487246088683605, 0.12773537635803223, 0.1455765962600708, -0.09494901448488235, 0.08082420378923416, -0.01681945100426674, -0.006435811053961515, -0.03182074427604675, -0.06794276833534241, -0.03411043807864189, 0.044835928827524185, 0.04862125590443611, 0.054694972932338715, -0.00020927679724991322, 0.08506274223327637, -0.2637648284435272, 0.019547412171959877, 0.049223218113183975, -0.007669814862310886, 0.08665230870246887, 0.0981588140130043, -0.044934503734111786, 0.1265098750591278, -0.028694309294223785, 0.1408674120903015, 0.08120559900999069, -0.094017893075943, -0.2289126217365265, -0.06515778601169586, 0.09092748910188675, 0.17774361371994019, 0.07613107562065125, -0.044072188436985016, 0.13494983315467834, -0.09061576426029205, 0.019411731511354446, 0.046639587730169296, -0.09750095009803772, -0.07027023285627365, 0.054167356342077255, 0.10083626955747604, 0.05716893821954727, -0.1299501359462738, -0.030166279524564743, 0.026966577395796776, 0.034798212349414825, 0.08069641888141632, 0.014283783733844757, 0.14616313576698303, 0.026912158355116844, -0.14686058461666107, -0.04396803304553032, 0.13829228281974792, 0.027194006368517876, -0.03661242499947548, -0.22585394978523254, 0.0005922320415265858, -0.0843546986579895, -0.025034263730049133, -0.05074915662407875, 0.03928638622164726, 0.0029324705246835947, 0.10066965967416763, -0.02967788092792034, -0.0899120420217514, -0.013487364165484905, 0.09153076261281967, 0.05131715163588524, 0.026260748505592346, -0.02087426371872425, 0.00470492709428072, 0.1279095560312271, 0.057402387261390686, -0.13031499087810516, -0.06031063199043274, -0.07162587344646454, -0.047255393117666245, -0.041367173194885254, 0.042213257402181625, 0.04280208796262741, 0.056872621178627014, 0.2534946799278259, -0.03676801919937134, 0.056562598794698715, 0.0627484917640686, 0.020530564710497856, 0.044361479580402374, 0.09421030431985855, -0.061966534703969955, -0.15374475717544556, -0.01008674968034029, 0.0980200245976448, -0.0030629239045083523, -0.022366363555192947, -0.04889882728457451, 0.04430406913161278, 0.03666825592517853, 0.10997398942708969, 0.09740255773067474, -0.0041997237130999565, -0.07796763628721237, -0.05332411453127861, 0.20161142945289612, -0.1514815092086792, 0.0387902669608593, 0.021236931905150414, -0.015054930001497269, -0.05037995055317879, 0.016125816851854324, 0.01459946483373642, -0.0291454941034317, 0.09464716166257858, -0.06804710626602173, -0.04150799661874771, -0.11483925580978394, -0.026674535125494003, 0.03268420696258545, 0.00800632406026125, -0.032027117908000946, -0.03625979274511337, -0.06560580432415009, -0.09274742007255554, 0.10053679347038269, -0.06182357296347618, -0.060888346284627914, -0.029948875308036804, -0.08954104036092758, 0.018951738253235817, 0.02496221289038658, 0.091240793466568, -0.02666437067091465, 0.04451191797852516, -0.012513834983110428, 0.06480924785137177, 0.08389968425035477, 0.03402974456548691, -0.0757552832365036, 0.06319116801023483, -0.20025759935379028, 0.08495914191007614, -0.0800900086760521, 0.029218247160315514, -0.160506933927536, -0.01921827718615532, 0.00264375121332705, 0.022457532584667206, 0.03197690099477768, 0.16344080865383148, -0.20236138999462128, -0.03563845902681351, 0.15991488099098206, -0.1076260432600975, -0.12177621573209763, 0.041881125420331955, -0.043930646032094955, 0.15965710580348969, 0.02390878275036812, -0.004042434971779585, 0.09795088320970535, -0.1475711613893509, -0.023964490741491318, -0.020299555733799934, -0.005359130911529064, 0.10204251110553741, 0.08630509674549103, -0.08433201164007187, 0.02343178167939186, 0.013953768648207188, -0.051182761788368225, -0.022443320602178574, -0.04937099292874336, -0.10647034645080566, 0.006505660247057676, -0.0814133957028389, 0.02591928280889988, -0.004806713666766882, -0.08173832297325134, -0.010987378656864166, -0.16521282494068146, -0.041013024747371674, 0.080232173204422, 0.008738456293940544, -0.02114478312432766, -0.10110321640968323, 0.046449657529592514, -0.026350514963269234, -0.019299786537885666, -0.14896205067634583, -0.03127721697092056, 0.017890403047204018, -0.13620184361934662, 0.007237366400659084, -0.11850045621395111, 0.06686335057020187, 0.012239331379532814, -0.06481441855430603, -0.03708382323384285, -0.006781097035855055, 0.006599651649594307, -0.048084404319524765, -0.24168920516967773, -0.0202324315905571, -0.052107587456703186, 0.15701216459274292, -0.22458578646183014, 0.0391354039311409, 0.049721769988536835, 0.13014259934425354, 0.006456756964325905, -0.06356898695230484, 0.032139912247657776, -0.068898506462574, -0.02791086956858635, -0.07639389485120773, -0.005945524666458368, -0.006731790490448475, -0.046518631279468536, 0.02069897949695587, -0.12090891599655151, -0.030663983896374702, 0.1015625149011612, 0.06832430511713028, -0.16517780721187592, -0.0136569207534194, -0.04512608423829079, -0.062012191861867905, -0.08360525965690613, -0.060295723378658295, 0.10952118784189224, 0.04995199665427208, 0.04032031074166298, -0.0758313238620758, -0.07104538381099701, 0.009362636134028435, -0.02175595425069332, -0.0211634561419487, 0.11645656824111938, 0.07379018515348434, -0.11224023252725601, 0.09695055335760117, 0.07176152616739273, 0.035716377198696136, 0.08370199799537659, -0.027154449373483658, -0.10487764328718185, -0.02695748582482338, 0.04962174966931343, 0.015528207644820213, 0.16051016747951508, -0.06683047115802765, 0.05271526053547859, 0.0450693741440773, -0.04085993394255638, 0.044434111565351486, -0.09803594648838043, 0.00832105241715908, 0.008406898938119411, -0.017184581607580185, 0.013662761077284813, -0.019376594573259354, 0.008572258055210114, 0.0873655304312706, 0.05158910155296326, 0.037572138011455536, 0.026796501129865646, -0.029371844604611397, -0.13465583324432373, 0.18668688833713531, -0.09550008922815323, -0.24468591809272766, -0.1556822806596756, 0.06773916631937027, 0.058110129088163376, -0.017237035557627678, 0.023191314190626144, -0.057986773550510406, -0.10515452176332474, -0.0838828906416893, 0.0027155010029673576, 0.03266002982854843, -0.056388355791568756, -0.07345637679100037, 0.046749480068683624, 0.047140903770923615, -0.11719518899917603, 0.03680480644106865, 0.056899264454841614, -0.019115187227725983, 0.004004235379397869, 0.05312405526638031, 0.08651448041200638, 0.18399770557880402, -0.008199452422559261, 0.00006906331691425294, 0.04601173475384712, 0.27532511949539185, -0.16007065773010254, 0.11383267492055893, 0.12536031007766724, -0.06417430192232132, 0.07746100425720215, 0.18959856033325195, 0.031976841390132904, -0.10131848603487015, 0.0320252887904644, 0.03005632571876049, -0.029746223241090775, -0.2683558762073517, -0.04848521575331688, -0.013314875774085522, -0.08886121958494186, 0.08073446154594421, 0.09178229421377182, 0.0766359344124794, 0.038712434470653534, -0.07187698781490326, -0.08621349185705185, 0.03627176582813263, 0.09938030689954758, -0.018959304317831993, 0.006506346166133881, 0.08340054750442505, -0.03540336340665817, 0.009885420091450214, 0.099531389772892, -0.015219200402498245, 0.16644342243671417, 0.0478072315454483, 0.10691328346729279, 0.08013132959604263, 0.09200374037027359, -0.002378062577918172, 0.02737002819776535, 0.016776610165834427, 0.025447269901633263, 0.013992567546665668, -0.08581594377756119, 0.02877195179462433, 0.11079565435647964, 0.03937005251646042, 0.03084411472082138, 0.013360953889787197, -0.036800604313611984, 0.0505000464618206, 0.18192875385284424, 0.010882833041250706, -0.20724409818649292, -0.08062963932752609, 0.05871279910206795, -0.07717147469520569, -0.13483645021915436, -0.011754429899156094, 0.03450503945350647, -0.16634777188301086, 0.026892883703112602, -0.04167970269918442, 0.09932820498943329, -0.08533335477113724, -0.039850715547800064, 0.1042175143957138, 0.06479167193174362, -0.02614145167171955, 0.05545124039053917, -0.19321802258491516, 0.13026118278503418, 0.025547225028276443, 0.0684945210814476, -0.08512773364782333, 0.09944665431976318, 0.004082023166120052, 0.0007930412539280951, 0.17098821699619293, 0.002292859135195613, -0.05475001409649849, -0.06718125194311142, -0.09588324278593063, -0.013017266988754272, 0.0973091647028923, -0.13776808977127075, 0.06693438440561295, -0.021276695653796196, -0.029727566987276077, -0.0023200875148177147, -0.08662791550159454, -0.1277594268321991, -0.167549729347229, 0.055525269359350204, -0.09974829107522964, 0.02867145836353302, -0.09424322098493576, -0.0666453018784523, 0.003903559409081936, 0.17671480774879456, -0.21068355441093445, -0.10195909440517426, -0.1505342572927475, -0.08614850044250488, 0.15814733505249023, -0.04455644637346268, 0.08548807352781296, 0.0014104091096669436, 0.1610710322856903, 0.016308892518281937, -0.01336646731942892, 0.10517074912786484, -0.09059996157884598, -0.19774620234966278, -0.05699877440929413, 0.1688915342092514, 0.13113462924957275, 0.03674261271953583, -0.014889905229210854, 0.024823851883411407, -0.048051510006189346, -0.11915593594312668, 0.025157729163765907, 0.1413734257221222, 0.06592490524053574, -0.00994897447526455, -0.02893890254199505, -0.10193102806806564, -0.061603520065546036, -0.046356551349163055, -0.0031502796337008476, 0.18692365288734436, -0.07656683027744293, 0.16103249788284302, 0.11005975306034088, -0.05400444194674492, -0.20917057991027832, 0.04756074771285057, 0.053044017404317856, 0.01437255460768938, 0.043122369796037674, -0.19219666719436646, 0.08327171951532364, -0.003923862706869841, -0.07270405441522598, 0.16494840383529663, -0.17118605971336365, -0.14347849786281586, 0.0994994193315506, 0.0373387485742569, -0.22214512526988983, -0.14080791175365448, -0.10224071890115738, -0.015406312420964241, -0.11477863788604736, 0.053649310022592545, -0.004148885142058134, 0.011037350632250309, 0.02653786540031433, 0.01175929419696331, 0.026956656947731972, -0.04655206948518753, 0.19828173518180847, -0.029319167137145996, 0.009108294732868671, -0.052098292857408524, -0.08659904450178146, 0.031239816918969154, -0.04703662917017937, 0.10410289466381073, -0.0016452833078801632, 0.030659452080726624, -0.15156042575836182, -0.042865440249443054, -0.055094558745622635, 0.03181813657283783, -0.09346551448106766, -0.08865223079919815, -0.047326475381851196, 0.09308157116174698, 0.09404002130031586, -0.028492121025919914, 0.0007937632617540658, -0.08968065679073334, 0.07125315070152283, 0.20003642141819, 0.1958054155111313, 0.06995762139558792, -0.06955525279045105, 0.020332643762230873, -0.03262636438012123, 0.04679378494620323, -0.23520320653915405, 0.04031446948647499, 0.056376293301582336, 0.022905737161636353, 0.08448328077793121, -0.00947054848074913, -0.15446849167346954, -0.07134253531694412, 0.08544494211673737, -0.05493659898638725, -0.17214451730251312, -0.030554642900824547, 0.021276414394378662, -0.20474179089069366, -0.040375497192144394, 0.023626498878002167, -0.024665040895342827, -0.03661192208528519, 0.024211054667830467, 0.0779169574379921, -0.016524162143468857, 0.10858429968357086, 0.08731520920991898, 0.09229105710983276, -0.10326189547777176, 0.07629965245723724, 0.07596059143543243, -0.04345763847231865, 0.028071438893675804, 0.11347384750843048, -0.051802754402160645, -0.034505415707826614, 0.07907052338123322, 0.08935019373893738, 0.029323481023311615, -0.05307482182979584, 0.008089395239949226, -0.0584036223590374, 0.06111699715256691, 0.11388109624385834, 0.027130126953125, -0.00020278082229197025, 0.05776326358318329, 0.03281862661242485, -0.08932865411043167, 0.11117307096719742, 0.060188181698322296, 0.01904141716659069, -0.05078946053981781, -0.03307927027344704, -0.004358852747827768, -0.0131904361769557, -0.02184063196182251, -0.004908058326691389, -0.09214063733816147, -0.006917557213455439, -0.08944131433963776, 0.02573596127331257, -0.07038608193397522, 0.010718019679188728, 0.0293679591268301, -0.053961168974637985, 0.0013292960356920958, 0.007214670069515705, -0.07190340757369995, -0.04803140461444855, -0.011441302485764027, 0.08513348549604416, -0.13166053593158722, 0.037109844386577606, 0.07266084849834442, -0.10374049842357635, 0.07896453887224197, -0.00801768247038126, 0.004869848024100065, 0.008466639555990696, -0.16143038868904114, 0.058227259665727615, -0.02001281827688217, -0.012775802053511143, 0.017872648313641548, -0.20288535952568054, -0.004133281297981739, -0.04838475584983826, -0.05909949913620949, 0.011297469027340412, -0.023427635431289673, -0.12421391904354095, 0.09559439867734909, -0.0003283861733507365, -0.06438656151294708, -0.018961846828460693, 0.039411962032318115, 0.09852207452058792, -0.027349920943379402, 0.13430365920066833, -0.028992118313908577, 0.07100006192922592, -0.17813576757907867, -0.00832216627895832, -0.013298123143613338, 0.03916066139936447, -0.030276689678430557, -0.022498425096273422, 0.06012347713112831, -0.022565262392163277, 0.177195206284523, -0.014079391956329346, 0.07326609641313553, 0.057004839181900024, 0.013610313646495342, 0.01719851791858673, 0.08236774802207947, 0.05616380646824837, -0.0032146251760423183, -0.00508534163236618, 0.02956998534500599, -0.010261778719723225, -0.04091853275895119, -0.1679808348417282, 0.06854455173015594, 0.1475587636232376, 0.04566395655274391, 0.02145567536354065, 0.02913753315806389, -0.11384984850883484, -0.08153904229402542, 0.12287963926792145, -0.017016654834151268, -0.03143095597624779, -0.06837926059961319, 0.16913944482803345, 0.14212028682231903, -0.19615541398525238, 0.0720026046037674, -0.05296690762042999, -0.048471394926309586, -0.13527421653270721, -0.1694951057434082, -0.061508890241384506, -0.05237546190619469, -0.019658921286463737, -0.06502756476402283, 0.04815830662846565, 0.05319412797689438, 0.005218065809458494, -0.01566176488995552, 0.11201795190572739, 0.009926238097250462, -0.027150634676218033, 0.052660975605249405, 0.06540547311306, 0.03390313684940338, -0.09579019248485565, 0.006719397846609354, -0.0022873827256262302, 0.014451303519308567, 0.06257788836956024, 0.017122022807598114, -0.05508671700954437, 0.018929796293377876, -0.019880075007677078, -0.11426179111003876, 0.040662821382284164, -0.013456898741424084, -0.04155447706580162, 0.143046572804451, 0.03334852680563927, 0.007098872680217028, -0.020553195849061012, 0.2302698791027069, -0.07772248238325119, -0.07001755386590958, -0.14995962381362915, 0.07686116546392441, -0.06487935781478882, 0.032928138971328735, 0.032033804804086685, -0.11856430768966675, 0.01550919096916914, 0.1656419038772583, 0.13252928853034973, -0.008097765035927296, 0.011568031273782253, 0.042752258479595184, 0.005230261944234371, -0.03105936385691166, 0.023806609213352203, 0.05132965371012688, 0.14989368617534637, -0.06714575737714767, 0.06154889613389969, -0.008624210953712463, -0.07780536264181137, -0.01748107373714447, 0.10768572241067886, 0.0005769342533312738, 0.0016409042291343212, -0.0717204213142395, 0.14423246681690216, -0.08263417333364487, -0.22356627881526947, 0.0642777606844902, -0.0733712837100029, -0.14614443480968475, -0.04904039204120636, 0.022068459540605545, -0.01365199126303196, 0.011048863641917706, 0.07550778239965439, -0.05101019889116287, 0.16948607563972473, 0.04437866061925888, -0.058202046900987625, -0.08523133397102356, 0.05581562966108322, -0.14652101695537567, 0.2837216556072235, 0.01849539205431938, 0.042694833129644394, 0.103920117020607, -0.018516987562179565, -0.14079231023788452, 0.01486047450453043, 0.10664543509483337, -0.06738100200891495, 0.05911201611161232, 0.168772891163826, 0.002296225633472204, 0.1290857493877411, 0.05629529803991318, -0.05242190510034561, 0.03785131126642227, -0.09357116371393204, -0.04735575616359711, -0.11206188052892685, 0.08278654515743256, -0.08431066572666168, 0.16105693578720093, 0.12590622901916504, -0.06802909076213837, -0.004265516065061092, -0.023764831945300102, 0.0821053758263588, 0.00932464562356472, 0.11213263124227524, 0.01829429343342781, -0.1835811585187912, 0.03447761386632919, 0.00750605296343565, 0.10142281651496887, -0.20156055688858032, -0.0582004189491272, 0.04190053418278694, -0.017465513199567795, -0.08174601197242737, 0.12342629581689835, 0.047131650149822235, 0.036319129168987274, -0.03962165489792824, -0.051282286643981934, 0.009981295093894005, 0.14727140963077545, -0.1147594228386879, -0.00775710865855217 ]
null
null
peft
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # mistral_5 This model is a fine-tuned version of [ybelkada/mistral-7b-instruct-v0.1-sharded](https://huggingface.co/ybelkada/mistral-7b-instruct-v0.1-sharded) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - num_epochs: 8 - mixed_precision_training: Native AMP ### Training results ### Framework versions - PEFT 0.8.2 - Transformers 4.37.2 - Pytorch 2.2.0 - Datasets 2.17.0 - Tokenizers 0.15.2
{"library_name": "peft", "tags": ["trl", "sft", "generated_from_trainer"], "base_model": "ybelkada/mistral-7b-instruct-v0.1-sharded", "model-index": [{"name": "mistral_5", "results": []}]}
null
anyiwang/mistral_5
[ "peft", "safetensors", "trl", "sft", "generated_from_trainer", "base_model:ybelkada/mistral-7b-instruct-v0.1-sharded", "region:us" ]
2024-02-15T03:24:16+00:00
[]
[]
TAGS #peft #safetensors #trl #sft #generated_from_trainer #base_model-ybelkada/mistral-7b-instruct-v0.1-sharded #region-us
# mistral_5 This model is a fine-tuned version of ybelkada/mistral-7b-instruct-v0.1-sharded on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - num_epochs: 8 - mixed_precision_training: Native AMP ### Training results ### Framework versions - PEFT 0.8.2 - Transformers 4.37.2 - Pytorch 2.2.0 - Datasets 2.17.0 - Tokenizers 0.15.2
[ "# mistral_5\n\nThis model is a fine-tuned version of ybelkada/mistral-7b-instruct-v0.1-sharded on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- num_epochs: 8\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.2.0\n- Datasets 2.17.0\n- Tokenizers 0.15.2" ]
[ "TAGS\n#peft #safetensors #trl #sft #generated_from_trainer #base_model-ybelkada/mistral-7b-instruct-v0.1-sharded #region-us \n", "# mistral_5\n\nThis model is a fine-tuned version of ybelkada/mistral-7b-instruct-v0.1-sharded on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- num_epochs: 8\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.2.0\n- Datasets 2.17.0\n- Tokenizers 0.15.2" ]
[ 50, 41, 6, 12, 8, 3, 103, 4, 36 ]
[ "passage: TAGS\n#peft #safetensors #trl #sft #generated_from_trainer #base_model-ybelkada/mistral-7b-instruct-v0.1-sharded #region-us \n# mistral_5\n\nThis model is a fine-tuned version of ybelkada/mistral-7b-instruct-v0.1-sharded on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- num_epochs: 8\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.2.0\n- Datasets 2.17.0\n- Tokenizers 0.15.2" ]
[ -0.10370275378227234, 0.007491067051887512, -0.0018324831034988165, 0.05869171395897865, 0.1549934595823288, 0.025059271603822708, 0.12172173708677292, 0.1042095497250557, -0.03827076405286789, 0.07295101135969162, 0.06020689010620117, 0.01928272284567356, 0.05440203472971916, 0.14056673645973206, -0.02875114418566227, -0.26104745268821716, 0.043311234563589096, -0.02186811901628971, -0.023110507056117058, 0.08874170482158661, 0.12288127839565277, -0.10791616886854172, 0.04730521887540817, 0.009474711492657661, -0.14741291105747223, 0.014503820799291134, -0.011190864257514477, -0.03878359869122505, 0.10562407970428467, 0.025371849536895752, 0.15513961017131805, -0.0020929588936269283, 0.13623380661010742, -0.228220134973526, 0.017150230705738068, 0.07503880560398102, 0.04186427593231201, 0.07798822969198227, 0.07225330919027328, 0.01498835813254118, 0.08664987981319427, -0.11635319888591766, 0.11663561314344406, 0.031806062906980515, -0.08172044157981873, -0.2092919945716858, -0.10594404488801956, 0.061568766832351685, 0.11860568076372147, 0.07681956887245178, 0.01003674790263176, 0.14261233806610107, -0.090386301279068, 0.05145413801074028, 0.1928728073835373, -0.22039994597434998, -0.08196921646595001, 0.05448925495147705, 0.051129188388586044, 0.08557743579149246, -0.10742834210395813, -0.041840262711048126, 0.06806954741477966, 0.04354388266801834, 0.06882862746715546, 0.007775298785418272, -0.06628602743148804, 0.00004370666283648461, -0.14646953344345093, -0.012051700614392757, 0.1339900940656662, 0.045615412294864655, -0.051695168018341064, -0.07529818266630173, -0.034249741584062576, -0.05154889076948166, -0.03542555496096611, -0.039752017706632614, 0.0022676815278828144, -0.023617573082447052, -0.011355970054864883, -0.0470612458884716, -0.09905724972486496, -0.09927985817193985, 0.0034891157411038876, 0.1296028196811676, 0.04690779373049736, 0.0018494761316105723, -0.043685171753168106, 0.10795623064041138, -0.002570199780166149, -0.08611118048429489, -0.009267172776162624, -0.018476707860827446, -0.08192320168018341, -0.07605300098657608, -0.04489767551422119, -0.041764214634895325, 0.006886588875204325, 0.1100323349237442, -0.10780517011880875, 0.08914642781019211, 0.021867290139198303, 0.03435597941279411, -0.024580011144280434, 0.08080019056797028, -0.04087071493268013, 0.04292081668972969, -0.010414537973701954, 0.09930004924535751, -0.018558606505393982, -0.014726337045431137, -0.07133742421865463, -0.051155392080545425, 0.08099149167537689, 0.04812273383140564, -0.054209765046834946, 0.006959719583392143, -0.0664895698428154, -0.021236125379800797, 0.013530678115785122, -0.12031181156635284, 0.03709886595606804, 0.01520154532045126, -0.058134060353040695, -0.027851199731230736, 0.030284689739346504, 0.029150396585464478, 0.0068596601486206055, 0.07233845442533493, -0.07433091849088669, 0.024647118523716927, -0.10035131871700287, -0.08005361258983612, 0.01347208209335804, -0.015175323002040386, -0.015326677821576595, -0.10605888068675995, -0.17886759340763092, -0.06003415584564209, 0.034945711493492126, -0.041453491896390915, 0.000787927710916847, -0.05267732962965965, -0.030863387510180473, 0.039674580097198486, -0.007036102004349232, 0.11639229953289032, -0.05351516231894493, 0.0833665281534195, -0.05597809702157974, -0.0021410321351140738, -0.02717958204448223, 0.028754903003573418, -0.06360001116991043, 0.04378356784582138, -0.10212573409080505, 0.05637028440833092, -0.10606078058481216, 0.03584624081850052, -0.13278794288635254, -0.09681180119514465, -0.01963750459253788, -0.01617477461695671, 0.11102405935525894, 0.11130549758672714, -0.20039542019367218, -0.010148471221327782, 0.12790422141551971, -0.11328456550836563, -0.08090481907129288, 0.10608014464378357, -0.06377117335796356, 0.08190369606018066, 0.038254715502262115, 0.17366717755794525, 0.12196267396211624, -0.14041514694690704, 0.043362487107515335, 0.0028738307300955057, 0.09317919611930847, 0.06623746454715729, 0.06206182762980461, -0.03218700364232063, -0.0850248783826828, 0.0065812175162136555, -0.06413642317056656, 0.04859953746199608, -0.09647709131240845, -0.07257739454507828, -0.047867730259895325, -0.060633447021245956, 0.07912104576826096, 0.020102083683013916, 0.027236836031079292, -0.0684979185461998, -0.05963239446282387, 0.166078120470047, 0.14908534288406372, -0.05288572981953621, 0.0037395451217889786, -0.0607539638876915, 0.03761474788188934, -0.01609429158270359, -0.038157228380441666, -0.17006392776966095, -0.11842156946659088, 0.030335308983922005, -0.04036131873726845, 0.001407609903253615, 0.027530429884791374, 0.07707017660140991, 0.07173266261816025, -0.0436006598174572, -0.051606398075819016, -0.11674706637859344, 0.013418681919574738, -0.11370473355054855, -0.1898653209209442, -0.05278740078210831, -0.06205497682094574, 0.16523359715938568, -0.2538967728614807, 0.021879801526665688, -0.013438801281154156, 0.13284164667129517, 0.03736969456076622, -0.04667099937796593, -0.009334729984402657, 0.0671095922589302, 0.00027187669184058905, -0.08402449637651443, 0.04773270711302757, 0.01810019090771675, -0.1282268613576889, -0.01821211725473404, -0.1487981081008911, 0.036468420177698135, 0.04767058044672012, 0.0379829928278923, -0.10690527409315109, -0.1399482786655426, -0.06229695305228233, -0.0529351606965065, -0.08522181957960129, -0.001605113036930561, 0.20084242522716522, 0.006095568649470806, 0.11764466762542725, -0.06318990141153336, -0.03800741955637932, -0.0046607982367277145, -0.02401522733271122, -0.01824287511408329, 0.08853047341108322, 0.021505534648895264, -0.11851536482572556, 0.0713140144944191, 0.1264573186635971, -0.06581529974937439, 0.1722499430179596, -0.05514077469706535, -0.10141848027706146, -0.012863804586231709, 0.05488596856594086, -0.00598198501393199, 0.10754531621932983, -0.06491713225841522, 0.03069116733968258, 0.017239131033420563, 0.04745741933584213, 0.040328774601221085, -0.1867060363292694, -0.021082939580082893, 0.014364331029355526, -0.019121550023555756, -0.021262910217046738, -0.016793975606560707, 0.028620030730962753, 0.08376390486955643, 0.01918650045990944, -0.010836265981197357, 0.01772211119532585, -0.006638075225055218, -0.11556515097618103, 0.19920337200164795, -0.1524747908115387, -0.08953877538442612, -0.11724942922592163, 0.07902940362691879, -0.0008035110658966005, -0.029718922451138496, 0.02746558003127575, -0.09578582644462585, -0.03749191761016846, -0.08343973010778427, 0.0007879487238824368, -0.05283297970890999, -0.024603359401226044, 0.024111062288284302, 0.007059300318360329, 0.09172283858060837, -0.12755407392978668, 0.015227767638862133, -0.004454921931028366, -0.08651036769151688, 0.022836701944470406, 0.027448436245322227, 0.07850559800863266, 0.14102767407894135, -0.011536093428730965, -0.004803434945642948, -0.048852402716875076, 0.22581495344638824, -0.05145657807588577, -0.025145960971713066, 0.10852089524269104, -0.0059711020439863205, 0.05948435515165329, 0.12183656543493271, 0.04584844037890434, -0.0880262553691864, 0.05038360878825188, 0.04515134170651436, -0.004672582261264324, -0.25460192561149597, -0.05188969895243645, -0.026197446510195732, -0.1005692258477211, 0.09408295154571533, 0.047392938286066055, -0.028448596596717834, 0.06331197172403336, -0.04560103267431259, 0.020785342901945114, 0.013966599479317665, 0.09159086644649506, 0.01974295638501644, 0.023643163964152336, 0.07019881159067154, -0.023387327790260315, -0.003972072619944811, 0.056212857365608215, 0.010483132675290108, 0.2703733444213867, -0.014516725204885006, 0.058389998972415924, 0.05969215929508209, 0.16290752589702606, -0.005492453463375568, 0.020103057846426964, 0.020880544558167458, -0.010074772872030735, -0.0031714041251689196, -0.06289361417293549, -0.040891338139772415, 0.05722014233469963, -0.04110366851091385, 0.07289788872003555, -0.12251461297273636, 0.0008387149428017437, 0.01943686231970787, 0.26589706540107727, 0.011791927739977837, -0.2600698471069336, -0.10410632193088531, 0.03304380923509598, -0.01703314296901226, -0.10158926248550415, 0.009195909835398197, 0.1645357757806778, -0.13545845448970795, 0.030031610280275345, -0.06293720006942749, 0.0974106416106224, 0.001301312237046659, -0.012464347295463085, 0.010016663931310177, 0.12789975106716156, -0.014837229624390602, 0.08317773044109344, -0.23209144175052643, 0.24625325202941895, 0.013558762148022652, 0.10430625826120377, -0.023348990827798843, 0.021772313863039017, 0.033938512206077576, 0.08363695442676544, 0.060378529131412506, 0.013736707158386707, -0.11217506229877472, -0.20788773894309998, -0.037839118391275406, 0.037375546991825104, 0.10656566917896271, -0.013974679633975029, 0.057690974324941635, -0.05094413086771965, 0.0426730252802372, 0.05201408639550209, -0.07664494216442108, -0.23126932978630066, -0.10373667627573013, 0.004495048429816961, 0.014628613367676735, 0.018200695514678955, -0.12885987758636475, -0.09655902534723282, -0.020529484376311302, 0.10545963048934937, -0.014890812337398529, -0.026625683531165123, -0.14305341243743896, 0.07234104722738266, 0.11547594517469406, -0.05583862587809563, 0.02504117414355278, 0.03497816622257233, 0.1310407668352127, 0.008086560294032097, -0.04764610901474953, 0.06925997138023376, -0.07627741992473602, -0.16321976482868195, -0.07257537543773651, 0.09171915799379349, 0.09831707924604416, 0.04666191712021828, 0.008434060961008072, 0.012491251341998577, 0.011005338281393051, -0.09552852064371109, 0.007189614698290825, 0.18468396365642548, 0.033469073474407196, 0.09275706112384796, -0.09869218617677689, -0.005852192640304565, -0.04046523943543434, -0.03803495690226555, 0.13114836812019348, 0.2421625852584839, -0.08926669508218765, 0.06800466030836105, 0.10279746353626251, -0.0942952036857605, -0.1698850393295288, 0.09951136261224747, 0.14679022133350372, 0.026923123747110367, 0.046624913811683655, -0.1836003214120865, 0.050546690821647644, 0.14419828355312347, -0.01897171325981617, 0.02314291149377823, -0.341582328081131, -0.11561784148216248, 0.07957324385643005, 0.14049088954925537, 0.032345496118068695, -0.11554763466119766, -0.040913790464401245, -0.027245668694376945, -0.11635813117027283, 0.03709633648395538, -0.12727117538452148, 0.07447786629199982, -0.01158178597688675, 0.0761537179350853, 0.03888992592692375, -0.033271193504333496, 0.1999153196811676, -0.02667349763214588, 0.11360204964876175, -0.05477451905608177, 0.057218387722969055, 0.02357971854507923, -0.07007239758968353, 0.039738889783620834, -0.006501294206827879, 0.066623255610466, -0.1330888271331787, -0.008608444593846798, -0.06364871561527252, 0.0720263347029686, -0.048492640256881714, -0.07391764968633652, -0.032590992748737335, 0.05806439742445946, 0.04354235902428627, -0.027811245992779732, 0.06394001841545105, -0.018789876252412796, 0.17662347853183746, 0.03173084929585457, 0.10366732627153397, -0.016721205785870552, -0.0577864907681942, 0.0014435237972065806, -0.02751953899860382, 0.08768343180418015, -0.1318284571170807, 0.02350354567170143, 0.12059172242879868, 0.03140996769070625, 0.1602693498134613, 0.04193072021007538, -0.07303474843502045, 0.038314372301101685, 0.04005015268921852, -0.06976406276226044, -0.16328342258930206, -0.021357020363211632, 0.1429073065519333, -0.14795564115047455, 0.001376443193294108, 0.11319497227668762, -0.07577525824308395, -0.021613383665680885, -0.023233085870742798, 0.012009136378765106, -0.04868293181061745, 0.17992213368415833, 0.04032432660460472, 0.06207119673490524, -0.05350066348910332, 0.0980229377746582, 0.07728397846221924, -0.08471378684043884, 0.05430625006556511, 0.0565822459757328, -0.0841103121638298, -0.040338873863220215, 0.0735551044344902, 0.16823206841945648, -0.01467415876686573, -0.04669822379946709, -0.049768056720495224, -0.104824960231781, 0.012309806421399117, 0.10767517983913422, 0.026678290218114853, -0.009223969653248787, -0.0030776658095419407, 0.025561563670635223, -0.10218451917171478, 0.06396003812551498, 0.04650748148560524, 0.08680082857608795, -0.1558523178100586, 0.15148788690567017, -0.0003663914103526622, 0.022836878895759583, -0.007482696790248156, -0.00510325375944376, -0.10929431766271591, 0.004241057671606541, -0.17683464288711548, 0.02466181293129921, -0.03683878481388092, 0.025283433496952057, 0.010054850950837135, -0.05249857157468796, -0.008069922216236591, 0.0448586530983448, -0.07875673472881317, -0.03980116546154022, 0.006059642415493727, 0.08990898728370667, -0.08435430377721786, -0.02439235895872116, 0.041798610240221024, -0.0810772031545639, 0.06324836611747742, 0.06573636084794998, 0.02820771187543869, 0.048913586884737015, -0.20473060011863708, 0.016969071701169014, 0.036060698330402374, 0.008407960645854473, 0.036180634051561356, -0.09795411676168442, -0.024332845583558083, -0.04353110119700432, 0.03905657306313515, 0.01269559096544981, 0.061690621078014374, -0.11910053342580795, -0.06347902864217758, -0.03848058357834816, -0.07116755098104477, -0.0802883431315422, 0.03657980635762215, 0.07214852422475815, 0.06840073317289352, 0.1231270581483841, -0.10596470534801483, 0.054387059062719345, -0.1756475567817688, -0.035706572234630585, -0.024136455729603767, 0.004704121500253677, -0.07027358561754227, -0.053584061563014984, 0.06889341026544571, -0.04999333620071411, 0.04791467636823654, -0.04847607761621475, 0.07408122718334198, 0.019084176048636436, -0.10939087718725204, 0.009109598584473133, 0.018177758902311325, 0.21766987442970276, 0.05292678251862526, -0.0019363450119271874, 0.053186386823654175, 0.0018465188331902027, 0.03526170924305916, 0.10852328687906265, 0.12982997298240662, 0.1756078153848648, -0.02936244197189808, 0.0604122094810009, 0.05070119723677635, -0.10190758109092712, -0.058967817574739456, 0.09034430235624313, 0.018080230802297592, 0.059569913893938065, -0.06475788354873657, 0.17642615735530853, 0.14402249455451965, -0.18782863020896912, 0.023147176951169968, -0.06526713073253632, -0.09863223135471344, -0.10587326437234879, -0.017577296122908592, -0.07695701718330383, -0.14617820084095, 0.007706783711910248, -0.11029241979122162, 0.011256511323153973, 0.06311064213514328, 0.00929349660873413, 0.043825190514326096, 0.13193821907043457, 0.0235553290694952, 0.008533081971108913, 0.050906892865896225, 0.0069090635515749454, 0.017709288746118546, -0.10173697024583817, -0.10726384073495865, 0.09177805483341217, -0.04040931537747383, 0.0418456606566906, -0.04580473527312279, 0.0077682919800281525, 0.02965027466416359, -0.0032525043934583664, -0.07755409181118011, 0.033442649990320206, 0.0095268115401268, 0.005221925210207701, 0.06881435960531235, 0.08543644845485687, -0.012106141075491905, -0.04538707062602043, 0.28468483686447144, -0.0722641870379448, -0.07545945048332214, -0.14622053503990173, 0.2188977599143982, 0.011969677172601223, 0.007004759274423122, 0.04043463617563248, -0.10835012793540955, 0.01850620098412037, 0.07279965281486511, 0.10231863707304001, -0.03563186526298523, 0.0056094154715538025, -0.025102170184254646, -0.025318410247564316, -0.0958128497004509, 0.14441031217575073, 0.09799153357744217, -0.007645345292985439, -0.06749843806028366, 0.007254555355757475, -0.022508665919303894, 0.0010132923489436507, -0.08434093743562698, 0.02287295088171959, -0.010667922906577587, 0.0012436230899766088, -0.0431959331035614, 0.09883204102516174, 0.01895003579556942, -0.12802009284496307, 0.0158951785415411, -0.10682602971792221, -0.13997133076190948, -0.04585731774568558, 0.058042265474796295, 0.0008565745083615184, 0.04521118476986885, -0.04548415169119835, 0.01498239953070879, 0.13517045974731445, -0.02150818333029747, -0.015144620090723038, -0.14433766901493073, 0.0906224474310875, -0.022792527452111244, 0.2231644243001938, -0.002694750903174281, 0.06895861029624939, 0.10002124309539795, 0.04291685298085213, -0.11961575597524643, 0.059095218777656555, 0.06707731634378433, -0.03919804096221924, 0.014556300826370716, 0.1471552550792694, -0.05492670461535454, 0.12718184292316437, 0.046468377113342285, -0.17628133296966553, 0.018357234075665474, -0.04878396913409233, -0.02646767906844616, -0.06851565837860107, 0.03957999497652054, -0.052439842373132706, 0.1496843546628952, 0.16837410628795624, -0.04141227528452873, -0.014092501252889633, -0.06399942189455032, 0.04342535138130188, 0.04605785757303238, 0.11215798556804657, -0.026226487010717392, -0.19901706278324127, 0.017033696174621582, 0.037571974098682404, 0.027747225016355515, -0.22667226195335388, -0.10203015804290771, 0.027505571022629738, -0.05840719863772392, -0.028382161632180214, 0.1206812933087349, 0.06259491294622421, 0.027411237359046936, -0.03687317296862602, -0.21048271656036377, -0.0355033352971077, 0.15905211865901947, -0.08888685703277588, -0.0475834459066391 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
automatic-speech-recognition
spsither/wav2vec2_run9.615
[ "transformers", "safetensors", "wav2vec2", "automatic-speech-recognition", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
2024-02-15T03:26:39+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #wav2vec2 #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 47, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.06877388060092926, 0.1546701192855835, -0.0037609888240695, 0.013798683881759644, 0.11170210689306259, 0.0049477447755634785, 0.07622946053743362, 0.1076156347990036, -0.024175573140382767, 0.12644733488559723, 0.04164152219891548, 0.09870775043964386, 0.11074616760015488, 0.18980292975902557, 0.0015578214079141617, -0.20271944999694824, 0.06667982041835785, -0.11557482928037643, 0.02210802026093006, 0.12125445902347565, 0.14131462574005127, -0.10717527568340302, 0.06805222481489182, -0.03453851491212845, -0.022604284808039665, -0.03256304934620857, -0.06200181692838669, -0.0628168061375618, 0.06936536729335785, 0.060818396508693695, 0.06474827229976654, 0.023958178237080574, 0.07868874818086624, -0.2985154092311859, 0.020363550633192062, 0.07747753709554672, 0.005190075840801001, 0.0596587099134922, 0.07716850191354752, -0.06847380846738815, 0.11357854306697845, -0.0553223080933094, 0.15529125928878784, 0.07729580253362656, -0.09200245141983032, -0.18732582032680511, -0.08171983063220978, 0.09086527675390244, 0.16344711184501648, 0.05807739868760109, -0.035454582422971725, 0.14257195591926575, -0.08119463175535202, 0.015228749252855778, 0.06432900577783585, -0.07448869198560715, -0.04995284602046013, 0.044303327798843384, 0.07393822818994522, 0.09027253836393356, -0.12936420738697052, -0.005840824451297522, 0.04285894334316254, 0.01751609519124031, 0.1045890524983406, 0.0271924901753664, 0.10937820374965668, 0.030452799052000046, -0.13982591032981873, -0.06308452039957047, 0.12294159829616547, 0.03608649969100952, -0.05978325754404068, -0.24299637973308563, -0.007494248915463686, -0.030862024053931236, -0.022421855479478836, -0.0449565127491951, 0.040200937539339066, -0.03043903410434723, 0.0803007185459137, 0.005218773614615202, -0.07346875220537186, -0.0566013865172863, 0.08528164029121399, 0.0660456046462059, 0.024965541437268257, -0.02511134371161461, 0.022877119481563568, 0.11602471768856049, 0.09200266003608704, -0.11191211640834808, -0.07020656764507294, -0.06118712201714516, -0.09110330045223236, -0.04440220445394516, 0.03338851034641266, 0.07138838618993759, 0.04954010248184204, 0.19076436758041382, 0.006971653085201979, 0.05134076997637749, 0.026316070929169655, 0.018496420234441757, 0.061533693224191666, 0.06859898567199707, -0.05315755307674408, -0.12085959315299988, -0.043275654315948486, 0.1195915937423706, 0.008576745167374611, -0.03422791138291359, -0.034871865063905716, 0.05920550227165222, 0.05124519392848015, 0.11922229826450348, 0.06299308687448502, 0.015805674716830254, -0.06944610923528671, -0.041848812252283096, 0.17807698249816895, -0.15696440637111664, 0.01886504516005516, 0.019594965502619743, -0.05179493874311447, -0.028022583574056625, 0.01927095092833042, 0.011918062344193459, -0.028684133663773537, 0.09848573058843613, -0.06384129822254181, -0.037289999425411224, -0.10494036227464676, -0.051826175302267075, 0.03436095267534256, -0.01885044015944004, -0.030469300225377083, -0.04276524484157562, -0.11668366193771362, -0.07342278957366943, 0.06446365267038345, -0.06070359796285629, -0.06312011927366257, -0.04004829749464989, -0.05974921956658363, 0.01184001937508583, -0.0018999426392838359, 0.12804386019706726, -0.03126852586865425, 0.04724927991628647, -0.05154479295015335, 0.07010733336210251, 0.13001501560211182, 0.0328618623316288, -0.06312436610460281, 0.06317896395921707, -0.20583610236644745, 0.10645388811826706, -0.0948607325553894, 0.026716187596321106, -0.16420963406562805, -0.024270139634609222, 0.02872021123766899, 0.03977278992533684, -0.014035328291356564, 0.13902691006660461, -0.1889396458864212, -0.037479519844055176, 0.1823769360780716, -0.1340419203042984, -0.09025664627552032, 0.06442771852016449, -0.056058306246995926, 0.1311984360218048, 0.051679398864507675, -0.016549112275242805, 0.050827931612730026, -0.14181455969810486, -0.021199021488428116, -0.05750836804509163, -0.01345672644674778, 0.14918801188468933, 0.06591099500656128, -0.060217004269361496, 0.03262941166758537, 0.02008114755153656, -0.02076314203441143, -0.052245598286390305, -0.03416990861296654, -0.09862805157899857, 0.003799794940277934, -0.08055862784385681, 0.018423959612846375, -0.026528598740696907, -0.08738208562135696, -0.0410190187394619, -0.1575777381658554, -0.001173238386400044, 0.1026405617594719, 0.0026203012093901634, -0.02646641992032528, -0.10305316001176834, 0.001408840762451291, 0.015838710591197014, -0.010245922021567822, -0.14677146077156067, -0.04217318072915077, 0.026863576844334602, -0.16719304025173187, 0.031281016767024994, -0.045817263424396515, 0.03617605194449425, 0.042714666575193405, -0.04341552406549454, -0.026187991723418236, 0.011214246973395348, 0.01926763355731964, -0.01759723760187626, -0.24584431946277618, -0.01623428985476494, -0.05088721215724945, 0.17665798962116241, -0.2476477026939392, 0.04387471452355385, 0.07402390241622925, 0.1185368224978447, 0.006659833248704672, -0.0473252609372139, 0.03859061002731323, -0.04956425726413727, -0.039547327905893326, -0.06162410229444504, -0.002731422893702984, -0.034249331802129745, -0.04925791174173355, 0.04766050726175308, -0.19274261593818665, -0.0254798773676157, 0.1145588755607605, 0.07196282595396042, -0.16417020559310913, -0.0721944123506546, -0.03388380631804466, -0.060263555496931076, -0.0855790227651596, -0.05511211231350899, 0.10627889633178711, 0.042532145977020264, 0.053568705916404724, -0.07193132489919662, -0.0538090355694294, 0.014475145377218723, -0.008023109287023544, -0.03674730286002159, 0.08616615831851959, 0.07892905920743942, -0.111492820084095, 0.0967666357755661, 0.06781410425901413, 0.06170906499028206, 0.10836543887853622, 0.0035758649464696646, -0.09838994592428207, -0.013410377316176891, 0.028753211721777916, 0.013008177280426025, 0.1445195972919464, -0.08268706500530243, 0.02993486076593399, 0.04475158452987671, -0.029572229832410812, 0.014260980300605297, -0.10948343575000763, 0.020612964406609535, 0.03188888356089592, -0.01410164125263691, 0.016051514074206352, -0.05129382014274597, 0.013738108798861504, 0.10363461822271347, 0.031123731285333633, 0.025897923856973648, 0.016665659844875336, -0.04273077845573425, -0.12888197600841522, 0.17441782355308533, -0.09573886543512344, -0.24906472861766815, -0.13649064302444458, 0.0033230632543563843, 0.04450872540473938, -0.01420661062002182, 0.019941311329603195, -0.06085766479372978, -0.10865217447280884, -0.10793688893318176, 0.02346382476389408, 0.04952440410852432, -0.08567548543214798, -0.05095811188220978, 0.05441328510642052, 0.03898037597537041, -0.12600500881671906, 0.024548007175326347, 0.04095667228102684, -0.07147589325904846, 0.005656755063682795, 0.061115942895412445, 0.08382482826709747, 0.1812773495912552, 0.012779363431036472, -0.015533777885138988, 0.01035984791815281, 0.21022020280361176, -0.14754468202590942, 0.08923394232988358, 0.142924964427948, -0.06379926204681396, 0.07994367927312851, 0.20067699253559113, 0.030222468078136444, -0.0959763154387474, 0.0354040265083313, 0.03157598897814751, -0.03929230570793152, -0.24485765397548676, -0.07799134403467178, 0.004727535881102085, -0.06941798329353333, 0.0999692752957344, 0.08970286697149277, 0.11357339471578598, 0.04878859966993332, -0.10688808560371399, -0.07536104321479797, 0.04997042194008827, 0.11770502477884293, -0.025654911994934082, 0.0004288276832085103, 0.09490229189395905, -0.032173965126276016, 0.024045821279287338, 0.09091470390558243, 0.01785297878086567, 0.1891387403011322, 0.045389045029878616, 0.13416282832622528, 0.08966030925512314, 0.05892613157629967, 0.02283613197505474, 0.020396918058395386, 0.022836502641439438, 0.028627371415495872, -0.02071341499686241, -0.08800762891769409, -0.01406664215028286, 0.1445012241601944, 0.03501417487859726, 0.03224355727434158, 0.005818283185362816, -0.03822546452283859, 0.07026989012956619, 0.16923215985298157, 0.01291902456432581, -0.22557523846626282, -0.06553208827972412, 0.07285686582326889, -0.07819344103336334, -0.10939628630876541, -0.00628721434623003, 0.039236925542354584, -0.1781243532896042, 0.0453440323472023, -0.016895415261387825, 0.09935811161994934, -0.11019659787416458, -0.022818224504590034, 0.03339223191142082, 0.06351818144321442, -0.033710017800331116, 0.07605454325675964, -0.20844414830207825, 0.14833855628967285, 0.007355031557381153, 0.06984888762235641, -0.10627210140228271, 0.07959222793579102, 0.018262188881635666, 0.0005360859213396907, 0.16532482206821442, -0.0075689139775931835, -0.07650822401046753, -0.08155251294374466, -0.07923656702041626, -0.010918287560343742, 0.10160883516073227, -0.10205793380737305, 0.08789419382810593, -0.006757213734090328, -0.030893130227923393, -0.00026032759342342615, -0.11519953608512878, -0.1342930644750595, -0.18055365979671478, 0.04992220178246498, -0.10558607429265976, 0.04552379995584488, -0.11181014776229858, -0.062069665640592575, -0.04111560434103012, 0.18840233981609344, -0.20550832152366638, -0.07671810686588287, -0.14316488802433014, -0.08166468888521194, 0.11773297190666199, -0.036535169929265976, 0.08007847517728806, 0.008441719226539135, 0.20702308416366577, -0.00666013965383172, 0.002528243465349078, 0.08686443418264389, -0.09668374806642532, -0.2072489857673645, -0.09340810775756836, 0.14340825378894806, 0.12398830056190491, 0.045563604682683945, -0.0001787850633263588, 0.021285003051161766, -0.004406071733683348, -0.11160994321107864, 0.036765191704034805, 0.1599014699459076, 0.08414851129055023, 0.041826896369457245, -0.023910723626613617, -0.15188267827033997, -0.1039518192410469, -0.06143968924880028, 0.022748636081814766, 0.18740743398666382, -0.06844107806682587, 0.17012163996696472, 0.157639279961586, -0.061386726796627045, -0.20854754745960236, 0.031976643949747086, 0.03363525867462158, -0.008795025758445263, 0.0332365483045578, -0.20113597810268402, 0.06802120804786682, 0.01531505398452282, -0.057996444404125214, 0.1332528293132782, -0.16826434433460236, -0.15160627663135529, 0.08843177556991577, 0.07692008465528488, -0.20126505196094513, -0.12921905517578125, -0.09711465984582901, -0.05218008533120155, -0.10807206481695175, 0.08772927522659302, -0.006655422504991293, 0.007214459590613842, 0.037578340619802475, 0.02635364979505539, 0.015357093885540962, -0.05328182876110077, 0.19721722602844238, 0.0011987579055130482, 0.044046565890312195, -0.07511261850595474, -0.077226422727108, 0.034381043165922165, -0.06312628090381622, 0.07982822507619858, -0.020660031586885452, 0.0017429457511752844, -0.11481664329767227, -0.06663372367620468, -0.05009456351399422, 0.029989875853061676, -0.08466581255197525, -0.09467059373855591, -0.051657307893037796, 0.09798348695039749, 0.09048279374837875, -0.03396918624639511, -0.06807554513216019, -0.10042613744735718, 0.06601390987634659, 0.22872091829776764, 0.18910692632198334, 0.06991440057754517, -0.06895517557859421, -0.0038870053831487894, -0.026509825140237808, 0.05879383906722069, -0.20851773023605347, 0.044600993394851685, 0.036500073969364166, 0.032537586987018585, 0.13215065002441406, -0.02442602440714836, -0.16357013583183289, -0.043075863271951675, 0.056227099150419235, -0.06633396446704865, -0.16863006353378296, 0.005107434932142496, 0.09075167030096054, -0.15091724693775177, -0.04752274975180626, 0.030901111662387848, -0.03220430761575699, -0.02397167682647705, 0.00030637482996098697, 0.08078145235776901, 0.020850084722042084, 0.1107739508152008, 0.06640642136335373, 0.11335843801498413, -0.10278842598199844, 0.08162284642457962, 0.08386309444904327, -0.11347422748804092, 0.04244251549243927, 0.05978094041347504, -0.06325716525316238, -0.03386267274618149, 0.016484335064888, 0.0787876546382904, 0.03214597329497337, -0.08122093230485916, 0.0026990212500095367, -0.11556044965982437, 0.06788678467273712, 0.14209748804569244, 0.03322440758347511, 0.007564007304608822, 0.04558844491839409, 0.031089849770069122, -0.09967122226953506, 0.10952559113502502, 0.0327114500105381, 0.03264835476875305, -0.052766215056180954, 0.007493352517485619, 0.044093240052461624, -0.012370331212878227, -0.01659340038895607, -0.04159332811832428, -0.062125492841005325, -0.004501889459788799, -0.15752804279327393, 0.029296958819031715, -0.06990371644496918, 0.009181820787489414, 0.0195058211684227, -0.03118128329515457, 0.001035416848026216, 0.014971627853810787, -0.0777391716837883, -0.03601877763867378, -0.00462498189881444, 0.10573451966047287, -0.15904870629310608, 0.012398114427924156, 0.0838126391172409, -0.12594857811927795, 0.0813586562871933, -0.0006106876535341144, -0.01206875778734684, 0.022131776437163353, -0.14767099916934967, 0.06096983700990677, -0.00651735020801425, 0.005330943502485752, 0.022080490365624428, -0.20231451094150543, 0.0010611782781779766, -0.046166326850652695, -0.0580565482378006, -0.006821162533015013, -0.034208331257104874, -0.10881488770246506, 0.10119375586509705, 0.01840946450829506, -0.0807829275727272, -0.019118202850222588, 0.049314580857753754, 0.10984907299280167, -0.05423201248049736, 0.13843025267124176, -0.022093484178185463, 0.05561875179409981, -0.17508383095264435, -0.015010466799139977, -0.01884511485695839, 0.01675039529800415, -0.032699406147003174, -0.0063448576256632805, 0.053761400282382965, -0.021795762702822685, 0.23006084561347961, -0.03329315781593323, 0.022746775299310684, 0.0662616565823555, -0.007395898457616568, -0.02466614730656147, 0.09141410142183304, 0.05831921473145485, 0.019823938608169556, 0.023462723940610886, 0.009678727947175503, -0.051977336406707764, -0.011846045032143593, -0.1287335902452469, 0.08032830059528351, 0.17006289958953857, 0.0832807645201683, -0.0011417492059990764, 0.05661620944738388, -0.11824764311313629, -0.08884397894144058, 0.10315068811178207, -0.03696487843990326, -0.008325101807713509, -0.05479050800204277, 0.14003127813339233, 0.16284166276454926, -0.1792466789484024, 0.06529472023248672, -0.06703231483697891, -0.054111137986183167, -0.1079135313630104, -0.1702733039855957, -0.06385406106710434, -0.04134172946214676, -0.003200325183570385, -0.056672241538763046, 0.07026970386505127, 0.10425727069377899, 0.015394158661365509, 0.007145122159272432, 0.08924684673547745, -0.034410521388053894, 0.003967431839555502, 0.04615078866481781, 0.05031316727399826, 0.015370454639196396, -0.06289559602737427, 0.003805057378485799, 0.012086667120456696, 0.03619912639260292, 0.05767577514052391, 0.03358588367700577, -0.015441972762346268, 0.00826429296284914, -0.019517268985509872, -0.0962890237569809, 0.0407244898378849, -0.028659315779805183, -0.04762914776802063, 0.14599058032035828, 0.023316938430070877, -0.005744231399148703, -0.019850272685289383, 0.22833019495010376, -0.06841307878494263, -0.08293036371469498, -0.13890130817890167, 0.1406106948852539, -0.04129096865653992, 0.054532211273908615, 0.048289187252521515, -0.10287833213806152, 0.031274814158678055, 0.14709845185279846, 0.14302049577236176, -0.028337303549051285, 0.01196619775146246, 0.009999874047935009, 0.005250520538538694, -0.026724260300397873, 0.052909236401319504, 0.049603480845689774, 0.12155342847108841, -0.06124946475028992, 0.09144628793001175, -0.0038096080534160137, -0.08695073425769806, -0.01940424181520939, 0.13583695888519287, -0.001434069243259728, 0.020704632624983788, -0.08129720389842987, 0.11675985902547836, -0.06527755409479141, -0.2561015188694, 0.060353249311447144, -0.06762448698282242, -0.14944049715995789, -0.018578823655843735, 0.027211744338274002, 0.0003355915832798928, 0.021279368549585342, 0.06146527826786041, -0.06275594234466553, 0.15064457058906555, 0.03758588433265686, -0.07729688286781311, -0.07095571607351303, 0.07545747607946396, -0.0798204317688942, 0.2952599823474884, 0.007051850203424692, 0.05692324787378311, 0.09223286807537079, -0.033274851739406586, -0.1323377937078476, 0.049896061420440674, 0.09064158797264099, -0.06194010376930237, 0.06410481035709381, 0.20840007066726685, -0.011975160799920559, 0.12260035425424576, 0.07416624575853348, -0.08735647797584534, 0.05223854258656502, -0.07405798882246017, -0.09430453926324844, -0.08655916899442673, 0.08934324234724045, -0.06278510391712189, 0.15317323803901672, 0.12562185525894165, -0.04725475609302521, 0.0027636797167360783, -0.025733815506100655, 0.054841578006744385, -0.0038393251597881317, 0.11300427466630936, 0.026762498542666435, -0.19724777340888977, 0.03347480297088623, -0.01826278306543827, 0.10099007189273834, -0.2592698633670807, -0.08135145157575607, 0.039587851613759995, -0.009570525959134102, -0.05378785356879234, 0.11855222284793854, 0.06144152209162712, 0.04968099668622017, -0.0558135025203228, -0.05388732627034187, 0.0009833982912823558, 0.1646765172481537, -0.10682281851768494, -0.0031281758565455675 ]
null
null
stable-baselines3
# **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
{"library_name": "stable-baselines3", "tags": ["LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "PPO", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "LunarLander-v2", "type": "LunarLander-v2"}, "metrics": [{"type": "mean_reward", "value": "249.46 +/- 23.16", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
jacobemmerson/ppo-LunarLander-v2
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
2024-02-15T03:30:52+00:00
[]
[]
TAGS #stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
# PPO Agent playing LunarLander-v2 This is a trained model of a PPO agent playing LunarLander-v2 using the stable-baselines3 library. ## Usage (with Stable-baselines3) TODO: Add your code
[ "# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.", "## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ "TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n", "# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.", "## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ 39, 41, 17 ]
[ "passage: TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ 0.03942384943366051, 0.04900386184453964, -0.005304091144353151, 0.026427261531352997, 0.107408307492733, -0.026511888951063156, 0.11188238859176636, 0.0814051404595375, 0.10722193866968155, 0.04762078449130058, 0.08338645845651627, 0.06030960753560066, 0.05080918222665787, 0.2571701407432556, 0.04754156619310379, -0.22987541556358337, 0.036159250885248184, -0.04869936779141426, 0.12395193427801132, 0.07178173214197159, -0.0038484656251966953, -0.06485428661108017, 0.020415637642145157, -0.013290755450725555, 0.05367108806967735, 0.04282612353563309, -0.01716216839849949, -0.08207534998655319, 0.07169748842716217, -0.06345846503973007, 0.06986866891384125, 0.07677983492612839, 0.13218913972377777, -0.17832116782665253, 0.029566360637545586, 0.02571309357881546, -0.07189024239778519, 0.01342033501714468, 0.008019951172173023, 0.05120139941573143, 0.17303818464279175, 0.019879888743162155, 0.07844575494527817, -0.0025605305563658476, -0.15412317216396332, -0.018950799480080605, 0.0436202734708786, 0.12546207010746002, 0.08808347582817078, 0.04605821147561073, 0.01970590092241764, 0.17503218352794647, -0.054352790117263794, -0.028833400458097458, 0.21759237349033356, -0.2881564497947693, -0.031460098922252655, 0.321048766374588, 0.06997483223676682, 0.09725230932235718, -0.07540661096572876, -0.03619609400629997, 0.007783263456076384, -0.013137873262166977, -0.028666524216532707, -0.07447073608636856, 0.17313385009765625, 0.05152064561843872, -0.05057951435446739, -0.09541505575180054, 0.16948209702968597, 0.006921638268977404, 0.0018855923553928733, -0.019282981753349304, 0.009060598909854889, 0.07402525842189789, -0.016097044572234154, -0.07255112379789352, 0.057438433170318604, 0.05330665782094002, 0.019649166613817215, -0.1435653269290924, -0.10762494057416916, -0.022740179672837257, -0.008012006990611553, 0.17786912620067596, -0.009255532175302505, 0.042902372777462006, 0.003065188182517886, 0.10384012013673782, -0.12480384111404419, -0.03354184702038765, -0.0454259067773819, -0.07565800100564957, -0.0223417766392231, -0.02058211714029312, -0.03580251708626747, 0.07184842973947525, 0.11971849203109741, 0.027368178591132164, 0.09350208193063736, 0.047715865075588226, -0.03206788748502731, 0.06343851238489151, 0.05555703118443489, 0.14222665131092072, 0.05807621404528618, 0.012854371219873428, 0.13179877400398254, 0.055213116109371185, 0.033023182302713394, -0.0613492950797081, -0.18252409994602203, 0.07489913702011108, -0.07031869143247604, 0.007941240444779396, 0.12051256000995636, -0.04480670019984245, -0.1183447614312172, -0.037500523030757904, -0.017392054200172424, -0.06224250793457031, -0.025395862758159637, 0.0547584593296051, -0.02883218228816986, -0.03973718360066414, 0.0011496668448671699, 0.09384800493717194, 0.00953749567270279, -0.1752052903175354, 0.03303423151373863, -0.025042934343218803, -0.10782608389854431, 0.009975161403417587, 0.0022444494534283876, 0.03394931182265282, 0.04408763721585274, -0.11822668462991714, -0.30899152159690857, -0.07652641832828522, 0.05490870401263237, -0.06516939401626587, -0.18425025045871735, -0.13193942606449127, 0.02454492449760437, -0.09037084132432938, -0.044885024428367615, -0.12759265303611755, -0.028549788519740105, 0.01743689924478531, 0.011519349180161953, 0.10758619755506516, -0.0106219332665205, -0.012188062071800232, -0.1571401208639145, 0.008273907005786896, -0.20951123535633087, 0.0890483483672142, -0.019150104373693466, 0.037884220480918884, -0.032381169497966766, -0.07404014468193054, 0.030707746744155884, 0.052499737590551376, -0.01474119070917368, 0.13510210812091827, -0.15592676401138306, -0.03691192343831062, -0.007996266707777977, -0.13611900806427002, -0.04786273464560509, -0.10358831286430359, -0.04357128217816353, 0.13354332745075226, 0.018664736300706863, 0.15356586873531342, -0.08709818124771118, -0.0722038671374321, 0.20489206910133362, -0.010411538183689117, -0.12820468842983246, -0.076752208173275, 0.10165707021951675, 0.021510310471057892, -0.056606587022542953, -0.02523270808160305, -0.1839766949415207, -0.0152357779443264, -0.04550420492887497, -0.047039128839969635, 0.01796751655638218, -0.010888241231441498, 0.13837894797325134, 0.08494598418474197, 0.05018039792776108, -0.06086122244596481, -0.006730288732796907, 0.10779471695423126, 0.08823856711387634, 0.008680110797286034, 0.023406028747558594, -0.05774238705635071, 0.09552932530641556, -0.04003755748271942, -0.0142367510125041, -0.08283266425132751, -0.036246106028556824, -0.026256313547492027, 0.17507147789001465, 0.09440762549638748, 0.2257927656173706, 0.09567736834287643, 0.039160262793302536, 0.031270865350961685, -0.13181598484516144, -0.1425403207540512, -0.0017254541162401438, 0.09020978957414627, -0.14270411431789398, -0.04119925573468208, -0.08974775671958923, -0.17768175899982452, -0.12202505767345428, 0.0006432619411498308, -0.17960017919540405, 0.06390921026468277, 0.05408334732055664, -0.035177867859601974, 0.03272094577550888, 0.13032332062721252, -0.011533179320394993, -0.03967514634132385, 0.0831870287656784, 0.0379033200442791, -0.041234664618968964, -0.021742934361100197, 0.11885567009449005, 0.15673065185546875, 0.13124459981918335, -0.03511447086930275, 0.004914294462651014, 0.07076404243707657, -0.02309088408946991, 0.06539414077997208, 0.0558244064450264, 0.20973342657089233, 0.188301220536232, 0.038996949791908264, 0.008822928182780743, -0.07048165798187256, 0.0855446457862854, -0.0742373839020729, -0.14302679896354675, -0.05579735338687897, 0.08729292452335358, 0.016605578362941742, 0.023469142615795135, 0.08711627870798111, 0.024545932188630104, 0.09132762253284454, 0.15968108177185059, 0.01990218088030815, -0.09659269452095032, -0.050218869000673294, 0.01175848301500082, 0.027713103219866753, 0.04794301092624664, -0.04514073207974434, -0.00937939714640379, 0.017020760104060173, -0.10303554683923721, 0.031789086759090424, -0.1413339376449585, -0.1358717679977417, 0.044326696544885635, 0.003906996920704842, 0.010907664895057678, 0.02786896750330925, -0.0038291432429105043, 0.019039705395698547, 0.04351753741502762, -0.06975466758012772, 0.047416772693395615, -0.024745507165789604, -0.020031947642564774, 0.03340689837932587, -0.057257164269685745, -0.205775648355484, -0.17696654796600342, 0.00013708483311347663, -0.09910997003316879, 0.10194740444421768, 0.018308809027075768, -0.12373185902833939, 0.047737859189510345, -0.05822649225592613, 0.027574289590120316, -0.01875593699514866, -0.049130141735076904, 0.10507171601057053, 0.1525275856256485, -0.016146350651979446, 0.018018173053860664, -0.04865182936191559, -0.10157987475395203, -0.19632206857204437, 0.0691583976149559, 0.04680244252085686, 0.014610917307436466, 0.10669491440057755, 0.018072687089443207, 0.02367905154824257, -0.007674071006476879, -0.016521066427230835, -0.011659215204417706, -0.08781040459871292, 0.31909599900245667, 0.04510033503174782, -0.025173069909214973, 0.02041010931134224, -0.0043001663871109486, -0.028083480894565582, 0.03263787180185318, -0.0985708013176918, -0.07548979669809341, -0.08774089068174362, -0.04367410019040108, -0.09784720093011856, 0.053299110382795334, 0.05916472524404526, 0.003188040340319276, -0.07727594673633575, 0.04221395403146744, 0.11369874328374863, -0.0923808291554451, -0.07137343287467957, 0.07477962225675583, 0.0972946360707283, -0.07331304252147675, 0.00012658814375754446, 0.00874367356300354, 0.023951783776283264, 0.037102166563272476, 0.06778035312891006, -0.03966575115919113, 0.08589404821395874, -0.19917890429496765, 0.0372927263379097, 0.106058269739151, 0.023754918947815895, 0.0638108178973198, 0.07643651217222214, -0.1058402881026268, -0.008500572293996811, -0.032518330961465836, -0.21341575682163239, 0.1668180525302887, 0.1355515867471695, 0.06788124144077301, -0.025637222453951836, -0.00461410591378808, -0.0649740919470787, 0.05773647129535675, 0.02723747305572033, -0.14758841693401337, 0.004883295856416225, 0.06064270809292793, 0.026899009943008423, 0.01614922471344471, 0.07971042394638062, 0.014697225764393806, -0.1801026314496994, -0.014406266622245312, 0.10730406641960144, 0.002390873385593295, 0.0053148469887673855, -0.03175045922398567, -0.1755964607000351, 0.0751047357916832, 0.004285442177206278, 0.07233936339616776, -0.1676585078239441, 0.14297930896282196, -0.10089799761772156, 0.07726949453353882, -0.004285062663257122, -0.021311495453119278, 0.02507244050502777, -0.0541163794696331, 0.15163759887218475, 0.01058570109307766, -0.021810131147503853, -0.1200498715043068, -0.1717042326927185, -0.019227758049964905, -0.11788936704397202, -0.11679866164922714, 0.050424277782440186, 0.062185097485780716, 0.04923136904835701, -0.061147067695856094, 0.1518532931804657, -0.047422297298908234, 0.060713399201631546, -0.06893875449895859, -0.06755045056343079, 0.03764858841896057, -0.12588608264923096, -0.08176055550575256, 0.05573027580976486, 0.19166934490203857, 0.15833087265491486, -0.02816431224346161, -0.03472423925995827, -0.047419581562280655, -0.006212298292666674, -0.007802055217325687, 0.0275666993111372, 0.023223137483000755, 0.07315318286418915, -0.07681374251842499, -0.11649256944656372, 0.033787861466407776, -0.06713802367448807, -0.055589709430933, -0.015439179725944996, 0.1513158082962036, 0.04671623185276985, 0.07720734924077988, -0.018946662545204163, 0.03887668624520302, -0.001724981120787561, -0.056474871933460236, 0.16197094321250916, 0.03885216265916824, -0.05193585529923439, 0.06837689876556396, 0.053174007683992386, 0.043745119124650955, 0.03011113777756691, -0.026783017441630363, 0.206032395362854, 0.1980147808790207, 0.014206883497536182, 0.2175983190536499, 0.03177616000175476, -0.03772832080721855, -0.1300560086965561, -0.065880686044693, -0.006372632458806038, 0.03559038043022156, 0.08070417493581772, -0.18207235634326935, -0.015011128038167953, -0.05689644813537598, -0.034518610686063766, -0.15059494972229004, -0.28553900122642517, -0.05957856774330139, 0.20075850188732147, 0.14706264436244965, 0.27519428730010986, -0.10432573407888412, 0.035197313874959946, 0.02663275972008705, -0.04912831634283066, -0.006501141935586929, 0.00018665487004909664, 0.10268618166446686, -0.15421873331069946, 0.1176437959074974, 0.08486983180046082, -0.019002694636583328, 0.01058861706405878, -0.1619086116552353, 0.00936629343777895, -0.12191236019134521, 0.05354422330856323, 0.1400289237499237, -0.048128653317689896, -0.054873593151569366, 0.14033560454845428, -0.024562934413552284, -0.22685599327087402, -0.04648222774267197, -0.043600670993328094, -0.010640020482242107, 0.026607351377606392, -0.1013401448726654, 0.04101909324526787, 0.1330099105834961, 0.009380043484270573, 0.1147187277674675, 0.11749245226383209, -0.052566803991794586, 0.10792597383260727, 0.2257719188928604, -0.018785694614052773, 0.04689010605216026, -0.12743118405342102, -0.0012336712097749114, -0.028270328417420387, 0.013657891191542149, -0.09504974633455276, -0.09938385337591171, 0.02366873063147068, 0.02872389927506447, 0.009118586778640747, 0.0921793207526207, -0.029922157526016235, 0.0759170651435852, 0.06817561388015747, -0.13014446198940277, -0.16288450360298157, 0.015828335657715797, -0.007344507612287998, 0.08354310691356659, 0.00027861111448146403, 0.08878035843372345, -0.11932205408811569, -0.018093237653374672, -0.03153328225016594, -0.03319635987281799, -0.130486860871315, -0.07138993591070175, 0.06156524643301964, 0.028095467016100883, -0.06602972000837326, 0.1398407518863678, 0.026440169662237167, 0.15942534804344177, 0.049197953194379807, 0.012499804608523846, 0.07227300107479095, -0.05345509201288223, 0.1283530443906784, 0.13818155229091644, -0.00868943240493536, -0.05460423603653908, -0.1013643890619278, -0.10236792266368866, 0.08925779908895493, -0.05773641914129257, 0.07476430386304855, -0.14885357022285461, -0.06675903499126434, 0.015772046521306038, 0.016141414642333984, -0.09562095999717712, 0.02571965754032135, -0.01625603251159191, -0.18119946122169495, 0.056570518761873245, -0.048285093158483505, 0.0440407395362854, -0.06347788125276566, -0.1110161691904068, -0.17226378619670868, 0.06091433763504028, 0.08593481779098511, -0.053876690566539764, -0.12229149043560028, 0.011023230850696564, -0.00012518465518951416, -0.06341652572154999, -0.05023367330431938, 0.09722746908664703, -0.11020902544260025, 0.031452205032110214, -0.012567701749503613, 0.08853451162576675, -0.03510405123233795, -0.011538895778357983, 0.044220831245183945, -0.08039166033267975, -0.009481523185968399, 0.03534642979502678, -0.026372017338871956, -0.04127239063382149, -0.2689029574394226, 0.0036654395516961813, 0.0341104120016098, 0.02497158572077751, 0.07856601476669312, 0.011906822212040424, 0.021174922585487366, 0.03993808850646019, -0.15396519005298615, -0.013395369984209538, 0.14574195444583893, -0.07689505815505981, -0.022186370566487312, 0.05703273415565491, -0.09054436534643173, 0.013882770203053951, -0.030287226662039757, 0.1345842480659485, 0.023923413828015327, 0.06404478847980499, -0.0851147472858429, 0.10106813907623291, -0.1451139897108078, -0.04998219385743141, -0.01244612317532301, 0.09761348366737366, 0.07019034773111343, -0.10272270441055298, 0.014697125181555748, 0.04210108891129494, 0.19416837394237518, 0.016384804621338844, -0.0356343574821949, -0.03396720811724663, 0.004015897400677204, 0.22076453268527985, 0.03044266067445278, 0.10457023978233337, 0.07281364500522614, -0.026583973318338394, 0.12624378502368927, 0.09929762035608292, 0.11280370503664017, -0.055645186454057693, 0.13904185593128204, 0.04667386785149574, 0.038641396909952164, 0.0614289753139019, 0.06836545467376709, 0.09098632633686066, -0.0008288522367365658, 0.1138714924454689, 0.013811973854899406, -0.02422109805047512, -0.021335409954190254, 0.17759373784065247, 0.10501719266176224, -0.14769648015499115, 0.029047364369034767, -0.01258957851678133, 0.039933037012815475, -0.014194529503583908, -0.15634691715240479, -0.07240267097949982, -0.3315149247646332, 0.1226184144616127, -0.07119352370500565, 0.019930170848965645, 0.007913772016763687, -0.037425633519887924, -0.03296699747443199, -0.04477746784687042, 0.13151589035987854, -0.013641550205647945, -0.006079165264964104, -0.04815853759646416, -0.015360191464424133, -0.11607866734266281, -0.11200575530529022, -0.013207737356424332, -0.13671602308750153, -0.010119039565324783, 0.05595948174595833, 0.003977729007601738, 0.01821410097181797, -0.03142618387937546, 0.0024383175186812878, 0.06541839241981506, -0.05751744285225868, 0.056182678788900375, 0.12097269296646118, 0.08766137808561325, -0.1058853268623352, 0.031048951670527458, 0.2011747509241104, 0.04359564557671547, -0.12483977526426315, 0.01449228823184967, 0.1819491684436798, 0.004885740112513304, 0.017068125307559967, -0.006097703706473112, -0.0540788508951664, -0.07554277032613754, 0.1251034289598465, 0.08296554535627365, -0.09985227137804031, 0.015833314508199692, -0.0726347416639328, -0.01594804972410202, -0.06374675035476685, 0.10130585730075836, 0.09538925439119339, 0.04440245032310486, -0.10621760785579681, -0.08487539738416672, -0.10891728103160858, 0.040588874369859695, -0.08629853278398514, -0.07311757653951645, 0.09629398584365845, -0.07057105004787445, -0.07029950618743896, 0.025521177798509598, -0.17978744208812714, -0.009467960335314274, 0.1711762249469757, -0.24654000997543335, -0.0916430801153183, -0.10857923328876495, 0.14477859437465668, 0.016497576609253883, 0.1013975441455841, -0.006207061931490898, -0.007889035157859325, -0.20577777922153473, 0.024890204891562462, -0.05293011665344238, -0.02073732763528824, 0.07814782857894897, -0.09476397186517715, 0.22629831731319427, -0.08276885002851486, 0.020940175279974937, 0.012659613974392414, 0.0870661810040474, -0.030675338581204414, 0.09283176809549332, -0.03660329803824425, -0.12576518952846527, -0.03620953485369682, 0.03001813031733036, 0.013904244638979435, 0.10071761906147003, 0.09772487729787827, -0.03414725139737129, 0.03389119729399681, 0.09747414290904999, 0.04172342270612717, -0.023843804374337196, 0.0360250361263752, -0.17077107727527618, 0.02182629331946373, -0.018498148769140244, -0.06935930997133255, 0.03687669709324837, -0.06603235751390457, 0.1639697551727295, 0.04022442549467087, 0.0670473501086235, -0.036152735352516174, 0.0073931049555540085, -0.014454689808189869, -0.013775371946394444, -0.026180334389209747, -0.17259705066680908, -0.10422050207853317, -0.1347656100988388, -0.012701659463346004, -0.034971047192811966, 0.04591470584273338, 0.023234914988279343, -0.0003200018545612693, -0.014577031135559082, -0.12090865522623062, 0.04360328987240791, 0.11146783083677292, -0.04631396010518074, -0.026193076744675636 ]
null
null
peft
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.7.1
{"library_name": "peft", "base_model": "codeparrot/codeparrot"}
null
adalib/megengine-cond-gen-codeparrot-prefix
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:codeparrot/codeparrot", "region:us" ]
2024-02-15T03:32:37+00:00
[ "1910.09700" ]
[]
TAGS #peft #safetensors #arxiv-1910.09700 #base_model-codeparrot/codeparrot #region-us
# Model Card for Model ID ## Model Details ### Model Description - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact ### Framework versions - PEFT 0.7.1
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.7.1" ]
[ "TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-codeparrot/codeparrot #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.7.1" ]
[ 35, 6, 3, 54, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4, 11 ]
[ "passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-codeparrot/codeparrot #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.7.1" ]
[ -0.10329236835241318, 0.1991094946861267, -0.0035496081691235304, 0.0320764034986496, 0.0945892333984375, 0.018722474575042725, 0.04824448376893997, 0.12525925040245056, -0.042864445596933365, 0.10972880572080612, 0.06623604893684387, 0.10955594480037689, 0.10282865911722183, 0.2034357488155365, 0.007155488710850477, -0.1959189474582672, 0.02770180068910122, -0.09508345276117325, -0.009301785379648209, 0.12406817823648453, 0.15033432841300964, -0.09707682579755783, 0.07696907222270966, -0.016842491924762726, -0.01898537389934063, -0.036891840398311615, -0.0783899649977684, -0.032675329595804214, 0.0423843115568161, 0.04027989134192467, 0.06320202350616455, -0.0010284704621881247, 0.08786033093929291, -0.2682170271873474, 0.01724882982671261, 0.046349991112947464, -0.005619754083454609, 0.08398384600877762, 0.09925679117441177, -0.04155835881829262, 0.11729296296834946, -0.03429543226957321, 0.1447087824344635, 0.07977385073900223, -0.09631329029798508, -0.20587898790836334, -0.0711367204785347, 0.07684213668107986, 0.1776040941476822, 0.07997488230466843, -0.044256117194890976, 0.13360482454299927, -0.09620600193738937, 0.018231961876153946, 0.04062868654727936, -0.0796104297041893, -0.07343902438879013, 0.06711498647928238, 0.10942767560482025, 0.053163494914770126, -0.13837280869483948, -0.03540616109967232, 0.024569101631641388, 0.04050806909799576, 0.0760534256696701, 0.019766200333833694, 0.1391512155532837, 0.02985544688999653, -0.1491759866476059, -0.04452425613999367, 0.12775851786136627, 0.026105333119630814, -0.03582258149981499, -0.22357040643692017, 0.0049335164949297905, -0.08705339580774307, -0.02808760292828083, -0.050879329442977905, 0.03442491590976715, 0.004562058951705694, 0.08835969120264053, -0.024290746077895164, -0.09167808294296265, -0.011855660937726498, 0.09715835005044937, 0.043023012578487396, 0.023720107972621918, -0.0279318206012249, 0.002449614927172661, 0.12286990135908127, 0.059708040207624435, -0.12925924360752106, -0.05743660032749176, -0.06734061986207962, -0.04403950646519661, -0.047377679497003555, 0.03163853660225868, 0.027819188311696053, 0.059605587273836136, 0.2507312297821045, -0.024732042104005814, 0.05848440155386925, 0.054291076958179474, 0.018252437934279442, 0.044734321534633636, 0.0977662056684494, -0.05256020277738571, -0.1497306376695633, -0.018936388194561005, 0.09864485263824463, -0.009359381161630154, -0.02077452838420868, -0.04808277264237404, 0.04318903386592865, 0.04737257957458496, 0.10851886123418808, 0.09459429979324341, -0.0035771166440099478, -0.07812892645597458, -0.04933774471282959, 0.21165241301059723, -0.15036462247371674, 0.042139485478401184, 0.022374821826815605, -0.014428223483264446, -0.06152331084012985, 0.007106795907020569, 0.018991848453879356, -0.024411004036664963, 0.09467687457799911, -0.06350524723529816, -0.03970067575573921, -0.12008386850357056, -0.020254241302609444, 0.037786852568387985, 0.012760885991156101, -0.02904166653752327, -0.029857967048883438, -0.05983944982290268, -0.09628355503082275, 0.10356172174215317, -0.06480321288108826, -0.0610169991850853, -0.03797745332121849, -0.09993628412485123, 0.020773757249116898, 0.03109990805387497, 0.10628335922956467, -0.02319858781993389, 0.04236413165926933, -0.013732589781284332, 0.06189176067709923, 0.08645255863666534, 0.036228246986866, -0.07309529185295105, 0.06289207190275192, -0.19817276298999786, 0.09132614731788635, -0.07704638689756393, 0.02864651195704937, -0.15782223641872406, -0.013861071318387985, 0.006248894613236189, 0.02022254467010498, 0.03507586196064949, 0.15412302315235138, -0.2035851627588272, -0.024532441049814224, 0.16218020021915436, -0.09985975176095963, -0.11708329617977142, 0.03965356573462486, -0.05680323764681816, 0.16186797618865967, 0.019650926813483238, -0.009927471168339252, 0.08559273183345795, -0.1474023312330246, -0.025457140058279037, -0.03177822753787041, -0.004023641347885132, 0.10740065574645996, 0.08121296018362045, -0.08140331506729126, 0.033494699746370316, 0.016097290441393852, -0.042079586535692215, -0.03229745477437973, -0.05116637796163559, -0.111270472407341, 0.0032467953860759735, -0.08391474932432175, 0.029999537393450737, -0.007374199107289314, -0.07686883956193924, -0.015222935006022453, -0.16282664239406586, -0.02819419465959072, 0.07947083562612534, 0.01443477813154459, -0.019197700545191765, -0.09322435408830643, 0.03170805424451828, -0.02483070269227028, -0.023633912205696106, -0.15835340321063995, -0.029325641691684723, 0.02153729274868965, -0.14053747057914734, 0.014486008323729038, -0.11843785643577576, 0.06604547053575516, 0.013066625222563744, -0.06505228579044342, -0.03554535657167435, -0.011915669776499271, 0.009852970018982887, -0.049514275044202805, -0.23495401442050934, -0.020242391154170036, -0.05747365579009056, 0.15570616722106934, -0.22586597502231598, 0.04150522127747536, 0.042241960763931274, 0.1306493878364563, 0.0035597607493400574, -0.0632866621017456, 0.02558288909494877, -0.06882250308990479, -0.02370600588619709, -0.0716695562005043, -0.005979429464787245, -0.002668083878234029, -0.03843466192483902, 0.009915197268128395, -0.11917467415332794, -0.05634033679962158, 0.10177865624427795, 0.05994154512882233, -0.17016547918319702, -0.015063108876347542, -0.04285911098122597, -0.06520961970090866, -0.08418621122837067, -0.06258635222911835, 0.1028047502040863, 0.05036497488617897, 0.037443552166223526, -0.07239457964897156, -0.07109133899211884, 0.00712712574750185, -0.023250464349985123, -0.02447434514760971, 0.11420147866010666, 0.06716551631689072, -0.11119965463876724, 0.09866861253976822, 0.07664984464645386, 0.029714545235037804, 0.081314817070961, -0.026460956782102585, -0.10333805531263351, -0.02987206168472767, 0.047711081802845, 0.012710303999483585, 0.1535472720861435, -0.07919952273368835, 0.04847134277224541, 0.04236757382750511, -0.03324655070900917, 0.0498337596654892, -0.09778207540512085, 0.012127798050642014, 0.005471101496368647, -0.010423904284834862, 0.019459696486592293, -0.02341809868812561, 0.012817231938242912, 0.08085740357637405, 0.05212326720356941, 0.0306490957736969, 0.03414905071258545, -0.03192135691642761, -0.12666679918766022, 0.180520698428154, -0.10406907647848129, -0.2341674566268921, -0.15631847083568573, 0.05662774667143822, 0.05440960079431534, -0.01599929668009281, 0.025788674131035805, -0.05863429233431816, -0.10053522139787674, -0.07932037115097046, -0.00039412869955413043, 0.02934226021170616, -0.06844586133956909, -0.0697588250041008, 0.05478376895189285, 0.04027644917368889, -0.11884526908397675, 0.03800419718027115, 0.06230232119560242, -0.01963106170296669, 0.00817993376404047, 0.05443860962986946, 0.08367152512073517, 0.1851641833782196, -0.009397785179316998, -0.005272079724818468, 0.05240032821893692, 0.27972644567489624, -0.15969666838645935, 0.11143927276134491, 0.11566600203514099, -0.0676729753613472, 0.079993836581707, 0.19276244938373566, 0.031111817806959152, -0.10343173891305923, 0.035727959126234055, 0.029416367411613464, -0.024281971156597137, -0.2687993347644806, -0.048981089144945145, -0.011912602931261063, -0.09494394809007645, 0.07778402417898178, 0.08550424128770828, 0.08882345259189606, 0.035288695245981216, -0.06378500908613205, -0.09618993103504181, 0.03923006355762482, 0.10156437009572983, -0.02127963677048683, 0.0045198858715593815, 0.08195331692695618, -0.028203364461660385, 0.0051971618086099625, 0.08904259651899338, -0.012480389326810837, 0.16496773064136505, 0.05493351072072983, 0.11151713877916336, 0.07972449064254761, 0.0957811251282692, -0.004405023064464331, 0.023521078750491142, 0.016099847853183746, 0.023084251210093498, 0.01289111003279686, -0.08184043318033218, 0.027636084705591202, 0.10693375021219254, 0.03995392471551895, 0.02337908186018467, 0.01462940126657486, -0.04491084814071655, 0.051204681396484375, 0.1860136240720749, 0.015520991757512093, -0.20027755200862885, -0.07545466721057892, 0.05730283632874489, -0.07728450745344162, -0.14053741097450256, -0.0176075492054224, 0.025777561590075493, -0.16871507465839386, 0.011927909217774868, -0.04473460465669632, 0.10090383887290955, -0.07111002504825592, -0.037563178688287735, 0.09582705795764923, 0.06932663172483444, -0.023215198889374733, 0.060129713267087936, -0.19539691507816315, 0.12666799128055573, 0.022838624194264412, 0.07298749685287476, -0.0895838588476181, 0.09916916489601135, -0.0012288594152778387, -0.011584814637899399, 0.1641952395439148, 0.001357188099063933, -0.07620725780725479, -0.06010977178812027, -0.08906254172325134, -0.014810220338404179, 0.10829322040081024, -0.13085880875587463, 0.06802394241094589, -0.0162927508354187, -0.030407987534999847, 0.004131825640797615, -0.08247561007738113, -0.12173120677471161, -0.1741877794265747, 0.05575304105877876, -0.10765613615512848, 0.03913215175271034, -0.09199798852205276, -0.0628759115934372, 0.007102193310856819, 0.1793217957019806, -0.18385601043701172, -0.09184932708740234, -0.14156349003314972, -0.09236942231655121, 0.16691496968269348, -0.03811681643128395, 0.08896732330322266, 0.0006371980998665094, 0.167173832654953, 0.013212242163717747, 0.0008053527562879026, 0.09846283495426178, -0.08838007599115372, -0.195245623588562, -0.060053374618291855, 0.16687048971652985, 0.1390485316514969, 0.03822357952594757, -0.008282543160021305, 0.028799647465348244, -0.049520693719387054, -0.10963472723960876, 0.02337036468088627, 0.130062997341156, 0.08295401930809021, -0.010147323831915855, -0.034744743257761, -0.10413122922182083, -0.06751642376184464, -0.05470778048038483, 0.004491682164371014, 0.19151616096496582, -0.07348428666591644, 0.15958790481090546, 0.12157517671585083, -0.05851850286126137, -0.2073076218366623, 0.04764602333307266, 0.05547701194882393, 0.010181306861341, 0.03388967365026474, -0.19430796802043915, 0.0895812064409256, -0.002211875980719924, -0.07187002897262573, 0.15506196022033691, -0.1751798391342163, -0.1419072151184082, 0.09645676612854004, 0.031242331489920616, -0.23179180920124054, -0.140110582113266, -0.10059773176908493, -0.020523587241768837, -0.11910218745470047, 0.06246551498770714, 0.004154243040829897, 0.013435576111078262, 0.03232221677899361, 0.022625509649515152, 0.0285483468323946, -0.05156245827674866, 0.20150147378444672, -0.021617839112877846, 0.0116057638078928, -0.05236828327178955, -0.09507346153259277, 0.03640175983309746, -0.050437383353710175, 0.09657271951436996, 0.0027802723925560713, 0.027580678462982178, -0.14331470429897308, -0.04037529602646828, -0.06119441241025925, 0.027803603559732437, -0.1000802218914032, -0.08828406035900116, -0.04697757214307785, 0.09533479064702988, 0.09622817486524582, -0.03175888583064079, 0.00981615949422121, -0.0842791348695755, 0.07180384546518326, 0.2082255482673645, 0.1852945238351822, 0.06956380605697632, -0.0607856884598732, 0.021769147366285324, -0.03267969191074371, 0.041049811989068985, -0.21977783739566803, 0.042970363050699234, 0.055601269006729126, 0.023724175989627838, 0.0868862122297287, -0.009627060033380985, -0.15081721544265747, -0.07427337765693665, 0.08015687763690948, -0.046447642147541046, -0.16000273823738098, -0.01915564388036728, 0.04663510248064995, -0.20957869291305542, -0.04341138154268265, 0.019573623314499855, -0.016891632229089737, -0.04244924709200859, 0.02295130304992199, 0.08179136365652084, -0.018676547333598137, 0.11014609783887863, 0.08830604702234268, 0.09282010793685913, -0.10115926712751389, 0.08067424595355988, 0.07793068140745163, -0.046769317239522934, 0.024064118042588234, 0.10837456583976746, -0.04830269142985344, -0.03586379438638687, 0.09512439370155334, 0.0946870818734169, 0.024960441514849663, -0.04589400440454483, 0.01367401797324419, -0.0520394966006279, 0.06197408214211464, 0.11672927439212799, 0.03141143545508385, -0.004231251776218414, 0.05709967017173767, 0.03799550607800484, -0.10058658570051193, 0.10852540284395218, 0.06427377462387085, 0.024908002465963364, -0.03996323049068451, -0.02511623315513134, -0.010927225463092327, -0.016850082203745842, -0.01710355468094349, -0.005908406805247068, -0.08984007686376572, -0.006262491457164288, -0.09741608798503876, 0.02910245582461357, -0.07385318726301193, 0.008803381584584713, 0.02880743332207203, -0.04584013670682907, 0.008316919207572937, 0.004094703122973442, -0.07695653289556503, -0.05422542989253998, -0.017286941409111023, 0.08707528561353683, -0.13027890026569366, 0.0341714583337307, 0.07813842594623566, -0.10886245220899582, 0.06746217608451843, -0.002395555842667818, 0.007175484672188759, 0.015258047729730606, -0.1680174022912979, 0.05064128711819649, -0.02307257056236267, -0.012751167640089989, 0.015237638726830482, -0.20685844123363495, -0.013277353718876839, -0.04568338766694069, -0.04917440563440323, 0.010579238645732403, -0.025905627757310867, -0.12403898686170578, 0.09993818402290344, -0.004685746040195227, -0.0767570436000824, -0.017627349123358727, 0.038374729454517365, 0.09996167570352554, -0.02670254558324814, 0.13602839410305023, -0.026359569281339645, 0.07776254415512085, -0.17095716297626495, -0.005339608062058687, -0.013490557670593262, 0.036647483706474304, -0.02355373091995716, -0.023600704967975616, 0.05970179662108421, -0.020983943715691566, 0.17308658361434937, -0.024301797151565552, 0.06980177015066147, 0.053884800523519516, 0.012619086541235447, 0.006580646615475416, 0.0890296995639801, 0.05994889512658119, -0.00224067154340446, -0.00451831566169858, 0.03458336740732193, -0.007273669820278883, -0.04424568638205528, -0.15858854353427887, 0.06816443055868149, 0.1642368882894516, 0.04969542473554611, 0.021598421037197113, 0.031035330146551132, -0.11431293189525604, -0.07415957748889923, 0.13565295934677124, -0.009394350461661816, -0.03296235576272011, -0.07538527995347977, 0.1824072301387787, 0.12872789800167084, -0.19966235756874084, 0.07852865755558014, -0.06783908605575562, -0.05593901127576828, -0.1232931911945343, -0.15602804720401764, -0.0648919865489006, -0.04340577498078346, -0.019624128937721252, -0.0639692172408104, 0.05613270401954651, 0.058813612908124924, 0.001425987109541893, -0.018078278750181198, 0.10630279779434204, 0.009662417694926262, -0.023386240005493164, 0.045132674276828766, 0.059963125735521317, 0.03376534953713417, -0.09632086008787155, 0.0096770990639925, -0.002340735634788871, 0.025244370102882385, 0.0662512481212616, 0.017782168462872505, -0.053059570491313934, 0.009303494356572628, -0.02173302322626114, -0.1160469576716423, 0.04477434605360031, -0.017120063304901123, -0.039619963616132736, 0.1449768990278244, 0.02904772013425827, 0.009236602112650871, -0.01978018693625927, 0.2358154058456421, -0.07457811385393143, -0.08457175642251968, -0.15541023015975952, 0.06076590716838837, -0.06565217673778534, 0.035359419882297516, 0.034571196883916855, -0.11856629699468613, 0.018512191250920296, 0.16088688373565674, 0.12946905195713043, -0.011538301594555378, 0.009198445826768875, 0.05150740593671799, 0.0027183787897229195, -0.03291062265634537, 0.017550481483340263, 0.054291024804115295, 0.14000749588012695, -0.07372044026851654, 0.06561611592769623, -0.01187864225357771, -0.07844166457653046, -0.021566620096564293, 0.10251908749341965, -0.0038683004677295685, 0.0015118604060262442, -0.06856850534677505, 0.14383848011493683, -0.084647536277771, -0.2348589301109314, 0.053169168531894684, -0.0701572448015213, -0.14949968457221985, -0.04537404328584671, 0.023992007598280907, -0.016428330913186073, 0.01799216866493225, 0.0795777291059494, -0.049674998968839645, 0.17259933054447174, 0.04154284670948982, -0.04437325522303581, -0.08154232054948807, 0.05796421319246292, -0.1414758861064911, 0.28577372431755066, 0.02069827914237976, 0.048000313341617584, 0.1059737354516983, -0.01667015440762043, -0.13934561610221863, 0.011938226409256458, 0.10723932832479477, -0.0637049525976181, 0.062433384358882904, 0.17652811110019684, -0.0022143926471471786, 0.12586885690689087, 0.0577392578125, -0.05767730250954628, 0.04133390262722969, -0.08329254388809204, -0.0484866127371788, -0.11132916063070297, 0.07939464598894119, -0.08110078424215317, 0.1582714021205902, 0.13221341371536255, -0.06497745960950851, -0.0005526122986339033, -0.02006004936993122, 0.08089762926101685, 0.0047617023810744286, 0.1109834760427475, 0.0031623311806470156, -0.1938789188861847, 0.03582930192351341, 0.013228845782577991, 0.10180895030498505, -0.2042633295059204, -0.06707853823900223, 0.050121136009693146, -0.018932392820715904, -0.07713542133569717, 0.11884412914514542, 0.04402478411793709, 0.03133934736251831, -0.03841261938214302, -0.04394432157278061, 0.0010597195941954851, 0.15094342827796936, -0.11137887090444565, -0.009867769666016102 ]
null
null
peft
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.8.2
{"library_name": "peft", "base_model": "distilbert-base-uncased"}
null
likhith231/distilbert-base-uncased-lora-text-classification
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:distilbert-base-uncased", "region:us" ]
2024-02-15T03:38:17+00:00
[ "1910.09700" ]
[]
TAGS #peft #safetensors #arxiv-1910.09700 #base_model-distilbert-base-uncased #region-us
# Model Card for Model ID ## Model Details ### Model Description - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact ### Framework versions - PEFT 0.8.2
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ "TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-distilbert-base-uncased #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ 37, 6, 3, 54, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4, 11 ]
[ "passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-distilbert-base-uncased #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2" ]
[ -0.10642906278371811, 0.20181091129779816, -0.003171114018186927, 0.033168528228998184, 0.09507043659687042, 0.018274547532200813, 0.0548069141805172, 0.12401581555604935, -0.02576358988881111, 0.11059535294771194, 0.06784230470657349, 0.09757009148597717, 0.10680077970027924, 0.21522867679595947, 0.004188282415270805, -0.20135881006717682, 0.029488584026694298, -0.09623736888170242, -0.005728109274059534, 0.12338162213563919, 0.1446412056684494, -0.0965484008193016, 0.07644246518611908, -0.018765466287732124, -0.007993046194314957, -0.033950090408325195, -0.07159552723169327, -0.02987894043326378, 0.04008631780743599, 0.04581599682569504, 0.05809962376952171, -0.0012055233819410205, 0.08583899587392807, -0.2676727771759033, 0.018071815371513367, 0.048218052834272385, -0.004933010786771774, 0.08598033338785172, 0.0956650972366333, -0.04368448629975319, 0.12286529690027237, -0.030787982046604156, 0.13996219635009766, 0.07888324558734894, -0.09355264902114868, -0.22597786784172058, -0.06760435551404953, 0.08847375959157944, 0.1777222603559494, 0.0742378681898117, -0.043942589312791824, 0.1265953928232193, -0.09065576642751694, 0.01613544300198555, 0.03740571439266205, -0.0929981991648674, -0.06976311653852463, 0.05096256360411644, 0.09874420613050461, 0.05627178028225899, -0.1344534009695053, -0.032968513667583466, 0.02470528893172741, 0.03357512503862381, 0.07723342627286911, 0.01369174662977457, 0.14301244914531708, 0.026139944791793823, -0.14684896171092987, -0.04074263200163841, 0.13414272665977478, 0.025651592761278152, -0.037009015679359436, -0.22616367042064667, 0.001814338262192905, -0.08017637580633163, -0.028750164434313774, -0.04971989989280701, 0.03853200376033783, 0.0029731655959039927, 0.10280165821313858, -0.026546383276581764, -0.08775345981121063, -0.009961317293345928, 0.09306603670120239, 0.05569102242588997, 0.02626071870326996, -0.019990984350442886, 0.008630656637251377, 0.12437444180250168, 0.054166510701179504, -0.13398972153663635, -0.05857621505856514, -0.07376765459775925, -0.04550224542617798, -0.04124477878212929, 0.04204322770237923, 0.044466860592365265, 0.05738210305571556, 0.2576381266117096, -0.03644370660185814, 0.05703691393136978, 0.0596281923353672, 0.018833592534065247, 0.0472378134727478, 0.09365973621606827, -0.05900712311267853, -0.1525956392288208, -0.013918329030275345, 0.10035113245248795, -0.004647709894925356, -0.02401573956012726, -0.0494675450026989, 0.050216738134622574, 0.03544064238667488, 0.10974179953336716, 0.09615515917539597, -0.0095453392714262, -0.07998356223106384, -0.053605325520038605, 0.20630496740341187, -0.1527041792869568, 0.04354526475071907, 0.024091683328151703, -0.01412655133754015, -0.04285199195146561, 0.011706131510436535, 0.016496645286679268, -0.02628028765320778, 0.09919089078903198, -0.06705133616924286, -0.03821440413594246, -0.11722107976675034, -0.026658009737730026, 0.031214091926813126, 0.009157467633485794, -0.030611516907811165, -0.03185737505555153, -0.06425211578607559, -0.09366980940103531, 0.10185589641332626, -0.0650702565908432, -0.05653331056237221, -0.03296969085931778, -0.08926854282617569, 0.021729907020926476, 0.026849959045648575, 0.08576913177967072, -0.025727059692144394, 0.04190802574157715, -0.011936672031879425, 0.06443411856889725, 0.08272659033536911, 0.03619356080889702, -0.07585514336824417, 0.06300991028547287, -0.19955949485301971, 0.08603131771087646, -0.08530089259147644, 0.028519131243228912, -0.15569865703582764, -0.015457026660442352, 0.006219884380698204, 0.022141801193356514, 0.03013981692492962, 0.1637963056564331, -0.20503884553909302, -0.035227734595537186, 0.15803740918636322, -0.10652005672454834, -0.11522184312343597, 0.039880555123090744, -0.04733574762940407, 0.1566607654094696, 0.021941838786005974, -0.00487925112247467, 0.09630709886550903, -0.1443040668964386, -0.025752970948815346, -0.02009568363428116, 0.001914201071485877, 0.10872694104909897, 0.08676699548959732, -0.08572939038276672, 0.027229517698287964, 0.015939727425575256, -0.05102347955107689, -0.029240470379590988, -0.04859967529773712, -0.10939077287912369, 0.007263127714395523, -0.079475536942482, 0.030086075887084007, -0.005909802857786417, -0.08101890981197357, -0.009804660454392433, -0.16427598893642426, -0.03456432744860649, 0.08188574761152267, 0.00747819896787405, -0.019764477387070656, -0.09755316376686096, 0.04408193379640579, -0.02614981308579445, -0.01907314918935299, -0.14950354397296906, -0.027902739122509956, 0.019898978993296623, -0.13543881475925446, 0.007094955071806908, -0.11675848066806793, 0.06520438194274902, 0.014599197544157505, -0.062338750809431076, -0.037288911640644073, -0.007358398754149675, 0.007872704416513443, -0.04671763256192207, -0.24220377206802368, -0.021076196804642677, -0.05476876720786095, 0.15700682997703552, -0.22456955909729004, 0.037365518510341644, 0.045232389122247696, 0.13221098482608795, 0.005514335352927446, -0.06407103687524796, 0.02784157358109951, -0.06464418023824692, -0.026861868798732758, -0.0737764984369278, -0.005808154586702585, -0.008678353391587734, -0.04358883574604988, 0.012248056009411812, -0.12053409963846207, -0.03639956936240196, 0.09942010045051575, 0.07044975459575653, -0.16162219643592834, -0.014850337989628315, -0.04183679446578026, -0.06100412830710411, -0.08166158199310303, -0.058178532868623734, 0.10895554721355438, 0.05198004096746445, 0.03864916041493416, -0.07245050370693207, -0.06819936633110046, 0.0075453962199389935, -0.02320593222975731, -0.02202790230512619, 0.11418948322534561, 0.07309940457344055, -0.11027286946773529, 0.10010849684476852, 0.07652165740728378, 0.03515646606683731, 0.08239977061748505, -0.02678629755973816, -0.10437919199466705, -0.02855013683438301, 0.0463380329310894, 0.015202938579022884, 0.15928642451763153, -0.06543345004320145, 0.052446555346250534, 0.044500190764665604, -0.03994492441415787, 0.04484312981367111, -0.09580379724502563, 0.010570613667368889, 0.008401244878768921, -0.013894101604819298, 0.01667194627225399, -0.022627925500273705, 0.009187576361000538, 0.08390801399946213, 0.05048397555947304, 0.038144368678331375, 0.026827530935406685, -0.03007824905216694, -0.12985895574092865, 0.18505901098251343, -0.1016729325056076, -0.24209021031856537, -0.15743498504161835, 0.06099282205104828, 0.0526563860476017, -0.016789134591817856, 0.021255699917674065, -0.05717365816235542, -0.10565439611673355, -0.08066041022539139, 0.0010020467452704906, 0.03207382187247276, -0.06029132008552551, -0.06816478818655014, 0.04478493332862854, 0.04737280681729317, -0.11628124117851257, 0.03611146658658981, 0.05645821616053581, -0.019917409867048264, 0.003947316203266382, 0.06104859709739685, 0.08649703115224838, 0.182611882686615, -0.0047232541255652905, -0.004901324398815632, 0.051133863627910614, 0.27613022923469543, -0.15885530412197113, 0.11113601922988892, 0.12105506658554077, -0.0699654296040535, 0.08003305643796921, 0.19354072213172913, 0.03655140474438667, -0.10087426751852036, 0.028004499152302742, 0.029024390503764153, -0.025858208537101746, -0.2630731761455536, -0.05187566950917244, -0.014416852965950966, -0.09166327118873596, 0.08017607033252716, 0.08952198177576065, 0.07505054026842117, 0.03694932162761688, -0.07088851928710938, -0.08914481103420258, 0.033991169184446335, 0.1004045382142067, -0.0189445149153471, 0.006192458793520927, 0.08409976959228516, -0.03308767452836037, 0.010179681703448296, 0.09776579588651657, -0.013416835106909275, 0.1659850925207138, 0.05307795852422714, 0.10575005412101746, 0.08123847097158432, 0.09194865822792053, -0.0030521738808602095, 0.025425586849451065, 0.014852665364742279, 0.022970914840698242, 0.013684424571692944, -0.08595077693462372, 0.0265091173350811, 0.11126265674829483, 0.039719607681035995, 0.03057960234582424, 0.014281084761023521, -0.04110376909375191, 0.05357813835144043, 0.17614562809467316, 0.014110242016613483, -0.20053185522556305, -0.07908113300800323, 0.06329239904880524, -0.07660087943077087, -0.1350620537996292, -0.014431549236178398, 0.03233424201607704, -0.16602414846420288, 0.025147894397377968, -0.04198816418647766, 0.09968524426221848, -0.08545852452516556, -0.03743759170174599, 0.09898914396762848, 0.0673462525010109, -0.023448187857866287, 0.05633781850337982, -0.192208930850029, 0.1246662512421608, 0.026139097288250923, 0.07037107646465302, -0.08278626203536987, 0.10046146810054779, 0.003525923006236553, 0.0008880468085408211, 0.1684543639421463, 0.002426974242553115, -0.060202792286872864, -0.06756467372179031, -0.09487012028694153, -0.013955187052488327, 0.0997828021645546, -0.13705089688301086, 0.06798042356967926, -0.01893150806427002, -0.028213167563080788, -0.0034978510811924934, -0.08678130060434341, -0.12728479504585266, -0.17321138083934784, 0.05607593059539795, -0.10304364562034607, 0.028876353055238724, -0.09206569939851761, -0.06289347261190414, 0.0017774669686332345, 0.17642326653003693, -0.21257634460926056, -0.09932458400726318, -0.14930382370948792, -0.08951093256473541, 0.15978512167930603, -0.044670723378658295, 0.08392763137817383, 0.000303533801343292, 0.16396617889404297, 0.013470211997628212, -0.012546534650027752, 0.10013853013515472, -0.09062289446592331, -0.19480407238006592, -0.05909644439816475, 0.16589365899562836, 0.1354351043701172, 0.03513338789343834, -0.01588025502860546, 0.02839021570980549, -0.04839238151907921, -0.11908963322639465, 0.024318436160683632, 0.14612165093421936, 0.06674274802207947, -0.00725489342585206, -0.033324845135211945, -0.10260003060102463, -0.06469228118658066, -0.04279598221182823, -0.002154805464670062, 0.18825852870941162, -0.07703451067209244, 0.16391661763191223, 0.11808646470308304, -0.05770591273903847, -0.21169929206371307, 0.04458777979016304, 0.05300869420170784, 0.014094795100390911, 0.04111030697822571, -0.19710838794708252, 0.08404745906591415, -0.0017245206981897354, -0.07244667410850525, 0.16619612276554108, -0.16870605945587158, -0.14085087180137634, 0.10117438435554504, 0.03114011324942112, -0.22067849338054657, -0.14122693240642548, -0.1030457466840744, -0.019498469308018684, -0.1135801300406456, 0.05505586788058281, 0.0011491802288219333, 0.011233721859753132, 0.026457449421286583, 0.013132721185684204, 0.02590205706655979, -0.04565008357167244, 0.20431210100650787, -0.02519236132502556, 0.009899839758872986, -0.05058844015002251, -0.08598287403583527, 0.02816963568329811, -0.047415487468242645, 0.09917830675840378, -0.0009602390346117318, 0.027781784534454346, -0.15344946086406708, -0.04022228345274925, -0.05285254493355751, 0.030747495591640472, -0.09488335251808167, -0.08714389055967331, -0.04619597643613815, 0.09410252422094345, 0.09447035938501358, -0.0279371477663517, 0.0028538312762975693, -0.0883597731590271, 0.07145485281944275, 0.20429152250289917, 0.18860343098640442, 0.07022745907306671, -0.06833101063966751, 0.020405905321240425, -0.03194950520992279, 0.04832802712917328, -0.23060695827007294, 0.04042166471481323, 0.057441458106040955, 0.02129385992884636, 0.08746905624866486, -0.011377048678696156, -0.15438617765903473, -0.07062256336212158, 0.0838589072227478, -0.05263335257768631, -0.16711686551570892, -0.025505101308226585, 0.028858793899416924, -0.20579379796981812, -0.04168994352221489, 0.02395344339311123, -0.02091827057301998, -0.037643495947122574, 0.02366614155471325, 0.0806458443403244, -0.018014954403042793, 0.10664575546979904, 0.08484898507595062, 0.09207899123430252, -0.10365059226751328, 0.07925679534673691, 0.07567957043647766, -0.045056380331516266, 0.026203760877251625, 0.11080143600702286, -0.049182258546352386, -0.03500420227646828, 0.07797082513570786, 0.09406089037656784, 0.029206477105617523, -0.051267821341753006, 0.011694391258060932, -0.05797962471842766, 0.06059711053967476, 0.10830764472484589, 0.02886979468166828, -0.00306878169067204, 0.05455809459090233, 0.03247230872511864, -0.09169252961874008, 0.11120402812957764, 0.060298267751932144, 0.01843075081706047, -0.04709484055638313, -0.03897586092352867, -0.006432607304304838, -0.017549943178892136, -0.02004108764231205, -0.00587001908570528, -0.09334774315357208, -0.008737697266042233, -0.09352020174264908, 0.02361086942255497, -0.07188036292791367, 0.010243745520710945, 0.02993376925587654, -0.05136583372950554, 0.003300947602838278, 0.00476469099521637, -0.07243837416172028, -0.04924301430583, -0.012973274104297161, 0.08386819064617157, -0.1333766132593155, 0.037369515746831894, 0.07378730922937393, -0.10521970689296722, 0.07364775985479355, -0.004519398789852858, 0.007868115790188313, 0.009385052137076855, -0.16200314462184906, 0.054981131106615067, -0.022758493199944496, -0.011731067672371864, 0.018162621185183525, -0.20100273191928864, -0.006914528086781502, -0.046346016228199005, -0.05719956010580063, 0.010202295146882534, -0.021600326523184776, -0.12367706745862961, 0.09591557830572128, -0.0010163950501009822, -0.06549273431301117, -0.018042253330349922, 0.03968352451920509, 0.10166357457637787, -0.025083132088184357, 0.13947436213493347, -0.028695717453956604, 0.07440628111362457, -0.17282968759536743, -0.009410020895302296, -0.013914361596107483, 0.03843492642045021, -0.025610219687223434, -0.023908289149403572, 0.0583910197019577, -0.0208873450756073, 0.17975354194641113, -0.01842864230275154, 0.07182954996824265, 0.05609835684299469, 0.01097019575536251, 0.00924734864383936, 0.08512815833091736, 0.05836591124534607, 0.002651509130373597, -0.0042807296849787235, 0.02488364651799202, -0.00937538780272007, -0.04169873520731926, -0.16244567930698395, 0.06580482423305511, 0.15236368775367737, 0.04796876013278961, 0.019513467326760292, 0.0301175806671381, -0.11302530020475388, -0.08041750639677048, 0.12689939141273499, -0.017520330846309662, -0.03338957577943802, -0.06816631555557251, 0.16909082233905792, 0.13617077469825745, -0.19893859326839447, 0.07581183314323425, -0.05424055457115173, -0.051029350608587265, -0.1339346021413803, -0.16724322736263275, -0.06132332980632782, -0.05156884714961052, -0.019458789378404617, -0.06648175418376923, 0.05232573673129082, 0.05664524808526039, 0.0025247661396861076, -0.014862428419291973, 0.10935600101947784, 0.005979019682854414, -0.024891825392842293, 0.05238298699259758, 0.06437598913908005, 0.032309357076883316, -0.09079143404960632, 0.008805512450635433, -0.001956923631951213, 0.01747080497443676, 0.06574554741382599, 0.01852973736822605, -0.052905768156051636, 0.017456458881497383, -0.020005786791443825, -0.11628905683755875, 0.04215237870812416, -0.013834195211529732, -0.04320378601551056, 0.14896385371685028, 0.0318441204726696, 0.007261664140969515, -0.02166076935827732, 0.23041598498821259, -0.07907618582248688, -0.07141468673944473, -0.15150022506713867, 0.07867314666509628, -0.06301809847354889, 0.03739922121167183, 0.030651435256004333, -0.11934886872768402, 0.015458629466593266, 0.15953953564167023, 0.13266761600971222, -0.012197006493806839, 0.011580032296478748, 0.0423787496984005, 0.005344280041754246, -0.03349979966878891, 0.021332580596208572, 0.05015050619840622, 0.1465596705675125, -0.06804472208023071, 0.06185370311141014, -0.009571387432515621, -0.0812913179397583, -0.021178606897592545, 0.104805126786232, -0.0016033141873776913, 0.0013996053021401167, -0.07416566461324692, 0.141665518283844, -0.08238261938095093, -0.2250383496284485, 0.06490501016378403, -0.07149378955364227, -0.14715908467769623, -0.048434432595968246, 0.025999682024121284, -0.013792021200060844, 0.010316209867596626, 0.07630991190671921, -0.052785634994506836, 0.17691689729690552, 0.04301740974187851, -0.04898673668503761, -0.08491872996091843, 0.05603376775979996, -0.14289546012878418, 0.27955886721611023, 0.021621407940983772, 0.042600829154253006, 0.10175727307796478, -0.016937239095568657, -0.1379525512456894, 0.012532075867056847, 0.10621775686740875, -0.06731253117322922, 0.05799940973520279, 0.17275801301002502, 0.0013690071646124125, 0.13053593039512634, 0.05946798250079155, -0.05745365470647812, 0.03824097663164139, -0.09482350945472717, -0.05388513579964638, -0.10701262205839157, 0.08242552727460861, -0.08393267542123795, 0.1611090451478958, 0.12774698436260223, -0.06627994030714035, -0.0048686182126402855, -0.02002224512398243, 0.08361531049013138, 0.008528433740139008, 0.1187901496887207, 0.014217148534953594, -0.18475468456745148, 0.03394443914294243, 0.010290835984051228, 0.10374479740858078, -0.20123466849327087, -0.059548329561948776, 0.04558917135000229, -0.01629858836531639, -0.07929091155529022, 0.12256954610347748, 0.04238631948828697, 0.034747932106256485, -0.0424092672765255, -0.05730847641825676, 0.009246394969522953, 0.14366647601127625, -0.1078328788280487, -0.004935360047966242 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-finetuned-squad This model is a fine-tuned version of [loganunger/bert-finetuned-squad](https://huggingface.co/loganunger/bert-finetuned-squad) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
{"tags": ["generated_from_trainer"], "base_model": "loganunger/bert-finetuned-squad", "model-index": [{"name": "bert-finetuned-squad", "results": []}]}
question-answering
loganunger/bert-finetuned-squad
[ "transformers", "tensorboard", "safetensors", "bert", "question-answering", "generated_from_trainer", "base_model:loganunger/bert-finetuned-squad", "endpoints_compatible", "region:us" ]
2024-02-15T03:42:37+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #bert #question-answering #generated_from_trainer #base_model-loganunger/bert-finetuned-squad #endpoints_compatible #region-us
# bert-finetuned-squad This model is a fine-tuned version of loganunger/bert-finetuned-squad on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
[ "# bert-finetuned-squad\n\nThis model is a fine-tuned version of loganunger/bert-finetuned-squad on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.2" ]
[ "TAGS\n#transformers #tensorboard #safetensors #bert #question-answering #generated_from_trainer #base_model-loganunger/bert-finetuned-squad #endpoints_compatible #region-us \n", "# bert-finetuned-squad\n\nThis model is a fine-tuned version of loganunger/bert-finetuned-squad on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.2" ]
[ 58, 41, 6, 12, 8, 3, 103, 4, 33 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #bert #question-answering #generated_from_trainer #base_model-loganunger/bert-finetuned-squad #endpoints_compatible #region-us \n# bert-finetuned-squad\n\nThis model is a fine-tuned version of loganunger/bert-finetuned-squad on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.2" ]
[ -0.057027775794267654, 0.060239601880311966, -0.002682821359485388, 0.06469865888357162, 0.14116479456424713, 0.026252180337905884, 0.14549508690834045, 0.0954384058713913, -0.07472638040781021, 0.06734488159418106, 0.05600982904434204, 0.03491803631186485, 0.03214433416724205, 0.09476780146360397, -0.030679985880851746, -0.2421957403421402, 0.018058814108371735, 0.007833367213606834, -0.08301606774330139, 0.0772789791226387, 0.10005083680152893, -0.11676333099603653, 0.053132619708776474, 0.01848912052810192, -0.15837636590003967, 0.03785720467567444, -0.020970439538359642, -0.03182212635874748, 0.10659661144018173, 0.03709046170115471, 0.14085708558559418, 0.006469447165727615, 0.12081526219844818, -0.24187670648097992, 0.010404841043055058, 0.07421304285526276, 0.03153643012046814, 0.060479361563920975, 0.04385938495397568, 0.02903980389237404, 0.06259872764348984, -0.13596211373806, 0.12184026837348938, 0.01642962358891964, -0.07853177189826965, -0.18055057525634766, -0.07846368849277496, 0.04541747644543648, 0.09676796942949295, 0.07992216944694519, -0.009016580879688263, 0.12910491228103638, -0.11302022635936737, 0.05720757693052292, 0.1908734291791916, -0.2631654143333435, -0.07244094461202621, 0.06190205737948418, 0.062955841422081, 0.03962339833378792, -0.11231960356235504, -0.01832621544599533, 0.04937494546175003, 0.03748879209160805, 0.07774923741817474, -0.014703142456710339, -0.07945806533098221, -0.0028520673513412476, -0.13648201525211334, 0.012860285118222237, 0.08408333361148834, 0.05428235977888107, -0.029294971376657486, -0.13358023762702942, -0.03556523472070694, -0.04120299965143204, -0.02583841234445572, -0.07582996785640717, 0.034843482077121735, -0.03767390176653862, -0.06709720194339752, -0.05014108493924141, -0.08246865123510361, -0.06685029715299606, -0.01943586766719818, 0.12559759616851807, 0.04565926641225815, -0.003761017695069313, -0.04310872033238411, 0.09251011908054352, -0.05265411362051964, -0.10085330903530121, 0.004537559114396572, 0.004418507684022188, -0.10619057714939117, -0.08239051699638367, -0.05335640907287598, -0.06428597867488861, 0.019895538687705994, 0.17423366010189056, -0.04248175770044327, 0.07964729517698288, 0.005516698118299246, -0.0003773765347432345, -0.011123865842819214, 0.11013741791248322, -0.03321128711104393, -0.014506908133625984, -0.019155319780111313, 0.08063030242919922, -0.006731762550771236, -0.011308057233691216, -0.07913096994161606, -0.009970289655029774, 0.08041784167289734, 0.03769250586628914, -0.06050053983926773, 0.03562784567475319, -0.033191680908203125, -0.02779904007911682, -0.007731051649898291, -0.11201664805412292, 0.06452973186969757, -0.001586581813171506, -0.051546674221754074, -0.024217670783400536, 0.00925343856215477, 0.03169247508049011, 0.005086437799036503, 0.12977787852287292, -0.07160665094852448, 0.011169979348778725, -0.08682744204998016, -0.09280432015657425, 0.02054499089717865, -0.05055432394146919, -0.004697412718087435, -0.07826582342386246, -0.13398995995521545, -0.04159229248762131, 0.03953535109758377, -0.037945494055747986, -0.004564897157251835, -0.051326774060726166, -0.027560411021113396, 0.021143388003110886, -0.00791768915951252, 0.12587903439998627, -0.04514507204294205, 0.047612257301807404, -0.01901274360716343, 0.024125168099999428, 0.002703715581446886, 0.03276629000902176, -0.07426168769598007, 0.02885473147034645, -0.14123675227165222, 0.06197729706764221, -0.09613554924726486, 0.01101717259734869, -0.11649832129478455, -0.08283136785030365, -0.0030448874458670616, -0.000504958676174283, 0.07308211922645569, 0.09057175368070602, -0.18351633846759796, -0.04927454888820648, 0.14094753563404083, -0.09160774201154709, -0.06974276900291443, 0.11748485267162323, -0.07550732791423798, 0.04441394284367561, 0.06481149047613144, 0.16029362380504608, 0.10567465424537659, -0.1490272879600525, -0.00020725223293993622, -0.04082949832081795, 0.07685162127017975, 0.06929837912321091, 0.03966675326228142, -0.018630068749189377, 0.019789554178714752, 0.0015903085004538298, -0.03388778120279312, 0.016364650800824165, -0.09247075766324997, -0.07874814420938492, -0.038611218333244324, -0.0693378821015358, 0.05380925536155701, 0.024542581290006638, 0.049319710582494736, -0.07753273099660873, -0.09649373590946198, 0.18002180755138397, 0.09774067997932434, -0.06906991451978683, 0.012798736803233624, -0.07525017857551575, 0.003747079288586974, -0.034100934863090515, -0.023429926484823227, -0.18502947688102722, -0.14143259823322296, 0.03303974121809006, -0.08208590000867844, 0.04647010937333107, 0.055590152740478516, 0.07679295539855957, 0.06986714899539948, -0.06534019857645035, 0.0019285759190097451, -0.09180260449647903, 0.020795373246073723, -0.1060042530298233, -0.19892384111881256, -0.043322887271642685, -0.038136545568704605, 0.15030308067798615, -0.2548837959766388, 0.020460061728954315, -0.02573709562420845, 0.16544540226459503, 0.039077866822481155, -0.05313779413700104, -0.03590666875243187, 0.07306716591119766, 0.002977445488795638, -0.07528328895568848, 0.04877041280269623, -0.011580582708120346, -0.11988353729248047, -0.08061324805021286, -0.14507578313350677, 0.036947235465049744, 0.07606276124715805, 0.02104026824235916, -0.10944285243749619, -0.06077239662408829, -0.06262582540512085, -0.0518079437315464, -0.10080864280462265, 0.0017128431936725974, 0.20994986593723297, 0.004333178512752056, 0.12072651833295822, -0.04343452677130699, -0.05452635511755943, -0.006417546421289444, 0.0024518424179404974, -0.003260940546169877, 0.08036015927791595, 0.06722967326641083, -0.12353678047657013, 0.082160584628582, 0.1123155951499939, -0.056087154895067215, 0.15808536112308502, -0.048148978501558304, -0.07887403666973114, -0.0318392850458622, 0.013175624422729015, -0.004124144557863474, 0.10435193032026291, -0.10442570596933365, 0.012816766276955605, 0.017559155821800232, 0.03556312248110771, 0.032651759684085846, -0.17316031455993652, -0.006650235969573259, 0.031919196248054504, -0.025665273889899254, 0.014760934747755527, -0.023349475115537643, 0.04298895224928856, 0.08547884225845337, 0.030074279755353928, -0.004195700865238905, 0.015633655712008476, -0.012377317994832993, -0.08477634936571121, 0.18023337423801422, -0.10449987649917603, -0.12045983970165253, -0.09788279980421066, 0.016940321773290634, -0.04749161750078201, -0.02943873219192028, 0.03722295165061951, -0.10243900865316391, -0.048984747380018234, -0.08011442422866821, 0.04290493577718735, -0.014760272577404976, -0.0010462981881573796, 0.04113617539405823, 0.004466845653951168, 0.07848826795816422, -0.13126440346240997, 0.011771111749112606, -0.03731551021337509, -0.09640481323003769, -0.004395541734993458, 0.055380381643772125, 0.09496589750051498, 0.12730540335178375, -0.023741303011775017, 0.0012600449845194817, -0.0395694337785244, 0.23316825926303864, -0.0638686865568161, -0.00813557393848896, 0.09321027994155884, -0.02326018735766411, 0.036624304950237274, 0.1275865137577057, 0.056148774921894073, -0.09551658481359482, 0.04047330096364021, 0.10521133244037628, -0.012715128250420094, -0.22921699285507202, -0.042532969266176224, -0.024522069841623306, -0.09201367199420929, 0.11141712218523026, 0.04030332714319229, -0.0028954704757779837, 0.04251112788915634, -0.004430883564054966, 0.046776603907346725, 0.0051027932204306126, 0.08561605215072632, 0.11606813222169876, 0.03824422508478165, 0.10206196457147598, -0.026890361681580544, -0.05949931964278221, 0.058704622089862823, -0.013123767450451851, 0.2606697678565979, 0.019061531871557236, 0.07742481678724289, 0.06369195133447647, 0.11149697750806808, -0.011614588089287281, 0.011907118372619152, 0.00962779764086008, -0.030091991648077965, 0.0005304917576722801, -0.05735713616013527, -0.00506433192640543, 0.031185604631900787, -0.0034626408014446497, 0.03955872356891632, -0.09998965263366699, -0.007364248391240835, 0.02439645305275917, 0.23671336472034454, 0.04021940380334854, -0.23782335221767426, -0.08588224649429321, 0.02014923095703125, -0.037733979523181915, -0.050496410578489304, 0.003860071999952197, 0.1471903920173645, -0.11334700137376785, 0.05199424922466278, -0.06085667014122009, 0.09617911279201508, -0.023779958486557007, 0.0027904994785785675, 0.049969710409641266, 0.12759627401828766, -0.015010404400527477, 0.06938567012548447, -0.2040264904499054, 0.20522311329841614, 0.018097274005413055, 0.1165410578250885, -0.056780461221933365, 0.02728315070271492, 0.014102384448051453, 0.04971139505505562, 0.07306711375713348, -0.007986062206327915, -0.08094970881938934, -0.17994549870491028, -0.051226601004600525, 0.04247039929032326, 0.12620288133621216, -0.030610211193561554, 0.080210842192173, -0.0380648598074913, 0.023649848997592926, 0.06960493326187134, -0.029121708124876022, -0.20565128326416016, -0.13566331565380096, 0.006022833753377199, 0.04010181874036789, -0.03949221223592758, -0.09211954474449158, -0.10975918918848038, -0.06045861169695854, 0.202521413564682, 0.020087366923689842, -0.014005823992192745, -0.13203342258930206, 0.1007288247346878, 0.11223191022872925, -0.049058347940444946, 0.021994763985276222, 0.020193642005324364, 0.12868823111057281, 0.00619266415014863, -0.07656031101942062, 0.0754670649766922, -0.0643555298447609, -0.15688514709472656, -0.07565978914499283, 0.11830650269985199, 0.08930603414773941, 0.04837993159890175, 0.004601144697517157, 0.029861250892281532, 0.029293563216924667, -0.0967528373003006, 0.012410514056682587, 0.09987671673297882, 0.05472288653254509, 0.07157070189714432, -0.1031355932354927, -0.005065985023975372, -0.026329051703214645, -0.02361881732940674, 0.1353703737258911, 0.23787041008472443, -0.09248937666416168, 0.0631168931722641, 0.0729193389415741, -0.07955391705036163, -0.1645282357931137, 0.08443669229745865, 0.11142456531524658, 0.011793055571615696, 0.05233263969421387, -0.16889622807502747, 0.13492782413959503, 0.13766056299209595, -0.006931770592927933, 0.038388125598430634, -0.31623443961143494, -0.15004876255989075, 0.08844625949859619, 0.11594270169734955, 0.0454348586499691, -0.13314871490001678, -0.03211697190999985, -0.028061464428901672, -0.15880827605724335, 0.13001112639904022, -0.1160370409488678, 0.10346729308366776, 0.010466896928846836, 0.0798596739768982, 0.022334016859531403, -0.036514684557914734, 0.14676503837108612, 0.014537619426846504, 0.09801579266786575, -0.0538388155400753, 0.0030133998952805996, 0.03904702141880989, -0.06290888041257858, 0.0108945919200778, -0.013881311751902103, 0.03507967293262482, -0.10995432734489441, -0.038792695850133896, -0.06306689977645874, 0.05830283463001251, -0.057527851313352585, -0.06917586922645569, -0.06331869214773178, 0.06193293631076813, 0.05416785553097725, -0.03503759950399399, 0.08677705377340317, -0.012389259412884712, 0.16022324562072754, 0.03331610932946205, 0.10247230529785156, -0.011362625285983086, -0.08701203763484955, 0.005344646982848644, -0.026721812784671783, 0.0723692923784256, -0.09980469942092896, 0.0449884794652462, 0.13695888221263885, 0.040694888681173325, 0.16116847097873688, 0.0497058741748333, -0.05015834420919418, 0.011605590581893921, 0.028863228857517242, -0.10139186680316925, -0.18532496690750122, 0.009127764031291008, -0.011935743503272533, -0.134214386343956, 0.013671474531292915, 0.1139930710196495, -0.04949445277452469, -0.01153903640806675, -0.006034382153302431, 0.0015689076390117407, -0.028907448053359985, 0.17652447521686554, 0.01788126491010189, 0.04842521250247955, -0.07220351696014404, 0.09694403409957886, 0.07436623424291611, -0.10639499127864838, 0.045527856796979904, 0.048831596970558167, -0.06956277042627335, -0.015086283907294273, 0.06103961914777756, 0.23368503153324127, -0.025616692379117012, -0.037925899028778076, -0.07919541001319885, -0.10492192208766937, 0.03945539891719818, 0.12304232269525528, 0.043478336185216904, -0.026064887642860413, -0.02226073108613491, 0.04900733008980751, -0.12666985392570496, 0.07456369698047638, 0.041630279272794724, 0.07092965394258499, -0.11479509621858597, 0.12074849754571915, 0.01068844459950924, 0.0038270989898592234, -0.01714545674622059, 0.018595004454255104, -0.12163498997688293, -0.02505938708782196, -0.14268799126148224, -0.010368775576353073, -0.034336864948272705, 0.013075149618089199, 0.0035710420925170183, -0.05556122958660126, -0.03690405189990997, 0.02703600749373436, -0.06326161324977875, -0.04297397658228874, 0.016983043402433395, 0.06277243793010712, -0.1378631889820099, -0.015802698209881783, 0.02782132849097252, -0.08682581782341003, 0.06379367411136627, 0.03726785629987717, 0.03468276932835579, 0.03198379650712013, -0.18034246563911438, -0.019077584147453308, 0.030724553391337395, 0.015379004180431366, 0.06227143108844757, -0.1142960637807846, -0.030278408899903297, -0.02367490530014038, 0.0694337785243988, 0.02041209675371647, 0.041658468544483185, -0.11363933980464935, -0.03753821551799774, -0.05731085315346718, -0.0871930867433548, -0.057018864899873734, 0.018866147845983505, 0.09514175355434418, 0.05187096819281578, 0.1654658168554306, -0.105890192091465, 0.07237979769706726, -0.19381675124168396, -0.037851396948099136, 0.004372410476207733, -0.011269744485616684, -0.030014188960194588, -0.04201053828001022, 0.061452291905879974, -0.05308510735630989, 0.12687407433986664, -0.0383896604180336, 0.08890581876039505, 0.04716489091515541, -0.08457884192466736, -0.030805520713329315, 0.020525358617305756, 0.20579181611537933, 0.06489352136850357, -0.026597533375024796, 0.06167764589190483, 0.0014483911218121648, 0.0560225248336792, 0.0942063257098198, 0.18217960000038147, 0.1836535781621933, -0.03512096777558327, 0.047347597777843475, 0.06713896989822388, -0.0960266962647438, -0.12429589033126831, 0.12940269708633423, -0.011356809176504612, 0.09566722810268402, -0.05237596482038498, 0.17793741822242737, 0.07958553731441498, -0.1811019778251648, 0.05453051999211311, -0.06472162902355194, -0.10523565113544464, -0.10754123330116272, -0.017709307372570038, -0.09047829359769821, -0.11783325672149658, 0.01907525584101677, -0.1155017539858818, 0.02096327766776085, 0.05742479860782623, 0.027675341814756393, 0.030233247205615044, 0.16289502382278442, -0.04822215437889099, 0.028478384017944336, 0.050375696271657944, 0.00961159449070692, -0.00016867394151631743, -0.09103870391845703, -0.06037099286913872, 0.05002930387854576, 0.0074614668264985085, 0.053411372005939484, -0.04430823773145676, 0.006553211249411106, 0.021534627303481102, -0.004584296140819788, -0.07104108482599258, 0.028957583010196686, 0.018623020499944687, 0.03766055032610893, 0.0376286506652832, 0.04781687259674072, -0.0010696904500946403, -0.046493761241436005, 0.25283607840538025, -0.06032680347561836, -0.1119961142539978, -0.14495524764060974, 0.18687434494495392, 0.008868485689163208, 0.013592471368610859, 0.03885645046830177, -0.1106252521276474, 0.0018060633447021246, 0.15220323204994202, 0.1469801664352417, -0.04973531514406204, 0.001000668155029416, 0.0027250645216554403, -0.021151898428797722, -0.045278288424015045, 0.09103096276521683, 0.08487946540117264, 0.026452241465449333, -0.043601106852293015, -0.021397938951849937, -0.007714195642620325, -0.022151274606585503, -0.06709342449903488, 0.0632820799946785, 0.04174207150936127, 0.01950833573937416, -0.050242360681295395, 0.07445551455020905, 0.03695287927985191, -0.19637492299079895, 0.06841864436864853, -0.15427668392658234, -0.16181734204292297, -0.024318566545844078, 0.09564857929944992, -0.031192908063530922, 0.05541829764842987, -0.020607007667422295, -0.008770536631345749, 0.09967141598463058, -0.01945059932768345, -0.02056671306490898, -0.1155204027891159, 0.0884346067905426, -0.02816888876259327, 0.22425907850265503, 0.003937881905585527, 0.07506918907165527, 0.11766703426837921, 0.03506116569042206, -0.09398757666349411, 0.04009123891592026, 0.06536846607923508, -0.08258745819330215, 0.004994632676243782, 0.14310212433338165, -0.05558902025222778, 0.12164333462715149, 0.05334028601646423, -0.19045986235141754, 0.021289274096488953, -0.04744880273938179, -0.06049743667244911, -0.07752031087875366, 0.023474546149373055, -0.06987375766038895, 0.15012942254543304, 0.19981935620307922, -0.01994832418859005, 0.009668583050370216, -0.06371590495109558, 0.038411371409893036, 0.05226544663310051, 0.12891963124275208, -0.04844217374920845, -0.21151408553123474, 0.03388861194252968, 0.0430116206407547, 0.019370168447494507, -0.2687264382839203, -0.11985465884208679, 0.0469977892935276, -0.045529481023550034, -0.03953538462519646, 0.10437746345996857, 0.08130767941474915, 0.04860009253025055, -0.04392588511109352, -0.24233877658843994, -0.032715052366256714, 0.14128145575523376, -0.11260191351175308, -0.046252280473709106 ]
null
null
peft
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.8.2
{"library_name": "peft", "base_model": "ybelkada/blip2-opt-2.7b-fp16-sharded"}
null
leoreigoto/Data3_V3_BLIP2_VQA
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:ybelkada/blip2-opt-2.7b-fp16-sharded", "region:us" ]
2024-02-15T03:45:51+00:00
[ "1910.09700" ]
[]
TAGS #peft #safetensors #arxiv-1910.09700 #base_model-ybelkada/blip2-opt-2.7b-fp16-sharded #region-us
# Model Card for Model ID ## Model Details ### Model Description - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact ### Framework versions - PEFT 0.8.2
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ "TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-ybelkada/blip2-opt-2.7b-fp16-sharded #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ 49, 6, 3, 54, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4, 11 ]
[ "passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-ybelkada/blip2-opt-2.7b-fp16-sharded #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2" ]
[ -0.1023959144949913, 0.20364031195640564, -0.003501187078654766, 0.02728327363729477, 0.08585690706968307, 0.020007897168397903, 0.06716983020305634, 0.12274689972400665, 0.012622345238924026, 0.12875720858573914, 0.05166934058070183, 0.10228630155324936, 0.12245386838912964, 0.22668033838272095, -0.012838003225624561, -0.19243070483207703, 0.0242630448192358, -0.0771036297082901, 0.005593413487076759, 0.12082977592945099, 0.13930879533290863, -0.09762415289878845, 0.07061418145895004, -0.022893469780683517, -0.006428177002817392, -0.03144235908985138, -0.06823558360338211, -0.026020023971796036, 0.05179716274142265, 0.052624523639678955, 0.04507560655474663, -0.0026365057565271854, 0.08946135640144348, -0.2719646394252777, 0.01013258844614029, 0.05144821107387543, -0.0014026554999873042, 0.08499309420585632, 0.0988464504480362, -0.026928776875138283, 0.1098974272608757, -0.04064032807946205, 0.13379088044166565, 0.07711464166641235, -0.09239108860492706, -0.22410325706005096, -0.07173959165811539, 0.08625253289937973, 0.1827836036682129, 0.06788639724254608, -0.03555542975664139, 0.1263314187526703, -0.07607069611549377, 0.015353861264884472, 0.07337884604930878, -0.10034869611263275, -0.07253215461969376, 0.06338474154472351, 0.1148637905716896, 0.08918241411447525, -0.1183159276843071, -0.03819655627012253, 0.03314727544784546, 0.04002053290605545, 0.08421246707439423, 0.011725823394954205, 0.16789239645004272, 0.030479831621050835, -0.14271286129951477, -0.052292387932538986, 0.13737669587135315, 0.023319685831665993, -0.040452320128679276, -0.23331835865974426, -0.015376968309283257, -0.06795480847358704, -0.03502865135669708, -0.05600921809673309, 0.038739193230867386, 0.003568066982552409, 0.11532311141490936, -0.035693444311618805, -0.08089860528707504, -0.016660982742905617, 0.11346282064914703, 0.07784787565469742, 0.013616815209388733, -0.015320400707423687, 0.011008908972144127, 0.12671732902526855, 0.05690603330731392, -0.12490561604499817, -0.04582716152071953, -0.06711649894714355, -0.04033141955733299, -0.027365777641534805, 0.05342256277799606, 0.0353180356323719, 0.042800284922122955, 0.25008848309516907, -0.025695038959383965, 0.05624420940876007, 0.05131547898054123, 0.01635528728365898, 0.032757923007011414, 0.09937477856874466, -0.052993740886449814, -0.18724198639392853, -0.014669667929410934, 0.10533948987722397, 0.008535911329090595, -0.02457994781434536, -0.04416726529598236, 0.029510188847780228, 0.030378494411706924, 0.11640799790620804, 0.10281682014465332, -0.024227039888501167, -0.07068724185228348, -0.059072207659482956, 0.216552734375, -0.15430507063865662, 0.04941238462924957, 0.01976020447909832, -0.02105451002717018, -0.05080863833427429, 0.015585738234221935, 0.011093972250819206, -0.03752332180738449, 0.10955710709095001, -0.06118832156062126, -0.049624186009168625, -0.11145076900720596, -0.04487142711877823, 0.03322191536426544, 0.005177776329219341, -0.03966242074966431, -0.031259022653102875, -0.08961821347475052, -0.09290562570095062, 0.09356793016195297, -0.059249572455883026, -0.07570087909698486, -0.022671813145279884, -0.06882917881011963, 0.02134208008646965, 0.017937064170837402, 0.08182405680418015, -0.025928014889359474, 0.04417520761489868, -0.031341515481472015, 0.0675363764166832, 0.08862745016813278, 0.03666501119732857, -0.07080328464508057, 0.06790003925561905, -0.19235196709632874, 0.08260858803987503, -0.0774281769990921, 0.027750061824917793, -0.15746945142745972, -0.007029620930552483, 0.007698461879044771, 0.02178480476140976, 0.037930868566036224, 0.15137243270874023, -0.1939772069454193, -0.03221370279788971, 0.165089413523674, -0.10368044674396515, -0.11016760021448135, 0.04177917167544365, -0.038249846547842026, 0.15739089250564575, 0.030280139297246933, -0.0012426129542291164, 0.09249915927648544, -0.14316007494926453, -0.013938642106950283, -0.024325892329216003, 0.017603334039449692, 0.0778016746044159, 0.07073507457971573, -0.08130092173814774, 0.017243625596165657, 0.017793981358408928, -0.061591293662786484, -0.0081405621021986, -0.03970023989677429, -0.0987352728843689, 0.0057478612288832664, -0.08744806051254272, 0.016670476645231247, 0.007478628307580948, -0.0831889808177948, -0.01675930991768837, -0.14190812408924103, -0.03510471433401108, 0.08182565867900848, 0.010915788821876049, -0.017329296097159386, -0.06898674368858337, 0.034451693296432495, -0.03666619583964348, -0.01648077555000782, -0.14489136636257172, -0.021427473053336143, 0.03630919009447098, -0.15304622054100037, -0.005145850591361523, -0.11835219711065292, 0.06576516479253769, 0.01511449459940195, -0.06323514133691788, -0.04222502559423447, 0.017977474257349968, -0.004212197847664356, -0.057308491319417953, -0.21832139790058136, -0.03378254547715187, -0.04404837638139725, 0.1459498107433319, -0.22241605818271637, 0.04439277574419975, 0.011004806496202946, 0.12492083758115768, 0.009400355629622936, -0.06328117102384567, 0.025281473994255066, -0.060985516756772995, -0.022151879966259003, -0.07330615818500519, -0.010349185205996037, -0.0023612214718014, -0.035151150077581406, 0.025524193421006203, -0.14404910802841187, -0.04832933098077774, 0.0898718312382698, 0.08889646828174591, -0.1480346918106079, 0.004582213703542948, -0.0479704849421978, -0.06432840973138809, -0.08677776902914047, -0.0746101588010788, 0.07303491979837418, 0.050404179841279984, 0.05474574863910675, -0.08046972006559372, -0.06728234142065048, 0.006895649246871471, -0.013016865588724613, -0.026044650003314018, 0.12051089853048325, 0.08064275234937668, -0.08795077353715897, 0.09453634172677994, 0.07926716655492783, 0.048489537090063095, 0.09307026863098145, -0.012605591677129269, -0.10079944878816605, -0.029974179342389107, 0.05817700922489166, 0.013242031447589397, 0.15151557326316833, -0.05331005901098251, 0.04880187660455704, 0.05229213833808899, -0.046881623566150665, 0.0392768569290638, -0.1003534123301506, 0.01598189026117325, 0.010466392152011395, -0.014099279418587685, 0.04221281781792641, -0.022992460057139397, 0.013768854551017284, 0.08918881416320801, 0.06303547322750092, 0.028075722977519035, 0.02467573806643486, -0.03831007331609726, -0.13685983419418335, 0.16914397478103638, -0.09373056143522263, -0.2350214421749115, -0.1508769690990448, 0.033380694687366486, 0.05175327509641647, -0.02225610613822937, 0.025787750259041786, -0.03416982665657997, -0.10775268822908401, -0.08076061308383942, 0.017707832157611847, 0.04551579803228378, -0.06924950331449509, -0.05629436671733856, 0.03717868775129318, 0.03978053480386734, -0.12887011468410492, 0.03030010685324669, 0.05566076934337616, 0.0030389325693249702, -0.0023474569898098707, 0.04472624510526657, 0.0838165283203125, 0.19098424911499023, 0.0038206656463444233, 0.0005792943411506712, 0.052969906479120255, 0.2804574966430664, -0.15404251217842102, 0.1257592886686325, 0.1265062391757965, -0.05037635564804077, 0.0915827602148056, 0.20057283341884613, 0.04166555032134056, -0.08155781030654907, 0.024662816897034645, 0.03081902116537094, -0.033339425921440125, -0.25858598947525024, -0.07089923322200775, -0.02134774811565876, -0.0718344897031784, 0.08356352150440216, 0.08594240248203278, 0.09141265600919724, 0.034850113093853, -0.07882706820964813, -0.06002916395664215, 0.048778217285871506, 0.10919934511184692, -0.031268373131752014, 0.01373243611305952, 0.08169406652450562, -0.03843279555439949, 0.00291964178904891, 0.0957426205277443, -0.016873784363269806, 0.1495479941368103, 0.043862324208021164, 0.10726974159479141, 0.06192391365766525, 0.07807627320289612, -0.00427240040153265, 0.05275103449821472, 0.014167661778628826, 0.027501339092850685, 0.009811957366764545, -0.0911397933959961, 0.027544016018509865, 0.12415897846221924, 0.019679436460137367, 0.03910135477781296, 0.02221125178039074, -0.0599311999976635, 0.03939114511013031, 0.2089223861694336, 0.011348971165716648, -0.19348034262657166, -0.07075284421443939, 0.06647099554538727, -0.08268588781356812, -0.14843600988388062, -0.007554139941930771, 0.02860885113477707, -0.1735040545463562, 0.021763663738965988, -0.04100511968135834, 0.10741151124238968, -0.07009892165660858, -0.04010515660047531, 0.09469404071569443, 0.06040317192673683, -0.030702605843544006, 0.04793144762516022, -0.17065903544425964, 0.1120685413479805, 0.030907707288861275, 0.06870585680007935, -0.09663672745227814, 0.10015694797039032, 0.004520023241639137, -0.019680628553032875, 0.16754156351089478, 0.004071482457220554, -0.046421702951192856, -0.07688678056001663, -0.068022720515728, -0.02427726984024048, 0.09113579243421555, -0.13575509190559387, 0.0659659206867218, -0.018054954707622528, -0.038992740213871, 0.0012381378328427672, -0.10723204910755157, -0.11051654070615768, -0.17094869911670685, 0.06820432096719742, -0.07434363663196564, 0.0032095236238092184, -0.09686926752328873, -0.055401481688022614, -0.009374987334012985, 0.1763092428445816, -0.17980854213237762, -0.11631330102682114, -0.1491946429014206, -0.10259010642766953, 0.17095473408699036, -0.04734873026609421, 0.08819764107465744, -0.0034766437020152807, 0.16787858307361603, -0.014214687049388885, -0.015718974173069, 0.087732695043087, -0.09375686198472977, -0.1954045295715332, -0.054642241448163986, 0.1798044741153717, 0.12645776569843292, 0.0340544693171978, -0.022557005286216736, 0.021083654835820198, -0.045462194830179214, -0.1131872907280922, 0.01341725792735815, 0.14102241396903992, 0.044991254806518555, 0.0011787780094891787, -0.026516659185290337, -0.11984778940677643, -0.05476224422454834, -0.050618626177310944, -0.007076145615428686, 0.2069881707429886, -0.08295559138059616, 0.16145899891853333, 0.12309258431196213, -0.048618387430906296, -0.20653872191905975, 0.0328410379588604, 0.04221039637923241, 0.015131350606679916, 0.03412385284900665, -0.18393175303936005, 0.08414941281080246, -0.011607940308749676, -0.07911363989114761, 0.1776852309703827, -0.18600867688655853, -0.13374063372612, 0.0897565707564354, 0.0228035319596529, -0.2260301113128662, -0.13407503068447113, -0.1137874573469162, -0.01737418957054615, -0.12759308516979218, 0.04896645247936249, 0.029215604066848755, 0.0038563767448067665, 0.017197856679558754, 0.017441323027014732, 0.04046647995710373, -0.05223904177546501, 0.1984073370695114, -0.026710696518421173, 0.006047636270523071, -0.050715576857328415, -0.09581901878118515, 0.02039787359535694, -0.05753452330827713, 0.110266774892807, -0.01032858807593584, 0.023703893646597862, -0.16006609797477722, -0.04355010390281677, -0.06762884557247162, 0.020070020109415054, -0.09522926807403564, -0.0928168073296547, -0.05127653107047081, 0.0819212943315506, 0.10820676386356354, -0.02295168861746788, 0.0005749896517954767, -0.07361534982919693, 0.07722114026546478, 0.22184693813323975, 0.1649458259344101, 0.048072416335344315, -0.054855767637491226, 0.010806005448102951, -0.03211841732263565, 0.0406072698533535, -0.21899348497390747, 0.0427355132997036, 0.06035930663347244, 0.036828190088272095, 0.08413819968700409, -0.012402283027768135, -0.1615157127380371, -0.07863713055849075, 0.07407596707344055, -0.06613833457231522, -0.16754086315631866, -0.03876882418990135, 0.04338647425174713, -0.19467578828334808, -0.0508623942732811, 0.030222972854971886, -0.021225454285740852, -0.03482480347156525, 0.015321789309382439, 0.08263038098812103, -0.006926113273948431, 0.10264649987220764, 0.07923399657011032, 0.09892527759075165, -0.105184406042099, 0.06824152916669846, 0.08597805351018906, -0.03393268212676048, 0.009138351306319237, 0.13618622720241547, -0.05014333128929138, -0.023080265149474144, 0.06525996327400208, 0.08105384558439255, 0.012546095997095108, -0.052424877882003784, 0.013543504290282726, -0.0725960060954094, 0.06149521470069885, 0.10263664275407791, 0.019078141078352928, -0.016350749880075455, 0.06600207090377808, 0.02000228501856327, -0.0930938571691513, 0.1252557784318924, 0.06951259076595306, 0.022252975031733513, -0.038538623601198196, -0.028527628630399704, -0.013999884948134422, -0.007938371039927006, -0.014650342985987663, -0.002483789110556245, -0.07650505006313324, -0.004464221652597189, -0.1101321205496788, 0.01693287119269371, -0.07970034331083298, 0.006229866296052933, 0.013589464128017426, -0.042377084493637085, -0.0007004099315963686, -0.0010941700311377645, -0.0825127363204956, -0.060410551726818085, -0.022047949954867363, 0.08035755902528763, -0.12847869098186493, 0.01676725037395954, 0.07004719227552414, -0.11456523090600967, 0.06635120511054993, -0.01037197932600975, 0.012879660353064537, 0.0023543727584183216, -0.13481667637825012, 0.05165598541498184, -0.02099967747926712, -0.0035665908362716436, 0.02452498860657215, -0.16873589158058167, -0.0032408160623162985, -0.05216484144330025, -0.0749099999666214, 0.0048598493449389935, -0.04050567373633385, -0.13309027254581451, 0.11099666357040405, -0.012723537161946297, -0.07056688517332077, -0.02372054010629654, 0.05077226459980011, 0.08550150692462921, -0.02510816790163517, 0.0938984677195549, -0.025579016655683517, 0.08201071619987488, -0.17812030017375946, -0.009961896575987339, -0.01465003564953804, 0.03416622057557106, -0.018006669357419014, -0.014842072501778603, 0.05195453017950058, -0.012286731973290443, 0.17104652523994446, -0.017782554030418396, 0.07721024751663208, 0.04898650571703911, -0.002305001951754093, 0.02977832779288292, 0.06967566907405853, 0.06522846221923828, -0.014649573713541031, -0.005505430977791548, 0.026262225583195686, -0.013115668669342995, -0.043591879308223724, -0.1496414840221405, 0.03997490927577019, 0.16483180224895477, 0.06666601449251175, 0.031087568029761314, 0.02089330367743969, -0.137874573469162, -0.08431115001440048, 0.10268014669418335, -0.016490649431943893, -0.017492152750492096, -0.07017472386360168, 0.1938580423593521, 0.12399069219827652, -0.19902944564819336, 0.07030948996543884, -0.047140613198280334, -0.03475939482450485, -0.12678919732570648, -0.1504630148410797, -0.057876620441675186, -0.0396706759929657, -0.023383179679512978, -0.061611108481884, 0.05963646247982979, 0.045422691851854324, -0.001120392931625247, -0.007043042685836554, 0.10047915577888489, 0.002273647114634514, -0.025097914040088654, 0.06634964048862457, 0.07210102677345276, 0.044173598289489746, -0.08350478112697601, 0.005473924335092306, -0.004407464060932398, 0.005279168486595154, 0.05772916600108147, 0.019375650212168694, -0.059642791748046875, 0.022227197885513306, -0.001713956706225872, -0.11705102026462555, 0.04011988639831543, -0.015719687566161156, -0.03440998122096062, 0.1458049863576889, 0.023999176919460297, 0.010189180262386799, -0.02722814306616783, 0.2215377539396286, -0.08876711875200272, -0.07316280901432037, -0.13400240242481232, 0.07494153827428818, -0.05017445236444473, 0.035624027252197266, 0.03759501129388809, -0.12485615164041519, 0.006820192094892263, 0.1633101850748062, 0.1287752240896225, 0.0015842759748920798, 0.009181773290038109, 0.05443078279495239, 0.005943849217146635, -0.03531435877084732, 0.022326532751321793, 0.04972163215279579, 0.18732118606567383, -0.07171057164669037, 0.07961983233690262, -0.014104119502007961, -0.07633235305547714, -0.02722921036183834, 0.1351630687713623, -0.0060691554099321365, -0.0005658621084876359, -0.061430297791957855, 0.13521265983581543, -0.05378999561071396, -0.22143587470054626, 0.061226099729537964, -0.08844279497861862, -0.14051449298858643, -0.03979843854904175, 0.0020295942667871714, -0.018312251195311546, 0.01529828179627657, 0.06830412894487381, -0.054241351783275604, 0.1824820339679718, 0.032676003873348236, -0.06139485538005829, -0.08912920951843262, 0.050898075103759766, -0.14380158483982086, 0.2871195673942566, 0.026256408542394638, 0.03674665838479996, 0.10091640055179596, -0.02551521733403206, -0.1445848047733307, 0.022934162989258766, 0.11640027910470963, -0.07552778720855713, 0.05333654582500458, 0.17266467213630676, -0.007869631983339787, 0.13195273280143738, 0.05329576134681702, -0.06672891974449158, 0.034466732293367386, -0.05582581087946892, -0.06128007173538208, -0.12193675339221954, 0.07341577857732773, -0.07046602666378021, 0.14958375692367554, 0.12877047061920166, -0.06340276449918747, -0.009213985875248909, -0.019985152408480644, 0.0771741271018982, 0.016045335680246353, 0.13332360982894897, 0.02260434255003929, -0.18001601099967957, 0.04458257183432579, -0.005987263284623623, 0.11051776260137558, -0.20083686709403992, -0.06018247827887535, 0.03471638634800911, -0.02889879234135151, -0.0835249274969101, 0.11544446647167206, 0.0454484261572361, 0.020198026672005653, -0.029829418286681175, -0.07667473703622818, -0.0012824061559513211, 0.15097537636756897, -0.10192548483610153, -0.005492199212312698 ]
null
null
null
# **Q-Learning** Agent playing1 **FrozenLake-v1** This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** . ## Usage ```python model = load_from_hub(repo_id="Noname08/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
{"tags": ["FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "q-FrozenLake-v1-4x4-noSlippery", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "FrozenLake-v1-4x4-no_slippery", "type": "FrozenLake-v1-4x4-no_slippery"}, "metrics": [{"type": "mean_reward", "value": "1.00 +/- 0.00", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
Noname08/q-FrozenLake-v1-4x4-noSlippery
[ "FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
2024-02-15T03:48:56+00:00
[]
[]
TAGS #FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
# Q-Learning Agent playing1 FrozenLake-v1 This is a trained model of a Q-Learning agent playing FrozenLake-v1 . ## Usage
[ "# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage" ]
[ "TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n", "# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage" ]
[ 40, 39 ]
[ "passage: TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage" ]
[ 0.04578453302383423, -0.08074592798948288, -0.00430759321898222, 0.10720831900835037, 0.05034215748310089, -0.040469273924827576, 0.11997015029191971, 0.018999949097633362, 0.20601962506771088, -0.010012076236307621, 0.1455274522304535, 0.007022971753031015, -0.006192410364747047, 0.1867983490228653, 0.04572829231619835, -0.26324528455734253, 0.01831899583339691, -0.09495259821414948, -0.07281816750764847, 0.11870454251766205, 0.05470194295048714, -0.01901467889547348, -0.0007633853238075972, 0.056141503155231476, -0.0673527717590332, 0.0007737681735306978, 0.031996939331293106, -0.012976245954632759, 0.19804789125919342, -0.02254498563706875, 0.06641989201307297, 0.054705578833818436, 0.0758768692612648, -0.1998077929019928, 0.0358855277299881, -0.04215473681688309, -0.09439758956432343, -0.03934839740395546, -0.018780618906021118, 0.05878105387091637, 0.053356342017650604, 0.03858819976449013, 0.058354366570711136, 0.09384993463754654, -0.0773480236530304, 0.04328357055783272, 0.04280758649110794, 0.024811049923300743, 0.04589218273758888, -0.0237203948199749, -0.027002155780792236, 0.08246652781963348, -0.22182892262935638, 0.10318073630332947, -0.010159241035580635, -0.5270710587501526, -0.00633762264624238, 0.24088262021541595, 0.11517096310853958, 0.05707438662648201, -0.06903956830501556, 0.10566288232803345, 0.03913382440805435, -0.007209456991404295, 0.03210983797907829, 0.02150118350982666, 0.12817370891571045, 0.06009242683649063, -0.09581366181373596, 0.040699947625398636, 0.13722525537014008, 0.012822695076465607, 0.020306183025240898, -0.08888901025056839, 0.0410032719373703, -0.03461858257651329, -0.007679527159780264, -0.09758518636226654, 0.05478060990571976, 0.012466507963836193, -0.0934976264834404, -0.09247440844774246, -0.04236573353409767, -0.06708304584026337, 0.11252415925264359, 0.046419668942689896, -0.0874939113855362, 0.03884070739150047, -0.06760413944721222, 0.05918780341744423, -0.16863860189914703, 0.02074250765144825, -0.06627868115901947, -0.09376336634159088, -0.11799788475036621, -0.01683047041296959, -0.07946427166461945, 0.009092256426811218, 0.056664444506168365, 0.1447116881608963, 0.22076484560966492, 0.06690320372581482, 0.09728849679231644, 0.07456006109714508, 0.06531001627445221, 0.1538129299879074, 0.10918238013982773, 0.019075315445661545, -0.015266558155417442, 0.0948706716299057, -0.06445580720901489, -0.1351388692855835, -0.15579092502593994, 0.005488025024533272, 0.0983937531709671, 0.08871900290250778, -0.044080477207899094, -0.006702381651848555, -0.024641724303364754, 0.08566431701183319, -0.11314457654953003, -0.024612564593553543, -0.002267979085445404, 0.06882024556398392, -0.024801667779684067, 0.020378148183226585, -0.06242705136537552, 0.12715265154838562, 0.04222423583269119, -0.059924717992544174, -0.055308472365140915, -0.03053177334368229, -0.014276440255343914, -0.027539284899830818, 0.02446848154067993, -0.07659092545509338, 0.04767750948667526, -0.16766095161437988, -0.042871296405792236, -0.04784649610519409, 0.025697942823171616, -0.03907240927219391, -0.13557587563991547, -0.17699143290519714, -0.048906855285167694, -0.022438718006014824, 0.03549358621239662, -0.038111843168735504, 0.006551501806825399, -0.006318534724414349, -0.1583600640296936, 0.09783563017845154, 0.09784027189016342, -0.03643378987908363, -0.02749447710812092, 0.056263517588377, -0.07194498926401138, 0.1561182290315628, -0.21054518222808838, -0.054014235734939575, -0.044764336198568344, -0.06595750898122787, 0.19673264026641846, 0.012690845876932144, -0.01202624011784792, 0.19873127341270447, -0.29073721170425415, -0.06078760325908661, 0.12533614039421082, -0.07834373414516449, -0.0936407670378685, 0.06941844522953033, -0.04206686094403267, 0.023345354944467545, 0.046047765761613846, 0.36345911026000977, -0.02069227211177349, -0.16197136044502258, -0.021782705560326576, 0.13971707224845886, -0.1184760183095932, 0.059895481914281845, 0.04240793362259865, 0.12543781101703644, -0.04250509291887283, -0.018672896549105644, -0.09023164212703705, 0.05999075248837471, -0.05241934582591057, -0.09016361832618713, -0.03393383324146271, -0.07645075023174286, 0.13294468820095062, -0.0629684180021286, 0.05601520463824272, -0.03255095332860947, -0.07133250683546066, -0.050324998795986176, -0.016492370516061783, 0.04460815340280533, 0.05951254442334175, -0.12794871628284454, 0.11029167473316193, 0.13025271892547607, -0.0006193425506353378, -0.07498852163553238, -0.17872096598148346, 0.003240168560296297, 0.009576505981385708, 0.039837226271629333, 0.17141658067703247, 0.12209978699684143, 0.033295199275016785, 0.008770671673119068, -0.06389404833316803, -0.18276847898960114, 0.058129217475652695, -0.056212130934000015, -0.14230976998806, -0.052409034222364426, -0.0728459507226944, 0.017381802201271057, -0.0859743058681488, -0.017379917204380035, 0.021926190704107285, 0.006908397190272808, 0.02990424446761608, -0.026645656675100327, -0.049561817198991776, 0.021254703402519226, 0.06490101665258408, -0.0037617047782987356, 0.12023693323135376, 0.008277264423668385, -0.18308481574058533, 0.07930773496627808, 0.08478537946939468, 0.09196605533361435, 0.013250201940536499, 0.02685922384262085, -0.021522263064980507, -0.08061408251523972, -0.054420311003923416, 0.02957955375313759, 0.11417073011398315, 0.1317172348499298, 0.2361993044614792, 0.08753683418035507, 0.04697408527135849, -0.02164587564766407, -0.016415923833847046, 0.002810494042932987, -0.06318057328462601, -0.029935607686638832, 0.10614971816539764, 0.05865858122706413, -0.067733034491539, -0.04576427489519119, 0.09590928256511688, 0.02732124738395214, 0.21205885708332062, -0.03342745825648308, 0.01286078616976738, -0.10957037657499313, -0.06550975888967514, -0.031982194632291794, 0.09201868623495102, 0.09498392790555954, 0.009755023755133152, -0.022056059911847115, -0.04259001836180687, 0.0012916827108711004, -0.1334889680147171, -0.10375088453292847, 0.026475343853235245, 0.013400445692241192, -0.11206940561532974, 0.11674030870199203, -0.11352457851171494, 0.039504457265138626, 0.06024791672825813, -0.13837239146232605, 0.04428480193018913, -0.029713207855820656, -0.07886212319135666, 0.16866780817508698, -0.11075661331415176, -0.094340018928051, -0.08831550180912018, 0.004082420375198126, 0.0075836325995624065, -0.03922267258167267, -0.009283260442316532, -0.19952571392059326, -0.005375816952437162, -0.03544965013861656, 0.013616434298455715, -0.06988783925771713, -0.11287739872932434, -0.010957922786474228, 0.07084179669618607, -0.043388739228248596, -0.07803605496883392, 0.007967432029545307, -0.08923084288835526, -0.10623309016227722, 0.028189711272716522, 0.019765101373195648, -0.022883659228682518, 0.16152891516685486, 0.01816628873348236, 0.05626589432358742, -0.03298520669341087, 0.30665266513824463, -0.038163769990205765, 0.08371731638908386, -0.02993497997522354, -0.07433546334505081, 0.06130730360746384, -0.022327827289700508, 0.06086638569831848, -0.020221687853336334, -0.02362890914082527, 0.0077952733263373375, -0.08579335361719131, -0.18365982174873352, -0.05417544022202492, 0.03724347800016403, 0.195254847407341, 0.031118987128138542, 0.01910330168902874, -0.0488768145442009, -0.010547760874032974, 0.1665220558643341, -0.10005921125411987, 0.04030545800924301, -0.05366240441799164, 0.11506262421607971, -0.08640182018280029, 0.06195629760622978, 0.020486772060394287, 0.04266135022044182, -0.04877188801765442, 0.09486009180545807, 0.0826394334435463, 0.1121082529425621, -0.02206910029053688, 0.046257395297288895, 0.019012698903679848, 0.07383184134960175, 0.11073657125234604, 0.0368414968252182, -0.0729052945971489, 0.001982470043003559, -0.006313489284366369, -0.039427030831575394, 0.11933320760726929, 0.17963355779647827, -0.11991413682699203, -0.05106910318136215, 0.27167606353759766, 0.0031242913100868464, 0.19481229782104492, -0.01315275114029646, 0.043591804802417755, -0.04484925419092178, 0.04572054371237755, -0.05338600277900696, -0.04086209088563919, 0.2094656229019165, 0.08045925945043564, -0.17165091633796692, -0.08549032360315323, -0.05912299454212189, 0.07081323862075806, 0.10728751868009567, 0.0013539529172703624, -0.04156802222132683, 0.0004610282776411623, 0.0014198932331055403, 0.08339415490627289, -0.14520122110843658, 0.11816094070672989, -0.03172019124031067, 0.05612684786319733, 0.017555562779307365, -0.045326150953769684, 0.04264266416430473, 0.07474290579557419, 0.26618310809135437, 0.0904107540845871, -0.040318213403224945, -0.0892091691493988, -0.12260187417268753, 0.010461576282978058, 0.029102616012096405, -0.03534553572535515, 0.0037547778338193893, -0.020087555050849915, 0.0318896509706974, 0.008264793083071709, 0.016230624169111252, -0.08987458795309067, -0.03175399824976921, -0.027736429125070572, -0.023839212954044342, 0.10733365267515182, -0.09495144337415695, -0.1444292515516281, -0.15713949501514435, 0.04191131144762039, -0.0766405463218689, -0.056593164801597595, -0.054507751017808914, -0.05239389091730118, -0.0311186034232378, -0.03773957118391991, 0.09099467098712921, -0.0021037792321294546, 0.14807306230068207, -0.1920108050107956, -0.04220759496092796, 0.051812779158353806, -0.07607918977737427, -0.08729588985443115, 0.03410962224006653, 0.12136995792388916, 0.05116051807999611, 0.11504370719194412, 0.013609255664050579, 0.09567681699991226, 0.0045484392903745174, -0.06713183224201202, 0.15302421152591705, -0.14069625735282898, -0.27875974774360657, -0.03836318850517273, 0.016946332529187202, 0.1615200787782669, -0.05613167956471443, 0.031766023486852646, 0.3335736393928528, 0.27782970666885376, -0.1428707242012024, 0.25916144251823425, 0.019178593531250954, 0.004398873541504145, -0.19130495190620422, -0.10125631093978882, 0.025324683636426926, 0.04740457236766815, 0.12032642960548401, -0.14564448595046997, -0.010732659138739109, -0.04543145373463631, -0.025908485054969788, 0.10386138409376144, -0.12300799041986465, -0.07263197749853134, 0.07765276730060577, 0.039809420704841614, 0.1808302253484726, 0.03932500258088112, 0.0014799144119024277, 0.13626977801322937, 0.06612244248390198, 0.019124457612633705, 0.05216038227081299, 0.08028066903352737, -0.018944554030895233, 0.14207926392555237, 0.05448179319500923, -0.02551644667983055, 0.052681710571050644, -0.0054580713622272015, -0.03219012916088104, 0.015605825930833817, -0.183198019862175, -0.10147556662559509, -0.0561356320977211, -0.10798973590135574, -0.04978342354297638, 0.056853994727134705, -0.12395523488521576, -0.007896827533841133, -0.03841273859143257, 0.03718273714184761, -0.07831971347332001, -0.09360362589359283, -0.036494381725788116, 0.1351792961359024, 0.07210618257522583, 0.04471297934651375, 0.035655103623867035, -0.07390819489955902, 0.07097936421632767, 0.21671734750270844, 0.08159157633781433, 0.028919655829668045, -0.19545674324035645, -0.024042490869760513, -0.0803457647562027, 0.06306298077106476, -0.08856996893882751, -0.016788700595498085, 0.11923003196716309, 0.08616556972265244, 0.05413002520799637, 0.09640096127986908, -0.045083072036504745, 0.021686913445591927, 0.02684609219431877, -0.15131035447120667, -0.18501274287700653, -0.08534606546163559, -0.03519878163933754, 0.11561143398284912, -0.06398691236972809, 0.10897188633680344, -0.13615410029888153, 0.010051886551082134, -0.006060056854039431, 0.02693452313542366, -0.03596206381917, -0.11251141875982285, 0.15348562598228455, 0.11999429017305374, -0.06767056882381439, 0.03127254918217659, -0.09527092427015305, -0.04423454403877258, 0.12686803936958313, -0.013623855076730251, -0.0371493324637413, -0.054547641426324844, -0.03628576174378395, 0.15247689187526703, -0.03436964750289917, 0.008244883269071579, -0.041229065507650375, -0.18217355012893677, 0.0798322781920433, 0.09045056998729706, 0.019827889278531075, -0.031874191015958786, -0.09797266125679016, -0.010231015272438526, -0.0011165260802954435, 0.11730700731277466, -0.10696814209222794, -0.10933240503072739, -0.15144047141075134, 0.06713984161615372, -0.0007159380475059152, 0.18502596020698547, -0.06394898891448975, -0.08904669433832169, -0.12429379671812057, 0.02344517596065998, -0.0027384376153349876, -0.042264558374881744, 0.01618490368127823, 0.07992301136255264, -0.04095321521162987, 0.02075677551329136, -0.06651144474744797, 0.06372585147619247, -0.11786920577287674, 0.09625071287155151, 0.01063506118953228, 0.016993753612041473, -0.0417880080640316, -0.01618220843374729, 0.039470795542001724, -0.057925306260585785, 0.07921463251113892, 0.011758086271584034, 0.0010938759660348296, 0.10196787863969803, -0.0034960443153977394, 0.06409632414579391, -0.05372481048107147, -0.023290161043405533, 0.06578411161899567, -0.05874887853860855, -0.03370826691389084, -0.1573946475982666, -0.0709633082151413, 0.020051732659339905, -0.04775108024477959, 0.002077929675579071, 0.03673801198601723, 0.062159497290849686, -0.06937079131603241, -0.12125655263662338, -0.043812792748212814, -0.028638383373618126, 0.021301284432411194, 0.10829301923513412, -0.07526551932096481, 0.1547859013080597, -0.052787959575653076, -0.00020603960729204118, 0.07437096536159515, 0.04048224538564682, 0.01393822580575943, -0.10422444343566895, -0.04698587954044342, -0.11035211384296417, 0.1502903699874878, -0.007902312092483044, -0.03533121198415756, 0.03719403222203255, -0.11946307867765427, -0.1572723090648651, 0.03418220207095146, 0.10199101269245148, 0.0448341928422451, 0.025807438418269157, 0.027079269289970398, -0.04042419046163559, -0.021270349621772766, -0.07034418731927872, 0.0882953479886055, -0.12085357308387756, -0.09669415652751923, 0.09555385261774063, 0.12178351730108261, -0.0036850625183433294, -0.07441367954015732, 0.11554073542356491, -0.021787192672491074, 0.05525410920381546, -0.02971339225769043, 0.10308072715997696, 0.0796005055308342, -0.12273547053337097, 0.005693064536899328, -0.036891788244247437, -0.0741485133767128, -0.12975730001926422, 0.019545545801520348, -0.061916105449199677, -0.13383042812347412, 0.12179028987884521, -0.09376577287912369, 0.030037038028240204, -0.10506992787122726, 0.021338803693652153, 0.01864001713693142, 0.061665527522563934, -0.10988292098045349, 0.08575301617383957, 0.13424484431743622, -0.043199893087148666, -0.07184189558029175, -0.12455986440181732, -0.05022053420543671, -0.04231856390833855, -0.13957437872886658, -0.11600435525178909, 0.0100301094353199, -0.023418782278895378, -0.05818291753530502, 0.0015462689334526658, -0.03659068048000336, 0.008594646118581295, 0.021907730028033257, 0.04032021388411522, -0.02693161368370056, 0.05134565755724907, -0.057569269090890884, -0.052510857582092285, 0.11489357799291611, 0.04113486409187317, -0.03561042994260788, -0.052359987050294876, 0.12997733056545258, -0.11959461867809296, 0.07662346214056015, -0.020313527435064316, 0.017129231244325638, -0.06435854732990265, 0.17131924629211426, 0.11673715710639954, -0.1367570012807846, -0.005008010193705559, -0.08210669457912445, 0.020409544929862022, 0.023555370047688484, 0.13693512976169586, -0.03411718085408211, -0.0012358218664303422, -0.1580323874950409, 0.018575575202703476, -0.18557456135749817, -0.03716109320521355, 0.04671547934412956, 0.09917585551738739, 0.15293832123279572, -0.0034432117827236652, -0.1263325810432434, 0.10424192249774933, -0.2118520885705948, 0.0907607227563858, 0.05121984705328941, -0.11874113976955414, -0.06765396893024445, -0.06795281916856766, 0.1198519766330719, 0.009196433238685131, 0.2040700763463974, -0.013615905307233334, -0.09132910519838333, -0.07060808688402176, -0.01980910450220108, -0.030524181202054024, 0.09714830666780472, 0.041414931416511536, 0.04653804749250412, 0.12821412086486816, 0.00368314771912992, 0.07533777505159378, 0.060310911387205124, 0.02759413793683052, -0.012300663627684116, 0.04076618701219559, 0.08261215686798096, -0.14588621258735657, -0.1659701019525528, 0.1326720416545868, 0.025149408727884293, 0.11792458593845367, 0.03658788278698921, -0.1549617499113083, 0.06687124073505402, 0.2523096203804016, -0.11147607117891312, 0.02505038119852543, 0.12737524509429932, -0.0366884209215641, 0.0672016367316246, 0.1144871786236763, -0.02633814327418804, -0.05217865854501724, -0.011363590136170387, 0.10233135521411896, 0.028660254552960396, -0.04646271467208862, -0.02340836264193058, -0.03373933956027031, -0.019070526584982872, -0.011738128960132599, -0.0909019410610199, -0.1543993502855301, -0.10471053421497345, -0.16619662940502167, 0.04399140924215317, -0.04626438021659851, 0.13418889045715332, 0.09469578415155411, -0.012723101302981377, 0.04568437114357948, 0.028575526550412178, 0.07275456190109253, 0.07916246354579926, -0.02939477376639843, -0.036159269511699677 ]
null
null
peft
# checkpoints This model is a fine-tuned version of [tiiuae/falcon-7b](https://huggingface.co/tiiuae/falcon-7b) on the Red Solar Sky dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure The following `bitsandbytes` quantization config was used during training: - quant_method: bitsandbytes - load_in_8bit: False - load_in_4bit: True - llm_int8_threshold: 6.0 - llm_int8_skip_modules: None - llm_int8_enable_fp32_cpu_offload: False - llm_int8_has_fp16_weight: False - bnb_4bit_quant_type: nf4 - bnb_4bit_use_double_quant: False - bnb_4bit_compute_dtype: bfloat16 ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results ### Framework versions - PEFT 0.6.0.dev0 - Transformers 4.37.2 - Pytorch 2.1.2 - Datasets 2.14.5 - Tokenizers 0.15.1
{"language": ["en"], "license": "apache-2.0", "library_name": "peft", "tags": ["generated_from_trainer", "pretrained", "lora"], "datasets": ["generator"], "base_model": "tiiuae/falcon-7b", "pipeline_tag": "text-generation", "widget": [{"text": "Hello how are you?"}, {"text": "Nice to meet you."}, {"text": "Good afternoon."}], "model-index": [{"name": "checkpoints", "results": []}]}
text-generation
KZMTx/RedSolarSkyAdapter
[ "peft", "safetensors", "generated_from_trainer", "pretrained", "lora", "text-generation", "en", "dataset:generator", "base_model:tiiuae/falcon-7b", "license:apache-2.0", "region:us" ]
2024-02-15T03:49:24+00:00
[]
[ "en" ]
TAGS #peft #safetensors #generated_from_trainer #pretrained #lora #text-generation #en #dataset-generator #base_model-tiiuae/falcon-7b #license-apache-2.0 #region-us
# checkpoints This model is a fine-tuned version of tiiuae/falcon-7b on the Red Solar Sky dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure The following 'bitsandbytes' quantization config was used during training: - quant_method: bitsandbytes - load_in_8bit: False - load_in_4bit: True - llm_int8_threshold: 6.0 - llm_int8_skip_modules: None - llm_int8_enable_fp32_cpu_offload: False - llm_int8_has_fp16_weight: False - bnb_4bit_quant_type: nf4 - bnb_4bit_use_double_quant: False - bnb_4bit_compute_dtype: bfloat16 ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results ### Framework versions - PEFT 0.6.0.dev0 - Transformers 4.37.2 - Pytorch 2.1.2 - Datasets 2.14.5 - Tokenizers 0.15.1
[ "# checkpoints\n\nThis model is a fine-tuned version of tiiuae/falcon-7b on the Red Solar Sky dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- quant_method: bitsandbytes\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: False\n- bnb_4bit_compute_dtype: bfloat16", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 30\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- PEFT 0.6.0.dev0\n- Transformers 4.37.2\n- Pytorch 2.1.2\n- Datasets 2.14.5\n- Tokenizers 0.15.1" ]
[ "TAGS\n#peft #safetensors #generated_from_trainer #pretrained #lora #text-generation #en #dataset-generator #base_model-tiiuae/falcon-7b #license-apache-2.0 #region-us \n", "# checkpoints\n\nThis model is a fine-tuned version of tiiuae/falcon-7b on the Red Solar Sky dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- quant_method: bitsandbytes\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: False\n- bnb_4bit_compute_dtype: bfloat16", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 30\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- PEFT 0.6.0.dev0\n- Transformers 4.37.2\n- Pytorch 2.1.2\n- Datasets 2.14.5\n- Tokenizers 0.15.1" ]
[ 62, 31, 6, 12, 8, 165, 102, 4, 40 ]
[ "passage: TAGS\n#peft #safetensors #generated_from_trainer #pretrained #lora #text-generation #en #dataset-generator #base_model-tiiuae/falcon-7b #license-apache-2.0 #region-us \n# checkpoints\n\nThis model is a fine-tuned version of tiiuae/falcon-7b on the Red Solar Sky dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- quant_method: bitsandbytes\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: False\n- bnb_4bit_compute_dtype: bfloat16### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 30\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- PEFT 0.6.0.dev0\n- Transformers 4.37.2\n- Pytorch 2.1.2\n- Datasets 2.14.5\n- Tokenizers 0.15.1" ]
[ -0.07995551079511642, 0.14239083230495453, -0.004142929334193468, 0.10216210782527924, 0.10531467944383621, 0.040955424308776855, 0.11152805387973785, 0.15043693780899048, -0.06970831006765366, 0.07949884235858917, 0.0594027116894722, 0.07064477354288101, 0.10581821203231812, 0.11432672291994095, -0.03401099517941475, -0.11089673638343811, 0.0201824102550745, -0.05040301755070686, -0.03397669643163681, 0.09465644508600235, 0.08845323324203491, -0.05797550454735756, 0.04609017074108124, -0.019286002963781357, -0.1330222636461258, 0.009220926091074944, -0.011676358990371227, -0.04045655578374863, 0.07588878273963928, 0.04929148405790329, 0.05958576500415802, 0.012667794711887836, 0.07543490082025528, -0.24727177619934082, -0.000018473670934326947, 0.0977557972073555, -0.008370870724320412, 0.09550055861473083, 0.04812287166714668, 0.03652159124612808, 0.04728017374873161, -0.08077943325042725, 0.09252920001745224, 0.052710164338350296, -0.11036942899227142, -0.18454323709011078, -0.11377417296171188, 0.0696391835808754, 0.05364227667450905, 0.10637961328029633, -0.006606051232665777, 0.16756321489810944, -0.0701218768954277, 0.08545128256082535, 0.12016710638999939, -0.21123087406158447, -0.056041594594717026, 0.03408707305788994, -0.0006246032426133752, 0.0495549812912941, -0.11757782846689224, -0.05359230563044548, 0.035591915249824524, 0.03364076465368271, 0.055290624499320984, -0.0005105460295453668, -0.1089278981089592, -0.0035019605420529842, -0.0866205096244812, -0.04574200510978699, 0.0870862826704979, 0.039165131747722626, -0.05010160431265831, -0.12057015299797058, -0.039343319833278656, -0.13550305366516113, 0.02719224989414215, -0.009080981835722923, -0.016231466084718704, -0.017498809844255447, -0.02439088001847267, -0.017372697591781616, -0.05335482582449913, -0.05122806131839752, 0.020012320950627327, 0.049429986625909805, 0.03436115384101868, 0.012796774506568909, -0.013061432167887688, 0.1315581500530243, 0.006840186193585396, -0.14493536949157715, -0.04242464527487755, -0.028247127309441566, -0.09615985304117203, -0.02467452548444271, -0.031133221462368965, 0.05441080033779144, -0.0009300839155912399, 0.20905420184135437, -0.05619245395064354, 0.10290546715259552, -0.02372003346681595, 0.0035446034744381905, -0.024605441838502884, 0.12649904191493988, -0.052106816321611404, -0.06246725469827652, 0.012825562618672848, 0.11414185166358948, -0.004462735261768103, -0.023595547303557396, -0.05706183612346649, 0.015530219301581383, 0.08209948241710663, 0.031153930351138115, -0.016395092010498047, 0.006234221626073122, -0.1048179417848587, -0.04054578021168709, 0.0967601016163826, -0.11904661357402802, 0.044563066214323044, 0.039876680821180344, -0.09120204299688339, -0.024110816419124603, 0.09422145038843155, -0.011512361466884613, -0.009529855102300644, 0.002282050671055913, -0.051725100725889206, -0.02044801414012909, -0.09347721934318542, -0.0811343565583229, 0.015842502936720848, -0.04960229992866516, -0.008030431345105171, -0.057993900030851364, -0.1478169709444046, -0.04757002741098404, 0.043631237000226974, -0.10166813433170319, -0.02634768933057785, -0.03685371205210686, -0.10540332645177841, 0.017926063388586044, 0.00691435020416975, 0.1251601278781891, -0.026830019429326057, 0.069893017411232, 0.027699213474988937, 0.033348824828863144, 0.011526337824761868, 0.005955612752586603, -0.059712085872888565, 0.034258998930454254, -0.10374985635280609, 0.11646199971437454, -0.08650178462266922, 0.00846999604254961, -0.1267906278371811, -0.08158987015485764, 0.010230137966573238, -0.04639109596610069, 0.11376416683197021, 0.11799854040145874, -0.1966673880815506, -0.0038487736601382494, 0.06842903792858124, -0.05835657939314842, -0.09345375746488571, 0.06447199732065201, -0.059246975928545, 0.10172399878501892, 0.02359057404100895, 0.04875198379158974, 0.12738817930221558, -0.13970886170864105, -0.04030817374587059, -0.023472756147384644, 0.07084213197231293, 0.09687823057174683, 0.04095398634672165, 0.03231525793671608, -0.01593243144452572, 0.011923306621611118, -0.02365465648472309, 0.017460960894823074, -0.0490763783454895, -0.08642655611038208, -0.03047015890479088, -0.06563489139080048, 0.023379122838377953, 0.03293919190764427, -0.015130642801523209, -0.04171856865286827, -0.13406944274902344, 0.06488025188446045, 0.12857793271541595, -0.0565253347158432, 0.013558004051446915, -0.09972450882196426, -0.0002818484208546579, 0.014461306855082512, -0.025123341009020805, -0.16765514016151428, -0.07901962101459503, 0.034650687128305435, -0.10343561321496964, 0.029907342046499252, -0.05055954307317734, 0.06101793423295021, 0.06553017348051071, -0.07252839207649231, -0.01959233731031418, -0.10366654396057129, 0.010763116180896759, -0.06048886105418205, -0.2022911161184311, -0.05019242689013481, -0.014608979225158691, 0.24690160155296326, -0.21686993539333344, 0.027584120631217957, 0.036457110196352005, 0.11465363204479218, 0.00489312969148159, -0.059118784964084625, -0.018918152898550034, 0.022796839475631714, -0.011628375388681889, -0.08413057029247284, 0.008240118622779846, 0.030598655343055725, -0.11064382642507553, -0.06307217478752136, -0.12227141112089157, -0.045048195868730545, 0.09337833523750305, 0.06268678605556488, -0.1348140835762024, -0.0020691361278295517, -0.037728287279605865, -0.06397917121648788, -0.02560908906161785, -0.019332708790898323, 0.17105776071548462, 0.020662331953644753, 0.10090760141611099, -0.07260983437299728, -0.07838556915521622, 0.01775972731411457, 0.017368244007229805, -0.018958337604999542, 0.1306152194738388, 0.024026809260249138, -0.1327284574508667, 0.08021335303783417, 0.07298188656568527, -0.03924411162734032, 0.0845046117901802, -0.010694948956370354, -0.06029002368450165, -0.07941778749227524, 0.038202039897441864, 0.04470326378941536, 0.13236482441425323, -0.03127780929207802, 0.08102931082248688, 0.03429574519395828, -0.02079903893172741, 0.026186535134911537, -0.16811057925224304, 0.0018645867239683867, 0.02259109914302826, -0.02754119411110878, -0.029685718938708305, -0.033668287098407745, 0.033535316586494446, 0.06290162354707718, -0.005244350526481867, -0.0074278865940868855, 0.03398391231894493, -0.030798304826021194, -0.08904534578323364, 0.1950514167547226, -0.14964929223060608, -0.15531709790229797, -0.13887561857700348, 0.03237447887659073, -0.056341543793678284, -0.04821416735649109, -0.026216747239232063, -0.06179383397102356, -0.04537288099527359, -0.09766040742397308, -0.0476115420460701, 0.005036858841776848, -0.01665058732032776, 0.03728140518069267, 0.030833018943667412, 0.13688164949417114, -0.12162309885025024, 0.02839583531022072, 0.03591347113251686, -0.06459593772888184, 0.022726619616150856, 0.03667069971561432, 0.04118013754487038, 0.10508023202419281, 0.017489394173026085, -0.00325835426338017, -0.0036041804123669863, 0.304442822933197, -0.08297661691904068, 0.033528752624988556, 0.12163584679365158, -0.014622541144490242, 0.056999869644641876, 0.13897180557250977, 0.05392010509967804, -0.09278444945812225, 0.044659942388534546, 0.0520569384098053, -0.015101183205842972, -0.2345050722360611, -0.013454627245664597, -0.03724418208003044, -0.06251861900091171, 0.09132478386163712, 0.03705697879195213, 0.09250625967979431, 0.04175009950995445, -0.023191625252366066, 0.03929177299141884, 0.0542694628238678, 0.08986900001764297, 0.04784798622131348, 0.05558393523097038, 0.11006224155426025, -0.011253072880208492, 0.040920283645391464, 0.06941839307546616, 0.04402249678969383, 0.17814642190933228, -0.06125156953930855, 0.08566191047430038, 0.05478106066584587, 0.18319785594940186, -0.02920132502913475, 0.01591038890182972, -0.01295425370335579, 0.005157790146768093, 0.011012195609509945, -0.07575777173042297, -0.013271655887365341, 0.04014860838651657, -0.030774645507335663, 0.08663841336965561, -0.030496934428811073, -0.061005622148513794, 0.04177051782608032, 0.14495153725147247, 0.09214961528778076, -0.2656891644001007, -0.094942606985569, 0.02154158428311348, -0.0014909333549439907, -0.07856015861034393, -0.035150881856679916, 0.11409846693277359, -0.13977161049842834, 0.04532118886709213, -0.06592842936515808, 0.08180531114339828, -0.02870710752904415, -0.007310418412089348, 0.1354215443134308, 0.09015744924545288, 0.0011604266474023461, 0.06489300727844238, -0.24457141757011414, 0.13241782784461975, 0.032307446002960205, 0.08913474529981613, -0.06256865710020065, 0.0568140484392643, -0.00829364638775587, -0.039275314658880234, 0.06589767336845398, -0.017840031534433365, -0.01221519149839878, -0.2247428447008133, -0.0811162143945694, -0.008712820708751678, 0.09763885289430618, -0.04025067761540413, 0.10061103850603104, 0.0006847570184618235, 0.007147453725337982, 0.04472479969263077, 0.009792443364858627, -0.1585841178894043, -0.15017324686050415, 0.0494396835565567, 0.0173011664301157, -0.03791050612926483, -0.06546099483966827, -0.07220711559057236, -0.07525473088026047, 0.12674590945243835, -0.1542128324508667, -0.03914562985301018, -0.11758219450712204, 0.06199127808213234, 0.17196017503738403, -0.04067009314894676, 0.023431725800037384, 0.007856402546167374, 0.06995387375354767, 0.010612302459776402, -0.06552856415510178, 0.050877220928668976, -0.060973163694143295, -0.21681582927703857, -0.05631459876894951, 0.12115569412708282, 0.10258551687002182, 0.04909311234951019, -0.016585953533649445, 0.03045867756009102, -0.016246573999524117, -0.11762534081935883, 0.002548014046624303, 0.06390372663736343, 0.08943760395050049, 0.04690581560134888, -0.04330361634492874, 0.046666570007801056, -0.07721127569675446, -0.003160295309498906, 0.023004259914159775, 0.2835082709789276, -0.07412221282720566, 0.06037922203540802, 0.0300588496029377, -0.06766057014465332, -0.1413816213607788, 0.05235502868890762, 0.11274361610412598, 0.034070271998643875, 0.07321377098560333, -0.18283186852931976, 0.06502973288297653, 0.09590088576078415, -0.019756365567445755, 0.04624541103839874, -0.34039291739463806, -0.13379621505737305, 0.027212155982851982, 0.06471578031778336, -0.05157070979475975, -0.09774310141801834, -0.03215101361274719, -0.031683918088674545, -0.11371345818042755, 0.135402649641037, -0.1120680719614029, 0.08969200402498245, 0.007604304235428572, 0.03884493187069893, 0.034635286778211594, -0.040891192853450775, 0.1991366446018219, 0.010137014091014862, 0.0689481794834137, -0.030533889308571815, 0.011268042027950287, 0.10579270869493484, -0.06906910985708237, 0.015956230461597443, -0.04250050708651543, 0.07003862410783768, -0.13854211568832397, 0.00601439643651247, -0.0800841674208641, 0.0075663127936422825, -0.09247030317783356, -0.02503778785467148, -0.04811255633831024, 0.11054224520921707, 0.0715700164437294, -0.027772337198257446, 0.048678264021873474, 0.006885121110826731, 0.1078956201672554, 0.24930153787136078, 0.03692946210503578, 0.08491621166467667, -0.05120879411697388, 0.02356136403977871, -0.006174883339554071, 0.05305543914437294, -0.11142227053642273, 0.04693052917718887, 0.1061774417757988, 0.020975740626454353, 0.1244841143488884, 0.04358965530991554, -0.1094159185886383, -0.003987821284681559, 0.02721717581152916, -0.09970896691083908, -0.05765214189887047, 0.004777231253683567, 0.02445359341800213, -0.12760253250598907, -0.022914448752999306, 0.12379662692546844, -0.049669813364744186, -0.013411055319011211, 0.04069501534104347, 0.04242792725563049, -0.04047289118170738, 0.1883370727300644, 0.031081628054380417, 0.06859687715768814, -0.09650326520204544, 0.12700088322162628, 0.10187714546918869, -0.01887699030339718, 0.029856529086828232, 0.047077443450689316, -0.05677705258131027, -0.001837127492763102, 0.04353821277618408, 0.022320253774523735, 0.0196000374853611, -0.033250633627176285, -0.07002044469118118, -0.091375932097435, 0.038098350167274475, 0.09295617789030075, 0.029689714312553406, 0.03338540717959404, -0.01793518103659153, 0.00495385704562068, -0.12267421185970306, 0.09530957788228989, 0.04571544751524925, 0.042188871651887894, -0.11241130530834198, 0.0997709110379219, 0.011759527027606964, 0.017920304089784622, 0.0013050950365141034, 0.00345103582367301, -0.13832098245620728, -0.0028585572727024555, -0.05671265721321106, 0.015371509827673435, 0.02139413356781006, 0.016302529722452164, -0.014212144538760185, -0.012474384158849716, -0.031025346368551254, 0.05542683228850365, -0.0546015165746212, -0.07910557836294174, 0.026565933600068092, 0.044966232031583786, -0.12053973227739334, -0.02786959521472454, 0.06477352976799011, -0.1416672319173813, 0.057833537459373474, 0.03259072080254555, 0.014668527990579605, 0.01657683774828911, -0.06446535140275955, 0.018125610426068306, -0.002885781927034259, 0.04014751315116882, 0.03665483742952347, -0.15059049427509308, 0.004987870343029499, -0.055661868304014206, -0.033019572496414185, 0.02542402781546116, 0.06669997423887253, -0.09080126881599426, -0.04644197225570679, -0.03147056698799133, -0.036041948944330215, -0.04234880581498146, 0.07000444084405899, 0.10229597240686417, 0.02959047071635723, 0.13064053654670715, -0.06638673692941666, 0.04574340581893921, -0.22354988753795624, -0.058416951447725296, 0.010147897526621819, -0.03781207278370857, -0.06315121799707413, -0.020561465993523598, 0.09778570383787155, -0.03635178133845329, 0.09449482709169388, -0.08497951179742813, 0.08158945292234421, 0.01443200372159481, -0.03375875577330589, -0.07157046347856522, 0.03374146670103073, 0.16170990467071533, 0.04626566544175148, 0.012537682428956032, 0.0670250654220581, -0.0062545714899897575, 0.006306260358542204, 0.08140718936920166, 0.11595147103071213, 0.13322652876377106, -0.005116344429552555, 0.0435832217335701, 0.036440879106521606, -0.10298970341682434, -0.0123938312754035, 0.2380334585905075, -0.07418429851531982, 0.06457998603582382, -0.09782178699970245, 0.12239311635494232, 0.04573361575603485, -0.1898219734430313, 0.06411019712686539, -0.030330149456858635, -0.10814933478832245, -0.09480655938386917, -0.05475331470370293, -0.05666667968034744, -0.09794282913208008, 0.012676986865699291, -0.10933855175971985, 0.03329257294535637, 0.08769166469573975, 0.023508500307798386, 0.00431171664968133, 0.09979449957609177, -0.023134203627705574, -0.007377606350928545, 0.03579946234822273, 0.02650391310453415, 0.002106876578181982, -0.05709259584546089, -0.04463430866599083, 0.08243833482265472, 0.045021913945674896, 0.09049684554338455, -0.015085356310009956, 0.06069697067141533, 0.028938017785549164, -0.0031194582115858793, -0.08190290629863739, 0.03217117488384247, -0.015228271484375, 0.016262110322713852, 0.10522028803825378, 0.04660531505942345, 0.025865809991955757, -0.04588954523205757, 0.2120314985513687, -0.07764272391796112, -0.05868354067206383, -0.16204681992530823, 0.1942339688539505, -0.011819841340184212, 0.008900475688278675, 0.05195912346243858, -0.10659244656562805, -0.03913862258195877, 0.12069597095251083, 0.09285201132297516, -0.10931751132011414, -0.021359402686357498, 0.046689439564943314, -0.018066853284835815, -0.059282947331666946, 0.10214445739984512, 0.07505198568105698, 0.0016246598679572344, -0.037280887365341187, 0.013873339630663395, -0.010530225932598114, -0.04059230163693428, -0.02668025717139244, 0.0796009823679924, -0.03285408765077591, 0.002466836478561163, -0.05783805251121521, 0.05828849598765373, 0.0005484339781105518, -0.21904166042804718, 0.11015769094228745, -0.17798742651939392, -0.17884910106658936, -0.021699853241443634, 0.03416718915104866, -0.03853961080312729, 0.02233388088643551, -0.00014802148507442325, 0.028373342007398605, 0.16132205724716187, -0.019821325317025185, 0.016192035749554634, -0.11807284504175186, 0.05659085139632225, -0.08973030745983124, 0.22763913869857788, -0.01836015097796917, 0.09373486042022705, 0.11419598013162613, -0.00734902685508132, -0.12319622188806534, 0.03492666408419609, 0.07953780889511108, -0.11059856414794922, 0.0038382266648113728, 0.1535676270723343, -0.012553988955914974, 0.1444181352853775, 0.05108914151787758, -0.08220577985048294, 0.009465365670621395, -0.07444005459547043, -0.021579621359705925, -0.07944262027740479, -0.02104620449244976, -0.057092420756816864, 0.15726380050182343, 0.20799264311790466, -0.06292811781167984, 0.019977297633886337, -0.046601150184869766, 0.02591272071003914, 0.010865329764783382, 0.12752538919448853, -0.013322061859071255, -0.19441910088062286, 0.06620054692029953, -0.05828890576958656, 0.07179035246372223, -0.18026548624038696, -0.09702061861753464, 0.06260616332292557, -0.07549349963665009, -0.05630403012037277, 0.13346487283706665, 0.016850169748067856, 0.03076377883553505, -0.03577665239572525, -0.16468334197998047, -0.054325200617313385, 0.14220626652240753, -0.09284841269254684, -0.045622192323207855 ]
null
null
null
# **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="Noname08/Taxi-v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
{"tags": ["Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "Taxi-v3", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "Taxi-v3", "type": "Taxi-v3"}, "metrics": [{"type": "mean_reward", "value": "7.54 +/- 2.73", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
Noname08/Taxi-v3
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
2024-02-15T03:50:03+00:00
[]
[]
TAGS #Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
# Q-Learning Agent playing1 Taxi-v3 This is a trained model of a Q-Learning agent playing Taxi-v3 . ## Usage
[ "# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ "TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n", "# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ 32, 33 ]
[ "passage: TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ 0.048862796276807785, -0.16549694538116455, -0.005485367961227894, 0.02960980497300625, 0.1345081776380539, -0.01784728653728962, 0.11895976960659027, 0.07759871333837509, -0.07461097836494446, -0.055395450443029404, 0.1418241262435913, 0.09088201075792313, 0.055222880095243454, 0.05699880048632622, 0.09511256217956543, -0.27440664172172546, 0.048217080533504486, -0.02918700873851776, 0.05621987581253052, 0.11878681182861328, 0.0670095682144165, -0.040441032499074936, 0.061956584453582764, 0.11818158626556396, -0.1018151044845581, -0.007344264071434736, 0.035402704030275345, -0.09440053254365921, 0.17413531243801117, 0.07204403728246689, 0.12337774783372879, 0.05132639780640602, 0.179361954331398, -0.12762396037578583, 0.024310702458024025, -0.0010275895474478602, -0.10138072073459625, -0.03909514099359512, -0.012415820732712746, -0.08349097520112991, 0.03230205550789833, 0.23522862792015076, 0.07199250161647797, 0.06632792949676514, -0.17707863450050354, -0.06584878265857697, -0.04375573247671127, 0.069611094892025, 0.14951466023921967, 0.03758616745471954, -0.033800311386585236, 0.1684885323047638, -0.2564343810081482, 0.05066783353686333, 0.037275806069374084, -0.42313119769096375, 0.017119819298386574, 0.1507398933172226, 0.15090937912464142, 0.06909667700529099, -0.10573802888393402, 0.013512322679162025, 0.051325585693120956, -0.0005318621988408267, 0.024325110018253326, 0.006554204970598221, 0.15601307153701782, 0.08537693321704865, -0.1487821787595749, -0.058576688170433044, 0.17441977560520172, -0.03788546845316887, -0.02613203600049019, -0.039745692163705826, 0.0067160045728087425, -0.06427708268165588, -0.004067842848598957, -0.1777995079755783, 0.00734262028709054, 0.06666424125432968, -0.014348524622619152, 0.014901017770171165, -0.035522811114788055, -0.0966939702630043, -0.023098144680261612, -0.08592145889997482, 0.01677769608795643, -0.006319406442344189, -0.10187895596027374, 0.05002119392156601, -0.061138734221458435, 0.0014382408699020743, -0.05123179033398628, -0.15047866106033325, -0.049055423587560654, -0.03481535613536835, 0.1474713832139969, -0.0044205985032022, -0.01873963139951229, -0.03164304047822952, 0.15474793314933777, 0.049551334232091904, -0.05370146036148071, 0.05625450983643532, 0.07605006545782089, 0.23867930471897125, 0.10401605814695358, 0.10196955502033234, -0.06798075139522552, 0.10180158913135529, -0.12330973148345947, -0.08915644884109497, -0.17508824169635773, 0.11820860952138901, 0.00015364694991149008, 0.1317785084247589, -0.12023144960403442, 0.07898581773042679, -0.067511186003685, 0.013453764840960503, 0.01636839471757412, 0.0820009782910347, -0.012399360537528992, 0.10676060616970062, -0.005061192903667688, -0.06941985338926315, 0.014177112840116024, 0.05935845896601677, 0.03754841163754463, -0.038601722568273544, -0.03192409873008728, -0.05762290954589844, -0.05065649375319481, -0.10128600150346756, -0.06447898596525192, 0.018573462963104248, -0.007677143905311823, -0.1833900660276413, -0.06407523155212402, 0.00897200871258974, 0.015712225809693336, -0.03988850116729736, -0.05148044601082802, -0.15265507996082306, -0.042461175471544266, -0.015450406819581985, -0.03500641882419586, -0.06214277446269989, -0.0383245050907135, 0.046435944736003876, -0.07560601085424423, 0.013364278711378574, 0.023342855274677277, 0.05405820533633232, -0.025881100445985794, 0.06068144738674164, -0.08357544988393784, 0.09493788331747055, -0.1540430635213852, -0.03271956741809845, -0.025445878505706787, -0.041183918714523315, 0.1752462536096573, 0.06099751964211464, -0.015994304791092873, 0.15260063111782074, -0.17141541838645935, -0.058121129870414734, 0.15596486628055573, 0.008629098534584045, -0.09967197477817535, -0.003560945624485612, -0.09397093951702118, 0.1428760588169098, 0.08571921288967133, 0.2478504776954651, 0.12005335837602615, -0.22748184204101562, 0.055358242243528366, 0.12515293061733246, -0.14365963637828827, 0.10365243256092072, 0.07344598323106766, 0.005470725707709789, -0.18886831402778625, -0.06843198090791702, -0.06121627986431122, 0.1053021252155304, -0.08522345870733261, -0.0776243582367897, 0.09323626756668091, -0.05086790770292282, 0.24641476571559906, -0.028281206265091896, 0.06174173951148987, -0.026681531220674515, -0.1389324963092804, -0.01723906397819519, 0.060955192893743515, 0.05258452147245407, -0.024835573509335518, -0.25895482301712036, 0.13646544516086578, 0.048650871962308884, 0.025074828416109085, 0.004106190986931324, -0.05691491439938545, 0.016934165731072426, 0.1511998474597931, 0.020012924447655678, 0.13717477023601532, 0.027723990380764008, 0.0706823319196701, -0.006239562761038542, -0.10560829937458038, -0.04169593006372452, 0.061916545033454895, -0.08518962562084198, -0.06641357392072678, 0.011197872459888458, -0.06935211271047592, -0.11783787608146667, -0.12166737765073776, -0.026334572583436966, -0.02980303019285202, -0.07444227486848831, 0.02368103712797165, 0.06536602973937988, -0.06702698022127151, -0.0023908785078674555, 0.007125476840883493, -0.011537045240402222, 0.16434046626091003, 0.011393417604267597, -0.007796820718795061, 0.1328643560409546, -0.11533161997795105, 0.12461213022470474, 0.049438029527664185, -0.024806302040815353, -0.04662557691335678, 0.0014137453399598598, -0.057529181241989136, 0.029044216498732567, -0.04390640929341316, 0.02774495631456375, 0.20111067593097687, 0.02772962674498558, 0.11389166116714478, -0.0656520202755928, 0.04385066404938698, -0.007961965166032314, -0.009693224914371967, 0.018563594669103622, 0.07608018070459366, 0.07813210040330887, -0.1324140727519989, 0.02262016013264656, 0.22455167770385742, 0.1385764330625534, 0.18313980102539062, -0.010877152904868126, 0.06325667351484299, -0.04875868931412697, 0.027505528181791306, 0.024100203067064285, 0.10314226150512695, -0.10732068121433258, -0.0322517491877079, -0.025407759472727776, 0.023599207401275635, -0.08197105675935745, -0.1055799350142479, -0.090115025639534, 0.01222382951527834, -0.03125503659248352, -0.15570329129695892, 0.13300658762454987, -0.10451057553291321, 0.01802753657102585, 0.04692702740430832, -0.22163605690002441, 0.11530312895774841, 0.014291439205408096, -0.10303618758916855, 0.11281087249517441, -0.12051989883184433, -0.08699832111597061, -0.05777236074209213, -0.18658851087093353, 0.05280197039246559, 0.04673841595649719, 0.05166793242096901, -0.18521739542484283, 0.024835903197526932, 0.05545609071850777, 0.13426995277404785, -0.09743253141641617, -0.07142634689807892, -0.15038461983203888, 0.016068490222096443, -0.033661190420389175, -0.16029728949069977, -0.005609163548797369, -0.032781440764665604, -0.18849676847457886, -0.04539939761161804, -0.15086813271045685, -0.034627582877874374, 0.20464378595352173, 0.026907702907919884, 0.09480511397123337, -0.07926445454359055, 0.3802889585494995, -0.042039383202791214, -0.06146497279405594, -0.01321389526128769, -0.07072482258081436, 0.02512686513364315, 0.13271741569042206, 0.0036099457647651434, -0.017886579036712646, -0.0037857077550143003, 0.0024592927657067776, -0.06234965845942497, -0.13400450348854065, 0.0028710351325571537, 0.03905198723077774, 0.1874423623085022, 0.004639793653041124, 0.06659388542175293, 0.03133883699774742, 0.057546284049749374, 0.07748064398765564, 0.030926106497645378, 0.0011591583024710417, -0.01591806672513485, 0.06604493409395218, -0.11684755235910416, 0.042466625571250916, -0.030429253354668617, -0.10143838077783585, -0.013183288276195526, 0.07950251549482346, 0.12755028903484344, 0.17849206924438477, -0.04790908098220825, 0.17489230632781982, 0.13580141961574554, 0.16576050221920013, 0.049315933138132095, -0.020801831036806107, -0.08773037046194077, -0.06118565797805786, 0.004774159751832485, -0.031952597200870514, 0.04869702458381653, 0.3231290578842163, 0.037619613111019135, -0.09036035090684891, 0.11149907857179642, 0.009480619803071022, 0.05359881371259689, 0.022797370329499245, -0.11162138730287552, 0.11170321702957153, 0.07968773692846298, -0.06341761350631714, -0.07602835446596146, 0.16758501529693604, -0.1109386757016182, -0.26646625995635986, -0.11410990357398987, -0.012305386364459991, 0.07903840392827988, 0.005651174578815699, 0.05498376116156578, -0.11829282343387604, -0.16034497320652008, -0.034191906452178955, 0.1335442066192627, -0.3077351450920105, 0.2065143585205078, -0.0198091771453619, 0.06707923114299774, -0.039657969027757645, -0.07026876509189606, 0.09694647043943405, 0.13174086809158325, 0.29124146699905396, 0.01396956667304039, 0.04841272905468941, -0.15176129341125488, -0.0976925864815712, 0.0018439020495861769, 0.015482662245631218, -0.02563396655023098, 0.028520405292510986, -0.0540912002325058, 0.008404579944908619, -0.018086453899741173, 0.2102297693490982, -0.11316607892513275, 0.004344627261161804, -0.06968966871500015, -0.11707738786935806, 0.19409789144992828, -0.07178345322608948, -0.04543264955282211, -0.14959357678890228, -0.15512511134147644, -0.004174166824668646, -0.02413962036371231, -0.019664527848362923, -0.17603960633277893, -0.18804074823856354, -0.05204557999968529, -0.005645004566758871, -0.003464865731075406, 0.05867868289351463, -0.07517234236001968, -0.04805335775017738, 0.1009904220700264, -0.07743175327777863, -0.056063808500766754, -0.1103200614452362, 0.1391381323337555, 0.06248528137803078, 0.16743235290050507, 0.05907081440091133, 0.0006117874872870743, 0.11471151560544968, -0.02913086675107479, 0.11103474348783493, -0.11291708797216415, -0.17145049571990967, -0.08334989100694656, -0.018775060772895813, 0.09519003331661224, -0.04789286106824875, 0.0028788831550627947, 0.2550160884857178, 0.14880181849002838, -0.0897710770368576, 0.27680760622024536, 0.04414956644177437, -0.09375058114528656, -0.18432219326496124, -0.15961645543575287, 0.03759992495179176, 0.060025621205568314, 0.13095876574516296, -0.057205069810152054, -0.08483537286520004, -0.08492398262023926, -0.07478608191013336, -0.13140805065631866, -0.24232175946235657, -0.030598774552345276, 0.22874866425991058, 0.08656918257474899, 0.08219650387763977, -0.012482990510761738, -0.01186054851859808, 0.00526038184762001, 0.02680150233209133, 0.12018456310033798, -0.13341329991817474, 0.11107480525970459, 0.022198403254151344, 0.044267985969781876, 0.009712530300021172, 0.07929777354001999, 0.03375575691461563, -0.003218587953597307, -0.0006439819699153304, -0.0988350659608841, -0.2596651017665863, 0.0816885456442833, -0.01623627357184887, -0.09960969537496567, 0.014988959766924381, 0.02061903104186058, -0.2089255303144455, 0.011128270998597145, -0.019883770495653152, -0.03150356933474541, -0.06483490765094757, -0.10664787143468857, -0.056551624089479446, 0.04928823933005333, 0.10853826254606247, 0.011660109274089336, 0.05354316532611847, -0.0404130220413208, 0.07917837053537369, 0.0826287642121315, 0.15132710337638855, 0.06795957684516907, -0.190711110830307, -0.10953907668590546, -0.0414445661008358, 0.12121522426605225, -0.12505418062210083, 0.036917757242918015, 0.053161121904850006, -0.016534561291337013, 0.14621229469776154, 0.1070784479379654, -0.07452095299959183, 0.11915595084428787, 0.08904775977134705, -0.04094788804650307, -0.23367151618003845, -0.07120766490697861, 0.11133213341236115, 0.07195597887039185, -0.03961895406246185, 0.018120890483260155, -0.04960581287741661, -0.013980977237224579, 0.048759616911411285, -0.0538676381111145, -0.07230538129806519, 0.004421027842909098, 0.1247575581073761, 0.1029362753033638, -0.04655474051833153, 0.01296416949480772, 0.037371400743722916, 0.003788623260334134, 0.04730486497282982, 0.0407949760556221, -0.08269952982664108, -0.04124005511403084, 0.02782733179628849, 0.37552911043167114, -0.010165480896830559, -0.020456433296203613, 0.018555615097284317, -0.19949445128440857, 0.09135842323303223, 0.13205479085445404, 0.04697350412607193, 0.004247748292982578, -0.08139242231845856, 0.026877427473664284, -0.010625290684401989, 0.09936143457889557, -0.07806670665740967, -0.05493134260177612, -0.21631066501140594, -0.025010565295815468, 0.017490221187472343, 0.24077683687210083, -0.08458559215068817, -0.12801732122898102, -0.20628872513771057, 0.13128381967544556, -0.11333390325307846, -0.03695881739258766, -0.024473199620842934, 0.03926658630371094, -0.01989821158349514, 0.06291737407445908, -0.0710630789399147, 0.006373001262545586, -0.11024709790945053, 0.055267609655857086, 0.04204455390572548, 0.1229788213968277, 0.014207782223820686, 0.02016810141503811, 0.05822525918483734, -0.01837925612926483, 0.07173580676317215, -0.06203491613268852, -0.04550490900874138, 0.14224006235599518, -0.020255116745829582, -0.04152837023139, -0.0483345128595829, -0.036874305456876755, 0.11981741338968277, -0.05059147998690605, -0.007141099311411381, -0.054929375648498535, -0.06906463205814362, 0.03462086617946625, -0.009175732731819153, -0.008798843249678612, 0.06801853328943253, 0.04024988040328026, -0.026994358748197556, 0.005263668950647116, 0.03447828069329262, -0.10330043733119965, -0.04955084249377251, 0.16955432295799255, -0.0749620869755745, 0.10274054110050201, -0.031069839373230934, 0.018015999346971512, 0.005847334861755371, -0.022399673238396645, -0.015360680408775806, -0.1457086056470871, -0.06137600541114807, -0.09489979594945908, 0.11565322428941727, 0.08146517723798752, 0.03358805552124977, 0.04274565726518631, 0.019532648846507072, -0.04414922371506691, -0.038583990186452866, 0.12961317598819733, 0.08133101463317871, 0.012996876612305641, 0.01137041300535202, 0.01941833831369877, -0.020302120596170425, 0.0028480992186814547, -0.01250747125595808, -0.07239153981208801, -0.05874783173203468, 0.09400010108947754, 0.1600283533334732, -0.06127211079001427, -0.13325586915016174, -0.020593497902154922, 0.04988488554954529, 0.0014717020094394684, -0.08777432143688202, 0.04833676666021347, 0.15805292129516602, -0.05623878911137581, 0.03216489031910896, -0.09984751045703888, -0.07263360917568207, -0.16060975193977356, -0.10029061883687973, -0.06092562898993492, -0.28350353240966797, 0.09752398729324341, 0.006392303854227066, -0.014731393195688725, 0.059529416263103485, 0.051305368542671204, -0.052508849650621414, 0.07068239152431488, -0.18146829307079315, -0.007054794579744339, 0.03497592359781265, -0.13212306797504425, 0.02475893869996071, -0.2378365397453308, 0.10198072344064713, -0.04623803123831749, -0.1519704908132553, -0.04004510119557381, 0.0641569048166275, -0.09540136158466339, -0.01822364516556263, -0.0475153923034668, -0.01922670193016529, 0.01624443754553795, -0.009348669089376926, -0.031147832050919533, 0.13716529309749603, 0.02827494591474533, -0.03268734738230705, 0.005254602525383234, 0.0223685409873724, 0.03955082967877388, -0.0969657450914383, -0.05986930429935455, 0.08311155438423157, -0.031056145206093788, 0.14728976786136627, 0.000341245875461027, 0.04181376099586487, -0.06758682429790497, 0.2593761384487152, 0.2023983597755432, -0.12479214370250702, 0.008118697442114353, -0.021801479160785675, 0.012670028023421764, -0.041751839220523834, 0.13110700249671936, 0.013386172242462635, 0.12186761200428009, -0.17513342201709747, -0.01036517322063446, -0.0818324014544487, -0.04501292482018471, 0.06702108681201935, 0.14714950323104858, 0.15742522478103638, 0.03436789661645889, -0.07328428328037262, 0.06722653657197952, -0.30119743943214417, 0.20540550351142883, -0.1346001923084259, -0.01498429011553526, -0.040251150727272034, -0.058389630168676376, 0.061147745698690414, 0.11309876292943954, 0.10832664370536804, -0.021150551736354828, -0.0905047357082367, -0.04486766457557678, -0.039378076791763306, -0.13019338250160217, -0.02718670479953289, 0.1654091775417328, 0.06799814850091934, 0.31520840525627136, -0.017577875405550003, 0.07702425122261047, 0.034410297870635986, 0.06451138854026794, 0.004519328009337187, 0.09537279605865479, 0.07960964739322662, -0.06345855444669724, -0.07373003661632538, -0.001637450186535716, 0.05033271387219429, 0.14567798376083374, -0.03826142102479935, -0.18691548705101013, 0.15858715772628784, 0.07192251086235046, -0.13762691617012024, -0.05777517706155777, 0.08409425616264343, -0.0739973932504654, 0.0550808347761631, 0.08115427941083908, 0.015876613557338715, -0.017793258652091026, -0.004664506763219833, 0.06074233725667, 0.024694660678505898, -0.02343848906457424, 0.003570882137864828, -0.08337053656578064, -0.04151543974876404, 0.07267895340919495, -0.0844460055232048, -0.20546193420886993, -0.0957019031047821, -0.07551700621843338, 0.030557552352547646, -0.0649830624461174, 0.12575586140155792, 0.1717868149280548, 0.0593598335981369, -0.03307248651981354, -0.10721943527460098, -0.035562749952077866, 0.07602505385875702, -0.044773899018764496, -0.09409699589014053 ]
null
null
diffusers
# Umi Sonoda <Gallery /> ## Model description This model was trained to generate high quality images based on SIFAS cards. To achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement. ## Trigger words You should use `id_umi_sonoda` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/theidoldaily/umi-sonoda/tree/main) them in the Files & versions tab.
{"license": "mit", "tags": ["text-to-image", "stable-diffusion", "lora", "diffusers", "template:sd-lora"], "widget": [{"text": "masterpiece, high quality, defined pupil, looking at viewer, rounded pupil, defined iris, (soft iris:1.2),", "parameters": {"negative_prompt": "bad_anatomy, deformation, amputation, deformity, deformed_nipples, duplicated_torso, deformed_torso, long_torso, large_torso, unproportioned_torso, (deformed_pussy:1.2), (deformed_hands:1.2), unproportioned_eyes, unproportioned_head, small_head, duplicated_nose, big_nose, fusioned_clothes, fusioned_arms, undefined_limbs, divided_pussy, red_pussy, duplicated_pussy, deformed_anus, deformed_pussy,"}, "output": {"url": "images/umi_final.png"}}], "base_model": "cagliostrolab/animagine-xl-3.0", "instance_prompt": "id_umi_sonoda"}
text-to-image
theidoldaily/umi-sonoda
[ "diffusers", "text-to-image", "stable-diffusion", "lora", "template:sd-lora", "base_model:cagliostrolab/animagine-xl-3.0", "license:mit", "region:us" ]
2024-02-15T03:51:54+00:00
[]
[]
TAGS #diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-cagliostrolab/animagine-xl-3.0 #license-mit #region-us
# Umi Sonoda <Gallery /> ## Model description This model was trained to generate high quality images based on SIFAS cards. To achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement. ## Trigger words You should use 'id_umi_sonoda' to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. Download them in the Files & versions tab.
[ "# Umi Sonoda\n\n<Gallery />", "## Model description \n\nThis model was trained to generate high quality images based on SIFAS cards.\n\nTo achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement.", "## Trigger words\n\nYou should use 'id_umi_sonoda' to trigger the image generation.", "## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab." ]
[ "TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-cagliostrolab/animagine-xl-3.0 #license-mit #region-us \n", "# Umi Sonoda\n\n<Gallery />", "## Model description \n\nThis model was trained to generate high quality images based on SIFAS cards.\n\nTo achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement.", "## Trigger words\n\nYou should use 'id_umi_sonoda' to trigger the image generation.", "## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab." ]
[ 56, 10, 68, 21, 28 ]
[ "passage: TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-cagliostrolab/animagine-xl-3.0 #license-mit #region-us \n# Umi Sonoda\n\n<Gallery />## Model description \n\nThis model was trained to generate high quality images based on SIFAS cards.\n\nTo achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement.## Trigger words\n\nYou should use 'id_umi_sonoda' to trigger the image generation.## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab." ]
[ -0.08195394277572632, -0.07577036321163177, -0.0018285684054717422, 0.04095853120088577, 0.12116023898124695, 0.03675244003534317, 0.25505033135414124, 0.07059222459793091, 0.14201951026916504, 0.04783869907259941, -0.009234510362148285, 0.04512912407517433, 0.06188968941569328, 0.20715805888175964, -0.046485535800457, -0.2567084729671478, 0.09093412011861801, -0.054640043526887894, -0.029842887073755264, 0.007877602241933346, 0.060100965201854706, -0.049286700785160065, 0.12573470175266266, -0.03361620754003525, -0.10296151041984558, 0.019962837919592857, -0.0011960952542722225, 0.008773531764745712, 0.00816673506051302, 0.056633636355400085, 0.023311233147978783, 0.08549617230892181, 0.08505554497241974, -0.12867876887321472, 0.06763244420289993, 0.004165761172771454, -0.02530420944094658, -0.0081029636785388, 0.0008208949584513903, -0.04511275142431259, 0.2217537760734558, -0.06929726153612137, -0.09648440778255463, -0.029050705954432487, -0.04841279983520508, -0.07418668270111084, -0.004580425098538399, -0.036817681044340134, 0.10972903668880463, 0.0034058312885463238, -0.0019130539149045944, -0.027514491230249405, 0.00997854769229889, 0.03245030716061592, 0.15187372267246246, -0.08714377880096436, -0.049617670476436615, 0.17644160985946655, 0.024251634255051613, 0.13516327738761902, -0.025234989821910858, 0.10734792798757553, 0.11033384501934052, -0.08018241822719574, -0.0017659786390140653, -0.06379486620426178, 0.12049257010221481, -0.04413529485464096, -0.0760892853140831, -0.003831158857792616, 0.29675331711769104, 0.03354305028915405, -0.051598597317934036, -0.10851573199033737, -0.08097471296787262, 0.18815992772579193, -0.08827561885118484, -0.003525989828631282, 0.012039844878017902, 0.030361101031303406, 0.08706838637590408, -0.08042366802692413, -0.08994858711957932, -0.07649888843297958, -0.04163133725523949, 0.18530835211277008, 0.054041411727666855, 0.05604808032512665, -0.03828626871109009, 0.04134979471564293, -0.118832066655159, -0.14228977262973785, 0.009681225754320621, -0.06290911883115768, 0.03873226419091225, 0.089133121073246, -0.0013247986789792776, -0.11308212578296661, 0.11208520829677582, 0.0013072364963591099, 0.057084619998931885, -0.003926832228899002, -0.003902137978002429, 0.10019206255674362, 0.0559350922703743, 0.01799481362104416, -0.00975793320685625, -0.039028093218803406, 0.033763986080884933, 0.06632216274738312, 0.0805571973323822, -0.08682405203580856, -0.14904356002807617, -0.012974947690963745, -0.09976474195718765, 0.017027845606207848, -0.0136481374502182, 0.0316871702671051, -0.02931329607963562, -0.01601671427488327, 0.09599561244249344, -0.05145417898893356, -0.038170523941516876, -0.06758413463830948, -0.020495271310210228, 0.23902402818202972, 0.0514221154153347, -0.005627509206533432, 0.028055334463715553, 0.06684428453445435, -0.017971251159906387, -0.013701876625418663, -0.04889264330267906, 0.009785063564777374, 0.002708208514377475, -0.12080079317092896, 0.04423099756240845, -0.1587047427892685, -0.2632210850715637, -0.022278979420661926, 0.06378301978111267, -0.026218436658382416, 0.06927226483821869, -0.02041987143456936, -0.0028581065125763416, -0.00839695893228054, 0.0019849406089633703, -0.042691200971603394, -0.07475056499242783, 0.05978234112262726, 0.02130127139389515, 0.10397407412528992, -0.13457056879997253, 0.018481435254216194, -0.06169736757874489, 0.07223589718341827, -0.12374747544527054, 0.055548280477523804, -0.04492483660578728, 0.0067251091822981834, -0.027917364612221718, 0.002250240882858634, -0.12570910155773163, 0.11537187546491623, -0.025632401928305626, 0.17935146391391754, -0.20731747150421143, -0.06413614004850388, 0.01128540001809597, -0.23545287549495697, -0.06547331809997559, 0.10980085283517838, -0.01621134951710701, 0.19062022864818573, 0.0846942588686943, 0.19138485193252563, 0.033952753990888596, -0.23608611524105072, 0.020253555849194527, 0.043899230659008026, -0.06369975209236145, -0.08336563408374786, 0.10171621292829514, 0.10886700451374054, 0.02703050896525383, 0.072203628718853, -0.08696644753217697, 0.06551587581634521, -0.04585148021578789, -0.048043392598629, -0.055527735501527786, -0.05500920116901398, -0.06547342985868454, 0.03711983934044838, -0.007299069780856371, 0.05136115476489067, 0.0046501485630869865, 0.04410288482904434, 0.14130480587482452, -0.0704772025346756, 0.014018482528626919, -0.06599175930023193, 0.2569795250892639, -0.13658924400806427, 0.023551255464553833, -0.04648697003722191, -0.05576777830719948, -0.030690208077430725, 0.0827823355793953, 0.043892424553632736, 0.013722585514187813, 0.06339918822050095, 0.05906498059630394, -0.051053788512945175, -0.0006491403910331428, 0.09208355844020844, -0.030628856271505356, 0.01426125317811966, -0.1337304711341858, -0.008518217131495476, -0.050747960805892944, 0.1175839900970459, -0.15451206266880035, 0.00894085131585598, 0.012725764885544777, 0.0790521577000618, 0.054141566157341, 0.06153209134936333, 0.01837047003209591, -0.022380849346518517, -0.06992722302675247, -0.02187274396419525, 0.04584822803735733, 0.00024821882834658027, -0.03912167623639107, 0.09681086987257004, -0.07328443974256516, 0.12601280212402344, 0.12583355605602264, 0.022352203726768494, 0.04497291147708893, -0.12845413386821747, 0.024657627567648888, -0.002313169650733471, -0.07596869021654129, -0.011768181808292866, -0.07487510144710541, -0.000639965059235692, 0.06388600170612335, -0.09012199938297272, 0.11864417046308517, 0.05752357095479965, -0.07420185208320618, -0.08025768399238586, 0.07064490765333176, 0.13458405435085297, 0.019133377820253372, 0.08507804572582245, 0.13264741003513336, -0.06596704572439194, 0.16835907101631165, -0.053983960300683975, -0.13240554928779602, 0.005274823401123285, 0.07438569515943527, 0.021527009084820747, 0.1469864696264267, 0.07325691729784012, -0.021745994687080383, 0.043927233666181564, -0.04431621730327606, 0.04896024242043495, -0.09660746157169342, -0.0511162094771862, 0.012190931476652622, -0.0571325458586216, 0.07081752270460129, 0.11463025957345963, -0.07419860363006592, 0.09011024981737137, -0.09357859939336777, -0.03773282468318939, -0.009086084552109241, -0.015154477208852768, -0.0010376987047493458, 0.08123935759067535, 0.04483793303370476, -0.11963072419166565, -0.18674643337726593, -0.022569041699171066, -0.11269331723451614, 0.02887512743473053, 0.04466026648879051, -0.07013615220785141, -0.05016792565584183, -0.07533443719148636, 0.013945632614195347, 0.07603712379932404, -0.05761894956231117, -0.08551649004220963, -0.026730284094810486, -0.044703830033540726, -0.05154918506741524, -0.0421174056828022, -0.04250284656882286, 0.029595473781228065, 0.08623071759939194, -0.10870369523763657, 0.19446106255054474, 0.036594416946172714, 0.03506298363208771, 0.032310064882040024, 0.012997247278690338, 0.13100360333919525, -0.076773501932621, 0.0595020055770874, 0.30503106117248535, 0.008235150016844273, 0.05188535526394844, 0.10696273297071457, -0.017202213406562805, -0.12923897802829742, -0.001379817840643227, 0.005271547939628363, -0.15364201366901398, -0.06817462295293808, -0.08763174712657928, -0.08952649682760239, -0.00662982976064086, 0.005633045453578234, 0.0403188057243824, 0.020858045667409897, 0.152674600481987, -0.010859165340662003, 0.01818210631608963, 0.03125596046447754, 0.07165646553039551, 0.0092904232442379, 0.0005258324672468007, 0.05247486010193825, -0.08197147399187088, -0.07309489697217941, 0.17357173562049866, 0.07763318717479706, 0.18164120614528656, -0.031891610473394394, 0.07334249466657639, 0.08083177357912064, 0.08246990293264389, 0.11021111160516739, 0.07226471602916718, -0.009864695370197296, -0.024951443076133728, -0.07243814319372177, -0.10972028225660324, -0.006147808860987425, 0.12240585684776306, -0.08255983144044876, -0.03831780329346657, -0.007876631803810596, 0.1342848688364029, 0.049668148159980774, 0.013706011697649956, -0.02336701937019825, -0.29527226090431213, 0.05041499808430672, 0.08437177538871765, 0.1213935986161232, -0.07290259003639221, 0.03490468114614487, 0.12450746446847916, -0.001955027226358652, 0.03112875297665596, -0.03765065222978592, 0.08712422102689743, -0.09312254190444946, -0.02913917787373066, -0.09347916394472122, 0.13251721858978271, -0.00297527271322906, 0.043749306350946426, -0.12443177402019501, 0.09507710486650467, -0.015191406942903996, 0.014976133592426777, -0.04316256195306778, -0.02650003507733345, 0.11074843257665634, 0.1443813294172287, 0.1712408810853958, 0.002115723444148898, -0.15343667566776276, -0.10751993209123611, -0.09655725210905075, 0.054038651287555695, 0.0563330352306366, -0.005520045757293701, 0.03901487588882446, -0.059774674475193024, 0.011755810119211674, 0.002044407883659005, 0.04878957197070122, -0.07906673848628998, -0.12190056592226028, 0.006803097203373909, 0.1083604022860527, 0.02305084466934204, -0.017945874482393265, -0.04582422599196434, -0.08590655773878098, 0.04311351105570793, 0.1723010390996933, -0.10054648667573929, -0.14740318059921265, -0.05565820634365082, 0.02508324570953846, -0.048373993486166, -0.05096299201250076, -0.015163003467023373, 0.12347457557916641, -0.04980403929948807, -0.14789891242980957, -0.02326122298836708, -0.03657105192542076, 0.004488189704716206, -0.010410018265247345, 0.029237790033221245, -0.01664523035287857, -0.067607581615448, 0.007815244607627392, -0.022884143516421318, 0.029570521786808968, -0.07113150507211685, 0.03510718420147896, 0.15082651376724243, 0.025055699050426483, 0.05673198774456978, -0.09065193682909012, -0.0183066688477993, -0.01457207091152668, -0.026441609486937523, -0.03134224936366081, 0.20282518863677979, 0.007901408709585667, 0.00766713498160243, 0.15015964210033417, -0.05114896595478058, -0.22294074296951294, -0.06297644972801208, -0.07487637549638748, 0.01783733069896698, -0.0004777694121003151, -0.06969704478979111, 0.12745241820812225, 0.10199658572673798, -0.0329953134059906, 0.14171633124351501, -0.2644931375980377, -0.09491550922393799, 0.02458851784467697, 0.16259293258190155, 0.38462796807289124, -0.2448662966489792, -0.020294461399316788, -0.14898379147052765, -0.1107330247759819, 0.03737306594848633, -0.10658825188875198, 0.06702146679162979, -0.0026023529935628176, 0.013261084444820881, 0.004113972652703524, -0.007134587038308382, 0.19805988669395447, -0.03193361684679985, 0.1258241981267929, -0.08945103734731674, -0.021880190819501877, 0.07166298478841782, -0.06739289313554764, 0.09593890607357025, -0.15381525456905365, 0.024402815848588943, -0.10302398353815079, -0.06984260678291321, 0.04474323242902756, 0.047362539917230606, 0.024354886263608932, -0.06395620852708817, -0.056703370064496994, 0.029316134750843048, 0.002621993189677596, 0.01943100057542324, 0.04222428798675537, -0.08785422146320343, -0.048072654753923416, 0.07183434069156647, -0.03190547972917557, 0.00818623322993517, -0.04080931469798088, -0.10992304980754852, -0.030462289229035378, 0.10913883149623871, -0.19396862387657166, 0.01133826281875372, 0.08677233755588531, -0.0052455877885222435, 0.12580451369285583, 0.0029200948774814606, -0.0043179853819310665, 0.09044279903173447, 0.12221376597881317, -0.07622653245925903, -0.14162370562553406, -0.043627604842185974, -0.1099768579006195, 0.10925974696874619, -0.0019797568675130606, 0.12262436747550964, -0.06128636375069618, 0.07015100866556168, -0.0018437417456880212, 0.021583009511232376, -0.02858887054026127, 0.022510195150971413, 0.03428608924150467, -0.04482711851596832, -0.09333056211471558, 0.09678436070680618, -0.03901295363903046, -0.015224740840494633, -0.023077525198459625, -0.026879403740167618, -0.0796852633357048, -0.05208441615104675, -0.02069111540913582, 0.20155902206897736, -0.11756563931703568, -0.12389922142028809, -0.10299567133188248, -0.08913638442754745, -0.019755758345127106, 0.01948920078575611, 0.10513466596603394, -0.07838261872529984, 0.01463041640818119, 0.02044539898633957, -0.08525654673576355, 0.03640098497271538, 0.07669620215892792, 0.06283008307218552, -0.1943252831697464, -0.12530581653118134, -0.024063752964138985, 0.033863771706819534, -0.10668405145406723, -0.052293166518211365, -0.06893568485975266, 0.022404810413718224, -0.1650790423154831, 0.1540296971797943, -0.07468089461326599, -0.020789263769984245, -0.00023806409444659948, -0.046778712421655655, -0.04222508519887924, 0.003308689221739769, -0.0648389607667923, 0.03602311387658119, 0.017962295562028885, 0.05684670805931091, -0.06110826134681702, -0.044839367270469666, 0.012059583328664303, -0.06578546017408371, -0.012531118467450142, 0.04635756462812424, -0.06646337360143661, 0.027532150968909264, -0.1885707676410675, -0.011669528670608997, 0.10888197273015976, 0.031068876385688782, -0.019478535279631615, 0.035596322268247604, 0.04834701120853424, 0.05233936384320259, -0.007010942790657282, -0.04048190265893936, -0.027085520327091217, -0.06568599492311478, 0.09291021525859833, -0.07734284549951553, -0.026119114831089973, -0.010715012438595295, 0.030146343633532524, 0.15644420683383942, 0.1437961310148239, 0.14090251922607422, -0.027073657140135765, -0.009464962407946587, -0.07578542828559875, 0.022445209324359894, 0.059223074465990067, -0.13424336910247803, -0.025462107732892036, -0.15707840025424957, -0.0159224234521389, -0.018956221640110016, 0.17330338060855865, 0.06317699700593948, -0.052577290683984756, -0.06402251124382019, 0.0483526885509491, 0.11106367409229279, -0.00834756251424551, 0.22491391003131866, 0.10617204755544662, 0.10471107065677643, -0.08473168313503265, 0.04734985530376434, 0.11108145117759705, 0.10262783616781235, 0.012809894047677517, 0.1434602290391922, -0.01774471253156662, 0.13108213245868683, 0.003558716271072626, 0.04331110417842865, 0.005103325936943293, 0.042546167969703674, -0.11253877729177475, 0.022492790594697, 0.007898389361798763, 0.003514269832521677, 0.22553245723247528, -0.10060997307300568, 0.007707991171628237, 0.05224232375621796, 0.002336165402084589, -0.10399985313415527, -0.29406672716140747, -0.1055968701839447, -0.18744796514511108, 0.03732512518763542, -0.049652330577373505, -0.006348864641040564, 0.16490551829338074, 0.005605914164334536, 0.02544364519417286, 0.06402955949306488, -0.1045946329832077, -0.024650145322084427, 0.07631737738847733, -0.07183344662189484, -0.09223821759223938, 0.0140782305970788, -0.10126917064189911, 0.07799112796783447, -0.05852053314447403, -0.026549577713012695, 0.0257797222584486, 0.1137688010931015, 0.05972393974661827, -0.10222119837999344, -0.03719107806682587, -0.04808284342288971, 0.012941857799887657, -0.024767715483903885, 0.05426584929227829, 0.042221229523420334, -0.030775506049394608, 0.028703756630420685, 0.12022560834884644, 0.017279036343097687, -0.08917296677827835, -0.09715230762958527, 0.054192785173654556, -0.08656296879053116, 0.03615482151508331, -0.044694844633340836, -0.05692066252231598, -0.0022403975017368793, 0.26415932178497314, 0.10882878303527832, -0.09687690436840057, -0.0018571047112345695, -0.049514904618263245, 0.015378814190626144, -0.0555010661482811, 0.13849855959415436, -0.007364078424870968, 0.1714693009853363, 0.00030672267894260585, -0.025094501674175262, -0.09730330109596252, -0.06610175967216492, -0.012765912339091301, 0.035285331308841705, -0.00967621710151434, -0.07824206352233887, -0.05880483239889145, 0.08079379051923752, -0.1399964541196823, -0.0788690447807312, 0.11842674762010574, -0.007245797198265791, -0.03337431326508522, -0.06101790815591812, 0.020980654284358025, 0.08602829277515411, -0.016225624829530716, -0.15107661485671997, 0.019249750301241875, -0.027932411059737206, 0.022536568343639374, -0.11566787213087082, -0.0031095652375370264, -0.02528942935168743, -0.034727856516838074, 0.08555625379085541, -0.004851132165640593, 0.02174323797225952, -0.008311713114380836, -0.03366873040795326, -0.03625234216451645, 0.129649817943573, -0.02826058492064476, -0.10882487893104553, -0.016044702380895615, 0.07833616435527802, -0.08796966075897217, -0.015061719343066216, 0.06038342043757439, -0.007682605180889368, 0.00422435300424695, 0.06182209774851799, -0.06678555905818939, -0.0511571429669857, 0.030905917286872864, -0.1598011702299118, 0.09201451390981674, 0.05402414873242378, 0.04537546634674072, -0.021616993471980095, -0.03435663878917694, 0.15580274164676666, 0.07321049273014069, -0.05723942816257477, 0.05688571557402611, -0.03934566676616669, -0.09059310704469681, 0.025782953947782516, 0.009663396514952183, -0.13699159026145935, -0.029393961653113365, -0.15279151499271393, -0.020897185429930687, -0.028276463970541954, 0.09426567703485489, 0.2288317233324051, 0.004854798316955566, -0.02865055948495865, -0.22934502363204956, 0.043126076459884644, 0.08374301344156265, -0.03596215695142746, -0.0473642498254776 ]
null
null
transformers
### SOLAR-10.7B-Instruct-Forest-DPO Introducing SOLAR-10.7B-Instruct-Forest-DPO, a LLM fine-tuned with base model upstage/SOLAR-10.7B-Instruct-v1.0, using direct preference optimization. This model showcases exceptional prowess across a spectrum of natural language processing (NLP) tasks. A mixture of the following datasets was used for fine-tuning. 1. Intel/orca_dpo_pairs 2. nvidia/HelpSteer 3. jondurbin/truthy-dpo-v0.1 💻 Usage ```python !pip install -qU transformers bitsandbytes accelerate from transformers import AutoTokenizer import transformers import torch model = "abhishekchohan/SOLAR-10.7B-Instruct-Forest-DPO" tokenizer = AutoTokenizer.from_pretrained(model) pipeline = transformers.pipeline( "text-generation", model=model, model_kwargs={"torch_dtype": torch.float16, "load_in_4bit": True}, ) messages = [{"role": "user", "content": "Explain what a Mixture of Experts is in less than 100 words."}] prompt = pipeline.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ```
{"language": ["en"], "license": "apache-2.0", "library_name": "transformers", "datasets": ["Intel/orca_dpo_pairs", "nvidia/HelpSteer", "jondurbin/truthy-dpo-v0.1"], "pipeline_tag": "text-generation"}
text-generation
abhishekchohan/SOLAR-10.7B-Instruct-Forest-DPO-v1
[ "transformers", "pytorch", "llama", "text-generation", "conversational", "en", "dataset:Intel/orca_dpo_pairs", "dataset:nvidia/HelpSteer", "dataset:jondurbin/truthy-dpo-v0.1", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T03:57:54+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #llama #text-generation #conversational #en #dataset-Intel/orca_dpo_pairs #dataset-nvidia/HelpSteer #dataset-jondurbin/truthy-dpo-v0.1 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
### SOLAR-10.7B-Instruct-Forest-DPO Introducing SOLAR-10.7B-Instruct-Forest-DPO, a LLM fine-tuned with base model upstage/SOLAR-10.7B-Instruct-v1.0, using direct preference optimization. This model showcases exceptional prowess across a spectrum of natural language processing (NLP) tasks. A mixture of the following datasets was used for fine-tuning. 1. Intel/orca_dpo_pairs 2. nvidia/HelpSteer 3. jondurbin/truthy-dpo-v0.1 Usage
[ "### SOLAR-10.7B-Instruct-Forest-DPO\nIntroducing SOLAR-10.7B-Instruct-Forest-DPO, a LLM fine-tuned with base model upstage/SOLAR-10.7B-Instruct-v1.0, using direct preference optimization.\nThis model showcases exceptional prowess across a spectrum of natural language processing (NLP) tasks. \n\nA mixture of the following datasets was used for fine-tuning.\n\n1. Intel/orca_dpo_pairs\n2. nvidia/HelpSteer\n3. jondurbin/truthy-dpo-v0.1\n\n\n Usage" ]
[ "TAGS\n#transformers #pytorch #llama #text-generation #conversational #en #dataset-Intel/orca_dpo_pairs #dataset-nvidia/HelpSteer #dataset-jondurbin/truthy-dpo-v0.1 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### SOLAR-10.7B-Instruct-Forest-DPO\nIntroducing SOLAR-10.7B-Instruct-Forest-DPO, a LLM fine-tuned with base model upstage/SOLAR-10.7B-Instruct-v1.0, using direct preference optimization.\nThis model showcases exceptional prowess across a spectrum of natural language processing (NLP) tasks. \n\nA mixture of the following datasets was used for fine-tuning.\n\n1. Intel/orca_dpo_pairs\n2. nvidia/HelpSteer\n3. jondurbin/truthy-dpo-v0.1\n\n\n Usage" ]
[ 102, 145 ]
[ "passage: TAGS\n#transformers #pytorch #llama #text-generation #conversational #en #dataset-Intel/orca_dpo_pairs #dataset-nvidia/HelpSteer #dataset-jondurbin/truthy-dpo-v0.1 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### SOLAR-10.7B-Instruct-Forest-DPO\nIntroducing SOLAR-10.7B-Instruct-Forest-DPO, a LLM fine-tuned with base model upstage/SOLAR-10.7B-Instruct-v1.0, using direct preference optimization.\nThis model showcases exceptional prowess across a spectrum of natural language processing (NLP) tasks. \n\nA mixture of the following datasets was used for fine-tuning.\n\n1. Intel/orca_dpo_pairs\n2. nvidia/HelpSteer\n3. jondurbin/truthy-dpo-v0.1\n\n\n Usage" ]
[ -0.06549474596977234, 0.08183326572179794, -0.007747066672891378, 0.06294268369674683, 0.05193234980106354, -0.016741644591093063, 0.06382233649492264, 0.1050294041633606, -0.008673394098877907, 0.03442110866308212, 0.07179339975118637, 0.22367659211158752, 0.06397757679224014, 0.09495411813259125, -0.050003182142972946, -0.2744574248790741, 0.06886465102434158, 0.0008099963306449354, -0.01055501401424408, 0.04989120364189148, 0.08368031680583954, -0.005307560320943594, 0.006063398905098438, -0.06438692659139633, -0.0875590592622757, -0.05211874470114708, -0.006084463093429804, -0.033042460680007935, 0.07965853810310364, 0.03217298537492752, 0.06064179167151451, 0.0787077248096466, 0.05465138331055641, -0.15758393704891205, 0.0006424979656003416, 0.041583526879549026, 0.026282187551259995, 0.05281642824411392, 0.07750366628170013, 0.08829042315483093, 0.17742234468460083, -0.08142150193452835, 0.060820963233709335, -0.013757663778960705, -0.04562530666589737, -0.2011718451976776, -0.03461278975009918, -0.027622276917099953, 0.07064637541770935, 0.05765991285443306, 0.03165499493479729, 0.13892093300819397, -0.0189657025039196, 0.0805724561214447, -0.004543184768408537, -0.09367148578166962, -0.00946887955069542, 0.12722238898277283, 0.03974062576889992, 0.04800748452544212, 0.007610463537275791, -0.04383452609181404, 0.005090354476124048, -0.025726670399308205, 0.06733694672584534, -0.019070962443947792, 0.02554018422961235, -0.04541604593396187, -0.14560219645500183, -0.06031161546707153, 0.20932459831237793, 0.07209371030330658, -0.04247748851776123, -0.09471294283866882, -0.11280450969934464, 0.03309091553092003, 0.036006052047014236, -0.05869140848517418, 0.021633153781294823, 0.0014621458249166608, 0.14234891533851624, 0.05710650235414505, -0.07599490135908127, -0.03921782970428467, -0.07202426344156265, 0.16747769713401794, 0.028979577124118805, 0.021323392167687416, 0.017183341085910797, 0.09216282516717911, -0.0024483425077050924, -0.11644987761974335, -0.032628677785396576, -0.06710422039031982, -0.012358829379081726, 0.005967200733721256, -0.07065833359956741, -0.011714829131960869, 0.04453771933913231, 0.13890254497528076, 0.12781207263469696, -0.011592851020395756, -0.02333994396030903, -0.004796381574124098, -0.010064046829938889, 0.0941072329878807, 0.003049275139346719, 0.055735405534505844, 0.06881512701511383, 0.017548594623804092, 0.09724703431129456, -0.01800365187227726, -0.03721090033650398, -0.053209010511636734, -0.05295414850115776, 0.056813277304172516, 0.13601289689540863, 0.024184072390198708, -0.0600212998688221, -0.09632399678230286, 0.19920192658901215, -0.10703246295452118, -0.023864977061748505, 0.010424787178635597, -0.0024206130765378475, 0.1176709309220314, 0.0501728318631649, 0.014742310158908367, -0.029841575771570206, 0.015214595943689346, 0.030004430562257767, -0.007025516591966152, -0.037649329751729965, -0.05216917023062706, 0.02485038712620735, 0.04413741081953049, 0.04030873626470566, -0.18724915385246277, -0.12784415483474731, 0.035398922860622406, 0.05554034933447838, -0.07231101393699646, -0.13225595653057098, 0.007496626581996679, -0.03951060399413109, 0.00012264600081834942, -0.00008312746649608016, -0.002264981623739004, -0.02531866729259491, 0.0463503934442997, 0.0475507527589798, 0.05747932940721512, -0.10159791260957718, 0.029240718111395836, -0.04254394397139549, 0.04374489188194275, -0.06248677149415016, 0.11837828159332275, -0.08727847784757614, 0.039978086948394775, -0.05324134603142738, -0.027432261034846306, 0.01837955042719841, -0.053616419434547424, 0.0057828230783343315, 0.13641873002052307, -0.23692215979099274, 0.0024457841645926237, 0.014528367668390274, -0.12858764827251434, -0.01171694416552782, 0.05893995240330696, 0.036630552262067795, -0.00948339607566595, 0.0017949845641851425, 0.12301281839609146, 0.18076102435588837, -0.1311962604522705, 0.03231246769428253, 0.01683686673641205, 0.019856946542859077, -0.023851679638028145, 0.11048681288957596, 0.06016082689166069, -0.07270728051662445, 0.04912959039211273, -0.1659853309392929, 0.05837031081318855, 0.009580080397427082, -0.08095065504312515, -0.08471545577049255, -0.011461494490504265, -0.003256069263443351, 0.03141378238797188, -0.03155713155865669, 0.011279641650617123, -0.03411591425538063, -0.04705341160297394, 0.15482400357723236, -0.06417953968048096, 0.03414900228381157, -0.044531043618917465, 0.030548684298992157, -0.06288184225559235, 0.009925387799739838, -0.10691088438034058, 0.011503557674586773, 0.08552327007055283, -0.06539075821638107, -0.09210656583309174, 0.08097460865974426, 0.017847027629613876, 0.0426381379365921, 0.0013186066644266248, -0.020201176404953003, -0.10454367101192474, 0.027475528419017792, 0.047030482441186905, -0.12122513353824615, 0.019673774018883705, -0.045823659747838974, 0.14017191529273987, -0.17710860073566437, -0.005570023786276579, 0.02449352853000164, 0.10062054544687271, -0.002279904903843999, -0.018698325380682945, -0.03792979195713997, 0.010973909869790077, -0.025347407907247543, -0.077997587621212, 0.05910690873861313, 0.03577204793691635, -0.0591544508934021, 0.05426153913140297, -0.08139809966087341, -0.060337867587804794, 0.09186695516109467, 0.03318115696310997, -0.022504325956106186, 0.03282980993390083, -0.06300456076860428, -0.0738435909152031, 0.00019180495291948318, -0.04684355854988098, -0.043589409440755844, -0.014744354411959648, 0.04017684981226921, -0.053975604474544525, -0.004338574130088091, 0.004142741207033396, 0.03129095956683159, -0.031045498326420784, 0.08212455362081528, 0.11183186620473862, 0.00994866993278265, 0.017822159454226494, 0.07048964500427246, -0.009712977334856987, 0.07145239412784576, -0.06127152964472771, -0.08585373312234879, -0.04644446447491646, 0.07745809108018875, 0.04384012892842293, 0.06483546644449234, -0.028603151440620422, 0.11016064882278442, 0.08458362519741058, 0.03542037680745125, 0.008300881832838058, -0.08936437219381332, -0.013617197051644325, 0.07012501358985901, -0.07064127922058105, -0.057450711727142334, 0.012197238393127918, -0.013581009581685066, 0.06231763958930969, -0.020481672137975693, -0.09029392153024673, 0.10001779347658157, -0.0009668566635809839, -0.06137271597981453, 0.13663241267204285, -0.09377135336399078, -0.2498503476381302, -0.11540282517671585, 0.010989280417561531, -0.07774166762828827, 0.0026236018165946007, 0.022137537598609924, 0.010720525868237019, -0.07794690877199173, -0.05579916760325432, 0.13078588247299194, -0.02886664867401123, 0.020175784826278687, 0.011088131926953793, 0.036752015352249146, 0.06743122637271881, -0.09116437286138535, 0.012152904644608498, 0.02835162915289402, -0.02564498782157898, 0.10718336701393127, -0.010072963312268257, 0.06815549731254578, 0.0492376871407032, 0.039991289377212524, -0.05135106295347214, 0.07387834787368774, 0.16970425844192505, -0.045276060700416565, 0.05919327959418297, 0.21508756279945374, 0.00016627332661300898, 0.013553138822317123, 0.18041104078292847, 0.010495737195014954, -0.09193558990955353, 0.0055621652863919735, -0.029142849147319794, -0.0319727323949337, -0.22563201189041138, -0.12765122950077057, -0.0739206001162529, 0.045211050659418106, 0.07458462566137314, 0.051805634051561356, 0.07144051045179367, 0.07894176989793777, -0.00273871049284935, 0.02566801942884922, -0.006557226646691561, 0.09926659613847733, 0.1923767626285553, 0.028380651026964188, 0.0878433883190155, -0.029709653928875923, 0.003575623035430908, 0.08598394691944122, 0.04464243724942207, 0.21738170087337494, 0.02472112886607647, 0.1010022833943367, 0.011585266329348087, 0.09912358969449997, 0.005822613835334778, 0.0735960602760315, 0.07371743768453598, -0.018525296822190285, 0.010649185627698898, -0.059620413929224014, -0.04393012449145317, 0.022597674280405045, -0.01637137122452259, -0.013813052326440811, -0.038670409470796585, 0.035310469567775726, -0.009356698021292686, 0.166560560464859, 0.03419911861419678, -0.19951672852039337, -0.017405688762664795, 0.009573778137564659, -0.04350293055176735, -0.0589001439511776, 0.021749230101704597, 0.04454154521226883, -0.07355295866727829, 0.07635155320167542, -0.0445430688560009, 0.05030469596385956, -0.1557837277650833, -0.02536277286708355, -0.04818271845579147, 0.06978976726531982, 0.10706347972154617, 0.1126699224114418, -0.2609574496746063, 0.09226245433092117, -0.04782805219292641, 0.03120400570333004, -0.04011225700378418, 0.07901724427938461, -0.056710030883550644, 0.09830896556377411, 0.09782478213310242, 0.033988650888204575, 0.01962844282388687, -0.009319713339209557, -0.12837457656860352, 0.024276016280055046, -0.019717028364539146, -0.0756833404302597, 0.010580461472272873, -0.016671033576130867, -0.019970344379544258, -0.014954987913370132, 0.07598578929901123, -0.10024035722017288, -0.12258566915988922, 0.11288192123174667, -0.01361492183059454, 0.03609388694167137, -0.059849999845027924, -0.028983870521187782, -0.0711890384554863, 0.14327512681484222, -0.050076231360435486, -0.061423368752002716, -0.08134929090738297, -0.019386833533644676, 0.10779012739658356, -0.03418167680501938, 0.013501991517841816, 0.004713777452707291, 0.08570386469364166, 0.031153196468949318, -0.16153956949710846, 0.029016010463237762, -0.1312558352947235, -0.06451056897640228, -0.0053772712126374245, 0.1092701405286789, 0.005267654545605183, 0.0016725948080420494, 0.019395766779780388, -0.004455422051250935, -0.08269388973712921, -0.10680172592401505, -0.02206515148282051, 0.07546970248222351, -0.026445504277944565, -0.011037351563572884, -0.016670051962137222, 0.02247006818652153, -0.027007652446627617, -0.07597171515226364, 0.04945947974920273, 0.20991063117980957, -0.07127352058887482, 0.12400715053081512, -0.00825080182403326, -0.048807159066200256, -0.12293177098035812, -0.09623689949512482, 0.08417116105556488, -0.00541149266064167, 0.001962251029908657, -0.17801913619041443, 0.12786123156547546, 0.11481332033872604, 0.00853472389280796, -0.014704215340316296, -0.20603395998477936, -0.09994679689407349, 0.06467598676681519, 0.10278692096471786, 0.11162356287240982, -0.09814409911632538, -0.04516033083200455, -0.02588002011179924, -0.22848519682884216, 0.2715160548686981, 0.008991201408207417, 0.05459500476717949, -0.04593253880739212, 0.018045907840132713, 0.03794766217470169, 0.012863950803875923, 0.27257758378982544, -0.030005687847733498, -0.025371473282575607, -0.07719892263412476, 0.13074280321598053, 0.012142657302320004, -0.024357952177524567, 0.04783404618501663, 0.04453544691205025, 0.06965257227420807, -0.11819080263376236, -0.041512612253427505, -0.061417486518621445, -0.010103638283908367, -0.04192489758133888, -0.08533744513988495, -0.006001140922307968, 0.018581820651888847, 0.11826344579458237, -0.01656019128859043, 0.03794310614466667, 0.012486929073929787, -0.03612590581178665, 0.04773658886551857, 0.08073116093873978, 0.026316363364458084, -0.06570453196763992, -0.09858855605125427, -0.0030639439355582, 0.04745420068502426, -0.08538728207349777, 0.03897654637694359, 0.09888250380754471, -0.0031292857602238655, 0.0010252133943140507, 0.013011595234274864, 0.0022251324262470007, -0.07879684865474701, 0.014375242404639721, -0.11317908018827438, -0.17740274965763092, 0.00004790313687408343, 0.0462370440363884, -0.011375576257705688, -0.005058594048023224, 0.11795324832201004, -0.048183001577854156, -0.011059386655688286, 0.061130158603191376, -0.011096375063061714, -0.027506865561008453, 0.09505218267440796, -0.005194994620978832, 0.06985671818256378, -0.10588416457176208, 0.05156686529517174, 0.04490526393055916, -0.0460733138024807, -0.07095850259065628, 0.027443470433354378, -0.08481558412313461, -0.03289458528161049, -0.03918520733714104, -0.055333979427814484, -0.03807966411113739, -0.08539865165948868, -0.05843536555767059, -0.09934764355421066, 0.06336513161659241, -0.140618696808815, 0.07048919796943665, 0.034497037529945374, -0.05820668488740921, -0.016917278990149498, -0.12176322191953659, 0.01991378329694271, -0.0378015860915184, 0.08944957703351974, -0.17646078765392303, -0.043235015124082565, -0.06672095507383347, 0.05384277552366257, -0.017601829022169113, -0.008189843036234379, -0.03631104156374931, -0.03661889582872391, -0.021917404606938362, 0.043381467461586, -0.05376977100968361, 0.04805872589349747, 0.017405860126018524, -0.017708443105220795, 0.027881667017936707, 0.053725823760032654, -0.04333636537194252, -0.03285176306962967, -0.058667656034231186, 0.007939287461340427, -0.006752134766429663, -0.04667589068412781, 0.0053693801164627075, -0.13745452463626862, 0.09471838921308517, 0.07453399151563644, 0.007052391301840544, -0.03694886714220047, -0.06850726902484894, 0.02407044544816017, 0.08858337253332138, 0.09722482413053513, -0.04231937974691391, -0.18237602710723877, -0.02768128737807274, 0.102047398686409, -0.08728723227977753, -0.03229738771915436, 0.08835423737764359, -0.14850299060344696, -0.06367561221122742, -0.0738370344042778, -0.08467824757099152, -0.03355317562818527, 0.0038482919335365295, -0.006739126052707434, 0.11643846333026886, 0.1280013620853424, -0.021704893559217453, 0.04188908264040947, -0.11570543795824051, -0.04512973874807358, -0.012125622481107712, -0.03617202490568161, 0.0729885920882225, 0.004105428233742714, 0.04746519401669502, 0.05110170692205429, 0.08966778963804245, -0.09013728052377701, 0.04404308646917343, -0.004608227405697107, 0.022451728582382202, 0.11871571838855743, -0.009995480068027973, 0.22328127920627594, 0.05625053867697716, 0.04288790747523308, 0.060155220329761505, -0.011148856952786446, 0.06143695488572121, 0.1446540355682373, 0.05519036576151848, 0.10928740352392197, -0.0016135071637108922, 0.008834145031869411, 0.016908328980207443, -0.03832124546170235, -0.09170696884393692, 0.06289297342300415, -0.03169407695531845, 0.07530099153518677, 0.013009550049901009, 0.08371932804584503, 0.0360591821372509, -0.10100546479225159, 0.0518583282828331, 0.019269529730081558, -0.06284304708242416, -0.07015899568796158, -0.1383807510137558, -0.06115161255002022, -0.1563183218240738, -0.01741057261824608, -0.08161409199237823, 0.0011578287230804563, 0.04465009644627571, 0.03754467889666557, -0.03904762491583824, 0.02233835682272911, 0.13665764033794403, -0.06270910054445267, 0.013321387581527233, -0.03354761376976967, -0.05482045188546181, 0.026554105803370476, -0.05471442639827728, -0.0012428106274455786, 0.06571980565786362, 0.002223389921709895, 0.013490298762917519, 0.056384604424238205, 0.034099165350198746, -0.04253331571817398, -0.0511542484164238, 0.00594507809728384, -0.0078014200553298, -0.04215938225388527, 0.23879697918891907, 0.04626583680510521, 0.021364988759160042, 0.032307323068380356, 0.20158909261226654, 0.0016754284733906388, -0.12126664817333221, -0.15605902671813965, -0.0012668933486565948, -0.09645337611436844, 0.02226346917450428, 0.0888085886836052, -0.04755767062306404, -0.08328308910131454, 0.10311947762966156, 0.013902584090828896, 0.050718050450086594, 0.003922854084521532, 0.07792061567306519, -0.0235140360891819, -0.0031298515386879444, 0.1048663780093193, 0.09743893891572952, 0.16201156377792358, -0.02330338954925537, -0.05024629831314087, -0.04067990928888321, 0.0206846222281456, -0.0805068165063858, 0.04109029471874237, -0.030253389850258827, -0.10092718154191971, 0.008354314602911472, 0.017315620556473732, 0.004695885814726353, -0.018680846318602562, 0.04290851950645447, -0.052452001720666885, -0.09600848704576492, -0.10772013664245605, 0.04734529182314873, -0.10414788126945496, -0.020196137949824333, -0.05963627249002457, 0.00472401175647974, 0.1770172417163849, -0.0098610520362854, -0.08888200670480728, 0.009925681166350842, 0.047897983342409134, 0.033564493060112, 0.03699610009789467, -0.004062249790877104, 0.1791088581085205, 0.04511360451579094, 0.08219607174396515, -0.17344868183135986, 0.08455773442983627, 0.03897765651345253, -0.09597333520650864, 0.09229598939418793, 0.11383039504289627, -0.03814055770635605, 0.04015315696597099, 0.04884852096438408, -0.06064300611615181, -0.07957272976636887, 0.1541920304298401, -0.001152162323705852, -0.05014890059828758, 0.0549820251762867, -0.12834103405475616, 0.06588909775018692, 0.09595739841461182, -0.04627969488501549, -0.0014925298746675253, -0.10023842006921768, 0.07343560457229614, -0.008826634846627712, 0.056102149188518524, -0.018823662772774696, -0.09405090659856796, -0.06387366354465485, 0.022904986515641212, 0.1046280488371849, -0.036813315004110336, -0.12260238081216812, -0.02721339650452137, -0.10934074968099594, -0.07625065743923187, 0.03565168008208275, -0.021556463092565536, -0.03477422147989273, -0.047299183905124664, -0.05334684997797012, -0.01985643431544304, 0.08344726264476776, -0.11331668496131897, -0.06654015928506851 ]
null
null
spacy
| Feature | Description | | --- | --- | | **Name** | `en_pipeline` | | **Version** | `0.0.0` | | **spaCy** | `>=3.7.3,<3.8.0` | | **Default Pipeline** | `transformer`, `ner` | | **Components** | `transformer`, `ner` | | **Vectors** | 0 keys, 0 unique vectors (0 dimensions) | | **Sources** | n/a | | **License** | n/a | | **Author** | [n/a]() | ### Label Scheme <details> <summary>View label scheme (17 labels for 1 components)</summary> | Component | Labels | | --- | --- | | **`ner`** | `Degree`, `Desc Responsibility`, `Edu Desc`, `Edu End Date`, `Edu Start Date`, `Email`, `Employer Names`, `Institution`, `Links`, `Location`, `Name`, `Phone`, `Position`, `Skills`, `Work End Date`, `Work Location`, `Work Start Date` | </details> ### Accuracy | Type | Score | | --- | --- | | `ENTS_F` | 83.33 | | `ENTS_P` | 81.16 | | `ENTS_R` | 85.63 | | `TRANSFORMER_LOSS` | 39026.84 | | `NER_LOSS` | 1290990.48 |
{"language": ["en"], "tags": ["spacy", "token-classification"]}
token-classification
drastimehta/en_pipeline
[ "spacy", "token-classification", "en", "model-index", "has_space", "region:us" ]
2024-02-15T04:00:56+00:00
[]
[ "en" ]
TAGS #spacy #token-classification #en #model-index #has_space #region-us
### Label Scheme View label scheme (17 labels for 1 components) ### Accuracy
[ "### Label Scheme\n\n\n\nView label scheme (17 labels for 1 components)", "### Accuracy" ]
[ "TAGS\n#spacy #token-classification #en #model-index #has_space #region-us \n", "### Label Scheme\n\n\n\nView label scheme (17 labels for 1 components)", "### Accuracy" ]
[ 25, 16, 5 ]
[ "passage: TAGS\n#spacy #token-classification #en #model-index #has_space #region-us \n### Label Scheme\n\n\n\nView label scheme (17 labels for 1 components)### Accuracy" ]
[ -0.035730354487895966, 0.08401022851467133, -0.001370814279653132, 0.0027297455817461014, 0.07593662291765213, 0.05745900049805641, 0.18905246257781982, 0.09987073391675949, 0.21252600848674774, 0.09154379367828369, 0.06766214966773987, 0.04411999508738518, 0.02542862296104431, 0.2755735516548157, -0.0773342177271843, -0.2569473683834076, 0.09113585203886032, -0.03324305638670921, 0.02299613691866398, 0.11707630753517151, 0.06928655505180359, -0.1717400699853897, 0.053834401071071625, -0.04956084489822388, -0.20302420854568481, 0.04966004937887192, 0.006860761437565088, -0.10585231333971024, 0.08707527071237564, -0.04982992261648178, 0.22799614071846008, 0.015778830274939537, 0.029654929414391518, -0.20028863847255707, 0.0011377915507182479, 0.010238832794129848, -0.04483124986290932, 0.0878683403134346, 0.019808365032076836, 0.0014592802617698908, -0.020317483693361282, -0.13015799224376678, 0.05147046968340874, 0.02505749650299549, -0.13883958756923676, -0.1822119802236557, -0.027799027040600777, 0.10766223073005676, 0.03308313339948654, -0.057624008506536484, -0.028620852157473564, 0.1032075360417366, -0.10047207772731781, 0.05863143503665924, 0.20717094838619232, -0.3179144859313965, -0.005423094145953655, 0.27708905935287476, -0.0031904212664812803, 0.11117063462734222, -0.05277175456285477, 0.14976060390472412, 0.13128529489040375, -0.03520581126213074, 0.026521045714616776, -0.002127371495589614, 0.06280646473169327, 0.039255402982234955, -0.13350889086723328, -0.08764754235744476, 0.4970542788505554, 0.05566052347421646, 0.014135677367448807, -0.13248059153556824, -0.10335352271795273, -0.15500544011592865, -0.07854212075471878, -0.04833723232150078, 0.0693935677409172, -0.0010493587469682097, 0.1285955011844635, 0.06893597543239594, -0.10848996788263321, -0.03672124072909355, -0.13858462870121002, 0.23714637756347656, -0.017668623477220535, 0.0795220285654068, -0.1710626184940338, -0.015113543719053268, -0.11678777635097504, -0.08759371936321259, 0.045749980956315994, -0.09341771900653839, -0.10081233084201813, -0.03342475742101669, -0.00618742173537612, 0.09755560755729675, 0.04935527220368385, 0.07662283629179001, -0.038349926471710205, 0.02024882845580578, -0.017999444156885147, 0.0834587961435318, 0.15250669419765472, 0.1849389523267746, -0.07149431854486465, -0.02306850627064705, -0.08253256976604462, -0.04556863754987717, 0.06163312494754791, -0.053518395870923996, -0.12820474803447723, -0.014675089158117771, 0.09279696643352509, 0.07555534690618515, -0.029592419043183327, -0.05188215151429176, -0.09724333137273788, -0.011723792180418968, 0.06289143860340118, -0.10488137602806091, 0.04840154945850372, -0.031917378306388855, -0.0049224719405174255, 0.04747140780091286, -0.10276813805103302, -0.018624460324645042, 0.08242756873369217, -0.0036733171436935663, -0.13875210285186768, 0.007893416099250317, -0.02493676170706749, -0.13807561993598938, 0.01752353273332119, -0.11451248079538345, 0.03754466772079468, -0.060027554631233215, -0.07598260790109634, -0.01566317304968834, -0.015610108152031898, -0.07181372493505478, 0.011648750863969326, 0.030958883464336395, -0.09757579863071442, 0.041252151131629944, 0.021072475239634514, -0.03241054713726044, -0.09200505912303925, -0.011555388569831848, -0.05615020915865898, 0.13516144454479218, -0.06392373144626617, 0.03304101154208183, -0.0630401223897934, 0.07500455528497696, -0.1587125062942505, 0.04138900339603424, -0.09006312489509583, 0.07807756215333939, -0.06588883697986603, -0.08988650888204575, -0.03040526993572712, 0.03901401907205582, -0.0813414454460144, 0.12104019522666931, -0.2589706480503082, -0.05716148763895035, 0.2021772414445877, -0.1725752353668213, -0.10442230850458145, 0.040328241884708405, -0.01571672409772873, 0.06756623089313507, 0.02432984672486782, 0.21176554262638092, -0.03243795037269592, -0.10929514467716217, -0.05717768147587776, 0.11139525473117828, -0.019684968516230583, -0.0245414637029171, 0.09767879545688629, 0.05834929272532463, 0.017761554569005966, 0.004433087073266506, 0.07077429443597794, -0.12279726564884186, -0.06188514456152916, -0.08665664494037628, -0.06440231204032898, 0.021990830078721046, 0.07626023888587952, 0.041797298938035965, 0.050026338547468185, -0.0717565268278122, -0.02289918065071106, 0.05307704210281372, 0.048790499567985535, 0.056848227977752686, -0.04571383818984032, 0.004100046586245298, 0.1554238200187683, -0.14572420716285706, -0.07204989343881607, -0.1701408475637436, -0.1473783254623413, 0.018352016806602478, 0.06459003686904907, 0.013353914953768253, 0.2369125783443451, -0.0007419293979182839, -0.03429606929421425, 0.015483362600207329, -0.004267841577529907, -0.003022572258487344, 0.08818550407886505, -0.07760375738143921, -0.14070254564285278, -0.07223067432641983, -0.10953684151172638, 0.009686083532869816, -0.06305254995822906, 0.027686554938554764, 0.1505061686038971, 0.0717858076095581, 0.03287636861205101, 0.04667750000953674, 0.05763440579175949, 0.034577857702970505, -0.06081431359052658, -0.0527830496430397, 0.04479653388261795, -0.1014096662402153, -0.038801223039627075, -0.031500689685344696, -0.14022915065288544, 0.1296258568763733, 0.15124541521072388, -0.11850369721651077, -0.0658019632101059, -0.03906989470124245, -0.024762533605098724, 0.03635694459080696, -0.1305474489927292, -0.019845526665449142, -0.15269246697425842, -0.06162025406956673, 0.024286525323987007, -0.0906018614768982, -0.025035260245203972, 0.029086699709296227, -0.04254851117730141, -0.1873973309993744, 0.1183289960026741, -0.061130963265895844, -0.18233221769332886, 0.19274771213531494, 0.2918400168418884, 0.12804433703422546, 0.13300618529319763, -0.012324503622949123, -0.03893885016441345, -0.0486643947660923, -0.057677559554576874, -0.10056804865598679, 0.1937611699104309, -0.1846938282251358, -0.06786041706800461, 0.0721031054854393, 0.029575290158391, 0.030175670981407166, -0.19244422018527985, -0.025734085589647293, 0.009012720547616482, 0.010669802315533161, -0.10098158568143845, 0.02171136811375618, 0.03945833817124367, 0.14611487090587616, 0.049637604504823685, -0.1562223583459854, 0.018909547477960587, -0.04815967381000519, -0.0025579072535037994, 0.13591797649860382, -0.08574357628822327, -0.24788136780261993, -0.11667381227016449, -0.008890260010957718, -0.041829969733953476, 0.042797308415174484, 0.0015271416632458568, -0.12877154350280762, -0.029531750828027725, -0.0003838702104985714, -0.034110505133867264, -0.19288387894630432, -0.01577647402882576, -0.03847494721412659, 0.07616744935512543, -0.14411285519599915, -0.06950205564498901, -0.11195182055234909, -0.08441121876239777, 0.07776983082294464, 0.0994582250714302, -0.12069186568260193, 0.07500903308391571, 0.27959415316581726, -0.06266063451766968, 0.07531826198101044, -0.013637400232255459, 0.152862548828125, -0.09221401810646057, 0.03168260306119919, 0.11134707927703857, 0.10679347068071365, 0.03847293183207512, 0.26849958300590515, 0.07463954389095306, -0.13589464128017426, -0.034563131630420685, -0.03337811678647995, -0.13350845873355865, -0.10541652888059616, -0.14545421302318573, -0.06908929347991943, -0.043902717530727386, 0.042821481823921204, 0.03361702337861061, 0.027777671813964844, 0.039625443518161774, 0.01491865050047636, -0.009645524434745312, -0.01744597777724266, 0.03953064605593681, 0.08416509628295898, -0.07769710570573807, 0.11946341395378113, -0.04700946807861328, -0.07593429088592529, 0.07391557097434998, 0.10393775254487991, 0.13911548256874084, 0.15673331916332245, 0.0018611922860145569, 0.0841442421078682, 0.019046247005462646, 0.13193051517009735, 0.0653538629412651, 0.13926155865192413, -0.01893642358481884, -0.036866262555122375, -0.07627138495445251, -0.0006345512229017913, 0.11638520658016205, 0.023908009752631187, -0.09024406224489212, -0.04309547320008278, -0.11335247755050659, 0.09468483179807663, -0.04163403809070587, 0.27550631761550903, -0.19709859788417816, 0.0031965794041752815, 0.14111845195293427, 0.08196230232715607, -0.08343565464019775, 0.09203430265188217, 0.10277429968118668, -0.055725160986185074, 0.04105160012841225, 0.026006612926721573, 0.09790785610675812, -0.06120933219790459, 0.017513776198029518, -0.11335743963718414, -0.07637059688568115, -0.025533869862556458, 0.07583096623420715, -0.047700297087430954, 0.3432633578777313, 0.029173947870731354, -0.10840581357479095, -0.08170633018016815, -0.008562419563531876, 0.030512657016515732, 0.2026945799589157, 0.20180480182170868, 0.06830795854330063, -0.2733030617237091, -0.2349873185157776, 0.002430245978757739, -0.052311304956674576, 0.1476309597492218, -0.04075627028942108, 0.016248727217316628, 0.012385661713778973, 0.015651732683181763, 0.002833073493093252, 0.15653552114963531, -0.04733268544077873, -0.028298314660787582, 0.03630431368947029, 0.1074860543012619, -0.13048167526721954, -0.020119350403547287, -0.08542976528406143, -0.2016579806804657, 0.12323221564292908, 0.05319290608167648, -0.06234263256192207, -0.11114714294672012, 0.04273366555571556, 0.1527707576751709, -0.007802230771631002, -0.005779482424259186, -0.05806112289428711, 0.16023023426532745, -0.019132595509290695, -0.08147048950195312, 0.12766797840595245, -0.03524721786379814, -0.01911110244691372, -0.057954564690589905, 0.12246137112379074, -0.07118473947048187, 0.013660728000104427, 0.047011762857437134, 0.12746259570121765, -0.07679858058691025, -0.10351362079381943, 0.12110581994056702, -0.007916886359453201, 0.023707594722509384, 0.29519951343536377, -0.053749702870845795, -0.1440654844045639, 0.015489156357944012, 0.08032556623220444, 0.12528684735298157, 0.24465681612491608, -0.11792483925819397, 0.057533517479896545, 0.006408830173313618, -0.017311420291662216, -0.22080527245998383, -0.004896612837910652, -0.17288385331630707, 0.045655205845832825, 0.013636433519423008, -0.02392730861902237, 0.10918763279914856, 0.027262436226010323, -0.0857284888625145, 0.019770028069615364, -0.24108806252479553, -0.06201697513461113, 0.17489776015281677, 0.0015624691732227802, 0.21426288783550262, -0.08134309202432632, -0.08356648683547974, -0.07899239659309387, -0.11232445389032364, 0.17912796139717102, -0.029030950739979744, 0.10214049369096756, -0.040171023458242416, -0.008776157163083553, 0.037969451397657394, -0.02678023837506771, 0.24292756617069244, 0.08661708980798721, 0.10759970545768738, 0.017100781202316284, -0.22041496634483337, 0.2273222804069519, -0.04777265712618828, 0.016123002395033836, 0.175497367978096, 0.027424130588769913, -0.18012350797653198, 0.011060175485908985, -0.027530891820788383, 0.02992699109017849, -0.05709552764892578, -0.0685325562953949, -0.11875714361667633, 0.0320165753364563, -0.031676385551691055, -0.055812373757362366, 0.3114497661590576, -0.06676589697599411, 0.18838132917881012, 0.1927529126405716, -0.06230166554450989, -0.11936625838279724, 0.002966652624309063, 0.0024522978346794844, -0.042624834924936295, 0.07085078954696655, -0.24641266465187073, 0.06007804349064827, 0.10296716541051865, 0.051688142120838165, 0.10865773260593414, 0.13159742951393127, -0.022135235369205475, -0.04006430506706238, 0.13434191048145294, -0.10554952174425125, -0.20189782977104187, -0.006461632903665304, -0.15634579956531525, 0.023090172559022903, 0.03889336809515953, 0.08152908086776733, 0.01916411519050598, -0.0055513037368655205, -0.012923804111778736, 0.012634036131203175, -0.06317035853862762, 0.0750458836555481, 0.04314029961824417, 0.08834587037563324, -0.12247855961322784, 0.08956455439329147, 0.10376633703708649, 0.038323428481817245, -0.05161623656749725, -0.04402806982398033, -0.07737570255994797, -0.0787849947810173, -0.03711368888616562, 0.07846508920192719, -0.08677921444177628, -0.07132866978645325, -0.04593663290143013, -0.15053430199623108, 0.005216867197304964, 0.12156575173139572, 0.15461358428001404, 0.09709376096725464, -0.0005492494092322886, -0.10194871574640274, 0.05895748734474182, 0.046590302139520645, -0.049918726086616516, 0.02111472561955452, -0.21560505032539368, 0.06911235302686691, -0.0555100254714489, 0.12492398172616959, -0.12576553225517273, -0.07842525094747543, -0.1520756632089615, -0.0007175910868681967, -0.02789420075714588, 0.03849748894572258, -0.056212350726127625, -0.03235528618097305, -0.007926161400973797, -0.013333777897059917, -0.08152531832456589, -0.047495923936367035, -0.12930433452129364, 0.045595910400152206, -0.019829852506518364, 0.15720343589782715, -0.10639157891273499, -0.01747502014040947, 0.08193890750408173, -0.027912568300962448, 0.05858449637889862, 0.0025420738384127617, 0.027635032311081886, 0.019061511382460594, -0.14541038870811462, -0.004164237063378096, 0.10392718762159348, 0.019954517483711243, 0.10951167345046997, -0.09635156393051147, -0.010881888680160046, -0.017038948833942413, -0.032536186277866364, 0.09028901159763336, -0.09527216106653214, -0.08863162249326706, -0.12738366425037384, -0.12168120592832565, -0.15714450180530548, -0.004003556445240974, 0.023480074480175972, 0.21484725177288055, 0.014530694112181664, -0.026873795315623283, 0.05855328217148781, -0.001972550991922617, -0.07951373606920242, -0.006601716857403517, -0.07787009328603745, -0.14319150149822235, 0.0477646067738533, -0.023916274309158325, 0.023329520598053932, -0.02372606471180916, 0.38997891545295715, 0.12645938992500305, 0.03046579658985138, 0.045613717287778854, 0.21315760910511017, -0.027921319007873535, 0.0393516942858696, 0.20615433156490326, 0.08708906918764114, -0.062042392790317535, 0.05803578346967697, 0.0439750961959362, 0.043965402990579605, 0.10406816005706787, 0.13599857687950134, 0.10105524212121964, -0.03991952911019325, 0.08565842360258102, 0.015938669443130493, -0.022934071719646454, -0.0611431710422039, 0.061340924352407455, 0.021704230457544327, 0.03260473161935806, -0.009017135947942734, -0.15786972641944885, 0.12860722839832306, -0.1852131485939026, 0.16157294809818268, -0.03398305922746658, -0.07065947353839874, -0.17352454364299774, -0.07799115031957626, -0.06820731610059738, -0.06578023731708527, -0.026643723249435425, -0.12816444039344788, -0.005826402455568314, 0.20992806553840637, 0.020335232838988304, 0.038737934082746506, 0.04761634021997452, -0.20950618386268616, 0.014733510091900826, 0.10835690051317215, 0.03540356457233429, 0.026196323335170746, -0.12652838230133057, -0.008597470819950104, 0.02371220663189888, -0.06567028909921646, -0.06375771015882492, -0.028382763266563416, 0.02807510830461979, -0.05546564608812332, -0.15506495535373688, -0.06897296011447906, -0.06953240185976028, -0.02481834590435028, -0.017527038231492043, -0.08879292756319046, 0.024092096835374832, -0.013767104595899582, -0.007004383951425552, 0.30056557059288025, -0.09974586963653564, 0.06919591128826141, -0.06034485250711441, 0.19807496666908264, -0.07375846803188324, 0.11343694478273392, 0.005280934274196625, -0.07480652630329132, -0.08623310923576355, 0.05529734119772911, 0.20506401360034943, 0.004187324550002813, 0.032590869814157486, 0.026282522827386856, 0.011436847038567066, 0.02318757399916649, -0.01384794618934393, -0.019452162086963654, 0.09397926926612854, -0.05936736613512039, 0.08821388334035873, -0.0770270898938179, -0.05830001458525658, -0.029339347034692764, -0.04682290926575661, 0.17002040147781372, -0.05777111276984215, -0.1104329451918602, 0.1548389345407486, -0.07691080123186111, 0.0031231422908604145, 0.27519291639328003, -0.21463342010974884, -0.11440262198448181, 0.029676098376512527, 0.030136296525597572, -0.030373938381671906, 0.07957658916711807, -0.12449422478675842, -0.00476092891767621, -0.030157532542943954, 0.010440468788146973, -0.144780695438385, -0.09498459100723267, 0.03476063534617424, -0.031340714544057846, 0.06767893582582474, -0.025570088997483253, 0.08357702940702438, 0.1090206429362297, -0.024625491350889206, -0.0452740304172039, 0.05582898482680321, 0.02045142464339733, 0.04840825870633125, -0.038328398019075394, 0.11208385229110718, -0.018170291557908058, -0.07737963646650314, 0.13183096051216125, -0.16965438425540924, -0.01219058409333229, 0.006637740880250931, -0.0484127476811409, -0.032562337815761566, 0.0015571898547932506, -0.08659150451421738, 0.09733925759792328, 0.10075158625841141, -0.021912401542067528, -0.010191395878791809, 0.026105836033821106, 0.05599471181631088, 0.058172691613435745, -0.03439399600028992, -0.02106679230928421, -0.016215888783335686, -0.008010377176105976, 0.11381177604198456, -0.0204181969165802, -0.12944312393665314, -0.02665938250720501, -0.033123455941677094, 0.05025516077876091, -0.011698556132614613, 0.08919557183980942, 0.1328991800546646, 0.018953615799546242, -0.04419571906328201, -0.24635735154151917, 0.07447201758623123, 0.10147132724523544, -0.08067146688699722, -0.03601837158203125 ]
null
null
diffusers
# Nozomi Tojo <Gallery /> ## Model description This model was trained to generate high quality images based on SIFAS cards. To achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement. ## Trigger words You should use `id_nozomi_tojo` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/theidoldaily/nozomi-tojo/tree/main) them in the Files & versions tab.
{"license": "mit", "tags": ["text-to-image", "stable-diffusion", "lora", "diffusers", "template:sd-lora"], "widget": [{"text": "masterpiece, high quality, defined pupil, looking at viewer, rounded pupil, defined iris, (soft iris:1.2),", "parameters": {"negative_prompt": "bad_anatomy, deformation, amputation, deformity, deformed_nipples, duplicated_torso, deformed_torso, long_torso, large_torso, unproportioned_torso, (deformed_pussy:1.2), (deformed_hands:1.2), unproportioned_eyes, unproportioned_head, small_head, duplicated_nose, big_nose, fusioned_clothes, fusioned_arms, undefined_limbs, divided_pussy, red_pussy, duplicated_pussy, deformed_anus, deformed_pussy,"}, "output": {"url": "images/nozomi_final.png"}}], "base_model": "cagliostrolab/animagine-xl-3.0", "instance_prompt": "id_nozomi_tojo"}
text-to-image
theidoldaily/nozomi-tojo
[ "diffusers", "text-to-image", "stable-diffusion", "lora", "template:sd-lora", "base_model:cagliostrolab/animagine-xl-3.0", "license:mit", "region:us" ]
2024-02-15T04:07:06+00:00
[]
[]
TAGS #diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-cagliostrolab/animagine-xl-3.0 #license-mit #region-us
# Nozomi Tojo <Gallery /> ## Model description This model was trained to generate high quality images based on SIFAS cards. To achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement. ## Trigger words You should use 'id_nozomi_tojo' to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. Download them in the Files & versions tab.
[ "# Nozomi Tojo\n\n<Gallery />", "## Model description \n\nThis model was trained to generate high quality images based on SIFAS cards.\n\nTo achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement.", "## Trigger words\n\nYou should use 'id_nozomi_tojo' to trigger the image generation.", "## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab." ]
[ "TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-cagliostrolab/animagine-xl-3.0 #license-mit #region-us \n", "# Nozomi Tojo\n\n<Gallery />", "## Model description \n\nThis model was trained to generate high quality images based on SIFAS cards.\n\nTo achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement.", "## Trigger words\n\nYou should use 'id_nozomi_tojo' to trigger the image generation.", "## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab." ]
[ 56, 11, 68, 22, 28 ]
[ "passage: TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-cagliostrolab/animagine-xl-3.0 #license-mit #region-us \n# Nozomi Tojo\n\n<Gallery />## Model description \n\nThis model was trained to generate high quality images based on SIFAS cards.\n\nTo achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement.## Trigger words\n\nYou should use 'id_nozomi_tojo' to trigger the image generation.## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab." ]
[ -0.07213572412729263, -0.07681828737258911, -0.0013226784067228436, 0.0344327948987484, 0.1241847351193428, 0.03882597014307976, 0.24897639453411102, 0.08074336498975754, 0.11302924901247025, 0.06267514079809189, -0.021133318543434143, 0.051501233130693436, 0.052938491106033325, 0.21533657610416412, -0.06847774237394333, -0.24597394466400146, 0.09711296111345291, -0.02793031744658947, -0.03220101818442345, 0.013135331682860851, 0.04046119377017021, -0.052344679832458496, 0.12213968485593796, -0.054489344358444214, -0.06641753017902374, 0.006319715175777674, -0.008538349531590939, -0.03577053174376488, -0.012350453995168209, 0.05232326686382294, -0.009764594957232475, 0.09287316352128983, 0.05545421317219734, -0.12525387108325958, 0.059744942933321, 0.01658121682703495, -0.011210990138351917, 0.017348596826195717, -0.0022832974791526794, -0.05119197070598602, 0.15316884219646454, -0.1216214969754219, -0.09895557165145874, -0.019372014328837395, -0.05954490974545479, -0.027153244242072105, -0.003653492545709014, -0.04275188967585564, 0.11599058657884598, 0.023978864774107933, 0.01727110520005226, -0.0015145490178838372, -0.014000876806676388, 0.03098941408097744, 0.2268344610929489, -0.06788662821054459, -0.08127962052822113, 0.1966104507446289, 0.05219227075576782, 0.15107376873493195, -0.0675358846783638, 0.1156679093837738, 0.10785921663045883, -0.09823984652757645, -0.00022302924480754882, -0.05360327288508415, 0.10087806731462479, -0.04663994908332825, -0.0464119128882885, 0.029891153797507286, 0.3209085464477539, 0.03934083133935928, -0.05275529623031616, -0.08199486881494522, -0.06995265185832977, 0.12525448203086853, -0.09324803203344345, -0.022832535207271576, 0.04544103145599365, 0.016106532886624336, 0.051389411091804504, -0.0998680368065834, -0.09065776318311691, -0.08265987783670425, -0.035867735743522644, 0.2573571801185608, 0.02520187944173813, 0.06429406255483627, -0.049525756388902664, 0.03513075038790703, -0.11814837902784348, -0.15605640411376953, -0.004796492401510477, -0.05912002548575401, 0.037461671978235245, 0.08154074102640152, -0.0097993528470397, -0.11804046481847763, 0.11809217184782028, -0.03650123253464699, 0.06132258102297783, 0.0031857287976890802, -0.01823882944881916, 0.1024598702788353, 0.0373818464577198, -0.039602916687726974, -0.013666915707290173, -0.06365920603275299, 0.026153450831770897, 0.02545028366148472, 0.08474738150835037, -0.09330728650093079, -0.11019664257764816, -0.04627475142478943, -0.08665936440229416, 0.0062111131846904755, 0.0022098878398537636, 0.026181558147072792, -0.023183198645710945, -0.03202610835433006, 0.07062426209449768, -0.03217737004160881, -0.04312993213534355, -0.06306862831115723, -0.01507557276636362, 0.27449554204940796, 0.03504889830946922, -0.025902502238750458, 0.024677492678165436, 0.08523571491241455, -0.04941036179661751, -0.0066085332073271275, -0.05532384291291237, -0.010188347660005093, 0.006209386046975851, -0.11008220165967941, 0.009486967697739601, -0.13036763668060303, -0.2708117961883545, -0.02145746536552906, 0.07306574285030365, -0.029010849073529243, 0.06653229147195816, -0.02503836154937744, -0.01905447617173195, 0.005939783528447151, 0.004377697594463825, 0.0067207165993750095, -0.07596185803413391, 0.0827862098813057, -0.005249630194157362, 0.10736235976219177, -0.13118433952331543, 0.005884529557079077, -0.08381334692239761, 0.06031256914138794, -0.15148068964481354, 0.06266230344772339, -0.05741909146308899, 0.002017169026657939, -0.05180186405777931, 0.0034199717920273542, -0.16600537300109863, 0.07095149159431458, -0.03747616335749626, 0.20320956408977509, -0.19427214562892914, -0.05443085730075836, 0.028215503320097923, -0.24486584961414337, -0.019772455096244812, 0.07635065168142319, -0.023095624521374702, 0.18721294403076172, 0.0975203886628151, 0.16833005845546722, 0.10491392016410828, -0.20923155546188354, 0.026735221967101097, 0.016722509637475014, -0.06929168850183487, -0.11436158418655396, 0.09482496231794357, 0.095787413418293, -0.04243575036525726, 0.09732045978307724, -0.11888203769922256, 0.04581645503640175, -0.06258254498243332, -0.04792733117938042, -0.05955904722213745, -0.06463897228240967, -0.014493979513645172, 0.016888810321688652, 0.036358583718538284, 0.027293605729937553, -0.018036890774965286, -0.03370805084705353, 0.13277782499790192, -0.05993080511689186, 0.00751302158460021, -0.02553139440715313, 0.23517274856567383, -0.2003994584083557, 0.014558638446033001, -0.024212509393692017, -0.029534660279750824, -0.022644342854619026, 0.07079721987247467, 0.05724037066102028, 0.05055595934391022, 0.052183229476213455, 0.04361328110098839, -0.07640264183282852, 0.02795172668993473, 0.10486498475074768, -0.023860344663262367, 0.023779088631272316, -0.13718153536319733, 0.020454639568924904, -0.03470691666007042, 0.09194505959749222, -0.12252744287252426, -0.004574556369334459, -0.015323041938245296, 0.06227089837193489, 0.037438977509737015, 0.05131339654326439, 0.02895963191986084, -0.013812362216413021, -0.04772450402379036, -0.014869138598442078, 0.0398140624165535, 0.004101384896785021, -0.06317329406738281, 0.10892981290817261, -0.10175883769989014, 0.11784526705741882, 0.12524305284023285, 0.03213386610150337, 0.014647235162556171, -0.15110284090042114, 0.032680388540029526, -0.014293886721134186, -0.04975898936390877, 0.021149374544620514, -0.12897342443466187, -0.014045336283743382, 0.053665775805711746, -0.09664880484342575, 0.13464456796646118, 0.08279693126678467, -0.08734019845724106, -0.06706472486257553, 0.05364134907722473, 0.15640148520469666, -0.04327718913555145, 0.09068884700536728, 0.08656217902898788, -0.09505816549062729, 0.20307965576648712, -0.026103759184479713, -0.13219289481639862, -0.02580512873828411, 0.07406289130449295, 0.014771952293813229, 0.14804759621620178, 0.08700079470872879, 0.0017273133853450418, 0.0354045033454895, -0.04318743944168091, 0.04586392641067505, -0.07350954413414001, -0.05679188296198845, -0.0007905831444077194, -0.05971821770071983, 0.09959230571985245, 0.1371707022190094, -0.07866532355546951, 0.09519679099321365, -0.07486561685800552, -0.0278803538531065, 0.0062003969214856625, -0.01309684757143259, -0.004474311601370573, 0.062042299658060074, 0.04082689806818962, -0.1066514328122139, -0.17173372209072113, -0.05831388756632805, -0.12537923455238342, 0.018951721489429474, 0.0721459910273552, -0.0641230046749115, -0.03053390234708786, -0.06575866788625717, -0.023550959303975105, 0.03780064731836319, -0.060869187116622925, -0.0722305104136467, -0.023994838818907738, -0.05602443218231201, -0.04733491316437721, -0.038453876972198486, -0.07218566536903381, -0.004555752966552973, 0.09129739552736282, -0.1328352391719818, 0.18778200447559357, 0.05440860614180565, 0.027838898822665215, 0.018686534836888313, -0.008982911705970764, 0.06515166908502579, -0.07613130658864975, 0.06452503800392151, 0.2870882451534271, 0.048515547066926956, 0.06345344334840775, 0.09389057755470276, -0.00676713278517127, -0.08982452750205994, -0.003579936223104596, 0.007338574156165123, -0.14742857217788696, -0.07052581012248993, -0.08851387351751328, -0.11112607270479202, -0.03232249245047569, -0.0040331678465008736, 0.04567179083824158, 0.11511695384979248, 0.17457692325115204, 0.015407745726406574, -0.02931671403348446, 0.055160705000162125, 0.10548295825719833, 0.022313237190246582, 0.044164758175611496, 0.0690644234418869, -0.06589063256978989, -0.05301441252231598, 0.1632911115884781, 0.08032886683940887, 0.2253638356924057, -0.02477142959833145, 0.04140852019190788, 0.08600059896707535, 0.0318458117544651, 0.10711169242858887, 0.0648968517780304, 0.03989953175187111, 0.0010644831927493215, -0.09636572748422623, -0.12147478014230728, -0.020554130896925926, 0.08598199486732483, -0.0558658204972744, -0.022404111921787262, -0.02759428136050701, 0.09896990656852722, 0.032543547451496124, 0.08801832795143127, -0.0326862633228302, -0.28061795234680176, 0.045232728123664856, 0.07695267349481583, 0.11735675483942032, -0.03538418188691139, 0.03681525960564613, 0.1041761115193367, 0.0009421851136721671, 0.04496532678604126, -0.023252999410033226, 0.08571191877126694, -0.052617013454437256, -0.014352771453559399, -0.13834598660469055, 0.14284127950668335, -0.017963020130991936, 0.043297138065099716, -0.15001021325588226, 0.09919603914022446, -0.021938413381576538, 0.033245231956243515, -0.05840262398123741, -0.049350205808877945, 0.11615776270627975, 0.058932479470968246, 0.17912904918193817, -0.003973599057644606, -0.05487505719065666, -0.07161571830511093, -0.14588361978530884, 0.04543330892920494, 0.051313865929841995, -0.024981042370200157, 0.029048750177025795, -0.0473591648042202, 0.00002306932583451271, -0.009492386132478714, 0.11747867614030838, -0.0923665463924408, -0.10041733831167221, 0.038793060928583145, 0.13782058656215668, -0.014738867990672588, -0.0067667714320123196, -0.05424042418599129, -0.0492212288081646, 0.07479297369718552, 0.22190631926059723, -0.06186041235923767, -0.14312468469142914, -0.06391780823469162, 0.06793031096458435, -0.036469560116529465, -0.046141888946294785, -0.03018493391573429, 0.11439930647611618, -0.030079638585448265, -0.14271780848503113, -0.01741124503314495, -0.02880118042230606, -0.013668842613697052, -0.027799777686595917, 0.0419594831764698, -0.007293186616152525, -0.06985048949718475, 0.0005784055683761835, -0.012327280826866627, 0.04031495004892349, -0.08582151681184769, 0.004928500857204199, 0.15631455183029175, 0.020275002345442772, 0.0739273726940155, -0.10666967183351517, -0.08047366887331009, -0.04465435445308685, -0.029700979590415955, -0.061598580330610275, 0.2248496264219284, 0.0028247737791389227, 0.04461359977722168, 0.15642289817333221, -0.07685696333646774, -0.18247158825397491, -0.04751787707209587, -0.08038323372602463, 0.02012043260037899, -0.01683724671602249, -0.030004749074578285, 0.11587697267532349, 0.12519341707229614, -0.04773114621639252, 0.15565969049930573, -0.23872138559818268, -0.08857216686010361, 0.016692036762833595, 0.14694522321224213, 0.4537327289581299, -0.24595481157302856, -0.02738776057958603, -0.13058315217494965, -0.0897328183054924, 0.07157308608293533, -0.07328633219003677, 0.096946120262146, -0.016323884949088097, -0.02127145417034626, 0.011331089772284031, -0.029652155935764313, 0.20162184536457062, -0.05046554282307625, 0.11329375952482224, -0.0693986788392067, -0.016841335222125053, 0.12218009680509567, -0.08951512724161148, 0.07741931080818176, -0.15570789575576782, 0.03301456198096275, -0.03403765335679054, -0.06600359827280045, 0.0036519740242511034, 0.04866120591759682, 0.020315831527113914, -0.05433768406510353, -0.06465814262628555, 0.05944480374455452, 0.022914623841643333, 0.038737256079912186, 0.04753197357058525, -0.107965387403965, 0.03020019643008709, 0.053230684250593185, -0.0008576505933888257, 0.06865202635526657, -0.045241910964250565, -0.09642357379198074, -0.055766407400369644, 0.12127038091421127, -0.1719105839729309, 0.0078207952901721, 0.09251511842012405, 0.007543096784502268, 0.0967380702495575, 0.02089065872132778, 0.01865559257566929, 0.0718972310423851, 0.1185075119137764, -0.08406058698892593, -0.07996302843093872, -0.035171885043382645, -0.07547501474618912, 0.111234150826931, -0.008769931271672249, 0.14345180988311768, -0.08802825957536697, 0.06427552551031113, 0.015550855547189713, 0.03222857788205147, -0.04280557110905647, 0.03148317709565163, -0.0052808611653745174, -0.029766393825411797, -0.08497351408004761, 0.10637524724006653, -0.008259416557848454, 0.026138082146644592, -0.028461823239922523, -0.02429720014333725, -0.08427426218986511, -0.05028751119971275, -0.028240442276000977, 0.18833118677139282, -0.10412394255399704, -0.097008116543293, -0.08217904716730118, -0.05395574867725372, -0.03224555030465126, 0.04657028988003731, 0.06976553052663803, -0.0711425170302391, -0.0050431834533810616, 0.025627968832850456, -0.10521725565195084, 0.01291944831609726, 0.0993943139910698, 0.068899005651474, -0.15540947020053864, -0.1096804067492485, -0.009226708672940731, 0.013526988215744495, -0.11086434870958328, -0.07183744758367538, -0.0687253549695015, 0.001450386713258922, -0.1235705316066742, 0.14932893216609955, -0.09738367050886154, -0.03456678241491318, -0.003398163942620158, -0.051760848611593246, -0.06128275394439697, -0.01165470015257597, -0.05499112606048584, 0.03951307386159897, 0.031832821667194366, 0.05926043167710304, -0.040755514055490494, -0.05467337369918823, 0.022649066522717476, -0.06907827407121658, -0.029341453686356544, 0.06190977618098259, -0.07443507760763168, 0.0024500712752342224, -0.2400490641593933, 0.017606081441044807, 0.10832569003105164, 0.015979759395122528, -0.058970674872398376, 0.0257949847728014, 0.043628010898828506, 0.0712103396654129, 0.03210585191845894, -0.04711724817752838, -0.06843167543411255, -0.08107980340719223, 0.09111817926168442, -0.10775187611579895, -0.040506210178136826, -0.01925293542444706, 0.0036740514915436506, 0.20120489597320557, 0.12314790487289429, 0.10467994213104248, -0.02359614707529545, 0.018020862713456154, -0.08041341602802277, 0.025930123403668404, 0.048855047672986984, -0.13150109350681305, -0.03358198329806328, -0.13501043617725372, 0.004730066750198603, -0.03092997334897518, 0.1653348058462143, 0.05501425266265869, -0.06604865938425064, -0.06947953999042511, -0.013880874961614609, 0.06731876730918884, -0.017005441710352898, 0.2754206359386444, 0.12278559803962708, 0.10321325808763504, -0.037395820021629333, 0.05306437611579895, 0.07804451137781143, 0.03606489300727844, 0.021333714947104454, 0.15422822535037994, -0.010633925907313824, 0.12496241927146912, -0.028561128303408623, 0.018886713311076164, 0.016924692317843437, 0.07580708712339401, -0.09072724729776382, 0.031690191477537155, 0.021859019994735718, -0.02094649337232113, 0.2142619490623474, -0.08996033668518066, 0.02258332073688507, 0.07124987244606018, 0.005456941667944193, -0.11422594636678696, -0.2887839674949646, -0.10356976836919785, -0.18004602193832397, 0.04302678629755974, -0.05595136061310768, 0.004572463221848011, 0.14367930591106415, 0.017719725146889687, 0.019190384075045586, 0.036328982561826706, -0.09217629581689835, -0.042188774794340134, 0.10568202286958694, -0.06901825964450836, -0.05327397212386131, -0.018671216443181038, -0.10507897287607193, 0.05042244866490364, -0.010984930209815502, -0.013188730925321579, 0.04758322611451149, 0.12676173448562622, 0.06880392879247665, -0.07176323980093002, -0.05679374933242798, -0.024725409224629402, 0.010383269749581814, -0.004207768011838198, 0.10875832289457321, 0.024277104064822197, -0.059210892766714096, 0.026536380872130394, 0.15364140272140503, -0.005736559629440308, -0.06914552301168442, -0.06231408193707466, 0.07587860524654388, -0.08364108204841614, 0.021568477153778076, -0.060390371829271317, -0.062306564301252365, -0.013707790523767471, 0.24369579553604126, 0.1519089788198471, -0.08796530961990356, 0.002706205705180764, -0.036342550069093704, 0.00367685966193676, -0.03338512405753136, 0.09952051192522049, -0.04038320854306221, 0.21345120668411255, -0.03644508123397827, 0.05119932070374489, -0.07789436727762222, -0.02767889015376568, -0.03951696678996086, 0.014674940146505833, -0.04454018548130989, -0.07577082514762878, -0.07160764187574387, 0.07407411187887192, -0.13543708622455597, -0.07017774134874344, 0.12360554933547974, -0.024666959419846535, -0.03835226222872734, -0.05194254219532013, -0.047552820295095444, 0.05401846766471863, -0.019221298396587372, -0.1493362933397293, 0.021662721410393715, -0.04821328446269035, 0.006319425534456968, -0.10173940658569336, -0.02510903961956501, -0.012599990703165531, 0.03835122659802437, 0.0754532739520073, -0.0003623466473072767, 0.013594083487987518, -0.006976701319217682, -0.020923061296343803, -0.028176426887512207, 0.13257862627506256, -0.05577809736132622, -0.14337094128131866, 0.01169323269277811, 0.05643641948699951, -0.0963752493262291, 0.029495006427168846, 0.07151266187429428, -0.057757411152124405, -0.010962729342281818, 0.1527152955532074, -0.03840531408786774, -0.035933274775743484, 0.026953965425491333, -0.16039623320102692, 0.10197768360376358, 0.043914150446653366, 0.04765534773468971, -0.04036213457584381, -0.039153024554252625, 0.16413834691047668, 0.04284438490867615, -0.09707621484994888, 0.028958791866898537, -0.04972374811768532, -0.09916674345731735, 0.04841947183012962, 0.03590292111039162, -0.18111860752105713, -0.02450616843998432, -0.14943066239356995, -0.02728099934756756, -0.011206702329218388, 0.04978489875793457, 0.19557778537273407, -0.0012976560974493623, -0.027155866846442223, -0.2541266977787018, 0.04015033692121506, 0.09818433970212936, -0.07154428958892822, -0.052393969148397446 ]
null
null
peft
# Low-rank decomposition of [valine/OpenSnark](https://huggingface.co/valine/OpenSnark) using [teknium/OpenHermes-2.5-Mistral-7B](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B) as base Created using [LoRD](https://github.com/thomasgauthier/LoRD)
{"library_name": "peft", "base_model": "teknium/OpenHermes-2.5-Mistral-7B"}
null
thomasgauthier/OpenSnark-LoRD
[ "peft", "safetensors", "base_model:teknium/OpenHermes-2.5-Mistral-7B", "region:us" ]
2024-02-15T04:07:21+00:00
[]
[]
TAGS #peft #safetensors #base_model-teknium/OpenHermes-2.5-Mistral-7B #region-us
# Low-rank decomposition of valine/OpenSnark using teknium/OpenHermes-2.5-Mistral-7B as base Created using LoRD
[ "# Low-rank decomposition of valine/OpenSnark using teknium/OpenHermes-2.5-Mistral-7B as base\n\nCreated using LoRD" ]
[ "TAGS\n#peft #safetensors #base_model-teknium/OpenHermes-2.5-Mistral-7B #region-us \n", "# Low-rank decomposition of valine/OpenSnark using teknium/OpenHermes-2.5-Mistral-7B as base\n\nCreated using LoRD" ]
[ 32, 36 ]
[ "passage: TAGS\n#peft #safetensors #base_model-teknium/OpenHermes-2.5-Mistral-7B #region-us \n# Low-rank decomposition of valine/OpenSnark using teknium/OpenHermes-2.5-Mistral-7B as base\n\nCreated using LoRD" ]
[ -0.0782826840877533, 0.015459190122783184, -0.005200248211622238, 0.045746177434921265, -0.007344833575189114, 0.02394048683345318, 0.0650351345539093, 0.1288812905550003, 0.17107923328876495, 0.07132041454315186, 0.06806416809558868, 0.08218362927436829, -0.08717890083789825, 0.20961549878120422, -0.08437418192625046, -0.20257626473903656, 0.07321972399950027, -0.0365288220345974, 0.022474762052297592, -0.00550417834892869, 0.05559211224317551, -0.02007284387946129, 0.00919974222779274, -0.05221220850944519, 0.048040635883808136, 0.05169510096311569, 0.009754159487783909, 0.03644561767578125, 0.11576516181230545, -0.009188276715576649, 0.049866896122694016, -0.00928640179336071, 0.053551070392131805, -0.1133948266506195, 0.07022105902433395, 0.011428339406847954, -0.03051416017115116, 0.08841881901025772, -0.004903467837721109, -0.080324187874794, 0.0742744505405426, -0.1840173453092575, -0.04260386526584625, 0.05113454908132553, -0.07307098060846329, -0.30411624908447266, -0.11657808721065521, -0.02859273925423622, 0.04269639775156975, 0.009493027813732624, 0.02397293411195278, 0.1820346564054489, 0.06904052942991257, 0.06555697321891785, 0.31592974066734314, -0.2782058119773865, -0.08602795004844666, 0.08729255944490433, -0.014919913373887539, 0.12469533830881119, -0.01953449659049511, 0.05578818544745445, 0.0626121237874031, -0.06952095031738281, -0.06980632990598679, 0.012320876121520996, -0.009136855602264404, -0.024789534509181976, -0.09862229228019714, 0.05796925351023674, 0.11151227355003357, 0.020418789237737656, -0.036173950880765915, 0.013362737372517586, -0.22368547320365906, -0.016015375033020973, -0.004381999373435974, 0.02547794207930565, -0.023548323661088943, 0.08647734671831131, 0.18527476489543915, -0.1207878589630127, -0.05550828576087952, -0.05805937573313713, -0.046456657350063324, 0.17235112190246582, -0.011257472448050976, 0.07042952626943588, 0.02043711021542549, 0.07816293835639954, -0.12745410203933716, -0.06487474590539932, 0.014017960987985134, -0.10324522107839584, -0.01834150031208992, -0.011013025417923927, -0.011121907271444798, -0.0037561804056167603, 0.1106758639216423, 0.2672174274921417, -0.07195325195789337, 0.035986363887786865, 0.01842390187084675, 0.0508367083966732, 0.002845409093424678, -0.1338343769311905, -0.01604951173067093, -0.10220393538475037, 0.11326450854539871, -0.036193203181028366, 0.12265585362911224, -0.007006563246250153, -0.027595454826951027, -0.1310071349143982, 0.05973281338810921, 0.033190712332725525, 0.06359432637691498, -0.052047062665224075, 0.014519918709993362, -0.018489215523004532, 0.2168305516242981, -0.025696970522403717, -0.038373664021492004, 0.0229296013712883, 0.03137102350592613, 0.18731689453125, 0.07431815564632416, -0.010481542907655239, 0.08682657778263092, 0.04070521518588066, -0.035652145743370056, 0.007352115586400032, -0.024863848462700844, -0.07291322201490402, -0.005628999322652817, -0.0953463762998581, 0.024598589166998863, -0.08380493521690369, -0.1582372784614563, 0.005258583463728428, 0.12751181423664093, -0.0652313083410263, 0.01643287017941475, 0.04563958942890167, -0.017847083508968353, 0.02278681844472885, 0.03588135540485382, -0.01199384219944477, 0.02055431343615055, -0.008439064025878906, -0.04537777975201607, 0.06674475222826004, -0.24915774166584015, -0.04720260575413704, -0.06302972882986069, 0.09548349678516388, -0.02915661595761776, -0.07862759381532669, -0.10852503031492233, 0.14925047755241394, -0.11301309615373611, -0.013004954904317856, -0.16511046886444092, -0.06996412575244904, 0.06081271916627884, 0.010709177702665329, -0.194297656416893, 0.06076514720916748, 0.05269002169370651, -0.11895564943552017, -0.020570412278175354, 0.0855569988489151, 0.014624412171542645, 0.04142313823103905, -0.010132203809916973, 0.0589480996131897, 0.13994009792804718, -0.09284944087266922, 0.050258561968803406, 0.09678825736045837, 0.08957377076148987, -0.049690041691064835, 0.0714368000626564, -0.05393308028578758, -0.0933515802025795, -0.009393705055117607, -0.030366579070687294, 0.025524267926812172, -0.038790877908468246, -0.08527067303657532, -0.09570349007844925, -0.036980003118515015, 0.1391879916191101, -0.04397943243384361, 0.07718191295862198, -0.022373370826244354, -0.01892782747745514, 0.030829178169369698, 0.11028335243463516, -0.004720508120954037, 0.057948071509599686, 0.007673509418964386, 0.20684613287448883, -0.0784807950258255, -0.040878888219594955, -0.12232127785682678, -0.13320837914943695, 0.011982862837612629, -0.05146045237779617, -0.07189469784498215, -0.010286114178597927, 0.09678337723016739, 0.051501382142305374, -0.03055271878838539, 0.020949965342879295, 0.04400237649679184, 0.045386794954538345, -0.055566005408763885, -0.1681513786315918, -0.0796775221824646, -0.09934481978416443, -0.0014066927833482623, -0.030069509521126747, -0.002670210786163807, 0.013952868059277534, 0.12968890368938446, 0.06275524944067001, 0.04394615441560745, 0.027546074241399765, 0.018545832484960556, 0.008598094806075096, -0.0789715051651001, 0.06507676094770432, -0.016290932893753052, -0.03703926131129265, -0.11031970381736755, -0.07137290388345718, 0.03142350912094116, 0.07804258912801743, 0.050395846366882324, 0.06352939456701279, -0.07740066945552826, -0.08433980494737625, 0.014816001057624817, 0.0374712236225605, 0.015016223303973675, -0.14402979612350464, -0.023298025131225586, -0.008318763226270676, -0.02385800890624523, -0.03320050239562988, -0.04796077311038971, -0.006935587152838707, -0.050108280032873154, 0.04778376221656799, 0.06098790466785431, -0.1638558954000473, 0.026288248598575592, 0.1867809295654297, -0.12750695645809174, 0.004324657376855612, 0.010524699464440346, -0.05434606969356537, -0.0075895520858466625, 0.12348977476358414, 0.002709613647311926, 0.10458620637655258, -0.03125002235174179, 0.07290440052747726, 0.006788702681660652, -0.03862670809030533, 0.05490247905254364, -0.0341871939599514, 0.008554402738809586, -0.07188329100608826, 0.014406396076083183, 0.04296311363577843, 0.21144452691078186, -0.08491215109825134, 0.07432648539543152, -0.005191479343920946, 0.029754221439361572, 0.014253593049943447, 0.010711616836488247, -0.028317412361502647, 0.11555984616279602, -0.1410592794418335, 0.13868336379528046, -0.0442008301615715, 0.04920310527086258, 0.036239899694919586, -0.026440806686878204, 0.03611019253730774, -0.05120029300451279, -0.06979475915431976, -0.04176328331232071, -0.052060890942811966, 0.04435189440846443, 0.07390245050191879, 0.0856829360127449, -0.032495398074388504, -0.008290792815387249, -0.06966450810432434, 0.0055904448963701725, -0.022467119619250298, 0.02445714734494686, 0.07932636141777039, -0.03464595600962639, -0.002560220193117857, -0.011864995583891869, 0.011789553798735142, 0.006486861500889063, 0.03116809017956257, 0.2687995433807373, 0.02834426611661911, 0.09466288238763809, 0.18489456176757812, -0.07925364375114441, 0.08052244037389755, 0.07956524193286896, 0.05856967717409134, -0.06728387624025345, -0.0014651857782155275, 0.023446349427103996, -0.06335180252790451, -0.21108536422252655, -0.0934653952717781, -0.11996009945869446, -0.04171927273273468, 0.012636023573577404, 0.1011168584227562, -0.0013678106479346752, 0.054957203567028046, -0.04466195032000542, 0.011565559543669224, -0.05007384344935417, 0.021114904433488846, 0.0379006452858448, 0.018733736127614975, -0.030520174652338028, -0.05267176032066345, 0.033345598727464676, 0.10453256219625473, 0.09151875227689743, 0.20972900092601776, 0.03946100175380707, 0.14487726986408234, 0.06600156426429749, 0.13342362642288208, 0.02706591598689556, 0.14615142345428467, -0.07325819134712219, 0.013054453767836094, -0.034455373883247375, -0.08727768063545227, 0.1146036684513092, 0.07029709219932556, -0.1574661135673523, 0.058361127972602844, 0.09497912973165512, 0.09637033194303513, 0.08070985972881317, 0.1386292427778244, 0.07982073724269867, -0.1393686980009079, -0.04311130940914154, 0.0911390483379364, 0.10862964391708374, -0.04348243400454521, -0.02794007770717144, -0.019938984885811806, 0.04803752154111862, 0.020329421386122704, -0.05629283934831619, 0.019560733810067177, 0.10828958451747894, -0.043641045689582825, 0.01431293599307537, 0.11059688031673431, -0.01060560718178749, 0.014352098107337952, -0.07192966341972351, 0.18790854513645172, 0.012769416905939579, -0.01711432635784149, 0.012742503546178341, -0.009759899228811264, 0.07733984291553497, 0.012669549323618412, 0.12947797775268555, 0.007957790978252888, -0.0952652171254158, 0.15484485030174255, -0.21749673783779144, 0.028803156688809395, 0.0391106940805912, -0.010505199432373047, 0.1035701185464859, -0.05368742346763611, -0.042648062109947205, -0.0003229541180189699, 0.1440889686346054, -0.1845390349626541, -0.07665269821882248, 0.05694150924682617, 0.018938954919576645, -0.08432064950466156, -0.02829274907708168, -0.06604131311178207, -0.04364212602376938, 0.11252281069755554, -0.027018636465072632, -0.10735996067523956, -0.14287570118904114, -0.03921220824122429, 0.20397980511188507, -0.05646641552448273, 0.10015621781349182, -0.013825546950101852, 0.020823944360017776, -0.0668688714504242, -0.1949608474969864, 0.1270993947982788, -0.07738814502954483, -0.07192222774028778, -0.025375498458743095, 0.09846815466880798, -0.03851412236690521, 0.013306332752108574, -0.03879205510020256, 0.012812790460884571, -0.05340836942195892, -0.04460044577717781, 0.004682794213294983, 0.2298867106437683, -0.12410906702280045, -0.06497570127248764, -0.09546501189470291, -0.05759183689951897, 0.019249876961112022, 0.0499589703977108, 0.10382813960313797, 0.2356729507446289, -0.10612841695547104, 0.07266739010810852, -0.0003930092207156122, -0.009655129164457321, -0.1261841356754303, 0.026631951332092285, -0.030282191932201385, -0.03119960054755211, -0.02887626551091671, -0.0927990972995758, 0.11157646775245667, 0.06704944372177124, -0.040090467780828476, 0.1538773626089096, -0.15319980680942535, -0.04119717329740524, 0.11464439332485199, 0.03744596242904663, 0.3075770139694214, -0.07888643443584442, -0.030929995700716972, -0.04058727249503136, -0.10732269287109375, 0.009754853323101997, -0.1874501258134842, -0.00020448780560400337, -0.06378880143165588, -0.06101067364215851, -0.017329111695289612, -0.036600179970264435, 0.1708625704050064, -0.03811792656779289, 0.09314421564340591, 0.0021247128024697304, -0.03124656341969967, 0.057028405368328094, -0.0592215396463871, 0.09125085920095444, 0.02612554281949997, 0.06031927838921547, 0.03638996556401253, 0.004587490577250719, -0.029055770486593246, 0.09349416196346283, -0.021953634917736053, -0.08899980783462524, 0.012300160713493824, 0.017330117523670197, -0.02388683706521988, -0.0012023817980661988, 0.16801971197128296, 0.10114891827106476, 0.15285363793373108, 0.08170217275619507, 0.017814498394727707, 0.025976402685046196, 0.14623452723026276, 0.1170257180929184, -0.03946954384446144, 0.06104299798607826, -0.10386674106121063, -0.021896284073591232, 0.1350654661655426, 0.05680247023701668, -0.0613517090678215, 0.016615115106105804, 0.00873594544827938, 0.053846459835767746, 0.09876088052988052, -0.18250975012779236, -0.14642491936683655, -0.02080896496772766, -0.008373197168111801, -0.046591274440288544, 0.07751097530126572, 0.13436946272850037, -0.007704684976488352, -0.03042883612215519, 0.04892267286777496, 0.017934992909431458, -0.01775335520505905, 0.014836288057267666, 0.03404788300395012, 0.01630444824695587, -0.03523460775613785, -0.015444879420101643, -0.005611943081021309, 0.004482519347220659, -0.08528132736682892, 0.08337678015232086, -0.09959439188241959, -0.1035202294588089, 0.010264459997415543, -0.08310720324516296, 0.0153043819591403, 0.07671108096837997, -0.08300860226154327, -0.02414707839488983, -0.021655362099409103, 0.1404290497303009, 0.09664379060268402, -0.013599298894405365, 0.020621340721845627, -0.001524349208921194, 0.03206381946802139, 0.07505851238965988, 0.003374602645635605, 0.08096358925104141, -0.18318428099155426, -0.05439360439777374, -0.14210115373134613, 0.007396833971142769, -0.01880863681435585, 0.018437407910823822, -0.02480982430279255, -0.03379586711525917, -0.18788811564445496, -0.004532921593636274, -0.09451259672641754, -0.0090512465685606, -0.027736879885196686, -0.00794444140046835, -0.010697384364902973, 0.08175037801265717, -0.06350492686033249, -0.025725113227963448, -0.010907297022640705, -0.017346419394016266, -0.0162598118185997, -0.02585817128419876, 0.05340622738003731, -0.05673671513795853, 0.10388874262571335, 0.08161803334951401, -0.025593847036361694, 0.052375659346580505, -0.18240314722061157, -0.03904619812965393, 0.09878984093666077, 0.000003861884124489734, 0.04227989539504051, -0.02188737317919731, -0.044199664145708084, -0.057719357311725616, 0.020117614418268204, -0.06064840033650398, 0.1235441267490387, -0.0652296394109726, -0.001757999649271369, -0.07457525283098221, 0.018789533525705338, 0.07227660715579987, -0.10142698884010315, 0.09189201891422272, 0.05599537119269371, 0.023393407464027405, 0.0010597399668768048, 0.0024498712737113237, -0.12316302210092545, 0.02395510859787464, -0.026437517255544662, 0.013575257733464241, -0.11788703501224518, -0.008319314569234848, 0.018929488956928253, -0.012307084165513515, 0.10323900729417801, -0.08215901255607605, -0.06030185893177986, -0.024318568408489227, -0.10284779965877533, 0.017124498263001442, 0.040048860013484955, 0.306253045797348, 0.027935177087783813, 0.05876081436872482, -0.11094389855861664, 0.033380135893821716, -0.0050774007104337215, 0.05101339519023895, 0.12985950708389282, 0.19004520773887634, -0.007020080927759409, 0.024822836741805077, 0.04376246780157089, -0.026666726917028427, -0.04310767725110054, 0.010958866216242313, 0.052975621074438095, -0.038407597690820694, -0.04564805328845978, -0.024648165330290794, 0.19113677740097046, -0.04283212870359421, -0.03840998187661171, 0.05488695204257965, -0.011800535023212433, -0.1686343401670456, -0.07863860577344894, -0.10764896869659424, -0.1585935354232788, -0.017625946551561356, -0.05664833262562752, -0.1281089037656784, -0.09783069044351578, 0.035289350897073746, 0.04385295882821083, 0.16805580258369446, 0.06701939553022385, -0.03872350603342056, 0.0165860578417778, 0.02131868340075016, -0.05161089822649956, 0.029487179592251778, -0.0728718712925911, -0.010970156639814377, -0.11318919062614441, -0.08250517398118973, -0.015544126741588116, -0.025113161653280258, 0.04578405246138573, -0.038574688136577606, -0.03634609282016754, -0.05629250779747963, 0.02557758241891861, -0.09702221304178238, 0.12992602586746216, 0.05066880211234093, -0.09465929865837097, -0.030695242807269096, 0.06190578266978264, -0.04057497903704643, -0.0391591340303421, -0.07701223343610764, 0.07672920823097229, -0.04106966033577919, 0.04482289403676987, -0.011073203757405281, -0.022682927548885345, 0.1038605123758316, -0.09893576800823212, 0.17061340808868408, 0.028316982090473175, 0.09838472306728363, 0.07051455974578857, 0.012430816888809204, -0.053300485014915466, 0.01216400507837534, 0.0532904714345932, 0.09047077596187592, -0.07595442235469818, -0.097231425344944, -0.041582897305488586, 0.026381853967905045, -0.16428521275520325, -0.06663468480110168, -0.0017649837536737323, -0.04928139969706535, -0.0203033909201622, 0.11890292167663574, -0.14628495275974274, 0.1156887337565422, 0.04730330407619476, -0.13402676582336426, -0.14028972387313843, -0.08091187477111816, -0.005303568206727505, 0.03434431180357933, -0.013678103685379028, -0.14252176880836487, -0.07682932913303375, 0.19971635937690735, -0.005156916566193104, -0.08647860586643219, -0.08690240234136581, 0.03727095574140549, 0.011031242087483406, -0.0008845196571201086, -0.0007764555048197508, -0.0353316031396389, 0.06496535986661911, -0.01579456217586994, -0.139401376247406, -0.0014964549336582422, 0.03578798100352287, -0.10483840107917786, 0.014181867241859436, -0.050240449607372284, -0.04611077159643173, 0.10709759593009949, 0.0432065948843956, -0.019167397171258926, -0.04250941053032875, 0.1994169056415558, -0.053236790001392365, -0.06540081650018692, 0.05158695578575134, -0.10685908049345016, 0.11204561591148376, 0.052386268973350525, -0.06575717031955719, 0.0005300151533447206, -0.014468123205006123, 0.05164604261517525, 0.0632985532283783, -0.050723787397146225, -0.07592758536338806, -0.0985807552933693, 0.04869639500975609, 0.009147529490292072, 0.031034700572490692, -0.02702365443110466, -0.05827951431274414, -0.11480752378702164, 0.014036428183317184, -0.08899455517530441, -0.0144398408010602, 0.1314389854669571, -0.021065212786197662, 0.015683995559811592, -0.06493327021598816, -0.006409002002328634, 0.03780005872249603, -0.12546734511852264, -0.030591590330004692 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
FINNUMBER/Yi-Ko-6B-Finch-QA-full
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T04:09:18+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 56, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.06061961501836777, 0.15481999516487122, -0.004844071343541145, 0.02074851468205452, 0.0983177199959755, 0.007407687604427338, 0.07119518518447876, 0.11185134947299957, -0.023851769044995308, 0.1167980208992958, 0.031993988901376724, 0.09781743586063385, 0.11217817664146423, 0.16186554729938507, 0.0015333457849919796, -0.22897611558437347, 0.049678247421979904, -0.125278040766716, -0.0294334813952446, 0.11977242678403854, 0.1422213912010193, -0.10954539477825165, 0.0752737894654274, -0.038042325526475906, -0.005828251596540213, -0.0323176346719265, -0.06205610930919647, -0.05266609415411949, 0.05311284959316254, 0.06794639676809311, 0.07308239489793777, 0.01171939354389906, 0.09106900542974472, -0.2724283039569855, 0.02348201349377632, 0.0805930644273758, -0.0006441773730330169, 0.07586129754781723, 0.04993962123990059, -0.08749990910291672, 0.07524524629116058, -0.060156844556331635, 0.1498761922121048, 0.07955671846866608, -0.09018243104219437, -0.19217631220817566, -0.07921334356069565, 0.09916994720697403, 0.1890910118818283, 0.05953684076666832, -0.026427440345287323, 0.11642678081989288, -0.08593545109033585, 0.013638701289892197, 0.06446459144353867, -0.06054406240582466, -0.055855002254247665, 0.06904532760381699, 0.08335285633802414, 0.08567540347576141, -0.12976622581481934, -0.010767064057290554, 0.015032444149255753, 0.008952446281909943, 0.08948688954114914, 0.017146794125437737, 0.1335189938545227, 0.040557652711868286, -0.13501930236816406, -0.043155476450920105, 0.09761431813240051, 0.03665134683251381, -0.04888195917010307, -0.2485782504081726, -0.023432478308677673, -0.04339504987001419, -0.03198111802339554, -0.03649339824914932, 0.043764639645814896, -0.014506848528981209, 0.07738617807626724, -0.004502781666815281, -0.0837155357003212, -0.04301247000694275, 0.07241875678300858, 0.06128999963402748, 0.02571401372551918, -0.015821760520339012, 0.0059297760017216206, 0.12327717989683151, 0.11431120336055756, -0.126715749502182, -0.052547648549079895, -0.06306339055299759, -0.08449548482894897, -0.044861067086458206, 0.030838407576084137, 0.037995077669620514, 0.045936476439237595, 0.23867325484752655, 0.007765117567032576, 0.053257301449775696, 0.04455438256263733, 0.014407169073820114, 0.06501194834709167, 0.11008983850479126, -0.05894824117422104, -0.09719445556402206, -0.028582042083144188, 0.10156717151403427, 0.007986726239323616, -0.04139331728219986, -0.05712985619902611, 0.07059531658887863, 0.018587570637464523, 0.12360043078660965, 0.08000938594341278, 0.003056557849049568, -0.0755772516131401, -0.062465377151966095, 0.17764076590538025, -0.15825673937797546, 0.04532013460993767, 0.03055616281926632, -0.0341108962893486, -0.009745313785970211, 0.012105142697691917, 0.025474950671195984, -0.021481726318597794, 0.09522198140621185, -0.05601342022418976, -0.034448131918907166, -0.11389608681201935, -0.03694311901926994, 0.030394554138183594, 0.011153047904372215, -0.02865210548043251, -0.03502652049064636, -0.08865131437778473, -0.06405586749315262, 0.09101516753435135, -0.07148737460374832, -0.04784895107150078, -0.016645915806293488, -0.07833752781152725, 0.021804187446832657, 0.01691517047584057, 0.09064167737960815, -0.0222476739436388, 0.03985358029603958, -0.0550384595990181, 0.061440225690603256, 0.11723454296588898, 0.027987057343125343, -0.05787884071469307, 0.061519939452409744, -0.2424532175064087, 0.10252492874860764, -0.07715212553739548, 0.04971238598227501, -0.15203025937080383, -0.02478341944515705, 0.03986154496669769, 0.01284773275256157, -0.008251311257481575, 0.14196595549583435, -0.21994100511074066, -0.030957341194152832, 0.16964265704154968, -0.10025953501462936, -0.08109250664710999, 0.060782887041568756, -0.05354252830147743, 0.11210215091705322, 0.04557164013385773, -0.02375967986881733, 0.05775221437215805, -0.14725260436534882, -0.011030761525034904, -0.041942402720451355, -0.0180682260543108, 0.16207332909107208, 0.0703711211681366, -0.06047816202044487, 0.07456906884908676, 0.01960151270031929, -0.014246034435927868, -0.04887177795171738, -0.02822130173444748, -0.1047162413597107, 0.01184528972953558, -0.06102835759520531, 0.018109694123268127, -0.021768750622868538, -0.09445013850927353, -0.029118487611413002, -0.17402999103069305, -0.0031633328180760145, 0.08821269869804382, -0.011630427092313766, -0.021509924903512, -0.11245372891426086, 0.009332616813480854, 0.030967719852924347, 0.0002618339203763753, -0.13677829504013062, -0.06033218279480934, 0.026970699429512024, -0.16097871959209442, 0.029791243374347687, -0.05741601809859276, 0.04530094936490059, 0.04005871340632439, -0.03433511033654213, -0.03489551320672035, 0.010874404571950436, 0.010431389324367046, -0.01894843392074108, -0.25422003865242004, -0.01882786676287651, -0.0234990194439888, 0.1751047968864441, -0.22956320643424988, 0.042598169296979904, 0.07489731162786484, 0.1460893303155899, 0.007349682506173849, -0.03550100699067116, 0.015185600146651268, -0.07262228429317474, -0.03268764168024063, -0.06316669285297394, -0.01207790058106184, -0.038400664925575256, -0.05820201337337494, 0.04906858503818512, -0.1686294972896576, -0.030321966856718063, 0.10717973858118057, 0.06342670321464539, -0.1473218947649002, -0.02780107781291008, -0.04056945815682411, -0.04624456167221069, -0.06676914542913437, -0.05461418256163597, 0.11812574416399002, 0.056411582976579666, 0.04860803112387657, -0.07140495628118515, -0.07455260306596756, 0.008036690764129162, -0.01956399530172348, -0.014917809516191483, 0.09334591031074524, 0.07554110884666443, -0.12264352291822433, 0.09177418053150177, 0.09668384492397308, 0.08576478064060211, 0.10314212739467621, -0.014663571491837502, -0.08914592862129211, -0.040637146681547165, 0.02245822176337242, 0.016187267377972603, 0.15129362046718597, -0.012961224652826786, 0.055492039769887924, 0.0358695350587368, -0.014034898020327091, 0.011105312965810299, -0.09736533463001251, 0.02655916102230549, 0.030835967510938644, -0.016302183270454407, 0.03745110332965851, -0.0447014644742012, 0.019208140671253204, 0.09039704501628876, 0.040895868092775345, 0.040978945791721344, 0.010155045427381992, -0.04354988783597946, -0.11037563532590866, 0.1787576973438263, -0.12389461696147919, -0.24818050861358643, -0.13812170922756195, 0.010281167924404144, 0.04737642779946327, -0.010411068797111511, 0.006690691225230694, -0.06616118550300598, -0.1175973042845726, -0.09878289699554443, 0.018617089837789536, 0.045352302491664886, -0.07590975612401962, -0.06842505931854248, 0.06414616107940674, 0.03875524550676346, -0.13939815759658813, 0.024007495492696762, 0.04662325978279114, -0.08205481618642807, -0.0029386086389422417, 0.0791812464594841, 0.06965780258178711, 0.17661017179489136, 0.013885351829230785, -0.023669935762882233, 0.026634456589818, 0.20819635689258575, -0.1436755359172821, 0.10975687950849533, 0.13545554876327515, -0.08767466992139816, 0.08120133727788925, 0.1998777538537979, 0.03777998685836792, -0.10680917650461197, 0.03608465939760208, 0.028374753892421722, -0.028325283899903297, -0.2502254545688629, -0.06958996504545212, 0.0019060121849179268, -0.05172049254179001, 0.07064855098724365, 0.08791537582874298, 0.09593888372182846, 0.016860228031873703, -0.09976044297218323, -0.07697858661413193, 0.046900223940610886, 0.10824491083621979, -0.00015424020239152014, -0.015208319760859013, 0.0904119610786438, -0.03033481352031231, 0.01743943803012371, 0.09215071052312851, 0.0030607767403125763, 0.17535938322544098, 0.051709048449993134, 0.17189906537532806, 0.07866133749485016, 0.06444311141967773, 0.02004685252904892, 0.007725914940237999, 0.021817529574036598, 0.017227526754140854, -0.0030957073904573917, -0.08709781616926193, -0.0034981227945536375, 0.1202581599354744, 0.049845851957798004, 0.029173865914344788, 0.012042860500514507, -0.030704669654369354, 0.08337877690792084, 0.1770893782377243, 0.0029054484330117702, -0.1893385946750641, -0.07169844210147858, 0.07795937359333038, -0.08648337423801422, -0.10729733109474182, -0.029470939189195633, 0.041069481521844864, -0.1729043871164322, 0.016882894560694695, -0.019335895776748657, 0.10788324475288391, -0.13190391659736633, -0.01772487722337246, 0.05657728388905525, 0.06932812184095383, -0.009677323512732983, 0.06694949418306351, -0.16090403497219086, 0.11770165711641312, 0.01751571334898472, 0.06636732816696167, -0.09608277678489685, 0.09618937969207764, -0.007830657996237278, 0.0041499207727611065, 0.1410749852657318, 0.010120149701833725, -0.05952107161283493, -0.09608154743909836, -0.10546442121267319, -0.009841260500252247, 0.1306990385055542, -0.14852415025234222, 0.08813067525625229, -0.02661319263279438, -0.044553373008966446, 0.003614129964262247, -0.12497276812791824, -0.13103094696998596, -0.18366187810897827, 0.05707118660211563, -0.12947207689285278, 0.04045100137591362, -0.10902881622314453, -0.045833900570869446, -0.02098964899778366, 0.20040063560009003, -0.23137451708316803, -0.06714103370904922, -0.1551055610179901, -0.08061286807060242, 0.14446212351322174, -0.046455029398202896, 0.08550118654966354, 0.0008278203313238919, 0.19068008661270142, 0.021319707855582237, -0.017237508669495583, 0.1072206199169159, -0.10052918642759323, -0.2010865956544876, -0.09273224323987961, 0.15895552933216095, 0.13766798377037048, 0.03809428587555885, -0.004381525795906782, 0.03171157464385033, -0.02098114788532257, -0.12076930701732635, 0.020226983353495598, 0.17317426204681396, 0.08982043713331223, 0.025265544652938843, -0.02972041629254818, -0.11267432570457458, -0.07061342149972916, -0.03774050623178482, 0.024755435064435005, 0.18072067201137543, -0.07222156971693039, 0.18405316770076752, 0.13775517046451569, -0.05534014105796814, -0.19904261827468872, 0.021996473893523216, 0.04293542355298996, 0.0070380112156271935, 0.0323902890086174, -0.20307663083076477, 0.09384101629257202, 0.0008334947633557022, -0.05131231248378754, 0.1379684954881668, -0.1823476254940033, -0.151598259806633, 0.06042521819472313, 0.043563615530729294, -0.19374065101146698, -0.12374074012041092, -0.08848230540752411, -0.04693066328763962, -0.15487661957740784, 0.10312657803297043, 0.0020827590487897396, 0.008401188999414444, 0.03778626397252083, 0.02252252586185932, 0.012139533646404743, -0.04198719933629036, 0.1914343535900116, -0.025891713798046112, 0.03347287327051163, -0.0790715217590332, -0.060851071029901505, 0.062408581376075745, -0.058187782764434814, 0.0755455270409584, -0.025226406753063202, 0.015947066247463226, -0.10598332434892654, -0.048235729336738586, -0.02852320298552513, 0.019321219995617867, -0.09431382268667221, -0.09348297864198685, -0.04829427972435951, 0.09367614984512329, 0.09042316675186157, -0.03652578964829445, -0.03649144619703293, -0.078715980052948, 0.038977332413196564, 0.17627815902233124, 0.18159319460391998, 0.04659178853034973, -0.07959239184856415, -0.001915142871439457, -0.014336181804537773, 0.04684065282344818, -0.22077152132987976, 0.060553863644599915, 0.04557652771472931, 0.016117896884679794, 0.11537692695856094, -0.0208132341504097, -0.16198977828025818, -0.06710557639598846, 0.061360616236925125, -0.06944561004638672, -0.17825035750865936, 0.0039279889315366745, 0.07344977557659149, -0.16578389704227448, -0.037031736224889755, 0.04200848564505577, -0.01189455483108759, -0.0403641052544117, 0.012352054007351398, 0.08063354343175888, 0.007078902795910835, 0.07699975371360779, 0.055281639099121094, 0.09124495089054108, -0.10227900743484497, 0.07410510629415512, 0.08149529248476028, -0.08644098788499832, 0.030720343813300133, 0.09573426842689514, -0.06469762325286865, -0.0346054881811142, 0.04237886518239975, 0.08354541659355164, 0.024281201884150505, -0.04682289808988571, 0.0023111123591661453, -0.09734189510345459, 0.05927345156669617, 0.11483542621135712, 0.03496333956718445, 0.011234734207391739, 0.03813567012548447, 0.04486291855573654, -0.08093374222517014, 0.11926916986703873, 0.023795632645487785, 0.020354853942990303, -0.04112942889332771, -0.040553025901317596, 0.035851649940013885, -0.026020776480436325, -0.011440055444836617, -0.035174157470464706, -0.0722682997584343, -0.014069457538425922, -0.16000694036483765, -0.0076758842915296555, -0.03660871088504791, 0.005114538595080376, 0.022510098293423653, -0.03652830421924591, 0.00792311318218708, 0.012217256240546703, -0.06868947297334671, -0.05553458258509636, -0.023233558982610703, 0.09422210603952408, -0.16494666039943695, 0.0220257006585598, 0.0823851153254509, -0.12121747434139252, 0.09289738535881042, 0.016782134771347046, 0.00412249518558383, 0.026962365955114365, -0.1545863002538681, 0.04763968288898468, -0.020152103155851364, 0.013473534025251865, 0.04222847521305084, -0.21637047827243805, -0.004404853098094463, -0.04015503451228142, -0.05566934496164322, -0.008993052877485752, -0.0319182425737381, -0.11338426172733307, 0.09645436704158783, 0.011025024577975273, -0.08443772792816162, -0.02965564839541912, 0.03353232145309448, 0.07690354436635971, -0.027447547763586044, 0.1498211771249771, -0.004663881380110979, 0.07559948414564133, -0.17581342160701752, -0.02282017655670643, -0.011197620071470737, 0.022367527708411217, -0.021871577948331833, -0.01622559316456318, 0.04623444378376007, -0.02704801969230175, 0.19120801985263824, -0.024701936170458794, 0.049393873661756516, 0.06364397704601288, 0.009232889860868454, -0.013832193799316883, 0.11151392012834549, 0.05708572641015053, 0.024334950372576714, 0.022262847051024437, 0.003451440716162324, -0.04008655622601509, -0.009981024079024792, -0.18596695363521576, 0.06803664565086365, 0.14585918188095093, 0.09060460329055786, -0.012669353745877743, 0.0707244873046875, -0.10161512345075607, -0.12005364894866943, 0.10127941519021988, -0.06415384262800217, -0.010188822634518147, -0.06542414426803589, 0.14027701318264008, 0.14953285455703735, -0.1886233240365982, 0.06583356112241745, -0.06602055579423904, -0.0566304549574852, -0.11457879096269608, -0.1930263340473175, -0.057075321674346924, -0.050602465867996216, -0.018466074019670486, -0.05384097993373871, 0.06939727067947388, 0.05750798434019089, 0.01126816775649786, 0.00868057832121849, 0.08568526059389114, -0.009656033478677273, 0.00248199631460011, 0.030120067298412323, 0.06713981181383133, 0.016768986359238625, -0.0321255661547184, 0.0179112758487463, -0.00597198773175478, 0.034156378358602524, 0.059282708913087845, 0.03608176112174988, -0.028436895459890366, 0.015559280291199684, -0.034912437200546265, -0.11309733241796494, 0.042801856994628906, -0.029640642926096916, -0.0749855786561966, 0.1347348988056183, 0.026981467381119728, 0.005015076603740454, -0.023140020668506622, 0.2503887414932251, -0.07436972856521606, -0.09334370493888855, -0.14373961091041565, 0.11701542884111404, -0.04212593287229538, 0.0635172426700592, 0.03596310690045357, -0.10810714215040207, 0.017985546961426735, 0.1320217251777649, 0.15442703664302826, -0.04732590913772583, 0.019251897931098938, 0.028577854856848717, 0.00439635943621397, -0.04075566306710243, 0.05177190154790878, 0.07100846618413925, 0.14500564336776733, -0.05157303810119629, 0.08530787378549576, 0.002609728369861841, -0.1021018698811531, -0.041973695158958435, 0.11415864527225494, -0.014296893030405045, 0.017620453611016273, -0.057136841118335724, 0.124222531914711, -0.05874236673116684, -0.23697422444820404, 0.06316976249217987, -0.0765061303973198, -0.1432730257511139, -0.024886758998036385, 0.071670763194561, -0.016632623970508575, 0.02605951391160488, 0.07167234271764755, -0.0754380151629448, 0.18880942463874817, 0.03957989811897278, -0.05233397334814072, -0.05954399332404137, 0.0744764655828476, -0.11850855499505997, 0.27879106998443604, 0.010482731275260448, 0.051307905465364456, 0.1042102724313736, -0.02021743729710579, -0.13270841538906097, 0.023401619866490364, 0.09579801559448242, -0.08917027711868286, 0.04087764397263527, 0.21448291838169098, -0.00629545608535409, 0.11935057491064072, 0.07611140608787537, -0.07468950748443604, 0.047562725841999054, -0.11468592286109924, -0.07639975845813751, -0.08699081838130951, 0.09244474768638611, -0.06785612553358078, 0.14258281886577606, 0.12599852681159973, -0.05530165135860443, 0.011584274470806122, -0.028389399871230125, 0.045467376708984375, 0.005578654818236828, 0.100032277405262, 0.011115525849163532, -0.18496567010879517, 0.024811718612909317, 0.016259413212537766, 0.10884406417608261, -0.18112654983997345, -0.09105053544044495, 0.046958595514297485, 0.0005061255069449544, -0.06443515419960022, 0.12483241409063339, 0.057313691824674606, 0.04654949903488159, -0.0451689288020134, -0.026830285787582397, -0.006042256020009518, 0.14264579117298126, -0.10707559436559677, -0.005129707511514425 ]
null
null
null
## Exllama v2 Quantizations of sparsetral-16x7B-v2-SPIN_iter0 Using <a href="https://github.com/turboderp/exllamav2/releases/tag/v0.0.13">turboderp's ExLlamaV2 v0.0.13</a> for quantization. <b>The "main" branch only contains the measurement.json, download one of the other branches for the model (see below)</b> Each branch contains an individual bits per weight, with the main one containing only the meaurement.json for further conversions. Original model: https://huggingface.co/serpdotai/sparsetral-16x7B-v2-SPIN_iter0 | Branch | Bits | lm_head bits | VRAM (4k) | VRAM (16k) | VRAM (32k) | Description | | ----- | ---- | ------- | ------ | ------ | ------ | ------------ | | [8_0](https://huggingface.co/bartowski/sparsetral-16x7B-v2-SPIN_iter0-exl2/tree/8_0) | 8.0 | 8.0 | 8.3 GB | 9.7 GB | 11.8 GB | Maximum quality that ExLlamaV2 can produce, near unquantized performance. | | [6_5](https://huggingface.co/bartowski/sparsetral-16x7B-v2-SPIN_iter0-exl2/tree/6_5) | 6.5 | 8.0 | 7.1 GB | 8.5 GB | 10.6 GB | Very similar to 8.0, good tradeoff of size vs performance, **recommended**. | | [5_0](https://huggingface.co/bartowski/sparsetral-16x7B-v2-SPIN_iter0-exl2/tree/5_0) | 5.0 | 6.0 | 5.7 GB | 7.1 GB | 9.2 GB | Slightly lower quality vs 6.5, but usable on 8GB cards. | | [4_25](https://huggingface.co/bartowski/sparsetral-16x7B-v2-SPIN_iter0-exl2/tree/4_25) | 4.25 | 6.0 | 5.1 GB | 6.5 GB | 8.6 GB | GPTQ equivalent bits per weight, slightly higher quality. | | [3_5](https://huggingface.co/bartowski/sparsetral-16x7B-v2-SPIN_iter0-exl2/tree/3_5) | 3.5 | 6.0 | 4.4 GB | 5.8 GB | 7.9 GB | Lower quality, only use if you have to. | ## Download instructions With git: ```shell git clone --single-branch --branch 6_5 https://huggingface.co/bartowski/sparsetral-16x7B-v2-SPIN_iter0-exl2 sparsetral-16x7B-v2-SPIN_iter0-exl2-6_5 ``` With huggingface hub (credit to TheBloke for instructions): ```shell pip3 install huggingface-hub ``` To download the `main` (only useful if you only care about measurement.json) branch to a folder called `sparsetral-16x7B-v2-SPIN_iter0-exl2`: ```shell mkdir sparsetral-16x7B-v2-SPIN_iter0-exl2 huggingface-cli download bartowski/sparsetral-16x7B-v2-SPIN_iter0-exl2 --local-dir sparsetral-16x7B-v2-SPIN_iter0-exl2 --local-dir-use-symlinks False ``` To download from a different branch, add the `--revision` parameter: Linux: ```shell mkdir sparsetral-16x7B-v2-SPIN_iter0-exl2-6_5 huggingface-cli download bartowski/sparsetral-16x7B-v2-SPIN_iter0-exl2 --revision 6_5 --local-dir sparsetral-16x7B-v2-SPIN_iter0-exl2-6_5 --local-dir-use-symlinks False ``` Windows (which apparently doesn't like _ in folders sometimes?): ```shell mkdir sparsetral-16x7B-v2-SPIN_iter0-exl2-6.5 huggingface-cli download bartowski/sparsetral-16x7B-v2-SPIN_iter0-exl2 --revision 6_5 --local-dir sparsetral-16x7B-v2-SPIN_iter0-exl2-6.5 --local-dir-use-symlinks False ``` Want to support my work? Visit my ko-fi page here: https://ko-fi.com/bartowski
{"language": ["en"], "license": "apache-2.0", "datasets": ["teknium/OpenHermes-2.5", "jondurbin/truthy-dpo-v0.1", "jondurbin/gutenberg-dpo-v0.1", "argilla/dpo-mix-7k"], "quantized_by": "bartowski", "pipeline_tag": "text-generation"}
text-generation
bartowski/sparsetral-16x7B-v2-SPIN_iter0-exl2
[ "text-generation", "en", "dataset:teknium/OpenHermes-2.5", "dataset:jondurbin/truthy-dpo-v0.1", "dataset:jondurbin/gutenberg-dpo-v0.1", "dataset:argilla/dpo-mix-7k", "license:apache-2.0", "region:us" ]
2024-02-15T04:11:22+00:00
[]
[ "en" ]
TAGS #text-generation #en #dataset-teknium/OpenHermes-2.5 #dataset-jondurbin/truthy-dpo-v0.1 #dataset-jondurbin/gutenberg-dpo-v0.1 #dataset-argilla/dpo-mix-7k #license-apache-2.0 #region-us
Exllama v2 Quantizations of sparsetral-16x7B-v2-SPIN\_iter0 ----------------------------------------------------------- Using <a href="URL ExLlamaV2 v0.0.13 for quantization. **The "main" branch only contains the URL, download one of the other branches for the model (see below)** Each branch contains an individual bits per weight, with the main one containing only the URL for further conversions. Original model: URL Download instructions --------------------- With git: With huggingface hub (credit to TheBloke for instructions): To download the 'main' (only useful if you only care about URL) branch to a folder called 'sparsetral-16x7B-v2-SPIN\_iter0-exl2': To download from a different branch, add the '--revision' parameter: Linux: Windows (which apparently doesn't like \_ in folders sometimes?): Want to support my work? Visit my ko-fi page here: URL
[]
[ "TAGS\n#text-generation #en #dataset-teknium/OpenHermes-2.5 #dataset-jondurbin/truthy-dpo-v0.1 #dataset-jondurbin/gutenberg-dpo-v0.1 #dataset-argilla/dpo-mix-7k #license-apache-2.0 #region-us \n" ]
[ 78 ]
[ "passage: TAGS\n#text-generation #en #dataset-teknium/OpenHermes-2.5 #dataset-jondurbin/truthy-dpo-v0.1 #dataset-jondurbin/gutenberg-dpo-v0.1 #dataset-argilla/dpo-mix-7k #license-apache-2.0 #region-us \n" ]
[ -0.10825076699256897, 0.22484834492206573, -0.004282933194190264, 0.04698605090379715, -0.005618900991976261, 0.028564633801579475, 0.1761474311351776, 0.1347513049840927, 0.04628629982471466, -0.05857393145561218, 0.14876136183738708, 0.10177440196275711, 0.045911047607660294, 0.1614788919687271, -0.09815645217895508, -0.1484382152557373, 0.05385097488760948, 0.0021751178428530693, -0.09885313361883163, 0.020084094256162643, 0.08661045134067535, 0.020490944385528564, 0.03693696856498718, -0.07446813583374023, 0.008845391683280468, 0.008898700587451458, -0.035553283989429474, -0.07354612648487091, 0.045201584696769714, -0.006202323362231255, 0.024522658437490463, 0.09192558377981186, -0.024899082258343697, -0.12293799221515656, 0.02576996013522148, 0.03266327455639839, -0.046071019023656845, 0.06136459857225418, 0.07195983827114105, 0.03398150950670242, 0.06403626501560211, -0.09169548749923706, -0.06752755492925644, 0.031712599098682404, -0.06013970822095871, -0.21328870952129364, -0.15908338129520416, -0.023730136454105377, -0.010800204239785671, 0.039917461574077606, 0.03721662983298302, 0.1133703738451004, 0.0053224116563797, 0.035916365683078766, 0.17593063414096832, -0.2044123411178589, -0.01578965038061142, 0.15733647346496582, 0.027966955676674843, 0.059646014124155045, 0.01569194532930851, -0.011278577148914337, 0.09117739647626877, -0.0022417958825826645, -0.034363798797130585, -0.05224643647670746, -0.09299280494451523, 0.03913378715515137, -0.05353829264640808, -0.025859791785478592, 0.45718705654144287, 0.046727266162633896, 0.005944403819739819, 0.026761481538414955, -0.08380990475416183, 0.0696106106042862, 0.03811686486005783, 0.03562905266880989, 0.05962410941720009, 0.0749499648809433, 0.09577469527721405, -0.0522746741771698, -0.10129068791866302, -0.022318869829177856, -0.17229993641376495, 0.047581229358911514, -0.03368600830435753, 0.07563144713640213, -0.12948530912399292, 0.033267054706811905, -0.056837908923625946, -0.14558175206184387, -0.000648156157694757, -0.09918256103992462, 0.13482436537742615, 0.010651222430169582, -0.06601186841726303, -0.09668566286563873, 0.21258440613746643, 0.1859954446554184, -0.010216766968369484, -0.023572102189064026, -0.09540095180273056, 0.05336606875061989, 0.053377795964479446, -0.06780563294887543, -0.053296931087970734, -0.0525687150657177, 0.1336444914340973, -0.061789482831954956, 0.098744697868824, -0.011599304154515266, -0.04826262965798378, -0.035829994827508926, -0.05093112960457802, 0.05646736919879913, 0.16781625151634216, 0.0021637415047734976, -0.03481622412800789, -0.07655372470617294, 0.13817012310028076, -0.08190394192934036, -0.06292684376239777, 0.005087447818368673, -0.08742178976535797, 0.08212589472532272, 0.007544206455349922, 0.08337944746017456, -0.014521857723593712, 0.017532723024487495, -0.027961984276771545, -0.08064991980791092, 0.007713889237493277, -0.022997120395302773, 0.13056734204292297, -0.12459561973810196, 0.05372095853090286, -0.10619297623634338, -0.27614137530326843, 0.0034930240362882614, 0.09181945770978928, -0.006455793511122465, -0.07696015387773514, 0.002017197199165821, -0.008809462189674377, 0.030814653262495995, -0.04732675850391388, 0.039240237325429916, -0.08888208866119385, 0.07195557653903961, -0.12176305800676346, 0.02818911336362362, -0.23693545162677765, 0.02944704331457615, -0.09779142588376999, -0.010496147908270359, 0.06390491873025894, 0.01099686324596405, -0.11192899942398071, 0.14225700497627258, -0.11234870553016663, 0.04300440847873688, -0.027881989255547523, -0.05736231058835983, 0.03261353820562363, 0.16886360943317413, -0.21043254435062408, 0.0019066184759140015, 0.10324842482805252, -0.06315729767084122, -0.18660937249660492, 0.052725907415151596, 0.016719499602913857, 0.062086425721645355, 0.07482299953699112, 0.21409158408641815, 0.09367331862449646, -0.07090243697166443, -0.0024918802082538605, 0.0856517106294632, 0.03950142487883568, -0.13781918585300446, 0.11834155023097992, -0.07523217797279358, -0.01748029515147209, 0.040815211832523346, -0.018212614580988884, 0.0439121276140213, 0.021832682192325592, -0.09362099319696426, -0.06179729849100113, -0.10284728556871414, 0.010994309559464455, -0.030483920127153397, 0.009793931618332863, -0.06237002834677696, 0.04257013648748398, -0.08907677233219147, 0.10968287289142609, 0.027236968278884888, 0.02293417416512966, 0.014441451989114285, 0.040032949298620224, -0.029143832623958588, 0.050654586404561996, -0.05651576817035675, -0.031002141535282135, 0.021019812673330307, -0.012914353981614113, -0.024798154830932617, 0.01863710768520832, 0.03959595412015915, -0.10492604225873947, -0.019519861787557602, 0.039656538516283035, 0.013687104918062687, 0.02327735163271427, 0.052092261612415314, -0.15964753925800323, 0.06113435700535774, -0.05239617079496384, 0.08843972533941269, 0.010159388184547424, 0.0036553870886564255, 0.04389719292521477, 0.08749295771121979, 0.014855429530143738, 0.043172817677259445, 0.04998461902141571, -0.045684412121772766, -0.013154813088476658, -0.04275445267558098, 0.06373419612646103, 0.06534447520971298, -0.11089784651994705, 0.02896452322602272, 0.06754756718873978, 0.1528182327747345, 0.15307633578777313, -0.010068618692457676, 0.11154232919216156, -0.06523067504167557, -0.02677677944302559, -0.03815541788935661, 0.022619685158133507, 0.009052731096744537, -0.1055319607257843, -0.013990215957164764, -0.003612364875152707, -0.08776509016752243, 0.0012364526046440005, 0.0002154160029022023, -0.0329752080142498, -0.03906382992863655, 0.02665812149643898, 0.10519248992204666, -0.15923476219177246, 0.12877430021762848, 0.2676759958267212, -0.04768966883420944, 0.14172989130020142, -0.11756148934364319, -0.06164664030075073, 0.005007212981581688, 0.010656865313649178, -0.03643162548542023, 0.09020505100488663, -0.07908204942941666, 0.10772819072008133, 0.11322145909070969, 0.033839352428913116, 0.04412543773651123, -0.05782948061823845, -0.08354056626558304, -0.02861827239394188, -0.046816714107990265, -0.11581917852163315, 0.054207853972911835, -0.04330188408493996, 0.11849770694971085, -0.0006747291190549731, -0.026132432743906975, 0.12711204588413239, -0.019905338063836098, -0.05413369461894035, 0.13019195199012756, -0.1871015727519989, -0.1042393147945404, 0.011950680986046791, -0.05240621790289879, -0.12901921570301056, -0.005057459231466055, 0.06782098859548569, -0.0633939802646637, -0.01764436438679695, -0.015544886700809002, -0.018240192905068398, -0.012018945068120956, 0.0054582031443715096, 0.062300994992256165, 0.05839751288294792, 0.011066826060414314, -0.14101627469062805, -0.005654155742377043, 0.031063023954629898, 0.03905940428376198, 0.11203962564468384, -0.09057963639497757, 0.07656220346689224, 0.04861436039209366, 0.08414322137832642, 0.008274043910205364, -0.02085631527006626, 0.12081737071275711, -0.024785779416561127, 0.006251120939850807, 0.14142563939094543, 0.025345031172037125, 0.0237033199518919, 0.16302305459976196, 0.04990953952074051, -0.07985144108533859, 0.012381543405354023, -0.028291266411542892, -0.029070083051919937, -0.3583139479160309, -0.08078110218048096, -0.09039674699306488, 0.20955973863601685, 0.02700120024383068, 0.03533821552991867, 0.07242799550294876, 0.06487344950437546, -0.05668189004063606, 0.07078249752521515, 0.0022712997160851955, 0.012503847479820251, 0.16503335535526276, 0.018457133322954178, 0.022167852148413658, -0.09086067229509354, 0.05509160831570625, 0.17897549271583557, 0.16388612985610962, 0.13800300657749176, 0.0618848092854023, 0.18627801537513733, 0.03225943073630333, 0.12760449945926666, -0.03861590102314949, 0.09969348460435867, 0.0661115050315857, 0.050703857094049454, -0.043055493384599686, -0.08434580266475677, -0.004392649512737989, 0.04755640774965286, -0.1154148206114769, -0.12031257152557373, 0.02806648053228855, -0.06591729819774628, 0.046694036573171616, 0.12340979278087616, 0.08429720252752304, -0.14752371609210968, 0.014638255350291729, 0.07387958467006683, 0.11325518786907196, -0.017195122316479683, 0.09779869765043259, -0.08570842444896698, -0.02365938015282154, 0.17739149928092957, -0.004159050062298775, 0.09988078474998474, -0.009962160140275955, -0.035776179283857346, -0.03702574968338013, -0.023034939542412758, 0.01598471775650978, 0.09305741637945175, -0.3123355209827423, 0.16120155155658722, 0.016038933768868446, -0.039506178349256516, -0.09016572684049606, -0.006145824212580919, 0.05594111606478691, 0.02467254549264908, 0.09643258899450302, 0.019227785989642143, 0.02689320594072342, 0.17771661281585693, -0.15692013502120972, 0.06503229588270187, -0.013416305184364319, -0.028089605271816254, -0.04460645467042923, 0.029616666957736015, 0.042471688240766525, 0.02242792397737503, 0.15933352708816528, -0.1899283528327942, -0.07978957146406174, 0.09475115686655045, 0.13188423216342926, -0.011382697150111198, -0.0443628691136837, -0.029886452481150627, -0.04546056687831879, 0.14103519916534424, 0.004841489251703024, -0.10656740516424179, -0.07448915392160416, 0.0032550019677728415, 0.11229614168405533, -0.006921922322362661, -0.00923927128314972, -0.02166254073381424, -0.027771245688199997, -0.07450597733259201, -0.2007501721382141, 0.10079739987850189, -0.1088159829378128, -0.019941816106438637, -0.05308296158909798, 0.10462012887001038, -0.021983519196510315, 0.018064342439174652, -0.020560061559081078, -0.007288477849215269, -0.05090729892253876, -0.04920465126633644, 0.037862252444028854, 0.06740281730890274, 0.02121387980878353, 0.030186478048563004, -0.024534182623028755, -0.05084587261080742, 0.015809422358870506, -0.11508412659168243, 0.13932019472122192, 0.23936378955841064, -0.03653893619775772, 0.15231120586395264, 0.11736973375082016, -0.08832330256700516, -0.20296114683151245, -0.06892517954111099, -0.058092471212148666, -0.07528772205114365, -0.01944585144519806, -0.23807409405708313, 0.06236262992024422, 0.1253160536289215, -0.08733472228050232, 0.19761469960212708, -0.24742603302001953, -0.055726997554302216, 0.15926243364810944, -0.04858454689383507, 0.29954707622528076, -0.16961897909641266, -0.08901017904281616, -0.11291242390871048, -0.21639321744441986, 0.21920841932296753, -0.23983100056648254, 0.025758877396583557, -0.00869054812937975, 0.002319150837138295, -0.011291258037090302, -0.01858515478670597, 0.155086487531662, 0.04149475321173668, 0.01628432422876358, -0.0992705300450325, 0.0810835137963295, 0.11581472307443619, -0.0400247685611248, 0.0673980712890625, -0.08218426257371902, 0.05123922973871231, -0.12100202590227127, -0.0031597355846315622, -0.033031679689884186, 0.08679180592298508, -0.02733512595295906, -0.117374487221241, -0.041604235768318176, -0.017923859879374504, 0.03670976683497429, -0.011023769155144691, 0.20562508702278137, 0.09406439960002899, 0.007229998242110014, 0.11797711998224258, 0.0454946868121624, -0.08070982992649078, 0.13136988878250122, -0.04937407374382019, -0.06532593816518784, 0.047100089490413666, -0.23140095174312592, -0.01879945583641529, 0.11373744159936905, 0.0007949915016070008, 0.001757325604557991, 0.008603273890912533, -0.051765669137239456, -0.026851842179894447, 0.08275699615478516, -0.17939485609531403, 0.027785098180174828, -0.04067618027329445, 0.0798850730061531, -0.019010521471500397, 0.08296376466751099, 0.17157775163650513, -0.0008167075575329363, -0.024548856541514397, 0.028907375410199165, 0.06590821593999863, -0.044171303510665894, 0.08363024145364761, 0.06494862586259842, -0.05146497115492821, -0.11625517904758453, 0.22034505009651184, 0.04800935089588165, -0.1401451826095581, -0.006455428898334503, 0.04887176677584648, -0.09526123851537704, -0.07578901946544647, 0.009776433929800987, 0.02579747699201107, -0.09078829735517502, -0.08126979321241379, -0.046327635645866394, -0.050062865018844604, 0.03750654309988022, -0.004214140120893717, 0.08468512445688248, 0.06644932925701141, 0.07876381278038025, -0.06466828286647797, 0.0539097785949707, 0.04098008573055267, -0.0012539959279820323, 0.023066554218530655, -0.08978074789047241, -0.15784016251564026, -0.08941728621721268, 0.03445300832390785, -0.002113426337018609, 0.02099868655204773, -0.04959726333618164, -0.02391681633889675, -0.16933323442935944, 0.010503108613193035, -0.04112052172422409, 0.0138707859441638, -0.05037321522831917, -0.02798476442694664, -0.022428514435887337, 0.017209136858582497, -0.07650329172611237, -0.030327288433909416, -0.032026227563619614, 0.07253287732601166, -0.10658573359251022, -0.09646930545568466, 0.025893107056617737, -0.006783757358789444, 0.11738037317991257, 0.10484029352664948, -0.04283881559967995, 0.011688641272485256, -0.11491116136312485, -0.04634086787700653, 0.06162368506193161, 0.04736422747373581, -0.026569409295916557, -0.08410917967557907, -0.07897598296403885, 0.10222368687391281, -0.05883987993001938, -0.0034978422336280346, 0.008906511589884758, -0.10667291283607483, -0.0943925753235817, -0.028550924733281136, -0.03841577470302582, -0.025345243513584137, -0.10928484052419662, 0.17640066146850586, 0.06263431161642075, 0.09665359556674957, 0.02177467942237854, 0.018115105107426643, -0.10582438111305237, 0.007978003472089767, -0.002119641751050949, -0.08461157977581024, -0.12736600637435913, 0.06275469809770584, 0.012068050913512707, -0.03447822853922844, 0.18226467072963715, -0.049154169857501984, -0.1179812103509903, 0.028485501185059547, -0.02094234898686409, 0.017471639439463615, 0.013375750742852688, 0.32619282603263855, 0.042995158582925797, -0.0038609642069786787, -0.08473605662584305, 0.02053404599428177, 0.0466894656419754, 0.11297047883272171, 0.06345024704933167, 0.10155703127384186, 0.15592007339000702, 0.030204299837350845, 0.020526351407170296, -0.06828457862138748, 0.04736775532364845, 0.04106014221906662, 0.13665805757045746, 0.09495194256305695, 0.051281556487083435, 0.06729023903608322, 0.07711559534072876, -0.06321866065263748, -0.0153909707441926, -0.046280935406684875, -0.016538947820663452, -0.09704017639160156, -0.16517406702041626, -0.06205831468105316, -0.1619621366262436, -0.020198477432131767, -0.13729554414749146, 0.026324748992919922, 0.04533889517188072, 0.03994954749941826, -0.07367968559265137, -0.02488522045314312, -0.035156767815351486, -0.11497083306312561, 0.08481008559465408, -0.018813835456967354, -0.09604582190513611, -0.02746199443936348, -0.07653307914733887, 0.005921598523855209, 0.03709224984049797, -0.05888809636235237, 0.01519270334392786, 0.03967544436454773, 0.05458987504243851, -0.06949160993099213, -0.054601699113845825, -0.06734165549278259, 0.02896827831864357, -0.010065626353025436, 0.18339617550373077, 0.04210137948393822, 0.008143913000822067, 0.13236185908317566, 0.16886666417121887, 0.0025270478799939156, -0.11596042662858963, -0.08454226702451706, -0.10665607452392578, -0.039357706904411316, 0.006024789065122604, 0.044117726385593414, -0.06209639832377434, -0.03353248909115791, 0.10643613338470459, 0.2765374779701233, -0.0431499183177948, -0.005356189329177141, -0.012021870352327824, 0.018783142790198326, 0.011939859949052334, 0.04707859084010124, 0.09675534069538116, 0.07227057218551636, -0.03102019429206848, -0.04015299305319786, -0.03065488487482071, 0.0420064702630043, -0.10406073927879333, 0.05911519378423691, 0.01655520871281624, -0.144989475607872, 0.017912982031702995, 0.15021218359470367, -0.10229459404945374, -0.009664087556302547, -0.08851590752601624, -0.11513011902570724, -0.06831681728363037, -0.06462877243757248, 0.028655176982283592, 0.05002741515636444, 0.003436828264966607, -0.06984024494886398, 0.005660810507833958, 0.1289599984884262, -0.03378574550151825, -0.1995798498392105, -0.08028727769851685, 0.05015116184949875, -0.06615881621837616, 0.09239468723535538, 0.003996691666543484, 0.13084810972213745, 0.05795453116297722, 0.03701706603169441, -0.15087835490703583, -0.0003166517417412251, 0.07638026028871536, 0.037692904472351074, -0.0046877446584403515, -0.10267703235149384, -0.08189156651496887, 0.010179836302995682, 0.11915186047554016, 0.030018219724297523, -0.04186120629310608, 0.31369102001190186, 0.0013463544892147183, -0.07743126899003983, 0.10237244516611099, -0.15439599752426147, 0.0693897157907486, -0.0029316830914467573, -0.06758007407188416, -0.023448709398508072, -0.05916494131088257, 0.01866641640663147, 0.025295216590166092, -0.15762931108474731, -0.029596177861094475, 0.008757563307881355, -0.042456597089767456, 0.04511064291000366, 0.05481184273958206, -0.06035862863063812, 0.01101012621074915, -0.10414763540029526, -0.007382539566606283, -0.0869925320148468, 0.002983568934723735, 0.01393928937613964, -0.07556755840778351, -0.030846228823065758, 0.09637870639562607, 0.004186707083135843, 0.0018374709179624915, -0.06091741845011711, -0.104821115732193 ]
null
null
peft
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.8.2
{"library_name": "peft", "base_model": "ybelkada/blip2-opt-2.7b-fp16-sharded"}
null
leoreigoto/Data2_V2_BLIP2_Finetune_Caption_First_Epoch
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:ybelkada/blip2-opt-2.7b-fp16-sharded", "has_space", "region:us" ]
2024-02-15T04:11:41+00:00
[ "1910.09700" ]
[]
TAGS #peft #safetensors #arxiv-1910.09700 #base_model-ybelkada/blip2-opt-2.7b-fp16-sharded #has_space #region-us
# Model Card for Model ID ## Model Details ### Model Description - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact ### Framework versions - PEFT 0.8.2
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ "TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-ybelkada/blip2-opt-2.7b-fp16-sharded #has_space #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ 53, 6, 3, 54, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4, 11 ]
[ "passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-ybelkada/blip2-opt-2.7b-fp16-sharded #has_space #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2" ]
[ -0.10039013624191284, 0.21657955646514893, -0.0024030706845223904, 0.0256516020745039, 0.09570260345935822, 0.014424755237996578, 0.052417729049921036, 0.1452466994524002, 0.023049617186188698, 0.12104405462741852, 0.05373246222734451, 0.09572810679674149, 0.10243304073810577, 0.23337650299072266, 0.02834252640604973, -0.18494552373886108, 0.006894040387123823, -0.08319985121488571, 0.021731829270720482, 0.12368126958608627, 0.14025932550430298, -0.09964630752801895, 0.06765026599168777, -0.03520560264587402, -0.0031814523972570896, -0.01613553799688816, -0.08129876106977463, -0.05055828392505646, 0.06214886158704758, 0.047316063195466995, 0.033082280308008194, -0.0033997977152466774, 0.087236687541008, -0.27969133853912354, 0.0125362453982234, 0.06085585057735443, 0.01203862577676773, 0.0641179084777832, 0.09931062161922455, -0.03897527605295181, 0.12662631273269653, -0.0493718720972538, 0.12789446115493774, 0.08257599174976349, -0.07108437269926071, -0.24258004128932953, -0.07879762351512909, 0.07382190972566605, 0.13893209397792816, 0.07019390165805817, -0.038212355226278305, 0.1519656628370285, -0.07368765026330948, 0.02208847925066948, 0.10206539928913116, -0.12156807631254196, -0.07267110049724579, 0.07076189666986465, 0.13043972849845886, 0.1049388125538826, -0.12642540037631989, -0.027906347066164017, 0.04348429664969444, 0.04636335000395775, 0.06364186853170395, 0.015285314060747623, 0.12886805832386017, 0.046696458011865616, -0.1402956247329712, -0.056097228080034256, 0.1489037722349167, 0.020678529515862465, -0.05512630566954613, -0.22690244019031525, -0.004610613454133272, -0.07790571451187134, -0.018987881019711494, -0.06138458475470543, 0.026149746030569077, -0.01585191674530506, 0.09918764978647232, -0.036500249058008194, -0.09142713993787766, -0.03794299438595772, 0.11442708224058151, 0.07099685817956924, 0.017569098621606827, -0.01625903509557247, 0.0368226021528244, 0.13838408887386322, 0.05176091566681862, -0.11190199851989746, -0.05818786844611168, -0.07140254229307175, -0.07870882749557495, -0.03547414019703865, 0.052074134349823, 0.0430535152554512, 0.0543309785425663, 0.24075549840927124, -0.0012998550664633512, 0.054489027708768845, 0.03126071020960808, 0.004462297074496746, 0.03572967275977135, 0.08437017351388931, -0.04933582618832588, -0.18604116141796112, -0.03748719021677971, 0.1140938401222229, -0.003911278210580349, -0.01980844885110855, -0.020390303805470467, 0.03073650598526001, 0.06823064386844635, 0.12009209394454956, 0.09605144709348679, -0.027003273367881775, -0.06341927498579025, -0.04062524437904358, 0.2160993069410324, -0.15406128764152527, 0.04000493139028549, -0.002761922078207135, -0.019594332203269005, -0.079726941883564, 0.019290266558527946, 0.01583869569003582, -0.006216653622686863, 0.09396684914827347, -0.055339306592941284, -0.050581514835357666, -0.111453115940094, -0.03585340082645416, 0.043069981038570404, -0.0014089044416323304, -0.030897041782736778, -0.06259839236736298, -0.10000528395175934, -0.07459314912557602, 0.06340686976909637, -0.05388692766427994, -0.06441399455070496, -0.0047172605991363525, -0.08242795616388321, 0.0072420137003064156, 0.0008718289900571108, 0.11031666398048401, -0.03697417676448822, 0.06008594110608101, -0.008484327234327793, 0.058513544499874115, 0.10597467422485352, 0.027767786756157875, -0.07601456344127655, 0.06216181814670563, -0.1921064853668213, 0.07992986589670181, -0.11940265446901321, 0.03005889616906643, -0.16931629180908203, -0.02637491002678871, 0.005713535472750664, 0.013775072991847992, 0.031809691339731216, 0.139659583568573, -0.2088581621646881, -0.013981545343995094, 0.15462680160999298, -0.09721225500106812, -0.1267915815114975, 0.07031789422035217, -0.026513325050473213, 0.1625058799982071, 0.02627221867442131, 0.00034044793574139476, 0.10421278327703476, -0.16374431550502777, -0.04727965220808983, -0.022161299362778664, 0.022395381703972816, 0.10738769918680191, 0.08512938022613525, -0.07287324219942093, 0.019486110657453537, 0.02769216150045395, -0.05885820463299751, -0.02240985631942749, -0.05410797521471977, -0.10463036596775055, -0.010983024723827839, -0.0950881689786911, 0.0383855439722538, 0.005023611709475517, -0.07816193997859955, -0.029721932485699654, -0.1553298681974411, -0.04630150645971298, 0.10683421790599823, 0.013428928330540657, -0.01928403414785862, -0.08577392995357513, 0.042430974543094635, -0.038688186556100845, -0.01853264309465885, -0.1672455370426178, -0.04265793040394783, 0.0548766553401947, -0.13821101188659668, -0.008626015856862068, -0.08321557193994522, 0.0678725615143776, 0.018922066316008568, -0.059243109077215195, -0.037998102605342865, -0.0004915246390737593, -0.0032141045667231083, -0.05715692415833473, -0.1993669867515564, -0.052292030304670334, -0.04968138411641121, 0.1607537865638733, -0.24297834932804108, 0.0352422334253788, 0.03289943188428879, 0.1363774836063385, 0.0058966949582099915, -0.06887710094451904, 0.024105742573738098, -0.049690697342157364, -0.031480301171541214, -0.06410276889801025, -0.004855535924434662, -0.01228423323482275, -0.02487354539334774, 0.007814871147274971, -0.1309639811515808, -0.036985814571380615, 0.083965964615345, 0.12220241874456406, -0.14506429433822632, -0.011172892525792122, -0.05952780693769455, -0.06801122426986694, -0.08420345187187195, -0.07493065297603607, 0.07536260038614273, 0.049253910779953, 0.0458417572081089, -0.0819956362247467, -0.06924261152744293, 0.017457325011491776, 0.004613655153661966, -0.017426228150725365, 0.11938370019197464, 0.06072527542710304, -0.06673262268304825, 0.08374842256307602, 0.07593663036823273, 0.05903845280408859, 0.0909430980682373, -0.003896536538377404, -0.11722707003355026, -0.028167778626084328, 0.06690454483032227, 0.014271051622927189, 0.16343623399734497, -0.06832274049520493, 0.04013467952609062, 0.05302174389362335, -0.04756931960582733, 0.03855845704674721, -0.09477328509092331, 0.01861688867211342, 0.003786276327446103, -0.01797379180788994, 0.05750349909067154, -0.006069931201636791, -0.0018295986810699105, 0.07608667016029358, 0.04906469210982323, 0.036879509687423706, 0.016423199325799942, -0.031098250299692154, -0.13429516553878784, 0.16098903119564056, -0.08634848892688751, -0.24564915895462036, -0.15840746462345123, 0.03228582441806793, 0.03610482066869736, -0.02986353263258934, 0.03273465484380722, -0.04697467386722565, -0.09861129522323608, -0.09470655769109726, -0.00644270284101367, 0.03635037690401077, -0.06407404690980911, -0.07429859787225723, 0.05781225860118866, 0.051477957516908646, -0.12834446132183075, 0.03511073440313339, 0.0664653480052948, -0.017233682796359062, 0.0007426906959153712, 0.0739726647734642, 0.10749318450689316, 0.14343838393688202, 0.0024507450871169567, -0.004398123826831579, 0.044044289737939835, 0.25428658723831177, -0.14874325692653656, 0.09937742352485657, 0.13538265228271484, -0.032471586018800735, 0.07272493839263916, 0.16842791438102722, 0.028236016631126404, -0.08506052941083908, 0.042820997536182404, 0.04097886383533478, -0.026646096259355545, -0.2597343623638153, -0.07186245173215866, -0.010376784019172192, -0.06384555250406265, 0.09980839490890503, 0.09774170815944672, 0.09468974173069, 0.024150844663381577, -0.06351613253355026, -0.052666813135147095, 0.013371585868299007, 0.1010521650314331, -0.037481509149074554, -0.005767601076513529, 0.07915021479129791, -0.04645397514104843, 0.009784448891878128, 0.09607289731502533, 0.009811758063733578, 0.1678275614976883, 0.02197352610528469, 0.0838027223944664, 0.07178521901369095, 0.10981250554323196, -0.011338148266077042, 0.030696362257003784, 0.04156498983502388, 0.028653357177972794, 0.015803711488842964, -0.1000349223613739, 0.00037310158950276673, 0.13143542408943176, 0.019680025056004524, 0.020477866753935814, 0.01608787663280964, -0.039162810891866684, 0.03834003955125809, 0.21798481047153473, 0.007780767511576414, -0.21053139865398407, -0.07911351323127747, 0.062164902687072754, -0.08506964147090912, -0.1432492434978485, -0.01401080284267664, 0.04644948989152908, -0.1812177300453186, 0.023421449586749077, -0.03851275146007538, 0.08921558409929276, -0.06780043244361877, -0.03306172043085098, 0.09014959633350372, 0.0592069998383522, -0.01142523717135191, 0.07288417220115662, -0.16962651908397675, 0.12276750057935715, 0.028014227747917175, 0.07706160843372345, -0.10877016931772232, 0.08445385098457336, 0.001987135037779808, -0.02883685566484928, 0.1874581128358841, -0.010979101061820984, -0.038437943905591965, -0.08018875122070312, -0.09492190927267075, -0.01215397473424673, 0.0979175940155983, -0.11549250036478043, 0.08574605733156204, -0.022721191868185997, -0.0328889936208725, -0.010112699121236801, -0.1124335378408432, -0.14824442565441132, -0.1710662543773651, 0.06656615436077118, -0.08567062765359879, 0.004873726051300764, -0.1165873110294342, -0.06541376560926437, -0.025362538173794746, 0.1980317384004593, -0.17798730731010437, -0.09025123715400696, -0.14303405582904816, -0.04597308114171028, 0.18141955137252808, -0.039617933332920074, 0.07049045711755753, -0.011732147075235844, 0.17738661170005798, -0.006743612699210644, -0.004253563936799765, 0.05601169914007187, -0.09308900684118271, -0.18214112520217896, -0.05485027655959129, 0.1657821536064148, 0.11794015020132065, 0.047387149184942245, -0.013942261226475239, 0.005683347582817078, -0.030596597120165825, -0.10601680725812912, 0.003344012424349785, 0.1500520259141922, 0.019793936982750893, 0.0026186774484813213, -0.045496921986341476, -0.08338036388158798, -0.06850090622901917, -0.07240230590105057, 0.03065483458340168, 0.21129590272903442, -0.09349125623703003, 0.16456225514411926, 0.10457519441843033, -0.054971423000097275, -0.21910879015922546, 0.030726579949259758, 0.06544775515794754, 0.01346807461231947, 0.03579012304544449, -0.18912886083126068, 0.0927852913737297, 0.004406351130455732, -0.07468047738075256, 0.1653127670288086, -0.17572912573814392, -0.13922834396362305, 0.07237043231725693, 0.020974406972527504, -0.2560601830482483, -0.14496318995952606, -0.11670807003974915, -0.018667979165911674, -0.12923860549926758, 0.03696313127875328, 0.026861319318413734, 0.007220887579023838, 0.018991822376847267, 0.007314010988920927, 0.04252808168530464, -0.06217929348349571, 0.20018890500068665, -0.027767540886998177, 0.0020654473919421434, -0.04718940705060959, -0.05009689927101135, 0.05979317054152489, -0.05202948674559593, 0.11435818672180176, 0.013019593432545662, 0.02782965824007988, -0.15940485894680023, -0.04327741637825966, -0.0758608803153038, 0.03470107167959213, -0.08294288069009781, -0.08526648581027985, -0.04686691239476204, 0.09271422028541565, 0.0778745710849762, -0.0246481541544199, -0.012410464696586132, -0.08147142827510834, 0.09767493605613708, 0.18193857371807098, 0.17111952602863312, 0.030020372942090034, -0.06551752984523773, 0.01773168332874775, -0.03535405918955803, 0.030244817957282066, -0.24362754821777344, 0.033789556473493576, 0.0752231702208519, 0.03979611024260521, 0.08015667647123337, -0.022341350093483925, -0.17499181628227234, -0.05346069484949112, 0.09011489897966385, -0.07878093421459198, -0.19668085873126984, -0.033720847219228745, 0.08894652128219604, -0.21097739040851593, -0.04455452039837837, 0.04044025391340256, -0.025475606322288513, -0.03495745360851288, 0.0002803383977152407, 0.07365503907203674, -0.010918328538537025, 0.11844710260629654, 0.06544296443462372, 0.10499735921621323, -0.09676411002874374, 0.0753079429268837, 0.09104850888252258, -0.06076887995004654, 0.01594753935933113, 0.11562642455101013, -0.053347039967775345, -0.0377999022603035, 0.03426124528050423, 0.05029534548521042, 0.013709468767046928, -0.057701632380485535, 0.004143511410802603, -0.036752648651599884, 0.050188783556222916, 0.07984692603349686, 0.03514023870229721, -0.017107566818594933, 0.07992835342884064, 0.017902344465255737, -0.09303557127714157, 0.11883639544248581, 0.06038394197821617, 0.03439304232597351, -0.06516654044389725, -0.014919014647603035, -0.0023769158869981766, 0.023254480212926865, -0.014539581723511219, -0.003506371285766363, -0.04150436818599701, -0.003050457453355193, -0.1264333426952362, -0.0035722425673156977, -0.07507935166358948, 0.008237270638346672, 0.025249751284718513, -0.03914192318916321, -0.009739532135426998, 0.008962051942944527, -0.08525317162275314, -0.07094574719667435, -0.017503373324871063, 0.09471937268972397, -0.13447150588035583, 0.010915354825556278, 0.07579963654279709, -0.11530116200447083, 0.07679224014282227, -0.004124960396438837, 0.018290283158421516, 0.009569688700139523, -0.1204732358455658, 0.045480500906705856, -0.007774645928293467, 0.01593817211687565, 0.015467488206923008, -0.17224904894828796, 0.009592992253601551, -0.04124117270112038, -0.04261110723018646, 0.017079971730709076, -0.022616703063249588, -0.13178414106369019, 0.08175484836101532, -0.014359836466610432, -0.05001497641205788, -0.02968703769147396, 0.0517333522439003, 0.08315660804510117, -0.005793449003249407, 0.09429702907800674, -0.02946334145963192, 0.05589986965060234, -0.176262766122818, -0.01945015974342823, -0.045160870999097824, 0.0364910252392292, -0.0004979937220923603, -0.002335125580430031, 0.05757569149136543, -0.0060677314177155495, 0.19162507355213165, -0.03349338471889496, 0.14212258160114288, 0.05008721351623535, -0.026548068970441818, 0.028822913765907288, 0.059149887412786484, 0.05867528170347214, 0.012772006914019585, 0.010172291658818722, 0.03997969254851341, -0.01605338789522648, -0.040060047060251236, -0.14042896032333374, 0.01069095078855753, 0.17461243271827698, 0.06576123833656311, 0.02550606057047844, 0.0360528789460659, -0.15343499183654785, -0.07904533296823502, 0.10781536996364594, -0.025404470041394234, 0.005563603714108467, -0.08263549208641052, 0.18332551419734955, 0.11789415031671524, -0.1755414605140686, 0.06211428344249725, -0.06549183279275894, -0.04733308404684067, -0.10945986211299896, -0.11803947389125824, -0.058913398534059525, -0.05049397423863411, -0.0014363524969667196, -0.05056574568152428, 0.05735341086983681, 0.03462890535593033, -0.002890799893066287, -0.010038130916655064, 0.11978176236152649, -0.00945291668176651, -0.016113094985485077, 0.06505990028381348, 0.05032568424940109, 0.032801467925310135, -0.07488272339105606, 0.009758695028722286, 0.011927608400583267, 0.01582201011478901, 0.05904164910316467, 0.01736876182258129, -0.07141341269016266, 0.026859238743782043, 0.002883747685700655, -0.10977957397699356, 0.0299705620855093, -0.015784157440066338, -0.05185519903898239, 0.14840343594551086, 0.046683602035045624, 0.016483623534440994, -0.03217694163322449, 0.23987185955047607, -0.07395137846469879, -0.07934325188398361, -0.14789079129695892, 0.08898080885410309, -0.03652782738208771, 0.03998464345932007, 0.026356291025877, -0.12178415060043335, 0.007244357373565435, 0.1616872102022171, 0.11593696475028992, 0.0022911066189408302, 0.0001365400676149875, 0.06299346685409546, 0.001504482002928853, -0.054684437811374664, 0.03226175531744957, 0.06286465376615524, 0.16691628098487854, -0.08119276911020279, 0.06880643218755722, -0.01200462132692337, -0.07179238647222519, -0.02308073826134205, 0.1136866882443428, -0.02499707043170929, 0.013842890970408916, -0.055233459919691086, 0.12882515788078308, -0.050725776702165604, -0.24679803848266602, 0.044503647834062576, -0.09192120283842087, -0.1555570662021637, -0.0248231440782547, 0.019719766452908516, -0.027176709845662117, 0.011683927848935127, 0.05994521453976631, -0.0427580252289772, 0.19436821341514587, 0.024646252393722534, -0.06617524474859238, -0.08211323618888855, 0.051299892365932465, -0.1134541854262352, 0.289729505777359, 0.00900786742568016, 0.05383238568902016, 0.11205587536096573, -0.04153810068964958, -0.1747211068868637, 0.012666660360991955, 0.12733502686023712, -0.0861736387014389, 0.06638690084218979, 0.18149487674236298, 0.00223703240044415, 0.13205626606941223, 0.05677587166428566, -0.032920196652412415, 0.03465012088418007, -0.01818814128637314, -0.041404206305742264, -0.1314304918050766, 0.06819728761911392, -0.06559377908706665, 0.1528960019350052, 0.12065953016281128, -0.06554459035396576, 0.002187412930652499, -0.046505481004714966, 0.07286950200796127, 0.010224172845482826, 0.13140636682510376, 0.02352953515946865, -0.19986788928508759, 0.048462674021720886, -0.026924382895231247, 0.09115619212388992, -0.2355252206325531, -0.0703015998005867, 0.07476413249969482, -0.02091049589216709, -0.058929335325956345, 0.12400282174348831, 0.039426855742931366, 0.020744655281305313, -0.028888216242194176, -0.10013705492019653, -0.007854988798499107, 0.15314337611198425, -0.12880705296993256, -0.020703166723251343 ]
null
null
transformers
# OpenBuddy - Open Multilingual Chatbot GitHub and Usage Guide: [https://github.com/OpenBuddy/OpenBuddy](https://github.com/OpenBuddy/OpenBuddy) Website and Demo: [https://openbuddy.ai](https://openbuddy.ai) Evaluation result of this model: [Evaluation.txt](Evaluation.txt) ![Demo](https://raw.githubusercontent.com/OpenBuddy/OpenBuddy/main/media/demo.png) # Copyright Notice Base model: https://huggingface.co/mistralai/Mixtral-8x7B-v0.1 License: Apache 2.0 ## Disclaimer All OpenBuddy models have inherent limitations and may potentially produce outputs that are erroneous, harmful, offensive, or otherwise undesirable. Users should not use these models in critical or high-stakes situations that may lead to personal injury, property damage, or significant losses. Examples of such scenarios include, but are not limited to, the medical field, controlling software and hardware systems that may cause harm, and making important financial or legal decisions. OpenBuddy is provided "as-is" without any warranty of any kind, either express or implied, including, but not limited to, the implied warranties of merchantability, fitness for a particular purpose, and non-infringement. In no event shall the authors, contributors, or copyright holders be liable for any claim, damages, or other liabilities, whether in an action of contract, tort, or otherwise, arising from, out of, or in connection with the software or the use or other dealings in the software. By using OpenBuddy, you agree to these terms and conditions, and acknowledge that you understand the potential risks associated with its use. You also agree to indemnify and hold harmless the authors, contributors, and copyright holders from any claims, damages, or liabilities arising from your use of OpenBuddy. ## 免责声明 所有OpenBuddy模型均存在固有的局限性,可能产生错误的、有害的、冒犯性的或其他不良的输出。用户在关键或高风险场景中应谨慎行事,不要使用这些模型,以免导致人身伤害、财产损失或重大损失。此类场景的例子包括但不限于医疗领域、可能导致伤害的软硬件系统的控制以及进行重要的财务或法律决策。 OpenBuddy按“原样”提供,不附带任何种类的明示或暗示的保证,包括但不限于适销性、特定目的的适用性和非侵权的暗示保证。在任何情况下,作者、贡献者或版权所有者均不对因软件或使用或其他软件交易而产生的任何索赔、损害赔偿或其他责任(无论是合同、侵权还是其他原因)承担责任。 使用OpenBuddy即表示您同意这些条款和条件,并承认您了解其使用可能带来的潜在风险。您还同意赔偿并使作者、贡献者和版权所有者免受因您使用OpenBuddy而产生的任何索赔、损害赔偿或责任的影响。
{"language": ["zh", "en", "fr", "de", "ja", "ko", "it", "ru"], "license": "apache-2.0", "library_name": "transformers", "pipeline_tag": "text-generation", "inference": false}
text-generation
OpenBuddy/openbuddy-mixtral-7bx8-v18.1-32k-gptq
[ "transformers", "mixtral", "text-generation", "zh", "en", "fr", "de", "ja", "ko", "it", "ru", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "4-bit", "region:us" ]
2024-02-15T04:11:48+00:00
[]
[ "zh", "en", "fr", "de", "ja", "ko", "it", "ru" ]
TAGS #transformers #mixtral #text-generation #zh #en #fr #de #ja #ko #it #ru #license-apache-2.0 #autotrain_compatible #text-generation-inference #4-bit #region-us
# OpenBuddy - Open Multilingual Chatbot GitHub and Usage Guide: URL Website and Demo: URL Evaluation result of this model: URL !Demo # Copyright Notice Base model: URL License: Apache 2.0 ## Disclaimer All OpenBuddy models have inherent limitations and may potentially produce outputs that are erroneous, harmful, offensive, or otherwise undesirable. Users should not use these models in critical or high-stakes situations that may lead to personal injury, property damage, or significant losses. Examples of such scenarios include, but are not limited to, the medical field, controlling software and hardware systems that may cause harm, and making important financial or legal decisions. OpenBuddy is provided "as-is" without any warranty of any kind, either express or implied, including, but not limited to, the implied warranties of merchantability, fitness for a particular purpose, and non-infringement. In no event shall the authors, contributors, or copyright holders be liable for any claim, damages, or other liabilities, whether in an action of contract, tort, or otherwise, arising from, out of, or in connection with the software or the use or other dealings in the software. By using OpenBuddy, you agree to these terms and conditions, and acknowledge that you understand the potential risks associated with its use. You also agree to indemnify and hold harmless the authors, contributors, and copyright holders from any claims, damages, or liabilities arising from your use of OpenBuddy. ## 免责声明 所有OpenBuddy模型均存在固有的局限性,可能产生错误的、有害的、冒犯性的或其他不良的输出。用户在关键或高风险场景中应谨慎行事,不要使用这些模型,以免导致人身伤害、财产损失或重大损失。此类场景的例子包括但不限于医疗领域、可能导致伤害的软硬件系统的控制以及进行重要的财务或法律决策。 OpenBuddy按“原样”提供,不附带任何种类的明示或暗示的保证,包括但不限于适销性、特定目的的适用性和非侵权的暗示保证。在任何情况下,作者、贡献者或版权所有者均不对因软件或使用或其他软件交易而产生的任何索赔、损害赔偿或其他责任(无论是合同、侵权还是其他原因)承担责任。 使用OpenBuddy即表示您同意这些条款和条件,并承认您了解其使用可能带来的潜在风险。您还同意赔偿并使作者、贡献者和版权所有者免受因您使用OpenBuddy而产生的任何索赔、损害赔偿或责任的影响。
[ "# OpenBuddy - Open Multilingual Chatbot\n\nGitHub and Usage Guide: URL\n\nWebsite and Demo: URL\n\nEvaluation result of this model: URL\n\n!Demo", "# Copyright Notice\n\nBase model: URL\n\nLicense: Apache 2.0", "## Disclaimer\n\nAll OpenBuddy models have inherent limitations and may potentially produce outputs that are erroneous, harmful, offensive, or otherwise undesirable. Users should not use these models in critical or high-stakes situations that may lead to personal injury, property damage, or significant losses. Examples of such scenarios include, but are not limited to, the medical field, controlling software and hardware systems that may cause harm, and making important financial or legal decisions.\n\nOpenBuddy is provided \"as-is\" without any warranty of any kind, either express or implied, including, but not limited to, the implied warranties of merchantability, fitness for a particular purpose, and non-infringement. In no event shall the authors, contributors, or copyright holders be liable for any claim, damages, or other liabilities, whether in an action of contract, tort, or otherwise, arising from, out of, or in connection with the software or the use or other dealings in the software.\n\nBy using OpenBuddy, you agree to these terms and conditions, and acknowledge that you understand the potential risks associated with its use. You also agree to indemnify and hold harmless the authors, contributors, and copyright holders from any claims, damages, or liabilities arising from your use of OpenBuddy.", "## 免责声明\n\n所有OpenBuddy模型均存在固有的局限性,可能产生错误的、有害的、冒犯性的或其他不良的输出。用户在关键或高风险场景中应谨慎行事,不要使用这些模型,以免导致人身伤害、财产损失或重大损失。此类场景的例子包括但不限于医疗领域、可能导致伤害的软硬件系统的控制以及进行重要的财务或法律决策。\n\nOpenBuddy按“原样”提供,不附带任何种类的明示或暗示的保证,包括但不限于适销性、特定目的的适用性和非侵权的暗示保证。在任何情况下,作者、贡献者或版权所有者均不对因软件或使用或其他软件交易而产生的任何索赔、损害赔偿或其他责任(无论是合同、侵权还是其他原因)承担责任。\n\n使用OpenBuddy即表示您同意这些条款和条件,并承认您了解其使用可能带来的潜在风险。您还同意赔偿并使作者、贡献者和版权所有者免受因您使用OpenBuddy而产生的任何索赔、损害赔偿或责任的影响。" ]
[ "TAGS\n#transformers #mixtral #text-generation #zh #en #fr #de #ja #ko #it #ru #license-apache-2.0 #autotrain_compatible #text-generation-inference #4-bit #region-us \n", "# OpenBuddy - Open Multilingual Chatbot\n\nGitHub and Usage Guide: URL\n\nWebsite and Demo: URL\n\nEvaluation result of this model: URL\n\n!Demo", "# Copyright Notice\n\nBase model: URL\n\nLicense: Apache 2.0", "## Disclaimer\n\nAll OpenBuddy models have inherent limitations and may potentially produce outputs that are erroneous, harmful, offensive, or otherwise undesirable. Users should not use these models in critical or high-stakes situations that may lead to personal injury, property damage, or significant losses. Examples of such scenarios include, but are not limited to, the medical field, controlling software and hardware systems that may cause harm, and making important financial or legal decisions.\n\nOpenBuddy is provided \"as-is\" without any warranty of any kind, either express or implied, including, but not limited to, the implied warranties of merchantability, fitness for a particular purpose, and non-infringement. In no event shall the authors, contributors, or copyright holders be liable for any claim, damages, or other liabilities, whether in an action of contract, tort, or otherwise, arising from, out of, or in connection with the software or the use or other dealings in the software.\n\nBy using OpenBuddy, you agree to these terms and conditions, and acknowledge that you understand the potential risks associated with its use. You also agree to indemnify and hold harmless the authors, contributors, and copyright holders from any claims, damages, or liabilities arising from your use of OpenBuddy.", "## 免责声明\n\n所有OpenBuddy模型均存在固有的局限性,可能产生错误的、有害的、冒犯性的或其他不良的输出。用户在关键或高风险场景中应谨慎行事,不要使用这些模型,以免导致人身伤害、财产损失或重大损失。此类场景的例子包括但不限于医疗领域、可能导致伤害的软硬件系统的控制以及进行重要的财务或法律决策。\n\nOpenBuddy按“原样”提供,不附带任何种类的明示或暗示的保证,包括但不限于适销性、特定目的的适用性和非侵权的暗示保证。在任何情况下,作者、贡献者或版权所有者均不对因软件或使用或其他软件交易而产生的任何索赔、损害赔偿或其他责任(无论是合同、侵权还是其他原因)承担责任。\n\n使用OpenBuddy即表示您同意这些条款和条件,并承认您了解其使用可能带来的潜在风险。您还同意赔偿并使作者、贡献者和版权所有者免受因您使用OpenBuddy而产生的任何索赔、损害赔偿或责任的影响。" ]
[ 61, 35, 13, 298, 234 ]
[ "passage: TAGS\n#transformers #mixtral #text-generation #zh #en #fr #de #ja #ko #it #ru #license-apache-2.0 #autotrain_compatible #text-generation-inference #4-bit #region-us \n# OpenBuddy - Open Multilingual Chatbot\n\nGitHub and Usage Guide: URL\n\nWebsite and Demo: URL\n\nEvaluation result of this model: URL\n\n!Demo# Copyright Notice\n\nBase model: URL\n\nLicense: Apache 2.0## Disclaimer\n\nAll OpenBuddy models have inherent limitations and may potentially produce outputs that are erroneous, harmful, offensive, or otherwise undesirable. Users should not use these models in critical or high-stakes situations that may lead to personal injury, property damage, or significant losses. Examples of such scenarios include, but are not limited to, the medical field, controlling software and hardware systems that may cause harm, and making important financial or legal decisions.\n\nOpenBuddy is provided \"as-is\" without any warranty of any kind, either express or implied, including, but not limited to, the implied warranties of merchantability, fitness for a particular purpose, and non-infringement. In no event shall the authors, contributors, or copyright holders be liable for any claim, damages, or other liabilities, whether in an action of contract, tort, or otherwise, arising from, out of, or in connection with the software or the use or other dealings in the software.\n\nBy using OpenBuddy, you agree to these terms and conditions, and acknowledge that you understand the potential risks associated with its use. You also agree to indemnify and hold harmless the authors, contributors, and copyright holders from any claims, damages, or liabilities arising from your use of OpenBuddy." ]
[ -0.07418109476566315, -0.04980161413550377, -0.004624125547707081, 0.061057813465595245, 0.06181827932596207, -0.05924529582262039, 0.1366567760705948, 0.027607202529907227, 0.13392826914787292, -0.01644248329102993, -0.006487189792096615, -0.0020540524274110794, 0.00392371229827404, -0.05869925767183304, 0.051415231078863144, -0.0967768207192421, -0.007239678408950567, -0.06711859256029129, 0.0980808362364769, -0.025968940928578377, 0.03562907502055168, 0.02972705103456974, 0.04534416273236275, 0.04391982778906822, 0.015687109902501106, -0.05496076121926308, 0.0688905119895935, 0.02010035328567028, 0.04622150957584381, 0.10822289437055588, -0.03696875274181366, -0.022807443514466286, -0.06981246918439865, -0.1608252227306366, 0.011009802110493183, 0.01319806557148695, -0.07503140717744827, -0.009850974194705486, 0.02988087572157383, 0.07100608944892883, 0.10694334656000137, 0.008398705162107944, -0.06528940796852112, 0.1085844337940216, -0.1352757066488266, -0.16786375641822815, -0.06051664426922798, 0.0847771167755127, 0.0022419027518481016, 0.07913287729024887, -0.036455657333135605, 0.05156036466360092, -0.00383800663985312, 0.043901875615119934, 0.1522333025932312, -0.19568774104118347, 0.034852512180805206, 0.0001357087749056518, 0.07568014413118362, 0.00657168822363019, -0.002527544042095542, -0.02235471084713936, 0.03320715203881264, -0.001998464809730649, 0.055289268493652344, -0.055048711597919464, 0.13823823630809784, -0.04781551659107208, -0.12089873850345612, -0.018987253308296204, 0.25145024061203003, 0.014382942579686642, -0.16950583457946777, -0.06521493196487427, 0.03784746676683426, 0.15947920083999634, 0.0674358457326889, 0.1370052546262741, -0.0003679021610878408, 0.015567830763757229, 0.1386207789182663, -0.0037200856022536755, -0.06231437623500824, 0.006934352684766054, -0.030440159142017365, 0.05757828429341316, -0.003191314171999693, 0.05033855140209198, -0.0928230807185173, -0.03713357821106911, -0.17819932103157043, -0.05485370382666588, -0.04797658324241638, -0.055988896638154984, 0.01259115245193243, 0.005243470426648855, -0.1149483248591423, -0.14568503201007843, 0.06939654052257538, 0.05454372614622116, -0.027476608753204346, -0.0711946189403534, -0.10055501013994217, 0.09750180691480637, 0.09066742658615112, -0.07895451039075851, -0.004569596610963345, 0.020431946963071823, 0.047192223370075226, 0.07301265001296997, 0.13139498233795166, -0.03999042883515358, -0.04072008281946182, 0.043051764369010925, -0.07533476501703262, 0.00556925917044282, 0.1647632122039795, 0.03593636304140091, -0.054697245359420776, -0.029953118413686752, 0.21320852637290955, -0.06355839222669601, -0.0346192792057991, -0.004170243162661791, -0.043390922248363495, -0.14189444482326508, 0.08346634358167648, -0.004389283712953329, -0.01902053691446781, 0.10025357455015182, -0.08259601891040802, 0.011768151074647903, -0.07942984998226166, -0.11792372167110443, 0.12474750727415085, -0.026252759620547295, -0.01097868476063013, -0.08773896098136902, -0.1366867870092392, 0.006520524155348539, -0.005853171925991774, 0.016634374856948853, -0.039605312049388885, 0.11734388768672943, 0.0032298394944518805, -0.05939245596528053, -0.05333252251148224, -0.17571420967578888, -0.06709738820791245, 0.049167122691869736, 0.015660181641578674, 0.06266506761312485, -0.15184862911701202, 0.022770488634705544, -0.12193004041910172, 0.010190867818892002, -0.020498326048254967, -0.010445608757436275, -0.09856251627206802, 0.04117422178387642, 0.0053840079344809055, -0.0031698027160018682, -0.10295786708593369, 0.08049800246953964, -0.0333501435816288, 0.11347851902246475, -0.025502080097794533, -0.07163295149803162, 0.24244017899036407, -0.14686940610408783, -0.06805016845464706, 0.13776785135269165, 0.0023890568409115076, 0.053384654223918915, 0.060741182416677475, 0.0630112886428833, 0.08946661651134491, -0.06138588488101959, -0.14748573303222656, -0.03243406489491463, 0.020294303074479103, 0.03940883278846741, -0.00881250761449337, -0.08209308236837387, -0.03619726002216339, -0.0057451156899333, 0.08383102715015411, -0.04540128633379936, 0.0636458471417427, -0.03696823492646217, -0.012422258034348488, -0.04582681506872177, 0.048386186361312866, -0.010495894588530064, -0.04530099406838417, 0.053754813969135284, 0.006299788132309914, 0.0020782749634236097, -0.03424989804625511, 0.04644663259387016, 0.05232451856136322, -0.0415295735001564, -0.009046141058206558, -0.03683258220553398, 0.023781338706612587, -0.1108103096485138, -0.06367018818855286, 0.05827198177576065, -0.18869344890117645, 0.1341463178396225, 0.08009397983551025, 0.005761178210377693, 0.03366367518901825, -0.07436224818229675, 0.07901401072740555, -0.03913448378443718, -0.030753815546631813, -0.00833340547978878, -0.2335866540670395, 0.005238962359726429, -0.06864709407091141, 0.040153868496418, -0.08686961233615875, 0.023604383692145348, 0.16766519844532013, -0.09943991899490356, 0.05601596459746361, -0.045444369316101074, 0.08888228237628937, 0.06334749609231949, 0.04931029677391052, 0.0530567429959774, 0.03804049268364906, 0.023688659071922302, -0.1718536615371704, 0.14395150542259216, -0.19488245248794556, 0.058141060173511505, 0.0774015560746193, -0.04566377401351929, -0.02052244171500206, -0.11851757764816284, -0.005794107913970947, -0.05383869260549545, -0.061902064830064774, -0.07556839287281036, -0.04122835770249367, -0.023258676752448082, 0.02162071503698826, -0.0737483948469162, 0.0015115566784515977, 0.044244442135095596, -0.14811868965625763, 0.10960667580366135, 0.05013424530625343, -0.02657218836247921, -0.10286712646484375, 0.08170786499977112, 0.01502177119255066, -0.14061102271080017, 0.20509207248687744, 0.08038058131933212, -0.04724219813942909, -0.043948426842689514, 0.035010926425457, 0.11575260758399963, 0.20742256939411163, -0.07840777188539505, 0.048767052590847015, 0.023303546011447906, 0.04178416728973389, -0.03008182905614376, -0.07339059561491013, -0.011831158772110939, 0.0036275091115385294, -0.007866550236940384, -0.1669219732284546, 0.011970733292400837, -0.09239254891872406, 0.07752688974142075, -0.02729969657957554, 0.07905245572328568, 0.1049402505159378, -0.05440053343772888, -0.1438167840242386, 0.08507727831602097, -0.03247284144163132, -0.2024252563714981, -0.12465626746416092, -0.02337336353957653, -0.09941893815994263, 0.03768692538142204, 0.08701710402965546, -0.06729412823915482, -0.04597945511341095, -0.06646952778100967, 0.01261751726269722, 0.06807882338762283, -0.08301113545894623, -0.0789860412478447, 0.025991924107074738, 0.04608225077390671, -0.032339662313461304, -0.02876073680818081, -0.016575638204813004, -0.11058788001537323, 0.06514635682106018, 0.009593487717211246, 0.07675877958536148, 0.044148944318294525, 0.12371918559074402, -0.08101235330104828, -0.0525088906288147, 0.05198565870523453, -0.010408950969576836, 0.021138986572623253, 0.18042218685150146, -0.07427126914262772, 0.08867676556110382, 0.10235095769166946, 0.036305468529462814, -0.011854973621666431, 0.05756522715091705, 0.07664431631565094, -0.09784667193889618, -0.12709839642047882, -0.037976618856191635, -0.1488044112920761, 0.13841275870800018, 0.049979306757450104, -0.008638639003038406, 0.22185644507408142, 0.02869303897023201, -0.051851384341716766, 0.141130730509758, 0.08913265913724899, 0.07126843929290771, 0.14358113706111908, -0.005245802458375692, 0.12186874449253082, -0.07097281515598297, 0.08675575256347656, 0.12252140790224075, 0.006881299428641796, 0.3157191574573517, 0.036675482988357544, 0.08667963743209839, 0.1380007416009903, 0.07213103771209717, -0.039630141109228134, -0.013857945799827576, -0.046685006469488144, 0.0574796237051487, -0.05148368328809738, -0.11223393678665161, -0.08105818927288055, 0.09753304719924927, -0.0810467004776001, -0.07012484222650528, 0.0034446357749402523, -0.047253359109163284, 0.026070157065987587, 0.06947930157184601, 0.009570426307618618, 0.010705891996622086, -0.05754600092768669, 0.013425437733530998, -0.01581498421728611, -0.002341330051422119, 0.0021399136167019606, 0.04928557947278023, -0.06124427914619446, 0.09133405238389969, -0.025340357795357704, 0.04116138815879822, 0.006673333700746298, 0.048851389437913895, -0.0634298175573349, -0.018835753202438354, -0.0220958162099123, 0.03177763149142265, -0.31193453073501587, 0.123006172478199, 0.02288416586816311, -0.005501382984220982, -0.05029004439711571, 0.002834923332557082, 0.07259845733642578, 0.06269627064466476, 0.09398283809423447, 0.07066049426794052, 0.01311841793358326, 0.06276217103004456, 0.032850317656993866, 0.021475480869412422, -0.009426584467291832, 0.01909862458705902, 0.02992131933569908, 0.022168578580021858, 0.05137096345424652, -0.02470533549785614, 0.17793042957782745, -0.14313510060310364, -0.07638087868690491, 0.14307035505771637, 0.015627888962626457, -0.005437773186713457, -0.10169779509305954, 0.03078860603272915, 0.11321322619915009, -0.030034663155674934, -0.2104979157447815, -0.02820873260498047, -0.0788232758641243, -0.0684695765376091, 0.008483869023621082, -0.017228979617357254, 0.030612409114837646, -0.007915016263723373, -0.004584131296724081, -0.0979773998260498, -0.0513598695397377, 0.04459371790289879, -0.12929518520832062, -0.1657659411430359, -0.04617227241396904, -0.05867939069867134, 0.1319243460893631, 0.05915996432304382, -0.028331918641924858, 0.038408681750297546, -0.016021402552723885, -0.10762117803096771, 0.04328793287277222, 0.09434978663921356, -0.006092083174735308, 0.10178317874670029, -0.08603353053331375, -0.12413300573825836, -0.053202491253614426, -0.08664482086896896, -0.07005725055932999, 0.21980519592761993, -0.012686616741120815, 0.10935485363006592, 0.2134256511926651, -0.1537928283214569, -0.23345915973186493, 0.02741590142250061, -0.052456095814704895, -0.03970591351389885, 0.06809626519680023, -0.15757043659687042, 0.0406036339700222, 0.017585111781954765, -0.06574604660272598, 0.1288229078054428, -0.04008174687623978, -0.07285819947719574, 0.02093944326043129, -0.018160037696361542, 0.2389136701822281, -0.15554289519786835, -0.0619300901889801, -0.08034648001194, -0.08568120002746582, 0.12086877226829529, -0.1812334507703781, 0.05790247395634651, -0.037801362574100494, -0.03208008036017418, -0.016283370554447174, -0.04180987551808357, 0.05739772319793701, 0.029471036046743393, 0.018582701683044434, -0.06431187689304352, 0.09611277282238007, 0.006774294655770063, -0.05621742457151413, 0.1057647243142128, -0.10317206382751465, -0.06521593779325485, -0.040418390184640884, -0.05264688655734062, -0.13525652885437012, 0.12686724960803986, -0.028664061799645424, -0.06835220009088516, -0.005390319973230362, 0.047399938106536865, 0.0459449402987957, 0.043775103986263275, -0.05975399166345596, -0.18075598776340485, 0.06920576840639114, 0.1582186222076416, 0.21997614204883575, -0.020619289949536324, -0.13872887194156647, 0.00998999085277319, -0.024098331108689308, 0.13469378650188446, -0.13041608035564423, 0.05693965405225754, -0.012699039652943611, 0.02511218935251236, 0.10393298417329788, 0.05414566397666931, -0.12592436373233795, 0.0626566931605339, 0.08077388256788254, -0.04522235319018364, 0.0009639691561460495, 0.03041842207312584, 0.20874160528182983, -0.11501145362854004, 0.00039992661913856864, 0.11771851778030396, -0.09337276965379715, 0.055690664798021317, -0.02732587605714798, 0.11903540790081024, -0.03066784143447876, -0.023051509633660316, 0.07641148567199707, 0.010813368484377861, -0.02829350158572197, 0.08366657048463821, 0.00846355501562357, -0.0706908330321312, 0.08321208506822586, -0.037159666419029236, -0.010389323346316814, -0.09615524858236313, -0.24951082468032837, -0.09795977175235748, -0.1316874623298645, -0.13294246792793274, 0.008781665936112404, -0.05844582989811897, -0.044711630791425705, 0.22022683918476105, -0.029611598700284958, 0.07247480750083923, 0.06484562903642654, 0.022618578746914864, -0.004868035204708576, 0.04738796502351761, -0.011002162471413612, 0.05912589281797409, 0.05451793596148491, -0.07828985899686813, 0.02583235502243042, -0.026556123048067093, -0.011819611303508282, 0.04894376918673515, -0.12631644308567047, -0.04025302827358246, -0.06690144538879395, -0.0022667553275823593, -0.0753726065158844, -0.026353346183896065, -0.020082157105207443, 0.042825013399124146, 0.068115234375, 0.02741640992462635, -0.027601318433880806, -0.005857436452060938, 0.004809322766959667, 0.07493273168802261, -0.15022066235542297, -0.033295899629592896, 0.08622245490550995, -0.018461784347891808, 0.06773503124713898, -0.07005943357944489, -0.12345384061336517, -0.0038376259617507458, -0.04275698959827423, 0.14073307812213898, -0.018908416852355003, -0.004299268126487732, -0.04412074387073517, -0.08526333421468735, -0.0014111119089648128, -0.005423273891210556, 0.002735723042860627, 0.03814510256052017, 0.10953918099403381, -0.1093057245016098, 0.07904110103845596, 0.05606004223227501, -0.023146336898207664, -0.057191431522369385, 0.022220782935619354, -0.003337798872962594, 0.00664290739223361, 0.05196838825941086, -0.02960864081978798, -0.01609576866030693, -0.1143772155046463, 0.007690093945711851, 0.057403627783060074, -0.007990879938006401, -0.06937870383262634, 0.003005277132615447, 0.016123443841934204, 0.014142371714115143, 0.16970114409923553, -0.006400390528142452, -0.0890580341219902, 0.03894863277673721, 0.0024884880986064672, 0.038595933467149734, -0.026408065110445023, -0.1415511518716812, -0.05817504599690437, -0.036787085235118866, -0.12825775146484375, 0.04175449162721634, -0.08021388202905655, -0.12305156141519547, 0.1220916137099266, -0.032676130533218384, 0.06871409714221954, -0.01107716653496027, 0.03888903185725212, -0.06948452442884445, 0.01324330735951662, -0.01877330057322979, 0.07217757403850555, 0.11254408955574036, -0.022636251524090767, 0.009506125934422016, 0.16153036057949066, -0.010575051419436932, 0.1651282012462616, -0.000562532339245081, -0.030324093997478485, -0.09862273186445236, -0.19359321892261505, 0.04154494032263756, -0.02103710174560547, -0.02596301957964897, -0.08478369563817978, 0.018238961696624756, 0.09250804036855698, -0.05705609545111656, 0.00607908982783556, 0.0788954570889473, -0.10850897431373596, 0.0028502712957561016, 0.01430240273475647, -0.0716940239071846, 0.0008310525445267558, 0.10709694027900696, 0.0735383853316307, 0.0396435372531414, 0.05483138561248779, 0.03811139240860939, 0.11445999890565872, 0.007985454984009266, 0.060929857194423676, -0.006943611428141594, -0.03660716488957405, -0.003622384276241064, -0.009321728721261024, 0.05127030983567238, 0.26851654052734375, 0.026374712586402893, -0.00676342798396945, 0.034678131341934204, 0.09714943170547485, -0.046613775193691254, -0.0374881736934185, -0.1088499054312706, 0.1868797093629837, -0.07052936404943466, -0.05837058648467064, -0.10669995099306107, -0.0164048969745636, 0.18047980964183807, 0.12614989280700684, 0.07586205005645752, -0.004404494538903236, 0.009159811772406101, -0.03017982468008995, 0.0011120359413325787, 0.08444683998823166, 0.040284063667058945, -0.03563428297638893, 0.3986421525478363, -0.0588410310447216, 0.16911089420318604, 0.0353124774992466, 0.0049275183118879795, -0.020733194425702095, 0.004073148127645254, 0.02478972263634205, -0.00039369642036035657, -0.07620558887720108, 0.04663703590631485, -0.0949665755033493, -0.09691037237644196, 0.02063221111893654, 0.05981457978487015, 0.0850871354341507, -0.03114454448223114, 0.07802361249923706, -0.02567627653479576, 0.09933395683765411, -0.025179727002978325, -0.04798863083124161, 0.06690208613872528, -0.048040296882390976, -0.06555188447237015, 0.01623559556901455, 0.04932183027267456, 0.09293711930513382, 0.08636609464883804, 0.0228279959410429, 0.1313459873199463, 0.03792545199394226, 0.025557776913046837, -0.04017684608697891, -0.01564512960612774, -0.05698613449931145, -0.07369332760572433, -0.0866437777876854, 0.011159343644976616, 0.05383296310901642, 0.07953815907239914, 0.04839764162898064, 0.04039843752980232, 0.02117151953279972, 0.0836162194609642, -0.0025740195997059345, -0.031349923461675644, 0.11156202852725983, -0.09993873536586761, 0.11219081282615662, 0.10297225415706635, -0.048200201243162155, -0.10428034514188766, -0.089984230697155, -0.028893692418932915, 0.004340540152043104, -0.05789059028029442, -0.03876614198088646, 0.05764724686741829, 0.06387253850698471, 0.12380342930555344, -0.006358661223202944, -0.2289784699678421, -0.057300955057144165, 0.11871376633644104, -0.003059549257159233, -0.022565999999642372, -0.04381498321890831, 0.0785200446844101, -0.014468526467680931, -0.06513266265392303, -0.13557013869285583, 0.04907532036304474, -0.0009303480619564652, -0.04512062668800354, -0.09777077287435532 ]
null
null
transformers
# Conformer CTC 4M parameters WanDB https://wandb.ai/huseinzol05/malaysian-conformer-ctc-tiny?workspace=user-huseinzol05
{"library_name": "transformers", "tags": []}
feature-extraction
mesolitica/conformer-4M-ctc
[ "transformers", "safetensors", "conformer", "feature-extraction", "custom_code", "region:us" ]
2024-02-15T04:12:07+00:00
[]
[]
TAGS #transformers #safetensors #conformer #feature-extraction #custom_code #region-us
# Conformer CTC 4M parameters WanDB URL
[ "# Conformer CTC 4M parameters\n\nWanDB URL" ]
[ "TAGS\n#transformers #safetensors #conformer #feature-extraction #custom_code #region-us \n", "# Conformer CTC 4M parameters\n\nWanDB URL" ]
[ 28, 12 ]
[ "passage: TAGS\n#transformers #safetensors #conformer #feature-extraction #custom_code #region-us \n# Conformer CTC 4M parameters\n\nWanDB URL" ]
[ -0.1427372843027115, 0.04513607546687126, -0.004642769228667021, 0.07217517495155334, 0.09229668974876404, 0.011074746027588844, 0.10109858214855194, 0.008746662177145481, -0.07774346321821213, 0.035846810787916183, 0.05778399482369423, 0.0459146648645401, -0.020424244925379753, 0.05541076511144638, -0.07519222050905228, -0.06683486700057983, 0.09685275703668594, 0.04370015487074852, -0.1270257979631424, 0.05405038222670555, 0.0030018293764442205, -0.07991565018892288, 0.12232675403356552, -0.044617582112550735, -0.22357453405857086, 0.024420160800218582, 0.03386535868048668, -0.08398150652647018, 0.06325464695692062, 0.05249852314591408, 0.21413631737232208, 0.10450165718793869, 0.0193166583776474, -0.054326381534338, 0.02517286129295826, -0.025191783905029297, -0.08307074755430222, 0.06860325485467911, -0.005500677973031998, 0.03626254200935364, -0.12322790920734406, -0.00824671145528555, 0.004217495210468769, 0.0360722690820694, -0.15780143439769745, 0.01982186734676361, -0.060761261731386185, 0.13854378461837769, 0.05396223068237305, 0.09713558852672577, 0.05614131689071655, 0.16926375031471252, -0.12534715235233307, 0.07805851846933365, 0.12070558220148087, -0.30993565917015076, 0.04759582504630089, 0.1929762214422226, 0.10415427386760712, 0.09172890335321426, -0.06987512856721878, 0.0379447378218174, 0.06678111851215363, -0.001565297949127853, 0.0036572664976119995, -0.1104048490524292, -0.11387169361114502, 0.04926718398928642, -0.08610120415687561, -0.037491586059331894, 0.2742551863193512, -0.022356810048222542, -0.021524205803871155, 0.049142539501190186, -0.09372227638959885, -0.06440282613039017, -0.01952313631772995, -0.02293592505156994, -0.009986411780118942, 0.08234742283821106, -0.12289491295814514, 0.021953260526061058, -0.10369303077459335, -0.05151597410440445, -0.1752425730228424, 0.11822643876075745, -0.03816036507487297, 0.04208157956600189, -0.17850126326084137, 0.03571914881467819, 0.028926758095622063, -0.12896420061588287, 0.006518788635730743, -0.04015904292464256, 0.019598113372921944, -0.03756919130682945, -0.0636935904622078, -0.010398984886705875, 0.12748028337955475, 0.09417934715747833, -0.13804635405540466, -0.0034323229920119047, -0.05275207385420799, 0.05712629482150078, -0.03576164320111275, 0.10556240379810333, -0.06419865787029266, 0.01645144633948803, 0.09389443695545197, -0.04647936671972275, 0.04355630651116371, -0.057000983506441116, -0.13052822649478912, -0.04108208790421486, 0.08020424097776413, 0.12950654327869415, 0.013037700206041336, 0.0817120224237442, -0.014755033887922764, 0.005177478771656752, 0.35967743396759033, -0.09218656271696091, -0.017099233344197273, 0.066130131483078, 0.06375029683113098, -0.016060497611761093, 0.048499464988708496, -0.04436121881008148, 0.022225020453333855, -0.05966305732727051, -0.13360358774662018, -0.029433060437440872, 0.010034945793449879, -0.11448050290346146, 0.059617653489112854, -0.029580427333712578, 0.030173534527420998, -0.1703357845544815, -0.106136754155159, 0.03611139953136444, -0.05340747535228729, 0.06159202754497528, 0.07311708480119705, 0.04327019304037094, -0.1509479135274887, 0.004035844001919031, -0.011226058006286621, -0.1915334165096283, -0.044018931686878204, 0.0514838732779026, 0.07889343053102493, 0.0357540063560009, -0.10335297882556915, -0.01784498617053032, -0.10450895130634308, 0.07492590695619583, -0.16582752764225006, -0.009123560041189194, -0.03453560173511505, 0.16281776130199432, -0.0178445465862751, -0.00610730005428195, -0.05817010626196861, 0.012099728919565678, 0.0018848262261599302, 0.18862996995449066, -0.10624883323907852, -0.05154132843017578, 0.25364211201667786, -0.17338970303535461, -0.17845425009727478, 0.010289148427546024, 0.015630269423127174, -0.07752077281475067, 0.0745173990726471, 0.06209105625748634, 0.1312369406223297, -0.09133761376142502, -0.054617740213871, 0.08881839364767075, -0.10432182997465134, -0.12888342142105103, -0.019686538726091385, 0.005627439823001623, -0.016277845948934555, -0.020255934447050095, -0.010341141372919083, 0.05656106397509575, -0.016259878873825073, -0.07899759709835052, -0.03758937492966652, -0.026890037581324577, 0.07117737084627151, -0.047486599534749985, 0.023113666102290154, 0.005355152301490307, 0.019343243911862373, 0.04498790577054024, 0.045290689915418625, -0.031005918979644775, 0.006890911143273115, -0.14552313089370728, 0.0396643802523613, -0.2635485827922821, 0.01693720929324627, -0.15173305571079254, -0.2299828976392746, 0.01798146776854992, 0.01608954556286335, 0.025244928896427155, 0.08769936114549637, 0.12410391122102737, 0.041471678763628006, 0.014167341403663158, -0.09495404362678528, 0.008464861661195755, 0.08271771669387817, -0.06313265115022659, -0.1408270001411438, -0.004069814458489418, -0.08007775992155075, 0.019062576815485954, -0.13895142078399658, 0.033016547560691833, 0.20310048758983612, 0.016698207706212997, 0.08212989568710327, 0.04197833687067032, 0.011829418130218983, 0.046776723116636276, -0.0182135421782732, -0.045939210802316666, -0.0031734106596559286, 0.007733152713626623, -0.14312110841274261, 0.12310320883989334, -0.21990303695201874, 0.2712210416793823, 0.19784599542617798, -0.13350777328014374, -0.011876631528139114, 0.01958959735929966, 0.02262580581009388, -0.0047981832176446915, 0.10084933787584305, -0.061449114233255386, 0.08404754847288132, -0.022485654801130295, 0.09617369621992111, -0.025166291743516922, -0.01536547765135765, 0.0433766208589077, -0.021752558648586273, -0.035983312875032425, -0.0300101637840271, 0.058831751346588135, -0.05265061929821968, 0.13478556275367737, 0.237101748585701, 0.024214792996644974, 0.08974400162696838, -0.1004975289106369, -0.042993370443582535, -0.05365787446498871, 0.05181713029742241, 0.04018554463982582, 0.09925775229930878, -0.01857217401266098, -0.07255404442548752, 0.001877806382253766, 0.01791214942932129, 0.0664706900715828, -0.1473628729581833, -0.015703940764069557, 0.020552704110741615, -0.057060111314058304, -0.12637636065483093, -0.032887812703847885, -0.03793300315737724, 0.01753726415336132, 0.011941052041947842, -0.05048541724681854, 0.10025355219841003, 0.0014372107107192278, -0.10791942477226257, 0.16376112401485443, -0.11464612931013107, -0.14610901474952698, -0.18010030686855316, 0.011748621240258217, -0.05301102250814438, 0.03788812458515167, 0.005095914006233215, -0.14304909110069275, -0.059228960424661636, -0.04226810112595558, -0.05427432060241699, 0.014434099197387695, 0.03522719815373421, 0.05654473230242729, 0.010045097209513187, 0.026150353252887726, -0.14386571943759918, -0.018510326743125916, 0.010424409061670303, -0.0682930052280426, 0.07638430595397949, -0.07214105874300003, 0.1444730907678604, 0.1720530092716217, 0.06042395904660225, -0.012178424745798111, -0.000684104161337018, 0.18053823709487915, -0.031289830803871155, 0.01859402097761631, 0.18006481230258942, -0.101444311439991, -0.00010591000318527222, 0.14436310529708862, 0.07951290160417557, -0.09586960822343826, -0.029705198481678963, -0.09351546317338943, -0.08341128379106522, -0.16654656827449799, -0.08349020034074783, -0.09865295141935349, 0.0019866458605974913, -0.03419117629528046, 0.07136090844869614, 0.19160781800746918, 0.06957221031188965, 0.03744586929678917, 0.08214292675256729, 0.018824975937604904, 0.03714465722441673, 0.05384795740246773, 0.052798543125391006, 0.04684726893901825, -0.10206320136785507, -0.06930207461118698, 0.049408771097660065, 0.06248965486884117, 0.21360817551612854, 0.11451742798089981, 0.2502962052822113, 0.045709844678640366, 0.0771397054195404, 0.14402729272842407, 0.15664729475975037, 0.004631725139915943, -0.0688498318195343, 0.05900634825229645, -0.06283822655677795, 0.021953962743282318, -0.0018152757547795773, 0.03679768741130829, -0.03641742095351219, -0.054993174970149994, -0.07520788162946701, 0.08008789271116257, 0.008547931909561157, -0.003565267426893115, -0.30662402510643005, -0.07647966593503952, 0.04555410519242287, 0.09362652897834778, 0.03363034501671791, 0.009281526319682598, 0.08431538939476013, 0.00691085634753108, -0.008473931811749935, -0.018001966178417206, 0.07104077935218811, -0.060616329312324524, 0.047461558133363724, -0.10181589424610138, -0.022651072591543198, 0.00793094839900732, 0.0022498422767966986, -0.14046494662761688, 0.17940253019332886, 0.08955815434455872, 0.06369271874427795, -0.001497539458796382, 0.018214384093880653, 0.021862538531422615, 0.27318036556243896, 0.05672755464911461, 0.016094036400318146, -0.014557015150785446, -0.19233188033103943, -0.08144931495189667, 0.050407495349645615, 0.10478299111127853, 0.030591309070587158, 0.08106572926044464, 0.03983531892299652, -0.03198153153061867, 0.035821523517370224, 0.015488519333302975, -0.07654465734958649, -0.06941178441047668, 0.03613751754164696, 0.0392533577978611, 0.09526454657316208, -0.0714007169008255, 0.00966603122651577, -0.05884568393230438, 0.20373114943504333, -0.17492948472499847, -0.021777508780360222, -0.12402549386024475, -0.1281309276819229, 0.099222332239151, -0.09619861096143723, 0.043641746044158936, -0.007661853451281786, 0.010882521979510784, -0.02104446478188038, -0.16980355978012085, 0.16830326616764069, -0.10109856724739075, -0.058370448648929596, -0.016324127092957497, 0.1337047964334488, -0.036769356578588486, -0.00472205551341176, 0.04062488675117493, 0.013056925497949123, 0.037607770413160324, -0.12321112304925919, -0.019981177523732185, 0.026558203622698784, -0.03572249412536621, 0.01880963332951069, -0.039389632642269135, -0.04278332367539406, 0.035529181361198425, 0.09750527888536453, 0.13294745981693268, 0.19973696768283844, -0.027386682108044624, 0.02540445141494274, 0.2678680717945099, -0.0009454986429773271, -0.23656874895095825, 0.03817637264728546, -0.16168440878391266, -0.04375558719038963, -0.06894189864397049, -0.07541553676128387, 0.20331686735153198, 0.0019917122554033995, 0.018754223361611366, 0.11182902008295059, -0.1836129128932953, -0.06371820718050003, 0.05877278000116348, 0.06431465595960617, 0.2241411805152893, -0.15573881566524506, -0.09670469909906387, -0.020548289641737938, -0.3464869260787964, 0.03979082405567169, -0.22254382073879242, 0.03112863190472126, 0.03583790361881256, -0.023157775402069092, -0.02152201160788536, -0.08235008269548416, 0.12338586151599884, 0.038100823760032654, 0.12135393917560577, -0.04135484993457794, -0.1255311369895935, 0.2465224713087082, -0.017769109457731247, 0.08415413647890091, -0.017348356544971466, 0.04467754065990448, 0.01980472356081009, 0.005365049000829458, -0.020178096368908882, 0.05988346412777901, -0.02041669934988022, -0.05254369601607323, 0.01162340585142374, -0.06579146534204483, -0.03354072943329811, -0.014621268957853317, 0.09626572579145432, 0.0011865104315802455, 0.03520052880048752, 0.09442555159330368, 0.10420534014701843, -0.1402333676815033, -0.030010836198925972, 0.03642372786998749, -0.0848127156496048, 0.15625356137752533, -0.09630092978477478, 0.10876057296991348, 0.08284670859575272, -0.007859086617827415, 0.08463089913129807, 0.1209767758846283, -0.0011834949254989624, 0.016563856974244118, 0.11886561661958694, -0.13204902410507202, 0.0647469013929367, -0.008932303637266159, -0.2084447294473648, -0.01779884658753872, 0.16172564029693604, 0.18327417969703674, -0.01660905033349991, 0.03420480340719223, -0.003693325212225318, -0.02053639106452465, -0.10731342434883118, 0.15424633026123047, 0.09901493042707443, 0.044810667634010315, -0.14315266907215118, 0.05080365017056465, 0.010055127553641796, -0.17136530578136444, -0.010589436627924442, 0.0472247451543808, -0.1720457524061203, -0.0939786434173584, -0.044320933520793915, 0.2624102532863617, -0.11351758986711502, -0.0773373395204544, -0.11636805534362793, -0.09099936485290527, 0.007236781995743513, 0.2617148160934448, 0.0462527759373188, 0.12787748873233795, 0.02226579748094082, 0.029507257044315338, -0.0964764654636383, 0.00018056143017020077, -0.046052418649196625, 0.11794376373291016, -0.17056556046009064, 0.114120714366436, -0.0005614215624518692, 0.06744695454835892, -0.08460090309381485, 0.002689415356144309, -0.15440751612186432, -0.01008311565965414, -0.0938635990023613, 0.005111384205520153, -0.12029168754816055, 0.002787387929856777, -0.0009672632440924644, -0.014901294372975826, -0.0843382254242897, 0.05539583042263985, -0.06252411007881165, 0.00350226741284132, 0.008305652998387814, 0.01442527212202549, -0.14066068828105927, 0.004609532188624144, -0.002693577203899622, -0.06348160654306412, 0.07035202533006668, 0.08795937150716782, -0.06353055685758591, 0.09785318374633789, -0.1414588838815689, -0.05677427351474762, 0.14577221870422363, 0.00835101306438446, 0.07514922320842743, 0.0795934870839119, 0.014816563576459885, 0.050003811717033386, -0.039536647498607635, 0.027316538617014885, 0.14951635897159576, -0.045143187046051025, -0.06148600950837135, -0.11273930221796036, -0.04561573639512062, -0.07026577740907669, -0.017267880961298943, 0.2431534230709076, -0.03955760598182678, 0.1557898074388504, -0.03924289345741272, 0.008539064787328243, -0.08488274365663528, -0.05651457607746124, -0.010887578129768372, -0.10434772819280624, -0.03276374191045761, -0.03818291798233986, 0.037529025226831436, -0.09794355183839798, 0.2160789519548416, -0.16547904908657074, -0.012832923792302608, 0.007522789761424065, 0.08947014063596725, -0.015566343441605568, 0.04582411050796509, 0.2790205180644989, 0.1170409619808197, -0.0477994903922081, -0.1307041049003601, 0.12512145936489105, 0.0069697205908596516, -0.1820828765630722, 0.037371158599853516, 0.04060882702469826, -0.11284470558166504, 0.12426169216632843, 0.1270657181739807, -0.020349251106381416, -0.030478239059448242, -0.14905981719493866, -0.13644589483737946, 0.054402053356170654, 0.08749116957187653, -0.06555908173322678, 0.24765323102474213, 0.009534111246466637, -0.0018734335899353027, 0.055928923189640045, -0.05183260142803192, -0.053240299224853516, 0.012137389741837978, -0.09789642691612244, -0.08860363811254501, -0.010113469325006008, -0.07419192790985107, -0.11397459357976913, 0.09636498242616653, 0.011815027333796024, 0.013510238379240036, 0.29551076889038086, 0.07482835650444031, -0.0014900260139256716, -0.059353411197662354, -0.032900355756282806, -0.03380594775080681, 0.11796698719263077, 0.043212685734033585, 0.016307909041643143, -0.053309787064790726, -0.007688459474593401, 0.053166184574365616, 0.014563717879354954, 0.10633613914251328, -0.09021821618080139, -0.11341860145330429, -0.02394654043018818, 0.02516501396894455, -0.11601624637842178, 0.0669836774468422, 0.024491406977176666, -0.045702751725912094, 0.045439526438713074, 0.00086750223999843, -0.06822650879621506, -0.12149380892515182, -0.11801140755414963, 0.1233380138874054, 0.05140954256057739, 0.0658821165561676, -0.05111820623278618, -0.07385369390249252, 0.04062141478061676, 0.2019611895084381, 0.27160727977752686, -0.07286275923252106, 0.09292228519916534, 0.03235185518860817, 0.017805926501750946, -0.027115510776638985, 0.14435864984989166, 0.014217422343790531, 0.23988467454910278, -0.019950561225414276, -0.1134013906121254, -0.08967872709035873, -0.04043864831328392, -0.06464219838380814, 0.0170455202460289, 0.023502923548221588, -0.06347376108169556, -0.058210745453834534, 0.05911913886666298, -0.030833084136247635, 0.061832476407289505, 0.11672969162464142, -0.12579315900802612, 0.01665537804365158, -0.06491256505250931, 0.07625878602266312, 0.019158296287059784, 0.02393277734518051, -0.019888196140527725, -0.05813899263739586, 0.06206991523504257, 0.038534823805093765, -0.2065810263156891, 0.024072682484984398, 0.010883728042244911, -0.0580938495695591, 0.011876672506332397, -0.022828776389360428, 0.06201808154582977, 0.08171693980693817, 0.07849366962909698, -0.06165911257266998, 0.1737167239189148, 0.046086668968200684, -0.1496693193912506, -0.04955015331506729, -0.024662591516971588, -0.021566001698374748, 0.0038400916382670403, 0.03532373905181885, -0.1315256506204605, 0.01309667807072401, 0.06743374466896057, -0.03702154755592346, -0.07328688353300095, -0.10901179909706116, -0.02588040940463543, 0.07572784274816513, -0.0016570131992921233, 0.00979749858379364, 0.049037788063287735, -0.0013063226360827684, 0.05264310538768768, 0.09546288847923279, -0.13421060144901276, -0.13405603170394897, -0.13539153337478638, 0.011391771025955677, 0.036083854734897614, 0.056935880333185196, -0.07913918048143387, -0.027680398896336555, -0.08606743067502975, 0.01788732409477234, -0.10114913433790207, 0.0937652513384819, 0.18016375601291656, 0.055355288088321686, 0.0004947049892507493, 0.04254951700568199, 0.03354810178279877, 0.0675627738237381, -0.20455624163150787, -0.14794732630252838 ]
null
null
peft
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.8.2
{"library_name": "peft", "base_model": "ybelkada/blip2-opt-2.7b-fp16-sharded"}
null
leoreigoto/Data2_V2_Blip2_Finetune_Caption
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:ybelkada/blip2-opt-2.7b-fp16-sharded", "region:us" ]
2024-02-15T04:12:11+00:00
[ "1910.09700" ]
[]
TAGS #peft #safetensors #arxiv-1910.09700 #base_model-ybelkada/blip2-opt-2.7b-fp16-sharded #region-us
# Model Card for Model ID ## Model Details ### Model Description - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact ### Framework versions - PEFT 0.8.2
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ "TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-ybelkada/blip2-opt-2.7b-fp16-sharded #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ 49, 6, 3, 54, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4, 11 ]
[ "passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-ybelkada/blip2-opt-2.7b-fp16-sharded #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2" ]
[ -0.1023959144949913, 0.20364031195640564, -0.003501187078654766, 0.02728327363729477, 0.08585690706968307, 0.020007897168397903, 0.06716983020305634, 0.12274689972400665, 0.012622345238924026, 0.12875720858573914, 0.05166934058070183, 0.10228630155324936, 0.12245386838912964, 0.22668033838272095, -0.012838003225624561, -0.19243070483207703, 0.0242630448192358, -0.0771036297082901, 0.005593413487076759, 0.12082977592945099, 0.13930879533290863, -0.09762415289878845, 0.07061418145895004, -0.022893469780683517, -0.006428177002817392, -0.03144235908985138, -0.06823558360338211, -0.026020023971796036, 0.05179716274142265, 0.052624523639678955, 0.04507560655474663, -0.0026365057565271854, 0.08946135640144348, -0.2719646394252777, 0.01013258844614029, 0.05144821107387543, -0.0014026554999873042, 0.08499309420585632, 0.0988464504480362, -0.026928776875138283, 0.1098974272608757, -0.04064032807946205, 0.13379088044166565, 0.07711464166641235, -0.09239108860492706, -0.22410325706005096, -0.07173959165811539, 0.08625253289937973, 0.1827836036682129, 0.06788639724254608, -0.03555542975664139, 0.1263314187526703, -0.07607069611549377, 0.015353861264884472, 0.07337884604930878, -0.10034869611263275, -0.07253215461969376, 0.06338474154472351, 0.1148637905716896, 0.08918241411447525, -0.1183159276843071, -0.03819655627012253, 0.03314727544784546, 0.04002053290605545, 0.08421246707439423, 0.011725823394954205, 0.16789239645004272, 0.030479831621050835, -0.14271286129951477, -0.052292387932538986, 0.13737669587135315, 0.023319685831665993, -0.040452320128679276, -0.23331835865974426, -0.015376968309283257, -0.06795480847358704, -0.03502865135669708, -0.05600921809673309, 0.038739193230867386, 0.003568066982552409, 0.11532311141490936, -0.035693444311618805, -0.08089860528707504, -0.016660982742905617, 0.11346282064914703, 0.07784787565469742, 0.013616815209388733, -0.015320400707423687, 0.011008908972144127, 0.12671732902526855, 0.05690603330731392, -0.12490561604499817, -0.04582716152071953, -0.06711649894714355, -0.04033141955733299, -0.027365777641534805, 0.05342256277799606, 0.0353180356323719, 0.042800284922122955, 0.25008848309516907, -0.025695038959383965, 0.05624420940876007, 0.05131547898054123, 0.01635528728365898, 0.032757923007011414, 0.09937477856874466, -0.052993740886449814, -0.18724198639392853, -0.014669667929410934, 0.10533948987722397, 0.008535911329090595, -0.02457994781434536, -0.04416726529598236, 0.029510188847780228, 0.030378494411706924, 0.11640799790620804, 0.10281682014465332, -0.024227039888501167, -0.07068724185228348, -0.059072207659482956, 0.216552734375, -0.15430507063865662, 0.04941238462924957, 0.01976020447909832, -0.02105451002717018, -0.05080863833427429, 0.015585738234221935, 0.011093972250819206, -0.03752332180738449, 0.10955710709095001, -0.06118832156062126, -0.049624186009168625, -0.11145076900720596, -0.04487142711877823, 0.03322191536426544, 0.005177776329219341, -0.03966242074966431, -0.031259022653102875, -0.08961821347475052, -0.09290562570095062, 0.09356793016195297, -0.059249572455883026, -0.07570087909698486, -0.022671813145279884, -0.06882917881011963, 0.02134208008646965, 0.017937064170837402, 0.08182405680418015, -0.025928014889359474, 0.04417520761489868, -0.031341515481472015, 0.0675363764166832, 0.08862745016813278, 0.03666501119732857, -0.07080328464508057, 0.06790003925561905, -0.19235196709632874, 0.08260858803987503, -0.0774281769990921, 0.027750061824917793, -0.15746945142745972, -0.007029620930552483, 0.007698461879044771, 0.02178480476140976, 0.037930868566036224, 0.15137243270874023, -0.1939772069454193, -0.03221370279788971, 0.165089413523674, -0.10368044674396515, -0.11016760021448135, 0.04177917167544365, -0.038249846547842026, 0.15739089250564575, 0.030280139297246933, -0.0012426129542291164, 0.09249915927648544, -0.14316007494926453, -0.013938642106950283, -0.024325892329216003, 0.017603334039449692, 0.0778016746044159, 0.07073507457971573, -0.08130092173814774, 0.017243625596165657, 0.017793981358408928, -0.061591293662786484, -0.0081405621021986, -0.03970023989677429, -0.0987352728843689, 0.0057478612288832664, -0.08744806051254272, 0.016670476645231247, 0.007478628307580948, -0.0831889808177948, -0.01675930991768837, -0.14190812408924103, -0.03510471433401108, 0.08182565867900848, 0.010915788821876049, -0.017329296097159386, -0.06898674368858337, 0.034451693296432495, -0.03666619583964348, -0.01648077555000782, -0.14489136636257172, -0.021427473053336143, 0.03630919009447098, -0.15304622054100037, -0.005145850591361523, -0.11835219711065292, 0.06576516479253769, 0.01511449459940195, -0.06323514133691788, -0.04222502559423447, 0.017977474257349968, -0.004212197847664356, -0.057308491319417953, -0.21832139790058136, -0.03378254547715187, -0.04404837638139725, 0.1459498107433319, -0.22241605818271637, 0.04439277574419975, 0.011004806496202946, 0.12492083758115768, 0.009400355629622936, -0.06328117102384567, 0.025281473994255066, -0.060985516756772995, -0.022151879966259003, -0.07330615818500519, -0.010349185205996037, -0.0023612214718014, -0.035151150077581406, 0.025524193421006203, -0.14404910802841187, -0.04832933098077774, 0.0898718312382698, 0.08889646828174591, -0.1480346918106079, 0.004582213703542948, -0.0479704849421978, -0.06432840973138809, -0.08677776902914047, -0.0746101588010788, 0.07303491979837418, 0.050404179841279984, 0.05474574863910675, -0.08046972006559372, -0.06728234142065048, 0.006895649246871471, -0.013016865588724613, -0.026044650003314018, 0.12051089853048325, 0.08064275234937668, -0.08795077353715897, 0.09453634172677994, 0.07926716655492783, 0.048489537090063095, 0.09307026863098145, -0.012605591677129269, -0.10079944878816605, -0.029974179342389107, 0.05817700922489166, 0.013242031447589397, 0.15151557326316833, -0.05331005901098251, 0.04880187660455704, 0.05229213833808899, -0.046881623566150665, 0.0392768569290638, -0.1003534123301506, 0.01598189026117325, 0.010466392152011395, -0.014099279418587685, 0.04221281781792641, -0.022992460057139397, 0.013768854551017284, 0.08918881416320801, 0.06303547322750092, 0.028075722977519035, 0.02467573806643486, -0.03831007331609726, -0.13685983419418335, 0.16914397478103638, -0.09373056143522263, -0.2350214421749115, -0.1508769690990448, 0.033380694687366486, 0.05175327509641647, -0.02225610613822937, 0.025787750259041786, -0.03416982665657997, -0.10775268822908401, -0.08076061308383942, 0.017707832157611847, 0.04551579803228378, -0.06924950331449509, -0.05629436671733856, 0.03717868775129318, 0.03978053480386734, -0.12887011468410492, 0.03030010685324669, 0.05566076934337616, 0.0030389325693249702, -0.0023474569898098707, 0.04472624510526657, 0.0838165283203125, 0.19098424911499023, 0.0038206656463444233, 0.0005792943411506712, 0.052969906479120255, 0.2804574966430664, -0.15404251217842102, 0.1257592886686325, 0.1265062391757965, -0.05037635564804077, 0.0915827602148056, 0.20057283341884613, 0.04166555032134056, -0.08155781030654907, 0.024662816897034645, 0.03081902116537094, -0.033339425921440125, -0.25858598947525024, -0.07089923322200775, -0.02134774811565876, -0.0718344897031784, 0.08356352150440216, 0.08594240248203278, 0.09141265600919724, 0.034850113093853, -0.07882706820964813, -0.06002916395664215, 0.048778217285871506, 0.10919934511184692, -0.031268373131752014, 0.01373243611305952, 0.08169406652450562, -0.03843279555439949, 0.00291964178904891, 0.0957426205277443, -0.016873784363269806, 0.1495479941368103, 0.043862324208021164, 0.10726974159479141, 0.06192391365766525, 0.07807627320289612, -0.00427240040153265, 0.05275103449821472, 0.014167661778628826, 0.027501339092850685, 0.009811957366764545, -0.0911397933959961, 0.027544016018509865, 0.12415897846221924, 0.019679436460137367, 0.03910135477781296, 0.02221125178039074, -0.0599311999976635, 0.03939114511013031, 0.2089223861694336, 0.011348971165716648, -0.19348034262657166, -0.07075284421443939, 0.06647099554538727, -0.08268588781356812, -0.14843600988388062, -0.007554139941930771, 0.02860885113477707, -0.1735040545463562, 0.021763663738965988, -0.04100511968135834, 0.10741151124238968, -0.07009892165660858, -0.04010515660047531, 0.09469404071569443, 0.06040317192673683, -0.030702605843544006, 0.04793144762516022, -0.17065903544425964, 0.1120685413479805, 0.030907707288861275, 0.06870585680007935, -0.09663672745227814, 0.10015694797039032, 0.004520023241639137, -0.019680628553032875, 0.16754156351089478, 0.004071482457220554, -0.046421702951192856, -0.07688678056001663, -0.068022720515728, -0.02427726984024048, 0.09113579243421555, -0.13575509190559387, 0.0659659206867218, -0.018054954707622528, -0.038992740213871, 0.0012381378328427672, -0.10723204910755157, -0.11051654070615768, -0.17094869911670685, 0.06820432096719742, -0.07434363663196564, 0.0032095236238092184, -0.09686926752328873, -0.055401481688022614, -0.009374987334012985, 0.1763092428445816, -0.17980854213237762, -0.11631330102682114, -0.1491946429014206, -0.10259010642766953, 0.17095473408699036, -0.04734873026609421, 0.08819764107465744, -0.0034766437020152807, 0.16787858307361603, -0.014214687049388885, -0.015718974173069, 0.087732695043087, -0.09375686198472977, -0.1954045295715332, -0.054642241448163986, 0.1798044741153717, 0.12645776569843292, 0.0340544693171978, -0.022557005286216736, 0.021083654835820198, -0.045462194830179214, -0.1131872907280922, 0.01341725792735815, 0.14102241396903992, 0.044991254806518555, 0.0011787780094891787, -0.026516659185290337, -0.11984778940677643, -0.05476224422454834, -0.050618626177310944, -0.007076145615428686, 0.2069881707429886, -0.08295559138059616, 0.16145899891853333, 0.12309258431196213, -0.048618387430906296, -0.20653872191905975, 0.0328410379588604, 0.04221039637923241, 0.015131350606679916, 0.03412385284900665, -0.18393175303936005, 0.08414941281080246, -0.011607940308749676, -0.07911363989114761, 0.1776852309703827, -0.18600867688655853, -0.13374063372612, 0.0897565707564354, 0.0228035319596529, -0.2260301113128662, -0.13407503068447113, -0.1137874573469162, -0.01737418957054615, -0.12759308516979218, 0.04896645247936249, 0.029215604066848755, 0.0038563767448067665, 0.017197856679558754, 0.017441323027014732, 0.04046647995710373, -0.05223904177546501, 0.1984073370695114, -0.026710696518421173, 0.006047636270523071, -0.050715576857328415, -0.09581901878118515, 0.02039787359535694, -0.05753452330827713, 0.110266774892807, -0.01032858807593584, 0.023703893646597862, -0.16006609797477722, -0.04355010390281677, -0.06762884557247162, 0.020070020109415054, -0.09522926807403564, -0.0928168073296547, -0.05127653107047081, 0.0819212943315506, 0.10820676386356354, -0.02295168861746788, 0.0005749896517954767, -0.07361534982919693, 0.07722114026546478, 0.22184693813323975, 0.1649458259344101, 0.048072416335344315, -0.054855767637491226, 0.010806005448102951, -0.03211841732263565, 0.0406072698533535, -0.21899348497390747, 0.0427355132997036, 0.06035930663347244, 0.036828190088272095, 0.08413819968700409, -0.012402283027768135, -0.1615157127380371, -0.07863713055849075, 0.07407596707344055, -0.06613833457231522, -0.16754086315631866, -0.03876882418990135, 0.04338647425174713, -0.19467578828334808, -0.0508623942732811, 0.030222972854971886, -0.021225454285740852, -0.03482480347156525, 0.015321789309382439, 0.08263038098812103, -0.006926113273948431, 0.10264649987220764, 0.07923399657011032, 0.09892527759075165, -0.105184406042099, 0.06824152916669846, 0.08597805351018906, -0.03393268212676048, 0.009138351306319237, 0.13618622720241547, -0.05014333128929138, -0.023080265149474144, 0.06525996327400208, 0.08105384558439255, 0.012546095997095108, -0.052424877882003784, 0.013543504290282726, -0.0725960060954094, 0.06149521470069885, 0.10263664275407791, 0.019078141078352928, -0.016350749880075455, 0.06600207090377808, 0.02000228501856327, -0.0930938571691513, 0.1252557784318924, 0.06951259076595306, 0.022252975031733513, -0.038538623601198196, -0.028527628630399704, -0.013999884948134422, -0.007938371039927006, -0.014650342985987663, -0.002483789110556245, -0.07650505006313324, -0.004464221652597189, -0.1101321205496788, 0.01693287119269371, -0.07970034331083298, 0.006229866296052933, 0.013589464128017426, -0.042377084493637085, -0.0007004099315963686, -0.0010941700311377645, -0.0825127363204956, -0.060410551726818085, -0.022047949954867363, 0.08035755902528763, -0.12847869098186493, 0.01676725037395954, 0.07004719227552414, -0.11456523090600967, 0.06635120511054993, -0.01037197932600975, 0.012879660353064537, 0.0023543727584183216, -0.13481667637825012, 0.05165598541498184, -0.02099967747926712, -0.0035665908362716436, 0.02452498860657215, -0.16873589158058167, -0.0032408160623162985, -0.05216484144330025, -0.0749099999666214, 0.0048598493449389935, -0.04050567373633385, -0.13309027254581451, 0.11099666357040405, -0.012723537161946297, -0.07056688517332077, -0.02372054010629654, 0.05077226459980011, 0.08550150692462921, -0.02510816790163517, 0.0938984677195549, -0.025579016655683517, 0.08201071619987488, -0.17812030017375946, -0.009961896575987339, -0.01465003564953804, 0.03416622057557106, -0.018006669357419014, -0.014842072501778603, 0.05195453017950058, -0.012286731973290443, 0.17104652523994446, -0.017782554030418396, 0.07721024751663208, 0.04898650571703911, -0.002305001951754093, 0.02977832779288292, 0.06967566907405853, 0.06522846221923828, -0.014649573713541031, -0.005505430977791548, 0.026262225583195686, -0.013115668669342995, -0.043591879308223724, -0.1496414840221405, 0.03997490927577019, 0.16483180224895477, 0.06666601449251175, 0.031087568029761314, 0.02089330367743969, -0.137874573469162, -0.08431115001440048, 0.10268014669418335, -0.016490649431943893, -0.017492152750492096, -0.07017472386360168, 0.1938580423593521, 0.12399069219827652, -0.19902944564819336, 0.07030948996543884, -0.047140613198280334, -0.03475939482450485, -0.12678919732570648, -0.1504630148410797, -0.057876620441675186, -0.0396706759929657, -0.023383179679512978, -0.061611108481884, 0.05963646247982979, 0.045422691851854324, -0.001120392931625247, -0.007043042685836554, 0.10047915577888489, 0.002273647114634514, -0.025097914040088654, 0.06634964048862457, 0.07210102677345276, 0.044173598289489746, -0.08350478112697601, 0.005473924335092306, -0.004407464060932398, 0.005279168486595154, 0.05772916600108147, 0.019375650212168694, -0.059642791748046875, 0.022227197885513306, -0.001713956706225872, -0.11705102026462555, 0.04011988639831543, -0.015719687566161156, -0.03440998122096062, 0.1458049863576889, 0.023999176919460297, 0.010189180262386799, -0.02722814306616783, 0.2215377539396286, -0.08876711875200272, -0.07316280901432037, -0.13400240242481232, 0.07494153827428818, -0.05017445236444473, 0.035624027252197266, 0.03759501129388809, -0.12485615164041519, 0.006820192094892263, 0.1633101850748062, 0.1287752240896225, 0.0015842759748920798, 0.009181773290038109, 0.05443078279495239, 0.005943849217146635, -0.03531435877084732, 0.022326532751321793, 0.04972163215279579, 0.18732118606567383, -0.07171057164669037, 0.07961983233690262, -0.014104119502007961, -0.07633235305547714, -0.02722921036183834, 0.1351630687713623, -0.0060691554099321365, -0.0005658621084876359, -0.061430297791957855, 0.13521265983581543, -0.05378999561071396, -0.22143587470054626, 0.061226099729537964, -0.08844279497861862, -0.14051449298858643, -0.03979843854904175, 0.0020295942667871714, -0.018312251195311546, 0.01529828179627657, 0.06830412894487381, -0.054241351783275604, 0.1824820339679718, 0.032676003873348236, -0.06139485538005829, -0.08912920951843262, 0.050898075103759766, -0.14380158483982086, 0.2871195673942566, 0.026256408542394638, 0.03674665838479996, 0.10091640055179596, -0.02551521733403206, -0.1445848047733307, 0.022934162989258766, 0.11640027910470963, -0.07552778720855713, 0.05333654582500458, 0.17266467213630676, -0.007869631983339787, 0.13195273280143738, 0.05329576134681702, -0.06672891974449158, 0.034466732293367386, -0.05582581087946892, -0.06128007173538208, -0.12193675339221954, 0.07341577857732773, -0.07046602666378021, 0.14958375692367554, 0.12877047061920166, -0.06340276449918747, -0.009213985875248909, -0.019985152408480644, 0.0771741271018982, 0.016045335680246353, 0.13332360982894897, 0.02260434255003929, -0.18001601099967957, 0.04458257183432579, -0.005987263284623623, 0.11051776260137558, -0.20083686709403992, -0.06018247827887535, 0.03471638634800911, -0.02889879234135151, -0.0835249274969101, 0.11544446647167206, 0.0454484261572361, 0.020198026672005653, -0.029829418286681175, -0.07667473703622818, -0.0012824061559513211, 0.15097537636756897, -0.10192548483610153, -0.005492199212312698 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-small-govReport-boardpapers-3072 This model is a fine-tuned version of [RMWeerasinghe/t5-small-finetuned-govReport-3072](https://huggingface.co/RMWeerasinghe/t5-small-finetuned-govReport-3072) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 3.6701 - Rouge1: 0.0443 - Rouge2: 0.0194 - Rougel: 0.0382 - Rougelsum: 0.0443 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 4e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 8 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 40 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:| | No log | 0.67 | 1 | 3.9496 | 0.0584 | 0.0214 | 0.0482 | 0.0572 | | No log | 2.0 | 3 | 3.9252 | 0.0562 | 0.0223 | 0.0463 | 0.0562 | | No log | 2.67 | 4 | 3.9121 | 0.0597 | 0.0223 | 0.0485 | 0.0596 | | No log | 4.0 | 6 | 3.8880 | 0.0597 | 0.0223 | 0.0485 | 0.0596 | | No log | 4.67 | 7 | 3.8755 | 0.0597 | 0.0223 | 0.0485 | 0.0596 | | No log | 6.0 | 9 | 3.8506 | 0.0597 | 0.0223 | 0.0485 | 0.0596 | | No log | 6.67 | 10 | 3.8395 | 0.0553 | 0.0197 | 0.0441 | 0.0541 | | No log | 8.0 | 12 | 3.8172 | 0.0582 | 0.0262 | 0.049 | 0.057 | | No log | 8.67 | 13 | 3.8065 | 0.0582 | 0.0262 | 0.049 | 0.057 | | No log | 10.0 | 15 | 3.7862 | 0.0582 | 0.0257 | 0.049 | 0.057 | | No log | 10.67 | 16 | 3.7769 | 0.057 | 0.0262 | 0.049 | 0.0556 | | No log | 12.0 | 18 | 3.7599 | 0.0577 | 0.0294 | 0.0495 | 0.0575 | | No log | 12.67 | 19 | 3.7522 | 0.0487 | 0.0174 | 0.042 | 0.0474 | | 4.3528 | 14.0 | 21 | 3.7378 | 0.048 | 0.0155 | 0.0406 | 0.0461 | | 4.3528 | 14.67 | 22 | 3.7310 | 0.0536 | 0.0206 | 0.0421 | 0.0511 | | 4.3528 | 16.0 | 24 | 3.7187 | 0.048 | 0.017 | 0.0394 | 0.0448 | | 4.3528 | 16.67 | 25 | 3.7132 | 0.043 | 0.017 | 0.0374 | 0.041 | | 4.3528 | 18.0 | 27 | 3.7031 | 0.043 | 0.017 | 0.0374 | 0.041 | | 4.3528 | 18.67 | 28 | 3.6985 | 0.043 | 0.017 | 0.0374 | 0.041 | | 4.3528 | 20.0 | 30 | 3.6905 | 0.043 | 0.017 | 0.0374 | 0.041 | | 4.3528 | 20.67 | 31 | 3.6869 | 0.043 | 0.017 | 0.0374 | 0.041 | | 4.3528 | 22.0 | 33 | 3.6807 | 0.0442 | 0.0194 | 0.0381 | 0.0423 | | 4.3528 | 22.67 | 34 | 3.6781 | 0.0442 | 0.0194 | 0.0381 | 0.0423 | | 4.3528 | 24.0 | 36 | 3.6740 | 0.0442 | 0.0194 | 0.0381 | 0.0423 | | 4.3528 | 24.67 | 37 | 3.6725 | 0.0442 | 0.0194 | 0.0381 | 0.0423 | | 4.3528 | 26.0 | 39 | 3.6705 | 0.0443 | 0.0194 | 0.0382 | 0.0443 | | 4.0602 | 26.67 | 40 | 3.6701 | 0.0443 | 0.0194 | 0.0382 | 0.0443 | ### Framework versions - Transformers 4.37.0 - Pytorch 2.1.2 - Datasets 2.17.0 - Tokenizers 0.15.1
{"license": "apache-2.0", "tags": ["Summarization", "generated_from_trainer"], "metrics": ["rouge"], "base_model": "RMWeerasinghe/t5-small-finetuned-govReport-3072", "pipeline_tag": "summarization", "model-index": [{"name": "t5-small-govReport-boardpapers-3072", "results": []}]}
summarization
RMWeerasinghe/t5-small-govReport-boardpapers-3072
[ "transformers", "safetensors", "t5", "text2text-generation", "Summarization", "generated_from_trainer", "summarization", "base_model:RMWeerasinghe/t5-small-finetuned-govReport-3072", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T04:15:21+00:00
[]
[]
TAGS #transformers #safetensors #t5 #text2text-generation #Summarization #generated_from_trainer #summarization #base_model-RMWeerasinghe/t5-small-finetuned-govReport-3072 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
t5-small-govReport-boardpapers-3072 =================================== This model is a fine-tuned version of RMWeerasinghe/t5-small-finetuned-govReport-3072 on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 3.6701 * Rouge1: 0.0443 * Rouge2: 0.0194 * Rougel: 0.0382 * Rougelsum: 0.0443 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 4e-05 * train\_batch\_size: 4 * eval\_batch\_size: 4 * seed: 42 * gradient\_accumulation\_steps: 8 * total\_train\_batch\_size: 32 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 40 ### Training results ### Framework versions * Transformers 4.37.0 * Pytorch 2.1.2 * Datasets 2.17.0 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 4e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* gradient\\_accumulation\\_steps: 8\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #safetensors #t5 #text2text-generation #Summarization #generated_from_trainer #summarization #base_model-RMWeerasinghe/t5-small-finetuned-govReport-3072 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 4e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* gradient\\_accumulation\\_steps: 8\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ 97, 126, 4, 30 ]
[ "passage: TAGS\n#transformers #safetensors #t5 #text2text-generation #Summarization #generated_from_trainer #summarization #base_model-RMWeerasinghe/t5-small-finetuned-govReport-3072 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 4e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* gradient\\_accumulation\\_steps: 8\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 40### Training results### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ -0.11094363778829575, 0.126925989985466, -0.003296265611425042, 0.08061610907316208, 0.12915977835655212, 0.014369962736964226, 0.1098797544836998, 0.1265239715576172, -0.12653425335884094, 0.08016200363636017, 0.12412302941083908, 0.09719577431678772, 0.03894544765353203, 0.1672595739364624, -0.046863000839948654, -0.2963344156742096, 0.010745645500719547, -0.012004783377051353, -0.12322009354829788, 0.12036848068237305, 0.10379541665315628, -0.1161709576845169, 0.07838846743106842, -0.005707672797143459, -0.13472828269004822, 0.0011428375728428364, -0.035209644585847855, -0.056840527802705765, 0.11037231236696243, 0.04087295010685921, 0.0810919851064682, 0.024648133665323257, 0.08452191948890686, -0.21467086672782898, 0.008458524011075497, 0.05257011204957962, 0.013690588995814323, 0.08112911880016327, 0.09007872641086578, -0.01330413669347763, 0.161099374294281, -0.10143350064754486, 0.04374498873949051, 0.036423686891794205, -0.1120074912905693, -0.22435860335826874, -0.08360203355550766, 0.056642770767211914, 0.13487349450588226, 0.08916257321834564, -0.035663969814777374, 0.10940588265657425, -0.080266572535038, 0.09964702278375626, 0.19675515592098236, -0.2791978418827057, -0.0741586685180664, 0.05773759260773659, 0.026828397065401077, 0.054750021547079086, -0.11021130532026291, -0.021851520985364914, 0.05572344362735748, 0.01353214681148529, 0.09746967256069183, 0.0006909138755872846, 0.0507393479347229, 0.012340093962848186, -0.15061964094638824, -0.04020269215106964, 0.19369125366210938, 0.0848793089389801, -0.028868479654192924, -0.1021033301949501, -0.05159677937626839, -0.14176702499389648, -0.0215078704059124, -0.003073803149163723, 0.034326035529375076, -0.03885834664106369, -0.0946861207485199, 0.029354043304920197, -0.07665271311998367, -0.05369803309440613, 0.02286330610513687, 0.08704067766666412, 0.05570143833756447, -0.029076475650072098, -0.008167915046215057, 0.1149168461561203, 0.029046494513750076, -0.15730667114257812, -0.0028190352022647858, 0.004125047009438276, -0.025595374405384064, -0.043039437383413315, -0.030800335109233856, -0.008885894902050495, 0.04172630235552788, 0.16013391315937042, -0.05640287324786186, 0.06299621611833572, 0.04685939848423004, 0.01963680423796177, -0.07346319407224655, 0.1406739503145218, -0.07141628116369247, -0.0988735556602478, -0.01051462534815073, 0.10304790735244751, 0.017911069095134735, 0.01190855074673891, -0.07598844915628433, 0.03520315885543823, 0.10322971642017365, 0.050660714507102966, -0.03842741623520851, 0.06160563603043556, -0.047023359686136246, -0.021845728158950806, 0.013110785745084286, -0.09442728757858276, 0.006299424916505814, 0.009106476791203022, -0.08334405720233917, -0.04985353350639343, -0.002120057586580515, 0.010421287268400192, -0.01390545628964901, 0.11333790421485901, -0.0893392488360405, -0.03078850358724594, -0.07899291068315506, -0.09187135100364685, 0.00952528603374958, -0.028003567829728127, 0.020278412848711014, -0.08828995376825333, -0.1789824515581131, -0.039823565632104874, 0.0432543084025383, -0.04886367917060852, -0.06620736420154572, -0.07280570268630981, -0.09409712255001068, 0.023076705634593964, 0.00012055298429913819, 0.11184193938970566, -0.06096106767654419, 0.0902055874466896, 0.037679754197597504, 0.04745767265558243, 0.06008418649435043, 0.03406035527586937, -0.09070831537246704, 0.04872312769293785, -0.15433476865291595, 0.0722881481051445, -0.058976780623197556, 0.0887189581990242, -0.111808180809021, -0.11430524289608002, -0.009506922215223312, -0.008824539370834827, 0.07459466904401779, 0.1461978405714035, -0.14812007546424866, -0.07173272967338562, 0.19242148101329803, -0.07920870929956436, -0.14457492530345917, 0.12508298456668854, -0.03666674345731735, 0.020290439948439598, 0.04774339869618416, 0.17349828779697418, 0.08176680654287338, -0.060823824256658554, -0.0014157393015921116, -0.03336619958281517, 0.0947558730840683, 0.033001016825437546, 0.10003075003623962, -0.017455244436860085, 0.0005714459693990648, 0.017343243584036827, -0.04744768142700195, 0.05417043715715408, -0.10438505560159683, -0.07526754587888718, -0.01715085655450821, -0.08435119688510895, 0.01563655398786068, 0.045151036232709885, 0.06849115341901779, -0.1075773760676384, -0.1069912239909172, 0.0025490073021501303, 0.11209650337696075, -0.07471293956041336, -0.003696161089465022, -0.06050385534763336, 0.08986963331699371, -0.025593824684619904, -0.00112742162309587, -0.14361046254634857, -0.05187321454286575, 0.027355175465345383, 0.0050268471240997314, -0.0009382785065099597, 0.0018201807979494333, 0.06529944390058517, 0.0961814671754837, -0.07274709641933441, -0.07741184532642365, -0.02502381056547165, 0.004293975420296192, -0.08518482744693756, -0.21569354832172394, -0.03486844524741173, -0.025585955008864403, 0.1678820252418518, -0.2618064284324646, 0.04191526025533676, -0.0012870969949290156, 0.1268022209405899, 0.04391893744468689, -0.032299865037202835, 0.0005466039874590933, 0.02997620590031147, -0.05297306180000305, -0.07963749021291733, 0.037994809448719025, 0.007674727588891983, -0.10097891092300415, -0.04352886602282524, -0.14170612394809723, 0.1710972785949707, 0.10038086771965027, -0.021979229524731636, -0.0945008173584938, -0.040646202862262726, -0.07195331901311874, -0.04046514257788658, -0.0346330851316452, -0.006844020914286375, 0.11096850037574768, -0.00826446246355772, 0.1445898711681366, -0.08341747522354126, -0.055617451667785645, 0.018141750246286392, -0.021130848675966263, 0.005436563398689032, 0.15391921997070312, 0.03387836366891861, -0.12493475526571274, 0.12711352109909058, 0.1279885470867157, -0.0370761938393116, 0.1585504114627838, -0.06412091851234436, -0.08019131422042847, -0.03034062683582306, 0.04230709373950958, 0.019845403730869293, 0.0871751680970192, -0.10392555594444275, 0.0025068270042538643, 0.021129857748746872, 0.032557375729084015, 0.02281828597187996, -0.17112885415554047, -0.023989364504814148, 0.04975498467683792, -0.039960816502571106, -0.015271837823092937, -0.0009314308990724385, -0.021317508071660995, 0.100294329226017, -0.000850585987791419, -0.05473228543996811, 0.0041646817699074745, -0.011548278853297234, -0.08589611202478409, 0.21360047161579132, -0.0851317048072815, -0.12549664080142975, -0.12277871370315552, -0.005389455705881119, -0.03846742957830429, 0.007122356444597244, 0.05537612736225128, -0.10293140262365341, -0.03552601486444473, -0.09696546196937561, 0.043776985257864, -0.01599741168320179, 0.018042461946606636, 0.020282026380300522, 0.02202560380101204, 0.06890331953763962, -0.10492973774671555, 0.01160325575619936, -0.01693546399474144, -0.039650384336709976, 0.04597965255379677, -0.004406370688229799, 0.11572358012199402, 0.1287747174501419, 0.012556646950542927, 0.027984393760561943, -0.04463530704379082, 0.22045938670635223, -0.07553491741418839, -0.011684330180287361, 0.11397381126880646, 0.014157073572278023, 0.0442480705678463, 0.1398359090089798, 0.03304468095302582, -0.10382483154535294, 0.04304731637239456, 0.0430433414876461, -0.015720902010798454, -0.2211015671491623, -0.026665786281228065, -0.054237283766269684, 0.020640874281525612, 0.10420653223991394, 0.032372232526540756, 0.007185904774814844, 0.06270184367895126, -0.015511855483055115, 0.027510156854987144, 0.02615755796432495, 0.09206458926200867, 0.057363059371709824, 0.03692959249019623, 0.11883141100406647, -0.039200086146593094, -0.054627083241939545, 0.04180612787604332, -0.04101504385471344, 0.2529999613761902, -0.020784595981240273, 0.14214780926704407, 0.05566788837313652, 0.1405867338180542, -0.006649029441177845, 0.06778672337532043, 0.005113921593874693, -0.028332971036434174, -0.013368773274123669, -0.0593787245452404, -0.022685300558805466, 0.04599270224571228, -0.030343376100063324, 0.07243639975786209, -0.1371522843837738, 0.06311309337615967, 0.06582047790288925, 0.2644226551055908, 0.07228721678256989, -0.32776424288749695, -0.10075021535158157, 0.03124217875301838, -0.044572990387678146, -0.026830386370420456, 0.04207393899559975, 0.11326643824577332, -0.091305673122406, 0.0642300471663475, -0.07137179374694824, 0.08232124149799347, -0.024126915261149406, 0.025405924767255783, 0.060380470007658005, 0.07588464766740799, -0.016340667381882668, 0.0835544615983963, -0.2386021912097931, 0.2847225069999695, -0.0055634984746575356, 0.06562182307243347, -0.04821513220667839, 0.02647264674305916, 0.03334372490644455, 0.048071134835481644, 0.0985093042254448, -0.007532014511525631, -0.08741527050733566, -0.1659613996744156, -0.06296742707490921, 0.029987793415784836, 0.12106098234653473, -0.09230093657970428, 0.1294054090976715, -0.04131676256656647, -0.018707633018493652, 0.05504611134529114, -0.04621290788054466, -0.09638596326112747, -0.09761395305395126, 0.002588865580037236, -0.0280542504042387, 0.03257598355412483, -0.10500868409872055, -0.0921454206109047, -0.07224199920892715, 0.1706337332725525, -0.04786232113838196, -0.03686239570379257, -0.12477820366621017, 0.10586142539978027, 0.11261218041181564, -0.08039962500333786, 0.040765974670648575, 0.012402983382344246, 0.1033143550157547, 0.04514506086707115, -0.029843442142009735, 0.11064046621322632, -0.0919543132185936, -0.21754321455955505, -0.053610265254974365, 0.1439804583787918, 0.02096210978925228, 0.05380129814147949, -0.016188859939575195, 0.019018081948161125, -0.013502947054803371, -0.08794434368610382, 0.023784611374139786, -0.0074294982478022575, 0.08468927443027496, 0.04859597608447075, -0.05184485390782356, 0.003729606280103326, -0.05522261932492256, -0.05750025436282158, 0.13635076582431793, 0.3174488842487335, -0.0741940289735794, -0.0036475956439971924, 0.040095798671245575, -0.05205301567912102, -0.1431618332862854, 0.023400744423270226, 0.069350466132164, 0.020063921809196472, 0.03318323567509651, -0.164321169257164, 0.0736471563577652, 0.0949910432100296, -0.010364926420152187, 0.07299824059009552, -0.31733372807502747, -0.13313686847686768, 0.09752612560987473, 0.1198115199804306, 0.018675563856959343, -0.17638510465621948, -0.04495127126574516, -0.010326530784368515, -0.09111886471509933, 0.05273735150694847, -0.0842810869216919, 0.10508163273334503, -0.008969640359282494, 0.026693833991885185, 0.01994260400533676, -0.0598599947988987, 0.14078646898269653, -0.0005555192474275827, 0.0859326496720314, -0.04172476753592491, 0.006633555516600609, 0.05680178478360176, -0.07502307742834091, 0.02351297438144684, -0.09693945944309235, 0.06274271011352539, -0.1259773075580597, -0.02649480476975441, -0.06884178519248962, 0.027530914172530174, -0.056080371141433716, -0.06839273124933243, -0.046367447823286057, 0.058710843324661255, 0.08172523230314255, -0.007899590767920017, 0.12804977595806122, -0.002921442501246929, 0.14586830139160156, 0.09836814552545547, 0.07703443616628647, -0.011703276075422764, -0.04109284281730652, -0.0138779915869236, -0.005885678809136152, 0.029566224664449692, -0.1483309417963028, 0.027695879340171814, 0.13763712346553802, 0.029986988753080368, 0.15576870739459991, 0.06864955276250839, -0.033563025295734406, 0.004793860483914614, 0.07609681040048599, -0.1311054229736328, -0.12165139615535736, -0.038230299949645996, -0.008363085798919201, -0.1349738985300064, 0.036206863820552826, 0.09709572046995163, -0.06274498999118805, -0.0020168826449662447, -0.010843324474990368, 0.032918717712163925, -0.023955069482326508, 0.20727188885211945, 0.04027548432350159, 0.08080321550369263, -0.08327731490135193, 0.09737563878297806, 0.040108587592840195, -0.09772320836782455, 0.016708500683307648, 0.07948492467403412, -0.09137861430644989, -0.031425900757312775, 0.041833020746707916, 0.14747832715511322, -0.019575465470552444, -0.034580592066049576, -0.14596720039844513, -0.12684187293052673, 0.07511270046234131, 0.07772473990917206, 0.07035896927118301, 0.008786173537373543, -0.029373742640018463, 0.02122802659869194, -0.12028782814741135, 0.1368873566389084, 0.10999123007059097, 0.07732407003641129, -0.16383793950080872, 0.13091376423835754, -0.01601017825305462, -0.010543313808739185, -0.013824705965816975, 0.029551874846220016, -0.125055730342865, -0.005291106645017862, -0.10693550109863281, -0.01033059973269701, -0.056384749710559845, 0.005185428075492382, -0.01362192165106535, -0.037247128784656525, -0.05796084925532341, 0.01629856415092945, -0.09566938877105713, -0.038891442120075226, 0.008821802213788033, 0.07532390207052231, -0.09523175656795502, -0.02149619534611702, 0.019233904778957367, -0.1089264377951622, 0.09232043474912643, 0.02584812045097351, 0.0270914938300848, 0.02466314099729061, -0.10851425677537918, 0.03270657733082771, 0.05353337153792381, -0.01017871592193842, 0.03494497388601303, -0.1417955458164215, 0.00029129674658179283, -0.03126514330506325, 0.007185298949480057, -0.00010966679110424593, 0.016434796154499054, -0.1277540922164917, -0.0057453494518995285, -0.04026474803686142, -0.055701132863759995, -0.06288432329893112, 0.03629469498991966, 0.04161061719059944, -0.007823183201253414, 0.18587759137153625, -0.08674650639295578, 0.030415482819080353, -0.23105522990226746, -0.011244707740843296, 0.001449459115974605, -0.0896592065691948, -0.0998031347990036, -0.03873702511191368, 0.07233928143978119, -0.06692453473806381, 0.10209948569536209, -0.037381429225206375, 0.03284560143947601, 0.03707648068666458, -0.0563850998878479, 0.080083467066288, 0.035208333283662796, 0.21636207401752472, 0.016026262193918228, -0.027443902567029, 0.021199824288487434, 0.027137042954564095, 0.09023363143205643, 0.05945682153105736, 0.16826175153255463, 0.16045859456062317, -0.04465508460998535, 0.07263462245464325, 0.018965337425470352, -0.0968058705329895, -0.15380437672138214, 0.08556260913610458, -0.019206270575523376, 0.11920005828142166, -0.011875934898853302, 0.14509914815425873, 0.14979559183120728, -0.18734146654605865, 0.00844531413167715, -0.06651849299669266, -0.07391472905874252, -0.09226292371749878, -0.0646006166934967, -0.08967619389295578, -0.15903134644031525, 0.013342306949198246, -0.13460610806941986, 0.016312357038259506, 0.062295470386743546, 0.03160814940929413, -0.0009487589122727513, 0.15809805691242218, 0.02120944857597351, 0.007208137772977352, 0.06790842115879059, 0.011541939340531826, -0.022489383816719055, -0.041987381875514984, -0.08195160329341888, 0.010825756005942822, -0.03974129632115364, 0.034767165780067444, -0.05832025781273842, -0.06849634647369385, 0.047532837837934494, -0.005605858750641346, -0.08582160621881485, 0.018290281295776367, 0.01855619251728058, 0.052130285650491714, 0.05070885643362999, 0.013704708777368069, -0.020350998267531395, -0.014875383116304874, 0.23832057416439056, -0.08594801276922226, -0.05962090566754341, -0.12113167345523834, 0.20820581912994385, 0.028361838310956955, -0.0008699373574927449, 0.03634415194392204, -0.09494409710168839, 0.027710216119885445, 0.16317309439182281, 0.1960013210773468, -0.03898352012038231, -0.025101235136389732, 0.0067087081260979176, -0.009600553661584854, -0.032975245267152786, 0.08277034014463425, 0.1277804970741272, 0.04562680423259735, -0.06598998606204987, -0.017421193420886993, -0.03884453698992729, -0.03555227816104889, -0.0367489717900753, 0.09162171930074692, 0.05139203742146492, -0.0020666185300797224, -0.013278700411319733, 0.0999380424618721, -0.0578133687376976, -0.10094417631626129, 0.02922038547694683, -0.17926433682441711, -0.1712462157011032, -0.039416853338479996, 0.05713935196399689, 0.0214877650141716, 0.07096057385206223, -0.016546091064810753, -0.025805065408349037, 0.06986475735902786, -0.0027362641412764788, -0.05967339873313904, -0.0953965038061142, 0.06691128760576248, -0.07323341816663742, 0.21296390891075134, -0.03869061544537544, 0.007223645690828562, 0.14511683583259583, 0.014627822674810886, -0.10360705852508545, 0.0603332482278347, 0.08055614680051804, -0.056818146258592606, 0.051891833543777466, 0.12561799585819244, -0.03294958546757698, 0.13539958000183105, 0.05673487111926079, -0.1281924843788147, 0.005518509075045586, -0.04931248351931572, -0.048090700060129166, -0.07310851663351059, 0.008288181386888027, -0.04223671928048134, 0.1460239589214325, 0.22101260721683502, -0.06612606346607208, 0.011308380402624607, -0.05237352475523949, 0.04456063359975815, 0.06666650623083115, 0.07575740665197372, -0.004270394332706928, -0.259553462266922, 0.014384109526872635, 0.04723064973950386, 0.010075110010802746, -0.2727937698364258, -0.08811656385660172, -0.006785271689295769, -0.04192791506648064, -0.1043493002653122, 0.12938764691352844, 0.09499973058700562, 0.044651325792074203, -0.051562823355197906, -0.08051800727844238, -0.07688066363334656, 0.17172972857952118, -0.1445232778787613, -0.08307136595249176 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": ["trl", "sft"]}
text-generation
ybelkada/test-automatic-tagging
[ "transformers", "safetensors", "llama", "text-generation", "trl", "sft", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T04:16:27+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #llama #text-generation #trl #sft #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #trl #sft #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 62, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #trl #sft #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.04139265418052673, 0.1722242832183838, -0.005511249881237745, 0.0189357940107584, 0.10351774841547012, 0.007488170173019171, 0.053782787173986435, 0.11686506122350693, -0.04136158898472786, 0.12049121409654617, 0.03978472948074341, 0.10854824632406235, 0.11710964888334274, 0.14666947722434998, 0.00195520487613976, -0.22164475917816162, 0.0471925288438797, -0.11311910301446915, -0.026472201570868492, 0.12014154344797134, 0.1483195424079895, -0.09877023100852966, 0.07405374199151993, -0.031364403665065765, -0.01976277492940426, -0.033224985003471375, -0.06165917590260506, -0.04188290610909462, 0.039299480617046356, 0.05756969004869461, 0.06539180129766464, 0.001487001427449286, 0.0872797742486, -0.2740541398525238, 0.018865173682570457, 0.06961553543806076, -0.0029641701839864254, 0.06348572671413422, 0.06653270870447159, -0.06446996331214905, 0.10695377737283707, -0.048933178186416626, 0.13937993347644806, 0.08843822777271271, -0.0936393141746521, -0.18266285955905914, -0.09263508766889572, 0.10108376294374466, 0.17314845323562622, 0.05274033546447754, -0.025193531066179276, 0.09762939065694809, -0.08472544699907303, 0.017721597105264664, 0.05270730331540108, -0.08688028901815414, -0.0563482865691185, 0.05796172469854355, 0.08986722677946091, 0.05228787660598755, -0.12480800598859787, -0.032590169459581375, 0.0026210385840386152, 0.01959611475467682, 0.07128550857305527, 0.018656587228178978, 0.14507532119750977, 0.0339556448161602, -0.13426542282104492, -0.04768969863653183, 0.10528716444969177, 0.03814857080578804, -0.041775867342948914, -0.2408970147371292, -0.03204251453280449, -0.03707160800695419, -0.03451048955321312, -0.04197743162512779, 0.043388884514570236, -0.0020766586530953646, 0.08929190039634705, -0.006048701703548431, -0.07451780885457993, -0.03613940253853798, 0.06299129128456116, 0.06394032388925552, 0.02911241725087166, -0.018013427034020424, 0.01671660505235195, 0.10972205549478531, 0.10487091541290283, -0.11634443700313568, -0.060189388692379, -0.0655711367726326, -0.07997546344995499, -0.0413973405957222, 0.030799981206655502, 0.01478270348161459, 0.06373008340597153, 0.2633737623691559, 0.02523009292781353, 0.05830543488264084, 0.03010810911655426, 0.007674427703022957, 0.052477747201919556, 0.10824281722307205, -0.06397172063589096, -0.1108320876955986, -0.023409530520439148, 0.08599712699651718, 0.0119717326015234, -0.037738144397735596, -0.04876318946480751, 0.06515824049711227, 0.035509053617715836, 0.11009177565574646, 0.09669167548418045, 0.01958681456744671, -0.07370802015066147, -0.06234503164887428, 0.18890081346035004, -0.1592905968427658, 0.03671742603182793, 0.039474599063396454, -0.038266152143478394, -0.010567347519099712, 0.014316120184957981, 0.020759545266628265, -0.03020627051591873, 0.08353850245475769, -0.05679228529334068, -0.0460251085460186, -0.11079160869121552, -0.03140000253915787, 0.03963953256607056, 0.010553130879998207, -0.03440975025296211, -0.03924540430307388, -0.07616686075925827, -0.08706684410572052, 0.08977017551660538, -0.0717964917421341, -0.05277155712246895, -0.02312835305929184, -0.08338230103254318, 0.02330220863223076, 0.022066133096814156, 0.0784529373049736, -0.02310224622488022, 0.052695564925670624, -0.050556499511003494, 0.056182533502578735, 0.1055482029914856, 0.038718923926353455, -0.059238966554403305, 0.05764951556921005, -0.23578405380249023, 0.08923166990280151, -0.07015169411897659, 0.06696770340204239, -0.15654432773590088, -0.02274715155363083, 0.0368778295814991, 0.00751019362360239, -0.006022947374731302, 0.13441745936870575, -0.2129882127046585, -0.024449266493320465, 0.16660422086715698, -0.09792499244213104, -0.07090940326452255, 0.052212733775377274, -0.044031333178281784, 0.11030184477567673, 0.031535279005765915, -0.005807520821690559, 0.0628088042140007, -0.1116349846124649, -0.004583585076034069, -0.05519717559218407, -0.022825121879577637, 0.14128947257995605, 0.07663483917713165, -0.07629034668207169, 0.06767985224723816, 0.024410447105765343, -0.025700828060507774, -0.055185865610837936, -0.016822461038827896, -0.10276598483324051, 0.01773652248084545, -0.06592164188623428, 0.0071457200683653355, -0.016769787296652794, -0.08997338265180588, -0.025631334632635117, -0.1714833676815033, -0.03606460615992546, 0.08107777684926987, -0.0026023148093372583, -0.014582614414393902, -0.11473827809095383, 0.01854405365884304, 0.040997736155986786, 0.005381781142205, -0.13550327718257904, -0.04257364198565483, 0.029827285557985306, -0.15436461567878723, 0.038500458002090454, -0.06505212187767029, 0.05161800608038902, 0.016728023067116737, -0.024154458194971085, -0.02382735162973404, 0.01990814134478569, 0.007024670485407114, -0.013208836317062378, -0.23930847644805908, -0.030523808673024178, -0.02743144892156124, 0.1687149852514267, -0.20244638621807098, 0.03475760295987129, 0.08543332666158676, 0.1551775187253952, 0.007221921347081661, -0.04917723312973976, 0.012014477513730526, -0.07092157006263733, -0.023330530151724815, -0.05615955963730812, 0.0012083996552973986, -0.020472567528486252, -0.04034966602921486, 0.030420562252402306, -0.1718258559703827, -0.04385634511709213, 0.09709953516721725, 0.047556761652231216, -0.1367577612400055, -0.01717122085392475, -0.03800658509135246, -0.052668675780296326, -0.03909987211227417, -0.06347472965717316, 0.10223289579153061, 0.05967874452471733, 0.04064521938562393, -0.05422886833548546, -0.07865968346595764, -0.0021571116521954536, -0.012350710108876228, -0.021148214116692543, 0.09663669019937515, 0.08365008980035782, -0.1379639357328415, 0.09477944672107697, 0.0873076543211937, 0.07930327951908112, 0.08826930075883865, -0.02064487896859646, -0.07811369746923447, -0.041565798223018646, 0.034856248646974564, 0.019707197323441505, 0.12474244832992554, -0.038920335471630096, 0.03861235827207565, 0.04033034294843674, -0.026701444759964943, 0.017970016226172447, -0.07758808881044388, 0.03313266485929489, 0.023121200501918793, -0.013193833641707897, 0.054333459585905075, -0.03507404774427414, 0.018521755933761597, 0.08736611157655716, 0.058756545186042786, 0.03708405792713165, 0.01708976738154888, -0.05427597835659981, -0.11087038367986679, 0.15839338302612305, -0.12608490884304047, -0.21765752136707306, -0.13357238471508026, 0.009490080177783966, 0.030329156666994095, -0.01694365404546261, 0.005587044171988964, -0.060498304665088654, -0.1157841682434082, -0.08712256699800491, 0.01551016140729189, 0.050134800374507904, -0.08472836762666702, -0.061171289533376694, 0.04956647753715515, 0.03838740661740303, -0.1420324593782425, 0.018942978233098984, 0.04319990426301956, -0.09420828521251678, -0.01079641841351986, 0.08009295165538788, 0.07644724100828171, 0.18227055668830872, 0.023805294185876846, -0.018289558589458466, 0.03283845633268356, 0.21312877535820007, -0.1373033970594406, 0.11198293417692184, 0.1346629410982132, -0.08706950396299362, 0.07853328436613083, 0.2040577083826065, 0.04083133116364479, -0.09917159378528595, 0.033185750246047974, 0.02999437414109707, -0.025952914729714394, -0.23641210794448853, -0.07004403322935104, -0.004025670699775219, -0.06254614144563675, 0.07735299319028854, 0.09651638567447662, 0.0784681960940361, 0.01303916610777378, -0.09250159561634064, -0.08996374160051346, 0.05852869525551796, 0.10616276413202286, 0.023758746683597565, -0.010819397866725922, 0.08845941722393036, -0.036525703966617584, 0.015782220289111137, 0.08483923971652985, 0.0005252371192909777, 0.1634787768125534, 0.05080989748239517, 0.18067625164985657, 0.0848069116473198, 0.07268007099628448, 0.0031657610088586807, 0.009335260838270187, 0.016666652634739876, 0.039801206439733505, -0.005358996335417032, -0.08148239552974701, -0.0270138680934906, 0.11122950166463852, 0.06578166037797928, 0.012782013043761253, 0.007697426714003086, -0.043564822524785995, 0.08215389400720596, 0.18700776994228363, -0.0023850479628890753, -0.18034884333610535, -0.05724934861063957, 0.07167387008666992, -0.09775814414024353, -0.10089543461799622, -0.005623296368867159, 0.016602298244833946, -0.16731193661689758, 0.031071249395608902, -0.024214431643486023, 0.10710599273443222, -0.13180649280548096, -0.017289698123931885, 0.08546080440282822, 0.07537814229726791, 0.00325193302705884, 0.05420098826289177, -0.1755291074514389, 0.09734869003295898, 0.010345921851694584, 0.06613057106733322, -0.09753984212875366, 0.09634924679994583, -0.0068987454287707806, -0.03153010457754135, 0.14413253962993622, -0.002784974640235305, -0.07470765709877014, -0.0690583810210228, -0.08543293178081512, -0.012737947516143322, 0.13091471791267395, -0.13898269832134247, 0.09071407467126846, -0.03909354284405708, -0.03882821649312973, -0.007163032423704863, -0.0880526602268219, -0.10614196211099625, -0.18218879401683807, 0.06416638940572739, -0.1361708790063858, 0.035792987793684006, -0.10682537406682968, -0.027188677340745926, -0.029113078489899635, 0.18549677729606628, -0.2437957525253296, -0.07211407274007797, -0.1444912850856781, -0.09461814165115356, 0.13351555168628693, -0.050014790147542953, 0.08923278003931046, -0.013159503228962421, 0.1546577662229538, 0.021921712905168533, -0.02579069882631302, 0.09421761333942413, -0.08544669300317764, -0.1926930993795395, -0.07177501916885376, 0.15786993503570557, 0.1244857907295227, 0.032881323248147964, -0.0029549715109169483, 0.03688029199838638, -0.017612045630812645, -0.11746013164520264, 0.020308375358581543, 0.15969213843345642, 0.06053909286856651, 0.01368645578622818, -0.02771882340312004, -0.10228043794631958, -0.07373582571744919, -0.025955528020858765, 0.032543864101171494, 0.17028437554836273, -0.07207192480564117, 0.17618514597415924, 0.13706210255622864, -0.05772562697529793, -0.21175675094127655, 0.0022410722449421883, 0.02623414434492588, -0.004848991986364126, 0.015538363717496395, -0.19550566375255585, 0.08806484937667847, -0.0011735234875231981, -0.052357956767082214, 0.11750499904155731, -0.1689670830965042, -0.1374114751815796, 0.08379434794187546, 0.04366026818752289, -0.18811742961406708, -0.13793399930000305, -0.09433916211128235, -0.03560960292816162, -0.16542401909828186, 0.09501351416110992, 0.029086610302329063, 0.01279810443520546, 0.030953075736761093, 0.019788993522524834, 0.02292264997959137, -0.04467814415693283, 0.17458944022655487, -0.01881459914147854, 0.02117486298084259, -0.09406382590532303, -0.07756610959768295, 0.025801025331020355, -0.04981527477502823, 0.07309889048337936, -0.012024049647152424, 0.008426345884799957, -0.10312525928020477, -0.03651638329029083, -0.04387262463569641, 0.017978128045797348, -0.0985637754201889, -0.08534010499715805, -0.04235544800758362, 0.09770673513412476, 0.09486958384513855, -0.027134086936712265, -0.02329561486840248, -0.07777043431997299, 0.05436117947101593, 0.20730538666248322, 0.18341942131519318, 0.04246291145682335, -0.0686640590429306, -0.0042101661674678326, -0.012240317650139332, 0.040951136499643326, -0.19176772236824036, 0.059671733528375626, 0.05790033936500549, 0.022321194410324097, 0.10411332547664642, -0.015301277860999107, -0.15725941956043243, -0.07588879764080048, 0.06935762614011765, -0.0633203312754631, -0.20109675824642181, 0.005668620578944683, 0.05525026097893715, -0.1749851405620575, -0.04218949377536774, 0.04785749316215515, -0.003059533890336752, -0.03765922039747238, 0.026610443368554115, 0.09476181119680405, 0.0019908808171749115, 0.07670184969902039, 0.06725285202264786, 0.07990311831235886, -0.10340512543916702, 0.08271146565675735, 0.09477692097425461, -0.07740912586450577, 0.027109019458293915, 0.10942762345075607, -0.05830185487866402, -0.03909740969538689, 0.025002168491482735, 0.08164498209953308, 0.01883842796087265, -0.036339592188596725, 0.011891803704202175, -0.10006634891033173, 0.06578171998262405, 0.0990314707159996, 0.030636711046099663, 0.020836731418967247, 0.0393795520067215, 0.04987800866365433, -0.07355048507452011, 0.12213615328073502, 0.030783897265791893, 0.015510526485741138, -0.040226925164461136, -0.03838871046900749, 0.009310542605817318, -0.025675995275378227, -0.005737519823014736, -0.02463916875422001, -0.08526302129030228, -0.01573621667921543, -0.1344650387763977, -0.014417117461562157, -0.05993317440152168, 0.012797602452337742, 0.03079844079911709, -0.03353812173008919, 0.004972065798938274, 0.005301079712808132, -0.07384614646434784, -0.06963427364826202, -0.014464515261352062, 0.09566465020179749, -0.16286233067512512, 0.02242831513285637, 0.08249443769454956, -0.11669174581766129, 0.09626647084951401, 0.01651747338473797, -0.005086736753582954, 0.023460952565073967, -0.1393972486257553, 0.032331958413124084, -0.040155962109565735, 0.009552007541060448, 0.031144948676228523, -0.20124459266662598, -0.0008879420929588377, -0.03673647716641426, -0.07063861936330795, -0.0090862475335598, -0.030402997508645058, -0.11478965729475021, 0.10718602687120438, 0.002245173556730151, -0.08107110112905502, -0.0336882546544075, 0.029499446973204613, 0.07790062576532364, -0.021243326365947723, 0.1477850377559662, -0.011712978594005108, 0.07237184047698975, -0.1603296995162964, -0.011160957626998425, -0.008894158527255058, 0.01940762810409069, -0.029238715767860413, -0.010890902951359749, 0.05012662708759308, -0.016858026385307312, 0.17031224071979523, -0.034122299402952194, 0.02129121869802475, 0.06898706406354904, 0.03831607103347778, -0.030093926936388016, 0.09930204600095749, 0.038610998541116714, 0.020207753404974937, 0.011195220984518528, 0.011295555159449577, -0.04223373532295227, -0.035372231155633926, -0.1934705525636673, 0.07545345276594162, 0.1758745312690735, 0.098245769739151, -0.017390236258506775, 0.0738692507147789, -0.10397400707006454, -0.0965336486697197, 0.1500321924686432, -0.038248948752880096, -0.006249778438359499, -0.07447223365306854, 0.12666769325733185, 0.14411661028862, -0.17948438227176666, 0.06849895417690277, -0.07055202126502991, -0.041767630726099014, -0.11033682525157928, -0.19594387710094452, -0.05765159800648689, -0.04832446947693825, -0.0191052109003067, -0.04547692462801933, 0.07126773148775101, 0.05765657126903534, 0.0018349154852330685, -0.005592921283096075, 0.06990435719490051, -0.038766175508499146, -0.00449542049318552, 0.03023001179099083, 0.06123078614473343, 0.00763664348050952, -0.03757287189364433, 0.016563747078180313, -0.012766031548380852, 0.05637874826788902, 0.07063084840774536, 0.04886922985315323, -0.029026931151747704, 0.019231151789426804, -0.04138490557670593, -0.10620179027318954, 0.04554767906665802, -0.026372171938419342, -0.07791831344366074, 0.1499139815568924, 0.020531395450234413, 0.004851186648011208, -0.013242626562714577, 0.2370923012495041, -0.06804169714450836, -0.09738621115684509, -0.14787611365318298, 0.08123983442783356, -0.03680659830570221, 0.0524895153939724, 0.039055727422237396, -0.10718115419149399, 0.022145681083202362, 0.14749126136302948, 0.15913428366184235, -0.04109198972582817, 0.022015046328306198, 0.03850727900862694, 0.00855081807821989, -0.027643553912639618, 0.04104061424732208, 0.0669441893696785, 0.16011744737625122, -0.047983769327402115, 0.08299979567527771, -0.0011984026059508324, -0.09032373875379562, -0.03606504574418068, 0.11040901392698288, -0.00786347035318613, 0.019113434478640556, -0.05761413648724556, 0.11980626732110977, -0.0732465460896492, -0.2267553061246872, 0.04876295477151871, -0.07038494199514389, -0.13377021253108978, -0.02632228657603264, 0.07991223782300949, -0.010571391321718693, 0.026576517149806023, 0.07522688060998917, -0.06974120438098907, 0.19860774278640747, 0.03984677419066429, -0.06001606211066246, -0.05303289368748665, 0.07825860381126404, -0.08528647571802139, 0.2845120429992676, 0.012923448346555233, 0.036362119019031525, 0.10918033123016357, -0.009960188530385494, -0.14107763767242432, 0.015298071317374706, 0.09498007595539093, -0.10316479951143265, 0.047996558248996735, 0.18568755686283112, 0.0008864350966177881, 0.12857073545455933, 0.07515165209770203, -0.08722532540559769, 0.04689505323767662, -0.07482599467039108, -0.06658551841974258, -0.09557390213012695, 0.10328887403011322, -0.0849103108048439, 0.14513863623142242, 0.1338893324136734, -0.05413280054926872, 0.01074116863310337, -0.036647967994213104, 0.04159361124038696, -0.001764152548275888, 0.11406978219747543, 0.007171210832893848, -0.18479910492897034, 0.025392163544893265, -0.022483542561531067, 0.10057500004768372, -0.16157642006874084, -0.09141390770673752, 0.04674063250422478, 0.010000529699027538, -0.07248106598854065, 0.12580238282680511, 0.057280778884887695, 0.03375311568379402, -0.04551668465137482, -0.025121787562966347, -0.008275391533970833, 0.14095145463943481, -0.10870447009801865, -0.00349931581877172 ]
null
null
peft
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # test-automatic-tagging-from-trainer This model is a fine-tuned version of [HuggingFaceM4/tiny-random-LlamaForCausalLM](https://huggingface.co/HuggingFaceM4/tiny-random-LlamaForCausalLM) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - training_steps: 1 ### Framework versions - PEFT 0.8.2 - Transformers 4.38.0.dev0 - Pytorch 2.2.0+cu118 - Datasets 2.16.1 - Tokenizers 0.15.0
{"library_name": "peft", "tags": ["trl", "sft", "generated_from_trainer"], "base_model": "HuggingFaceM4/tiny-random-LlamaForCausalLM", "model-index": [{"name": "test-automatic-tagging-from-trainer", "results": []}]}
null
ybelkada/test-automatic-tagging-from-trainer
[ "peft", "safetensors", "trl", "sft", "generated_from_trainer", "base_model:HuggingFaceM4/tiny-random-LlamaForCausalLM", "region:us" ]
2024-02-15T04:16:28+00:00
[]
[]
TAGS #peft #safetensors #trl #sft #generated_from_trainer #base_model-HuggingFaceM4/tiny-random-LlamaForCausalLM #region-us
# test-automatic-tagging-from-trainer This model is a fine-tuned version of HuggingFaceM4/tiny-random-LlamaForCausalLM on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - training_steps: 1 ### Framework versions - PEFT 0.8.2 - Transformers 4.38.0.dev0 - Pytorch 2.2.0+cu118 - Datasets 2.16.1 - Tokenizers 0.15.0
[ "# test-automatic-tagging-from-trainer\n\nThis model is a fine-tuned version of HuggingFaceM4/tiny-random-LlamaForCausalLM on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- training_steps: 1", "### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.2.0+cu118\n- Datasets 2.16.1\n- Tokenizers 0.15.0" ]
[ "TAGS\n#peft #safetensors #trl #sft #generated_from_trainer #base_model-HuggingFaceM4/tiny-random-LlamaForCausalLM #region-us \n", "# test-automatic-tagging-from-trainer\n\nThis model is a fine-tuned version of HuggingFaceM4/tiny-random-LlamaForCausalLM on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- training_steps: 1", "### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.2.0+cu118\n- Datasets 2.16.1\n- Tokenizers 0.15.0" ]
[ 50, 48, 6, 12, 8, 3, 89, 44 ]
[ "passage: TAGS\n#peft #safetensors #trl #sft #generated_from_trainer #base_model-HuggingFaceM4/tiny-random-LlamaForCausalLM #region-us \n# test-automatic-tagging-from-trainer\n\nThis model is a fine-tuned version of HuggingFaceM4/tiny-random-LlamaForCausalLM on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- training_steps: 1### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.2.0+cu118\n- Datasets 2.16.1\n- Tokenizers 0.15.0" ]
[ -0.09988157451152802, 0.02801988460123539, -0.0015998071758076549, 0.10382170975208282, 0.16170957684516907, 0.02465004101395607, 0.10545257478952408, 0.11071889102458954, -0.07589567452669144, 0.07415518164634705, 0.0725250169634819, 0.014282269403338432, 0.04062383621931076, 0.17715522646903992, -0.03283233940601349, -0.27707183361053467, -0.0013036472955718637, -0.03027874045073986, -0.0294471587985754, 0.10485120862722397, 0.10014479607343674, -0.08359089493751526, 0.06376086175441742, -0.0052636731415987015, -0.16467176377773285, 0.026089303195476532, -0.0004337840073276311, -0.035643890500068665, 0.12362536042928696, 0.006450592540204525, 0.10551959276199341, -0.03065178170800209, 0.1436469852924347, -0.20728032290935516, 0.008018897846341133, 0.0996452271938324, 0.04982771724462509, 0.07107329368591309, 0.07594817131757736, -0.00961293838918209, 0.09725544601678848, -0.05403629690408707, 0.09130219370126724, 0.042997777462005615, -0.10196543484926224, -0.18036234378814697, -0.09628114849328995, 0.05948522686958313, 0.09149714559316635, 0.0876777246594429, 0.012685137800872326, 0.13931059837341309, -0.07440288364887238, 0.061184145510196686, 0.24246709048748016, -0.23175077140331268, -0.10379046946763992, 0.08554474264383316, 0.06537949293851852, 0.08095994591712952, -0.11891303956508636, -0.02961115352809429, 0.08889502286911011, 0.0656459853053093, 0.1277313232421875, -0.03678075224161148, -0.057548150420188904, -0.04531695321202278, -0.12648384273052216, 0.008929243311285973, 0.11153175681829453, 0.023983098566532135, -0.06216301769018173, -0.09803538024425507, -0.06116917356848717, -0.056945137679576874, -0.033876921981573105, -0.02829546108841896, 0.0245190542191267, -0.03627699986100197, -0.004429823253303766, -0.07216645777225494, -0.050314683467149734, -0.10745221376419067, 0.026702074334025383, 0.1644057333469391, 0.028781665489077568, 0.015538748353719711, -0.010815593414008617, 0.11898978799581528, -0.05982150882482529, -0.08224907517433167, -0.0075770230032503605, -0.03366253525018692, -0.04591972380876541, -0.04765503481030464, -0.0639929473400116, 0.034037329256534576, -0.009912021458148956, 0.16121651232242584, -0.14717617630958557, 0.07034805417060852, -0.004546092823147774, 0.009601661935448647, -0.07628466188907623, 0.0754089429974556, -0.04559031501412392, -0.006286690942943096, -0.003869087668135762, 0.1341477632522583, 0.000812763231806457, 0.004322882741689682, -0.05562995746731758, -0.00781102292239666, 0.006545285694301128, 0.04963558539748192, -0.08370715379714966, 0.022751517593860626, -0.02472562901675701, -0.026695558801293373, 0.020228195935487747, -0.11871897429227829, 0.04652995988726616, 0.004759145900607109, -0.07061585783958435, -0.07925637811422348, 0.03085356578230858, 0.008200437761843204, -0.009722966700792313, 0.12335126101970673, -0.08529344946146011, 0.031479090452194214, -0.1025921106338501, -0.08636755496263504, -0.024297690019011497, -0.08477466553449631, -0.02611246146261692, -0.0853205993771553, -0.20741797983646393, -0.05211836099624634, 0.02504895068705082, -0.052352361381053925, -0.013271128758788109, -0.04537418857216835, -0.08134881407022476, -0.009528698399662971, -0.0015562797198072076, 0.10247530788183212, -0.05493390932679176, 0.0762374997138977, -0.02264866791665554, 0.026438040658831596, 0.007556915283203125, 0.026878971606492996, -0.0745450109243393, 0.037170760333538055, -0.12128549069166183, 0.06799358129501343, -0.10056982934474945, 0.03408045321702957, -0.08055603504180908, -0.10173800587654114, -0.05165664479136467, 0.01796177588403225, 0.07762421667575836, 0.13974244892597198, -0.18599069118499756, -0.037040263414382935, 0.2207115888595581, -0.10451939702033997, -0.07960309833288193, 0.0703638345003128, -0.04514714330434799, 0.1024332121014595, 0.032122429460287094, 0.1946285516023636, 0.10394635051488876, -0.1659555286169052, 0.055856868624687195, -0.02066945843398571, 0.06424808502197266, 0.04309313744306564, 0.01329098641872406, -0.04963311925530434, -0.04238095134496689, 0.017966002225875854, -0.09614120423793793, 0.010845281183719635, -0.07721120119094849, -0.05725245550274849, -0.05320628359913826, -0.09666216373443604, 0.06654927134513855, 0.03582711145281792, 0.03519811853766441, -0.08296781778335571, -0.0653943121433258, 0.11963627487421036, 0.1308179795742035, -0.040553610771894455, 0.007065735291689634, -0.037271205335855484, 0.02229445055127144, 0.008400795981287956, -0.03149213269352913, -0.1851508617401123, -0.09589783847332001, 0.034173883497714996, -0.003432461293414235, 0.0011193393729627132, -0.030088940635323524, 0.08860398828983307, 0.08211209625005722, -0.06903701275587082, -0.025216832756996155, -0.09276971966028214, -0.018540246412158012, -0.0988534465432167, -0.1602656990289688, -0.05017821863293648, -0.03630124777555466, 0.16329306364059448, -0.2557857036590576, 0.04592754691839218, -0.07224742323160172, 0.11747664958238602, 0.060128841549158096, -0.06519296020269394, -0.01591424085199833, 0.07892908900976181, -0.002296420047059655, -0.09979922324419022, 0.06566623598337173, 0.008111159317195415, -0.031163416802883148, -0.06906729191541672, -0.14342138171195984, 0.057618413120508194, 0.06859103590250015, 0.09881517291069031, -0.1393836885690689, -0.059645142406225204, -0.0914945974946022, -0.017689868807792664, -0.11937546730041504, 0.0033280975185334682, 0.12362643331289291, 0.0150525514036417, 0.15591725707054138, -0.07469300925731659, -0.04107452929019928, 0.010796546936035156, -0.02176813967525959, 0.030902432277798653, 0.07538070529699326, 0.0526299774646759, 0.04155328869819641, 0.0861375480890274, 0.039524126797914505, -0.054218895733356476, 0.13619227707386017, -0.03945746272802353, -0.07763722538948059, 0.005113037768751383, 0.053688641637563705, 0.012774437665939331, 0.13612796366214752, 0.0036847018636763096, 0.013745282776653767, 0.016096139326691628, 0.034453049302101135, 0.02352106384932995, -0.2074512243270874, -0.0415152981877327, 0.006656024605035782, -0.0332222618162632, -0.017373399809002876, -0.01016890350729227, 0.03176961466670036, 0.09504387527704239, 0.00979529321193695, -0.02600713260471821, 0.00006221525836735964, 0.0016855646390467882, -0.09691672772169113, 0.17495658993721008, -0.11224222183227539, -0.11722279340028763, -0.1273220181465149, 0.046703144907951355, -0.07361441105604172, -0.009689277037978172, 0.025590810924768448, -0.0664920061826706, -0.025348149240016937, -0.11712746322154999, -0.010250497609376907, 0.03291483223438263, -0.02912714146077633, 0.03083202615380287, 0.01387733779847622, 0.11871226876974106, -0.11379474401473999, 0.014929922297596931, -0.024857647716999054, -0.09276361763477325, 0.01800891011953354, 0.03198800981044769, 0.10868103057146072, 0.12684845924377441, -0.000049466532800579444, 0.023276539519429207, -0.04437405616044998, 0.23869211971759796, -0.08294578641653061, -0.0007781587773934007, 0.11787894368171692, 0.01850242353975773, 0.06088859215378761, 0.10769428312778473, 0.052634432911872864, -0.12513139843940735, 0.05945044010877609, 0.09550125151872635, -0.03830830380320549, -0.1987132877111435, -0.05014587938785553, -0.014655929990112782, -0.0313873365521431, 0.11029736697673798, 0.049505945295095444, 0.032195281237363815, 0.051519282162189484, -0.0075074732303619385, 0.012542295269668102, 0.013909423723816872, 0.0818055272102356, 0.01987895928323269, 0.019438043236732483, 0.09317705780267715, -0.022968824952840805, -0.027251973748207092, 0.0656709223985672, -0.02015269733965397, 0.26726233959198, -0.041175905615091324, -0.029180580750107765, 0.04641861468553543, 0.150969997048378, -0.05135996639728546, 0.0640387237071991, -0.0009062413591891527, -0.03371889516711235, 0.009616807103157043, -0.058708325028419495, -0.0160183347761631, 0.04755930230021477, -0.10861825942993164, 0.09756559133529663, -0.10303780436515808, -0.008508941158652306, 0.04201589524745941, 0.34072890877723694, 0.01584787480533123, -0.26467880606651306, -0.08771323412656784, -0.007161501795053482, -0.015273287892341614, -0.09536504745483398, 0.033659208565950394, 0.16235359013080597, -0.13769248127937317, 0.06552591174840927, -0.09046702831983566, 0.09250553697347641, 0.05555761232972145, 0.016805503517389297, 0.0772014781832695, 0.15151160955429077, -0.03135767951607704, 0.034964703023433685, -0.2158765345811844, 0.21183691918849945, 0.012306059710681438, 0.12546272575855255, -0.03045899234712124, 0.021335655823349953, 0.05691884830594063, 0.10201221704483032, 0.10815287381410599, 0.0018433061195537448, 0.00015735447232145816, -0.21201570332050323, -0.027867654338479042, 0.01443956047296524, 0.13187702000141144, -0.04617546871304512, 0.07866205275058746, -0.06092724949121475, 0.02471538633108139, 0.029663046821951866, -0.08553048968315125, -0.15329469740390778, -0.07310549914836884, -0.010536529123783112, 0.019071491435170174, -0.03617827966809273, -0.10036803781986237, -0.09462346136569977, -0.02223355881869793, 0.08384450525045395, 0.005960887763649225, -0.03811094909906387, -0.15563741326332092, 0.017807530239224434, 0.09228253364562988, -0.025801440700888634, 0.030001984909176826, 0.027089715003967285, 0.1270042508840561, -0.0013943470548838377, -0.05504094436764717, 0.06186261400580406, -0.08647291362285614, -0.1746513545513153, -0.04596596956253052, 0.1444062739610672, 0.06415688991546631, 0.03681829199194908, -0.023872600868344307, 0.03369072452187538, 0.044084951281547546, -0.11104048788547516, 0.007363969460129738, 0.11057005822658539, 0.01364559680223465, 0.05691739544272423, -0.11217298358678818, 0.05881712585687637, -0.04751848801970482, -0.006624583620578051, 0.06008872389793396, 0.24579419195652008, -0.10556371510028839, 0.061920374631881714, 0.00041061683441512287, -0.07289520651102066, -0.20746493339538574, 0.1105492040514946, 0.1332196593284607, 0.0011079153046011925, 0.09144464880228043, -0.16556954383850098, 0.081833615899086, 0.14078070223331451, -0.03261110186576843, 0.1373748630285263, -0.32590341567993164, -0.12556438148021698, 0.028264440596103668, 0.12554721534252167, 0.010294083505868912, -0.1447698026895523, -0.04516918212175369, -0.011964193545281887, -0.09891755878925323, 0.04840194061398506, -0.1226787269115448, 0.09382858127355576, -0.01665598154067993, 0.07544457912445068, 0.033586129546165466, -0.011270391754806042, 0.14641763269901276, -0.006928345654159784, 0.12572087347507477, -0.04833703860640526, 0.05129426717758179, 0.005526657681912184, -0.06889335811138153, 0.04473045468330383, -0.002757070353254676, 0.061177950352430344, -0.13499125838279724, -0.011558996513485909, -0.10532193630933762, 0.07647549360990524, -0.052033744752407074, -0.06506367772817612, -0.029441120103001595, 0.05257236212491989, 0.033612947911024094, -0.019985876977443695, 0.021715324372053146, -0.023495670408010483, 0.21686851978302002, 0.12556399405002594, 0.05651644989848137, -0.04781504347920418, -0.06908762454986572, 0.022909382358193398, -0.018279721960425377, 0.06174101680517197, -0.15520580112934113, 0.0284146498888731, 0.11156094819307327, 0.0619189627468586, 0.13188673555850983, 0.0483718141913414, -0.04367399960756302, -0.00031110274721868336, 0.041722141206264496, -0.11326702684164047, -0.15110276639461517, -0.02065347321331501, -0.014162329025566578, -0.11851003766059875, 0.02726220339536667, 0.11255759000778198, -0.07000575959682465, -0.003361272159963846, -0.023418448865413666, 0.007290079724043608, -0.030313560739159584, 0.17735011875629425, 0.04524320363998413, 0.042081620544195175, -0.0808497816324234, 0.11374083161354065, 0.018971005454659462, -0.027115831151604652, 0.023783253505825996, 0.09665083885192871, -0.08863567560911179, -0.012028135359287262, 0.05479706823825836, 0.1303689032793045, -0.026951059699058533, -0.037367820739746094, -0.0901583656668663, -0.11253606528043747, 0.02234010212123394, 0.12114262580871582, 0.0426754467189312, -0.043843794614076614, -0.010099273175001144, 0.0459134578704834, -0.15116119384765625, 0.07129483669996262, 0.03286455571651459, 0.08019489049911499, -0.14902997016906738, 0.1658705174922943, 0.02206159569323063, 0.017401333898305893, -0.02256227843463421, 0.029473870992660522, -0.09850471466779709, -0.012125499546527863, -0.13646309077739716, -0.02216457575559616, -0.035468779504299164, 0.011041597463190556, -0.007933816872537136, -0.0323040708899498, -0.0362115241587162, 0.04250146076083183, -0.07616220414638519, -0.056658025830984116, 0.00529859121888876, 0.04519084095954895, -0.11771250516176224, -0.008694532327353954, 0.01954679936170578, -0.09104575216770172, 0.05979050695896149, 0.04833395406603813, 0.012433189898729324, 0.029252002015709877, -0.1030053049325943, 0.015376610681414604, 0.047994621098041534, -0.01595000922679901, 0.05604355037212372, -0.06538472324609756, -0.01031490694731474, -0.056600574404001236, 0.05212561786174774, 0.031453538686037064, 0.06730236113071442, -0.1343042105436325, 0.012800503522157669, -0.03760944679379463, -0.053135938942432404, -0.04952255263924599, 0.014255720190703869, 0.09533862769603729, 0.04747738689184189, 0.1295165866613388, -0.10605233907699585, 0.05835102126002312, -0.20503829419612885, -0.05466274917125702, -0.018448345363140106, -0.02565068192780018, -0.06333168596029282, -0.007317978423088789, 0.0742013230919838, -0.03168681263923645, 0.08937270194292068, 0.007409828715026379, 0.11865340173244476, 0.0284064132720232, -0.09673397243022919, 0.004742861725389957, -0.006653213407844305, 0.15294305980205536, 0.04511589556932449, -0.030499527230858803, 0.07466268539428711, 0.02089625783264637, 0.047692250460386276, 0.04206511378288269, 0.17333245277404785, 0.13896554708480835, -0.05420858412981033, 0.051064372062683105, 0.06425996124744415, -0.11084936559200287, -0.09076535701751709, 0.05916048586368561, 0.015642808750271797, 0.07548646628856659, -0.07194669544696808, 0.16154465079307556, 0.10948886722326279, -0.19694499671459198, 0.05346917733550072, -0.050157446414232254, -0.0806504637002945, -0.13374114036560059, 0.025262819603085518, -0.06066436320543289, -0.16516956686973572, 0.01719577983021736, -0.11335159838199615, 0.01820780523121357, 0.11651502549648285, -0.018564973026514053, 0.029769517481327057, 0.19469282031059265, -0.06302327662706375, 0.007441941183060408, 0.0811486691236496, 0.015799839049577713, 0.013491678051650524, -0.07783686369657516, -0.09674079716205597, 0.0529058463871479, -0.052376069128513336, 0.04784424602985382, -0.06222204491496086, -0.030631205067038536, 0.010205211117863655, -0.02786438912153244, -0.041962914168834686, 0.046972617506980896, 0.0329354964196682, 0.004632242023944855, 0.052450116723775864, 0.07092887908220291, -0.012575841508805752, -0.042003560811281204, 0.3257961571216583, -0.1024424359202385, -0.04400232806801796, -0.15921510756015778, 0.2716485559940338, 0.01815987378358841, 0.00881986878812313, 0.020729172974824905, -0.1116730123758316, -0.020772580057382584, 0.20712199807167053, 0.1287730485200882, -0.08393807709217072, -0.0033709818962961435, -0.018672628328204155, -0.015186619944870472, -0.05520973354578018, 0.14390914142131805, 0.11007390916347504, 0.0026447377167642117, -0.039976950734853745, -0.0036757434718310833, 0.0016716477693989873, -0.015529833734035492, -0.08763454109430313, 0.07093456387519836, 0.0010801829630509019, 0.008065992966294289, -0.05267859995365143, 0.07614660263061523, 0.019731057807803154, -0.14351676404476166, 0.0852738544344902, -0.14462633430957794, -0.16048096120357513, -0.022529365494847298, -0.0052796620875597, -0.016482418403029442, 0.07462076842784882, -0.017998719587922096, 0.00612834095954895, 0.15263642370700836, -0.046998776495456696, -0.005115969106554985, -0.1643505096435547, 0.0505831278860569, -0.13108563423156738, 0.23022085428237915, -0.02269081212580204, 0.000058876394177787006, 0.07629937678575516, 0.011317628435790539, -0.12320160120725632, 0.06909354776144028, 0.05997845157980919, -0.05667663738131523, -0.025107048451900482, 0.1421917825937271, -0.05796753242611885, 0.07510463893413544, 0.03812742978334427, -0.17698349058628082, 0.00717900088056922, 0.02292674593627453, -0.0561482273042202, -0.08481750637292862, 0.0014734938740730286, -0.06247951462864876, 0.12435121834278107, 0.22388601303100586, -0.023768335580825806, 0.03117443062365055, -0.06661474704742432, 0.0377926230430603, 0.048284292221069336, 0.1002175435423851, -0.02354658767580986, -0.22385138273239136, 0.017923325300216675, 0.03277762979269028, -0.028435563668608665, -0.284750759601593, -0.0836203396320343, 0.00950611848384142, -0.06339745968580246, -0.021592790260910988, 0.10502063482999802, 0.030283315107226372, 0.04722929000854492, -0.023495269939303398, -0.19376887381076813, -0.026530027389526367, 0.19446147978305817, -0.09284002333879471, -0.0384848415851593 ]
null
null
diffusers
# Rin Hoshizora <Gallery /> ## Model description This model was trained to generate high quality images based on SIFAS cards. To achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement. ## Trigger words You should use `id_rin_hoshizora` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/theidoldaily/rin-hoshizora/tree/main) them in the Files & versions tab.
{"license": "mit", "tags": ["text-to-image", "stable-diffusion", "lora", "diffusers", "template:sd-lora"], "widget": [{"text": "masterpiece, high quality, defined pupil, looking at viewer, rounded pupil, defined iris, (soft iris:1.2),", "parameters": {"negative_prompt": "bad_anatomy, deformation, amputation, deformity, deformed_nipples, duplicated_torso, deformed_torso, long_torso, large_torso, unproportioned_torso, (deformed_pussy:1.2), (deformed_hands:1.2), unproportioned_eyes, unproportioned_head, small_head, duplicated_nose, big_nose, fusioned_clothes, fusioned_arms, undefined_limbs, divided_pussy, red_pussy, duplicated_pussy, deformed_anus, deformed_pussy,"}, "output": {"url": "images/rinchan.png"}}], "base_model": "cagliostrolab/animagine-xl-3.0", "instance_prompt": "id_rin_hoshizora"}
text-to-image
theidoldaily/rin-hoshizora
[ "diffusers", "text-to-image", "stable-diffusion", "lora", "template:sd-lora", "base_model:cagliostrolab/animagine-xl-3.0", "license:mit", "region:us" ]
2024-02-15T04:23:27+00:00
[]
[]
TAGS #diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-cagliostrolab/animagine-xl-3.0 #license-mit #region-us
# Rin Hoshizora <Gallery /> ## Model description This model was trained to generate high quality images based on SIFAS cards. To achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement. ## Trigger words You should use 'id_rin_hoshizora' to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. Download them in the Files & versions tab.
[ "# Rin Hoshizora\n\n<Gallery />", "## Model description \n\nThis model was trained to generate high quality images based on SIFAS cards.\n\nTo achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement.", "## Trigger words\n\nYou should use 'id_rin_hoshizora' to trigger the image generation.", "## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab." ]
[ "TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-cagliostrolab/animagine-xl-3.0 #license-mit #region-us \n", "# Rin Hoshizora\n\n<Gallery />", "## Model description \n\nThis model was trained to generate high quality images based on SIFAS cards.\n\nTo achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement.", "## Trigger words\n\nYou should use 'id_rin_hoshizora' to trigger the image generation.", "## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab." ]
[ 56, 10, 68, 22, 28 ]
[ "passage: TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-cagliostrolab/animagine-xl-3.0 #license-mit #region-us \n# Rin Hoshizora\n\n<Gallery />## Model description \n\nThis model was trained to generate high quality images based on SIFAS cards.\n\nTo achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement.## Trigger words\n\nYou should use 'id_rin_hoshizora' to trigger the image generation.## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab." ]
[ -0.042151596397161484, -0.10972780734300613, -0.0010352557292208076, 0.018069125711917877, 0.08698146045207977, 0.044452935457229614, 0.2414029836654663, 0.06479998677968979, 0.15730713307857513, 0.058585479855537415, -0.006779599469155073, 0.08691844344139099, 0.05924077704548836, 0.2329532355070114, -0.05202837288379669, -0.2739579379558563, 0.10806330293416977, -0.039385322481393814, -0.00025025461218319833, -0.00515741715207696, 0.07033749669790268, -0.05617137625813484, 0.12867408990859985, -0.03535938635468483, -0.03019089810550213, 0.012039349414408207, -0.02242392674088478, -0.011964578181505203, 0.03428177908062935, 0.07813587039709091, 0.02929895929992199, 0.10028225928544998, 0.10109786689281464, -0.148649662733078, 0.0694982185959816, 0.011838669888675213, -0.023715512827038765, 0.006927106995135546, -0.0015941555611789227, -0.030112048611044884, 0.23648382723331451, -0.12241067737340927, -0.11492166668176651, -0.030346769839525223, -0.03226734697818756, -0.03583990037441254, -0.04077407717704773, -0.0409262552857399, 0.09139715880155563, -0.008964882232248783, 0.011478288099169731, 0.008747831918299198, -0.008608407340943813, 0.04109865054488182, 0.21433135867118835, -0.12267066538333893, -0.06398601830005646, 0.21728669106960297, 0.025734011083841324, 0.15867386758327484, -0.04906909167766571, 0.14284898340702057, 0.12459485977888107, -0.07432709634304047, 0.015824411064386368, -0.05931594595313072, 0.05171306058764458, -0.025450821965932846, -0.06731721013784409, 0.017272744327783585, 0.2427680939435959, 0.03047759085893631, -0.041948772966861725, -0.10699236392974854, -0.07364758104085922, 0.12802743911743164, -0.08234599977731705, -0.0020940566901117563, 0.026617543771862984, 0.015443684533238411, 0.07323715090751648, -0.08788648992776871, -0.09269276261329651, -0.0518764965236187, -0.03139957785606384, 0.18229277431964874, 0.06353769451379776, 0.06358391791582108, -0.019187547266483307, 0.06637915968894958, -0.07881922274827957, -0.15351490676403046, 0.0032938688527792692, -0.0728263184428215, 0.03095090575516224, 0.06754519045352936, 0.04033231735229492, -0.07465027272701263, 0.14086173474788666, -0.0617145337164402, 0.04996052756905556, 0.0063062445260584354, -0.0214975718408823, 0.0874674841761589, 0.06343740224838257, -0.012515133246779442, -0.030202412977814674, -0.050495415925979614, 0.04130303114652634, 0.03191579878330231, 0.09800957888364792, -0.08096638321876526, -0.11860133707523346, -0.03819599375128746, -0.07121556252241135, 0.012760470621287823, -0.05684211850166321, 0.03134658932685852, 0.008092762902379036, -0.004105428233742714, 0.03951068967580795, -0.04474151134490967, -0.06922779232263565, -0.06771327555179596, -0.015794888138771057, 0.24200047552585602, 0.0299025047570467, -0.010826562531292439, 0.034168682992458344, 0.10095444321632385, -0.03504808992147446, -0.018717002123594284, -0.06931745260953903, -0.017745325341820717, 0.0007887209067121148, -0.1713458150625229, 0.026414548978209496, -0.14631769061088562, -0.24811725318431854, -0.018084179610013962, 0.06962009519338608, -0.04674723744392395, 0.07920520752668381, -0.006909096613526344, 0.009610896930098534, -0.004468481056392193, 0.0024132465478032827, -0.004359448328614235, -0.091025710105896, 0.07102286070585251, -0.018453547731041908, 0.09895289689302444, -0.1568995863199234, 0.004143227357417345, -0.05214466154575348, 0.06919097155332565, -0.1494475156068802, 0.033143993467092514, -0.08077087998390198, 0.04442476108670235, -0.04781491681933403, -0.022168796509504318, -0.1279173195362091, 0.09257509559392929, -0.027301494032144547, 0.144606351852417, -0.19272467494010925, -0.0206560418009758, 0.028875943273305893, -0.23870696127414703, -0.06356024742126465, 0.10257083922624588, -0.03313395380973816, 0.21410322189331055, 0.0744778960943222, 0.194702610373497, 0.06264302134513855, -0.23547594249248505, 0.056831348687410355, 0.05008736997842789, -0.0611923485994339, -0.07678374648094177, 0.07415587455034256, 0.11375647783279419, 0.03382620960474014, 0.09200803935527802, -0.06872542947530746, 0.0582352839410305, -0.05831482633948326, -0.04145945608615875, -0.04351523891091347, -0.054068710654973984, -0.08931262046098709, 0.026541011407971382, 0.03623437136411667, 0.02642078325152397, 0.02593960613012314, -0.05442218482494354, 0.15892532467842102, -0.054727304726839066, 0.00259710312820971, -0.03469758853316307, 0.18851065635681152, -0.09157469123601913, 0.018748829141259193, -0.005536951124668121, -0.046252693980932236, -0.025934293866157532, 0.023750342428684235, 0.03919629752635956, 0.0757443979382515, 0.045067593455314636, 0.04822777956724167, -0.04683726653456688, 0.028537558391690254, 0.1120942011475563, -0.027967575937509537, -0.005567153915762901, -0.13355140388011932, -0.01603061705827713, -0.03224720433354378, 0.11459267884492874, -0.16197866201400757, -0.014906466007232666, -0.01125500351190567, 0.0988125428557396, 0.016921471804380417, 0.07582076638936996, 0.022352030500769615, -0.02951502613723278, -0.06706082820892334, -0.03203084319829941, 0.03267358988523483, 0.002445578807964921, -0.01866019144654274, 0.11309874057769775, -0.07711716741323471, 0.11106676608324051, 0.12198521196842194, 0.06189202144742012, 0.04041766747832298, -0.15986473858356476, 0.02055053599178791, -0.0017340626800432801, -0.07048612833023071, 0.0003494397969916463, -0.11961710453033447, -0.01844903640449047, 0.05483027175068855, -0.08079670369625092, 0.11943406611680984, 0.04556598886847496, -0.0790332704782486, -0.05724845826625824, 0.07785012573003769, 0.11134898662567139, -0.07521864771842957, 0.07348289340734482, 0.06308136135339737, -0.0680849701166153, 0.18737885355949402, -0.05039338394999504, -0.13656450808048248, 0.003874585498124361, 0.07451124489307404, -0.002166648395359516, 0.16768793761730194, 0.07784747332334518, -0.04575373977422714, 0.04110271483659744, -0.05541335046291351, 0.044600386172533035, -0.09365478903055191, -0.05055226758122444, -0.0009433514787815511, -0.05433907359838486, 0.10387129336595535, 0.1394999772310257, -0.06551052629947662, 0.0932585820555687, -0.07846410572528839, -0.08215255290269852, -0.033299192786216736, -0.01981886476278305, -0.0064890217036008835, 0.06623595207929611, 0.0498172789812088, -0.08633535355329514, -0.1849565953016281, -0.0713089257478714, -0.08886074274778366, 0.003265797859057784, 0.062176208943128586, -0.059344831854104996, -0.0806671753525734, -0.08266735076904297, 0.029936401173472404, 0.053625766187906265, -0.07605122774839401, -0.06128014624118805, -0.010779928416013718, -0.09394655376672745, -0.08143704384565353, -0.05583053082227707, -0.08262196183204651, 0.025126254186034203, 0.1143919825553894, -0.13292686641216278, 0.1795720010995865, 0.03292020037770271, 0.037172503769397736, 0.057098109275102615, -0.0017691891407594085, 0.05504591017961502, -0.06771038472652435, 0.07320291548967361, 0.27555012702941895, 0.05186796560883522, 0.058270879089832306, 0.09423505514860153, -0.00965026579797268, -0.100356824696064, 0.026569226756691933, 0.007404942065477371, -0.15192793309688568, -0.09961015731096268, -0.10447663068771362, -0.09502202272415161, -0.04583040624856949, -0.003546270076185465, 0.03295904025435448, 0.08194633573293686, 0.17615891993045807, -0.002262425608932972, -0.008443801663815975, 0.04249770939350128, 0.07983306050300598, 0.007518578786402941, 0.019265592098236084, 0.059410519897937775, -0.09908249974250793, -0.07128145545721054, 0.15514187514781952, 0.0376909002661705, 0.233827605843544, -0.04147394746541977, 0.0903027355670929, 0.1006845235824585, 0.08616535365581512, 0.09600800275802612, 0.09151235222816467, 0.034588899463415146, -0.012420409359037876, -0.08285600692033768, -0.1145472601056099, -0.021045120432972908, 0.10349877923727036, -0.1244719848036766, -0.05466293543577194, 0.012868757359683514, 0.15453065931797028, 0.028368128463625908, 0.12775008380413055, -0.029842741787433624, -0.27415573596954346, 0.0699746236205101, 0.09213630110025406, 0.1217295452952385, -0.06358781456947327, 0.04462926834821701, 0.1074788048863411, -0.025052135810256004, 0.016859007999300957, -0.01633867621421814, 0.0844518169760704, -0.09973065555095673, -0.014746708795428276, -0.11782804131507874, 0.14931099116802216, -0.01589067280292511, 0.05079610273241997, -0.11990015208721161, 0.14461424946784973, -0.013102284632623196, 0.015336807817220688, -0.05563565343618393, -0.04914926737546921, 0.1270451843738556, 0.06401657313108444, 0.17951031029224396, 0.001955924089998007, -0.09196093678474426, -0.09545614570379257, -0.1016286313533783, 0.04746845364570618, 0.06331196427345276, 0.0015071263769641519, 0.041294604539871216, -0.06253524124622345, 0.011243690736591816, -0.008256150409579277, 0.09049271047115326, -0.09341483563184738, -0.11169794946908951, 0.015670739114284515, 0.10923541337251663, 0.012530087493360043, -0.025403015315532684, -0.04681117460131645, -0.12680131196975708, 0.04503647983074188, 0.27261653542518616, -0.08237151056528091, -0.13831445574760437, -0.04175515100359917, 0.03677548095583916, -0.050046201795339584, -0.05714991316199303, -0.01713026873767376, 0.1137678399682045, -0.07292767614126205, -0.1474672257900238, -0.03340566158294678, -0.02576088346540928, 0.026885386556386948, -0.002614798257127404, 0.058994267135858536, 0.010969498194754124, -0.058874864131212234, 0.02862399071455002, -0.03596241772174835, 0.04214966297149658, -0.07791498303413391, -0.00813040230423212, 0.13610418140888214, 0.00031306667369790375, 0.052543122321367264, -0.11599265038967133, -0.007021816447377205, -0.048407383263111115, -0.03978806361556053, -0.01831652596592903, 0.1715995818376541, -0.0043804831802845, 0.024257954210042953, 0.1414816826581955, -0.05612844228744507, -0.2224946767091751, -0.06553157418966293, -0.09840173274278641, 0.01515877339988947, -0.04349065572023392, -0.050183095037937164, 0.1230502724647522, 0.1177784651517868, -0.03837120160460472, 0.16241762042045593, -0.23243050277233124, -0.08204521238803864, 0.0661601722240448, 0.15519598126411438, 0.40235960483551025, -0.24345150589942932, -0.020255839452147484, -0.14627885818481445, -0.19037826359272003, -0.018538974225521088, -0.128115713596344, 0.06234662979841232, -0.010525890626013279, -0.0015757792862132192, -0.0007256612880155444, -0.02549390122294426, 0.20431318879127502, -0.0262411180883646, 0.1423650085926056, -0.09409469366073608, -0.00898868776857853, 0.11359146982431412, -0.08646571636199951, 0.10255569964647293, -0.12466764450073242, 0.008645644411444664, -0.06463419646024704, -0.048737186938524246, 0.01247725822031498, 0.05117752030491829, 0.021243080496788025, -0.07367721945047379, -0.06729509681463242, 0.05802738294005394, 0.02188165858387947, 0.025889264419674873, 0.11545072495937347, -0.07628477364778519, -0.00865296833217144, 0.025876881554722786, -0.030436890199780464, 0.0017767002573236823, -0.0003558089956641197, -0.09161306172609329, -0.03234736993908882, 0.08569493889808655, -0.14897537231445312, -0.012841630727052689, 0.10549064725637436, 0.01704888977110386, 0.11164118349552155, 0.006734191440045834, -0.0013897534226998687, 0.0722828134894371, 0.11658583581447601, -0.07910587638616562, -0.1202765479683876, -0.033214665949344635, -0.045866675674915314, 0.134446382522583, 0.005733699072152376, 0.1260499209165573, -0.05201656371355057, 0.05886933580040932, -0.0072722360491752625, 0.029054775834083557, -0.023148201406002045, 0.016397766768932343, -0.011972345411777496, -0.02676512859761715, -0.11001468449831009, 0.0910327285528183, -0.018881723284721375, 0.04085796698927879, -0.048979293555021286, 0.002315604593604803, -0.10446786135435104, -0.05055272951722145, -0.022893816232681274, 0.14354735612869263, -0.1127210482954979, -0.09994199126958847, -0.07677941769361496, -0.1190633550286293, -0.07065140455961227, 0.07559701800346375, 0.10356102883815765, -0.06732269376516342, 0.029418639838695526, 0.0030662326607853174, -0.06879004091024399, 0.007315218914300203, 0.08030415326356888, 0.07292461395263672, -0.17459529638290405, -0.10242121666669846, -0.026671139523386955, 0.023401889950037003, -0.10718473792076111, -0.07024568319320679, -0.052264150232076645, 0.008476298302412033, -0.17978370189666748, 0.1456298828125, -0.11742715537548065, -0.044603656977415085, -0.010402833111584187, -0.051509495824575424, -0.038193561136722565, 0.004957279656082392, -0.07151569426059723, 0.037290431559085846, 0.013184582814574242, 0.08194459974765778, -0.029223531484603882, -0.06479854136705399, 0.010705606080591679, -0.08377495408058167, -0.013608965091407299, 0.07836734503507614, -0.05151084437966347, 0.03833889216184616, -0.22411172091960907, -0.012947266921401024, 0.11979059875011444, 0.02439911849796772, -0.0164994727820158, 0.007828923873603344, 0.017533201724290848, 0.04379505291581154, 0.022294213995337486, -0.042455073446035385, -0.08844765275716782, -0.0682540237903595, 0.116053007543087, -0.10837282985448837, -0.01603107713162899, -0.008633355610072613, 0.023446254432201385, 0.18814179301261902, 0.1317286640405655, 0.08821170777082443, -0.027927789837121964, -0.01500711403787136, -0.06739875674247742, 0.028416471555829048, 0.03682277351617813, -0.12115687876939774, -0.019219806417822838, -0.1513419896364212, -0.015746671706438065, -0.018421441316604614, 0.16309486329555511, 0.06643565744161606, -0.1099763736128807, -0.05587783455848694, 0.024043245241045952, 0.07077492028474808, -0.012421702966094017, 0.24308453500270844, 0.085372194647789, 0.11180058866739273, -0.06516328454017639, 0.03301634639501572, 0.09791591763496399, 0.0504804365336895, 0.047171302139759064, 0.17551948130130768, 0.015478530898690224, 0.12455129623413086, 0.012472246773540974, 0.020907603204250336, 0.04187211021780968, 0.05800075829029083, -0.07665877044200897, 0.009974086657166481, 0.005932461936026812, -0.01144969742745161, 0.23412978649139404, -0.09863413125276566, 0.008092425763607025, 0.04973486810922623, 0.0014975789235904813, -0.11257314682006836, -0.2710762321949005, -0.10076542943716049, -0.14913786947727203, 0.05443352460861206, -0.0448492094874382, 0.012408728711307049, 0.21659301221370697, 0.041845161467790604, 0.0333249494433403, 0.04954161494970322, -0.07659101486206055, -0.048339854925870895, 0.08445844799280167, -0.03998350352048874, -0.08921349793672562, -0.011369295418262482, -0.09559211134910583, 0.04828908294439316, -0.045581307262182236, -0.03690090402960777, 0.026921061798930168, 0.13030461966991425, 0.03349332511425018, -0.07398250699043274, -0.036035701632499695, -0.043399401009082794, 0.012404314242303371, -0.01135692186653614, 0.09857281297445297, 0.049792397767305374, -0.051031023263931274, 0.01032213680446148, 0.15223298966884613, 0.0058388374745845795, -0.08127976953983307, -0.06566605716943741, 0.09329886734485626, -0.07479051500558853, 0.03855515643954277, -0.03824761137366295, -0.0424814373254776, 0.006591080222278833, 0.2385985553264618, 0.15797534584999084, -0.06962862610816956, 0.004658734891563654, -0.06992925703525543, 0.011802742257714272, -0.0480877123773098, 0.11088643223047256, 0.013039768673479557, 0.18567170202732086, -0.04345686733722687, 0.029513750225305557, -0.10123678296804428, -0.04682226851582527, -0.05799466371536255, -0.0004778676084242761, -0.017914926633238792, -0.06928133219480515, -0.05455678328871727, 0.06933967769145966, -0.12195398658514023, -0.10973766446113586, 0.08310303837060928, -0.013920875266194344, -0.01550461258739233, -0.05243989825248718, -0.04149205982685089, 0.06047332286834717, 0.006121805869042873, -0.1563483327627182, 0.026560556143522263, -0.028116578236222267, 0.02684410661458969, -0.12100311368703842, -0.00580228166654706, 0.00968519039452076, -0.05080074444413185, 0.07131903618574142, -0.012967747636139393, 0.012165895663201809, -0.0032073857728391886, -0.03227753937244415, -0.014302113093435764, 0.12794312834739685, -0.024437308311462402, -0.12681181728839874, -0.03498923406004906, 0.0362810418009758, -0.08248140662908554, -0.014384903013706207, 0.061894066631793976, -0.07896562665700912, -0.0030938792042434216, 0.1547117680311203, -0.054650500416755676, -0.04794582724571228, 0.038487326353788376, -0.15757779777050018, 0.10718952119350433, 0.05982556194067001, 0.04024488478899002, -0.016918132081627846, -0.031056685373187065, 0.11895900964736938, 0.052509408444166183, -0.09417164325714111, 0.036830171942710876, -0.020621823146939278, -0.10046625137329102, -0.05857933685183525, 0.013301604427397251, -0.1827915608882904, -0.0268184095621109, -0.16236679255962372, -0.030645934864878654, -0.019573304802179337, 0.06836811453104019, 0.23495447635650635, 0.018604546785354614, -0.013319271616637707, -0.23601162433624268, 0.03207838162779808, 0.08013457804918289, -0.044972170144319534, -0.0533745214343071 ]
null
null
null
# Lora of georgia/ジョージア/佐治亚 (Azur Lane) ## What Is This? This is the LoRA model of waifu georgia/ジョージア/佐治亚 (Azur Lane). ## How Is It Trained? * This model is trained with [HCP-Diffusion](https://github.com/7eu7d7/HCP-Diffusion). * The [auto-training framework](https://github.com/deepghs/cyberharem) is maintained by [DeepGHS Team](https://huggingface.co/deepghs). * The base model used for training is [deepghs/animefull-latest](https://huggingface.co/deepghs/animefull-latest). * Dataset used for training is the `stage3-p480-800` in [CyberHarem/georgia_azurlane](https://huggingface.co/datasets/CyberHarem/georgia_azurlane), which contains 76 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 800 steps, 40 checkpoints were saved and evaluated. * **Trigger word is `georgia_azurlane`.** * Pruned core tags for this waifu are `breasts, blue_eyes, earrings, black_hair, large_breasts, bangs, heterochromia, hair_ornament, yellow_eyes, long_hair, hair_between_eyes, star_earrings`. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. ## How to Use It? ### If You Are Using A1111 WebUI v1.7+ **Just use it like the classic LoRA**. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 620, you need to download [`620/georgia_azurlane.pt`](https://huggingface.co/CyberHarem/georgia_azurlane/resolve/main/620/georgia_azurlane.pt) as the embedding and [`620/georgia_azurlane.safetensors`](https://huggingface.co/CyberHarem/georgia_azurlane/resolve/main/620/georgia_azurlane.safetensors) for loading Lora. By using both files together, you can generate images for the desired characters. ## Which Step Should I Use? We selected 5 good steps for you to choose. The best one is step 620. 1520 images (1.52 GiB) were generated for auto-testing. ![Metrics Plot](metrics_plot.png) The base model used for generating preview images is [Meina/MeinaMix_V11](https://huggingface.co/Meina/MeinaMix_V11). Here are the preview of the recommended steps: | Step | Epoch | CCIP | AI Corrupt | Bikini Plus | Score | Download | pattern_0 | pattern_1 | pattern_2 | portrait_0 | portrait_1 | portrait_2 | full_body_0 | full_body_1 | profile_0 | profile_1 | free_0 | free_1 | shorts | maid_0 | maid_1 | miko | yukata | suit | china | bikini_0 | bikini_1 | bikini_2 | sit | squat | kneel | jump | crossed_arms | angry | smile | cry | grin | n_lie_0 | n_lie_1 | n_stand_0 | n_stand_1 | n_stand_2 | n_sex_0 | n_sex_1 | |-------:|--------:|:----------|:-------------|:--------------|:----------|:-----------------------------------------------------------------------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------------|:-------------------------------------------|:-------------------------------------------|:-------------------------------------------|:---------------------------------------------|:---------------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-------------------------------|:-----------------------------------|:-------------------------------|:---------------------------------|:---------------------------------------|:---------------------------------------|:---------------------------------------|:-----------------------------|:---------------------------------|:---------------------------------|:-------------------------------|:-----------------------------------------------|:---------------------------------|:---------------------------------|:-----------------------------|:-------------------------------|:-------------------------------------|:-------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------------|:-------------------------------------|:-------------------------------------| | 620 | 33 | **0.877** | 0.939 | 0.827 | **0.807** | [Download](https://huggingface.co/CyberHarem/georgia_azurlane/resolve/main/620/georgia_azurlane.zip) | ![pattern_0](620/previews/pattern_0.png) | ![pattern_1](620/previews/pattern_1.png) | ![pattern_2](620/previews/pattern_2.png) | ![portrait_0](620/previews/portrait_0.png) | ![portrait_1](620/previews/portrait_1.png) | ![portrait_2](620/previews/portrait_2.png) | ![full_body_0](620/previews/full_body_0.png) | ![full_body_1](620/previews/full_body_1.png) | ![profile_0](620/previews/profile_0.png) | ![profile_1](620/previews/profile_1.png) | ![free_0](620/previews/free_0.png) | ![free_1](620/previews/free_1.png) | ![shorts](620/previews/shorts.png) | ![maid_0](620/previews/maid_0.png) | ![maid_1](620/previews/maid_1.png) | ![miko](620/previews/miko.png) | ![yukata](620/previews/yukata.png) | ![suit](620/previews/suit.png) | ![china](620/previews/china.png) | ![bikini_0](620/previews/bikini_0.png) | ![bikini_1](620/previews/bikini_1.png) | ![bikini_2](620/previews/bikini_2.png) | ![sit](620/previews/sit.png) | ![squat](620/previews/squat.png) | ![kneel](620/previews/kneel.png) | ![jump](620/previews/jump.png) | ![crossed_arms](620/previews/crossed_arms.png) | ![angry](620/previews/angry.png) | ![smile](620/previews/smile.png) | ![cry](620/previews/cry.png) | ![grin](620/previews/grin.png) | ![n_lie_0](620/previews/n_lie_0.png) | ![n_lie_1](620/previews/n_lie_1.png) | ![n_stand_0](620/previews/n_stand_0.png) | ![n_stand_1](620/previews/n_stand_1.png) | ![n_stand_2](620/previews/n_stand_2.png) | ![n_sex_0](620/previews/n_sex_0.png) | ![n_sex_1](620/previews/n_sex_1.png) | | 520 | 28 | 0.802 | 0.935 | 0.837 | 0.766 | [Download](https://huggingface.co/CyberHarem/georgia_azurlane/resolve/main/520/georgia_azurlane.zip) | ![pattern_0](520/previews/pattern_0.png) | ![pattern_1](520/previews/pattern_1.png) | ![pattern_2](520/previews/pattern_2.png) | ![portrait_0](520/previews/portrait_0.png) | ![portrait_1](520/previews/portrait_1.png) | ![portrait_2](520/previews/portrait_2.png) | ![full_body_0](520/previews/full_body_0.png) | ![full_body_1](520/previews/full_body_1.png) | ![profile_0](520/previews/profile_0.png) | ![profile_1](520/previews/profile_1.png) | ![free_0](520/previews/free_0.png) | ![free_1](520/previews/free_1.png) | ![shorts](520/previews/shorts.png) | ![maid_0](520/previews/maid_0.png) | ![maid_1](520/previews/maid_1.png) | ![miko](520/previews/miko.png) | ![yukata](520/previews/yukata.png) | ![suit](520/previews/suit.png) | ![china](520/previews/china.png) | ![bikini_0](520/previews/bikini_0.png) | ![bikini_1](520/previews/bikini_1.png) | ![bikini_2](520/previews/bikini_2.png) | ![sit](520/previews/sit.png) | ![squat](520/previews/squat.png) | ![kneel](520/previews/kneel.png) | ![jump](520/previews/jump.png) | ![crossed_arms](520/previews/crossed_arms.png) | ![angry](520/previews/angry.png) | ![smile](520/previews/smile.png) | ![cry](520/previews/cry.png) | ![grin](520/previews/grin.png) | ![n_lie_0](520/previews/n_lie_0.png) | ![n_lie_1](520/previews/n_lie_1.png) | ![n_stand_0](520/previews/n_stand_0.png) | ![n_stand_1](520/previews/n_stand_1.png) | ![n_stand_2](520/previews/n_stand_2.png) | ![n_sex_0](520/previews/n_sex_0.png) | ![n_sex_1](520/previews/n_sex_1.png) | | 560 | 30 | 0.786 | 0.925 | 0.832 | 0.747 | [Download](https://huggingface.co/CyberHarem/georgia_azurlane/resolve/main/560/georgia_azurlane.zip) | ![pattern_0](560/previews/pattern_0.png) | ![pattern_1](560/previews/pattern_1.png) | ![pattern_2](560/previews/pattern_2.png) | ![portrait_0](560/previews/portrait_0.png) | ![portrait_1](560/previews/portrait_1.png) | ![portrait_2](560/previews/portrait_2.png) | ![full_body_0](560/previews/full_body_0.png) | ![full_body_1](560/previews/full_body_1.png) | ![profile_0](560/previews/profile_0.png) | ![profile_1](560/previews/profile_1.png) | ![free_0](560/previews/free_0.png) | ![free_1](560/previews/free_1.png) | ![shorts](560/previews/shorts.png) | ![maid_0](560/previews/maid_0.png) | ![maid_1](560/previews/maid_1.png) | ![miko](560/previews/miko.png) | ![yukata](560/previews/yukata.png) | ![suit](560/previews/suit.png) | ![china](560/previews/china.png) | ![bikini_0](560/previews/bikini_0.png) | ![bikini_1](560/previews/bikini_1.png) | ![bikini_2](560/previews/bikini_2.png) | ![sit](560/previews/sit.png) | ![squat](560/previews/squat.png) | ![kneel](560/previews/kneel.png) | ![jump](560/previews/jump.png) | ![crossed_arms](560/previews/crossed_arms.png) | ![angry](560/previews/angry.png) | ![smile](560/previews/smile.png) | ![cry](560/previews/cry.png) | ![grin](560/previews/grin.png) | ![n_lie_0](560/previews/n_lie_0.png) | ![n_lie_1](560/previews/n_lie_1.png) | ![n_stand_0](560/previews/n_stand_0.png) | ![n_stand_1](560/previews/n_stand_1.png) | ![n_stand_2](560/previews/n_stand_2.png) | ![n_sex_0](560/previews/n_sex_0.png) | ![n_sex_1](560/previews/n_sex_1.png) | | 320 | 17 | 0.758 | **0.958** | **0.848** | 0.741 | [Download](https://huggingface.co/CyberHarem/georgia_azurlane/resolve/main/320/georgia_azurlane.zip) | ![pattern_0](320/previews/pattern_0.png) | ![pattern_1](320/previews/pattern_1.png) | ![pattern_2](320/previews/pattern_2.png) | ![portrait_0](320/previews/portrait_0.png) | ![portrait_1](320/previews/portrait_1.png) | ![portrait_2](320/previews/portrait_2.png) | ![full_body_0](320/previews/full_body_0.png) | ![full_body_1](320/previews/full_body_1.png) | ![profile_0](320/previews/profile_0.png) | ![profile_1](320/previews/profile_1.png) | ![free_0](320/previews/free_0.png) | ![free_1](320/previews/free_1.png) | ![shorts](320/previews/shorts.png) | ![maid_0](320/previews/maid_0.png) | ![maid_1](320/previews/maid_1.png) | ![miko](320/previews/miko.png) | ![yukata](320/previews/yukata.png) | ![suit](320/previews/suit.png) | ![china](320/previews/china.png) | ![bikini_0](320/previews/bikini_0.png) | ![bikini_1](320/previews/bikini_1.png) | ![bikini_2](320/previews/bikini_2.png) | ![sit](320/previews/sit.png) | ![squat](320/previews/squat.png) | ![kneel](320/previews/kneel.png) | ![jump](320/previews/jump.png) | ![crossed_arms](320/previews/crossed_arms.png) | ![angry](320/previews/angry.png) | ![smile](320/previews/smile.png) | ![cry](320/previews/cry.png) | ![grin](320/previews/grin.png) | ![n_lie_0](320/previews/n_lie_0.png) | ![n_lie_1](320/previews/n_lie_1.png) | ![n_stand_0](320/previews/n_stand_0.png) | ![n_stand_1](320/previews/n_stand_1.png) | ![n_stand_2](320/previews/n_stand_2.png) | ![n_sex_0](320/previews/n_sex_0.png) | ![n_sex_1](320/previews/n_sex_1.png) | | 440 | 24 | 0.778 | 0.906 | 0.832 | 0.739 | [Download](https://huggingface.co/CyberHarem/georgia_azurlane/resolve/main/440/georgia_azurlane.zip) | ![pattern_0](440/previews/pattern_0.png) | ![pattern_1](440/previews/pattern_1.png) | ![pattern_2](440/previews/pattern_2.png) | ![portrait_0](440/previews/portrait_0.png) | ![portrait_1](440/previews/portrait_1.png) | ![portrait_2](440/previews/portrait_2.png) | ![full_body_0](440/previews/full_body_0.png) | ![full_body_1](440/previews/full_body_1.png) | ![profile_0](440/previews/profile_0.png) | ![profile_1](440/previews/profile_1.png) | ![free_0](440/previews/free_0.png) | ![free_1](440/previews/free_1.png) | ![shorts](440/previews/shorts.png) | ![maid_0](440/previews/maid_0.png) | ![maid_1](440/previews/maid_1.png) | ![miko](440/previews/miko.png) | ![yukata](440/previews/yukata.png) | ![suit](440/previews/suit.png) | ![china](440/previews/china.png) | ![bikini_0](440/previews/bikini_0.png) | ![bikini_1](440/previews/bikini_1.png) | ![bikini_2](440/previews/bikini_2.png) | ![sit](440/previews/sit.png) | ![squat](440/previews/squat.png) | ![kneel](440/previews/kneel.png) | ![jump](440/previews/jump.png) | ![crossed_arms](440/previews/crossed_arms.png) | ![angry](440/previews/angry.png) | ![smile](440/previews/smile.png) | ![cry](440/previews/cry.png) | ![grin](440/previews/grin.png) | ![n_lie_0](440/previews/n_lie_0.png) | ![n_lie_1](440/previews/n_lie_1.png) | ![n_stand_0](440/previews/n_stand_0.png) | ![n_stand_1](440/previews/n_stand_1.png) | ![n_stand_2](440/previews/n_stand_2.png) | ![n_sex_0](440/previews/n_sex_0.png) | ![n_sex_1](440/previews/n_sex_1.png) | ## Anything Else? Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. ## All Steps We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * [Steps From 620 to 800](all/0.md) * [Steps From 420 to 600](all/1.md) * [Steps From 220 to 400](all/2.md) * [Steps From 20 to 200](all/3.md)
{"license": "mit", "tags": ["art", "not-for-all-audiences"], "datasets": ["CyberHarem/georgia_azurlane"], "pipeline_tag": "text-to-image"}
text-to-image
CyberHarem/georgia_azurlane
[ "art", "not-for-all-audiences", "text-to-image", "dataset:CyberHarem/georgia_azurlane", "license:mit", "region:us" ]
2024-02-15T04:24:51+00:00
[]
[]
TAGS #art #not-for-all-audiences #text-to-image #dataset-CyberHarem/georgia_azurlane #license-mit #region-us
Lora of georgia/ジョージア/佐治亚 (Azur Lane) ===================================== What Is This? ------------- This is the LoRA model of waifu georgia/ジョージア/佐治亚 (Azur Lane). How Is It Trained? ------------------ * This model is trained with HCP-Diffusion. * The auto-training framework is maintained by DeepGHS Team. * The base model used for training is deepghs/animefull-latest. * Dataset used for training is the 'stage3-p480-800' in CyberHarem/georgia\_azurlane, which contains 76 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 800 steps, 40 checkpoints were saved and evaluated. * Trigger word is 'georgia\_azurlane'. * Pruned core tags for this waifu are 'breasts, blue\_eyes, earrings, black\_hair, large\_breasts, bangs, heterochromia, hair\_ornament, yellow\_eyes, long\_hair, hair\_between\_eyes, star\_earrings'. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. How to Use It? -------------- ### If You Are Using A1111 WebUI v1.7+ Just use it like the classic LoRA. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 620, you need to download '620/georgia\_azurlane.pt' as the embedding and '620/georgia\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters. Which Step Should I Use? ------------------------ We selected 5 good steps for you to choose. The best one is step 620. 1520 images (1.52 GiB) were generated for auto-testing. !Metrics Plot The base model used for generating preview images is Meina/MeinaMix\_V11. Here are the preview of the recommended steps: Anything Else? -------------- Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. All Steps --------- We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * Steps From 620 to 800 * Steps From 420 to 600 * Steps From 220 to 400 * Steps From 20 to 200
[ "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 620, you need to download '620/georgia\\_azurlane.pt' as the embedding and '620/georgia\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 620.\n\n\n1520 images (1.52 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 620 to 800\n* Steps From 420 to 600\n* Steps From 220 to 400\n* Steps From 20 to 200" ]
[ "TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/georgia_azurlane #license-mit #region-us \n", "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 620, you need to download '620/georgia\\_azurlane.pt' as the embedding and '620/georgia\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 620.\n\n\n1520 images (1.52 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 620 to 800\n* Steps From 420 to 600\n* Steps From 220 to 400\n* Steps From 20 to 200" ]
[ 45, 38, 471 ]
[ "passage: TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/georgia_azurlane #license-mit #region-us \n### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file." ]
[ 0.010862302966415882, -0.011728858575224876, -0.004239538684487343, 0.07879064232110977, 0.08760734647512436, 0.08625081181526184, 0.22327478229999542, 0.06784671545028687, 0.16002801060676575, -0.0659618228673935, 0.10004977881908417, 0.046322356909513474, -0.013703989796340466, 0.0407056100666523, -0.02483169175684452, -0.15382656455039978, -0.0759129598736763, -0.018453439697623253, 0.005333933513611555, 0.018293047323822975, 0.07556047290563583, 0.0035304485354572535, 0.10306437313556671, -0.05644423887133598, -0.05764574185013771, 0.061910152435302734, -0.023808317258954048, -0.030258087441325188, 0.02875382825732231, 0.062274061143398285, 0.10396822541952133, 0.009891609661281109, 0.04939156025648117, -0.1602475941181183, 0.0635356530547142, -0.015929225832223892, -0.10557965934276581, -0.009192450903356075, 0.017850253731012344, -0.0494173988699913, 0.13020603358745575, 0.0033100470900535583, -0.11332870274782181, 0.054754842072725296, -0.1428319811820984, -0.03565269336104393, -0.06066851690411568, 0.026950445026159286, 0.1380605250597, 0.06450054794549942, 0.015523559413850307, 0.060023967176675797, -0.0353703498840332, 0.09673546999692917, 0.1250012218952179, -0.145332932472229, -0.06192299351096153, 0.09596985578536987, 0.000219187859329395, 0.13171890377998352, -0.054201848804950714, 0.11361858993768692, 0.06496554613113403, -0.04334321618080139, -0.14888405799865723, -0.0949966162443161, -0.21747183799743652, -0.014047835022211075, 0.0057954019866883755, 0.023404907435178757, 0.4296875298023224, 0.06734006851911545, 0.040557052940130234, 0.06349457055330276, -0.06461716443300247, 0.01713140681385994, -0.09511173516511917, 0.15049350261688232, 0.039311643689870834, 0.10753902047872543, -0.04992346838116646, -0.12462880462408066, -0.12694057822227478, -0.07392899692058563, -0.08967676013708115, 0.011964181438088417, 0.016891686245799065, 0.12522409856319427, -0.19705483317375183, 0.00870585534721613, -0.07171551138162613, -0.13483640551567078, 0.036805447190999985, -0.09268907457590103, 0.15476953983306885, 0.07030090689659119, -0.009235437959432602, -0.001429879805073142, 0.24466025829315186, 0.10985825955867767, 0.19565097987651825, 0.05241139605641365, -0.0932459682226181, 0.12964177131652832, 0.03270012512803078, -0.07925219088792801, 0.010236085392534733, -0.1049681156873703, 0.14213256537914276, -0.07027579098939896, 0.10847721993923187, -0.052298083901405334, -0.10635171830654144, 0.005298549309372902, -0.10195857286453247, 0.06839609891176224, 0.05402541533112526, 0.008939109742641449, -0.06414242088794708, 0.05851219967007637, 0.006680905353277922, -0.033549509942531586, -0.016012748703360558, -0.014777460135519505, -0.040906090289354324, 0.024409234523773193, 0.10725698620080948, 0.030735284090042114, 0.05904308333992958, -0.019304677844047546, -0.024815967306494713, -0.014574356377124786, -0.028177907690405846, 0.021752629429101944, 0.029871800914406776, 0.025482747703790665, 0.0901700034737587, -0.15959236025810242, -0.09558666497468948, -0.006199934054166079, 0.04708002507686615, 0.005465536843985319, 0.0906016007065773, 0.012479747645556927, 0.05608850717544556, -0.004839938133955002, -0.02115429751574993, 0.026761768385767937, -0.10351159423589706, 0.08951040357351303, -0.009526897221803665, 0.09004183113574982, -0.18327942490577698, -0.0045598335564136505, -0.05531845986843109, -0.00042769964784383774, 0.0700240507721901, 0.01406739093363285, -0.10715505480766296, 0.1514371782541275, -0.0053770472295582294, 0.07323580235242844, -0.10434188693761826, 0.050008706748485565, 0.02356085367500782, 0.09577637165784836, -0.08984317630529404, -0.00859237089753151, 0.1215413510799408, -0.13959185779094696, -0.15495190024375916, 0.09103041887283325, -0.009029830805957317, 0.027898086234927177, 0.0723869577050209, 0.1816074252128601, 0.17672161757946014, -0.18457774817943573, -0.013081535696983337, 0.04316376894712448, -0.033699799329042435, -0.08240663260221481, -0.02133910544216633, 0.09143927693367004, 0.0038571686018258333, 0.04502376168966293, -0.04050043970346451, 0.13690395653247833, -0.03568096086382866, -0.07213340699672699, -0.04282690957188606, -0.07982304692268372, -0.07867421209812164, 0.0528247095644474, -0.010266684927046299, -0.07130611687898636, 0.0022701395209878683, -0.18521347641944885, 0.1701304167509079, 0.014756116084754467, 0.013643642887473106, -0.07055293768644333, 0.12838158011436462, 0.0419381745159626, -0.0024486940819770098, -0.005121717695146799, -0.04976068064570427, -0.10391389578580856, 0.2537987530231476, 0.10006741434335709, 0.08024589717388153, 0.06162000447511673, -0.05173656716942787, -0.06312110275030136, 0.019534790888428688, 0.013206143863499165, -0.038999222218990326, 0.01886267028748989, -0.10110437124967575, 0.06850890815258026, -0.0082999924197793, 0.03302677348256111, -0.0061130160465836525, -0.026820527389645576, 0.10963346809148788, 0.02369130589067936, -0.009485751390457153, 0.0742349624633789, 0.04583358019590378, -0.02811560593545437, -0.054213616997003555, -0.004456402733922005, 0.07341726124286652, -0.007892805151641369, -0.058674246072769165, 0.012050087563693523, -0.014398042112588882, 0.014354374259710312, 0.19632168114185333, -0.20087029039859772, 0.035185348242521286, 0.0217722300440073, 0.048207901418209076, 0.04591609165072441, 0.01667894795536995, -0.02801278606057167, 0.04839324951171875, -0.02971990406513214, 0.07366682589054108, -0.002816743217408657, 0.06875661015510559, -0.023954974487423897, -0.1321287900209427, -0.02245093695819378, -0.030096232891082764, 0.14442911744117737, -0.20794276893138885, 0.06557916849851608, 0.17312227189540863, -0.10077211260795593, 0.16693523526191711, 0.000998211675323546, -0.006120848935097456, 0.008307072333991528, 0.01917971856892109, -0.001291630556806922, 0.10988418757915497, -0.09336941689252853, -0.03298589587211609, 0.0183913242071867, -0.08914722502231598, 0.0442374013364315, -0.11651704460382462, -0.10574705898761749, -0.06067498028278351, -0.03929588943719864, -0.01739506796002388, 0.03438844904303551, -0.05598807707428932, 0.06489110738039017, -0.09670037031173706, -0.06008235737681389, -0.037114471197128296, -0.08233651518821716, 0.018840404227375984, 0.017416777089238167, -0.06046386808156967, -0.13004249334335327, -0.12696342170238495, -0.09401969611644745, -0.12849467992782593, -0.004873942118138075, 0.06922013312578201, -0.12287286669015884, -0.041841644793748856, 0.014721178449690342, -0.039023492485284805, 0.08169415593147278, -0.06878986209630966, 0.022083397954702377, 0.06437397748231888, -0.038162656128406525, -0.1588316112756729, -0.012599078938364983, -0.07534530013799667, -0.056888118386268616, 0.15560570359230042, -0.16533894836902618, 0.1644727736711502, -0.0324273444712162, 0.0579853393137455, 0.06890828907489777, 0.02602308616042137, 0.12543614208698273, -0.12118884176015854, 0.06623318791389465, 0.19510957598686218, 0.05540318414568901, 0.07374102622270584, 0.11463374644517899, 0.07467583566904068, -0.1195870116353035, 0.04473424702882767, 0.06930898874998093, -0.09817248582839966, -0.08886615186929703, -0.07720191776752472, -0.11314830929040909, -0.043707769364118576, 0.060688555240631104, 0.05676994100213051, 0.04116207733750343, 0.12511689960956573, -0.05698107182979584, -0.018937239423394203, 0.10858088731765747, 0.05003659427165985, 0.11628860235214233, 0.0116971330717206, 0.0498000830411911, -0.149108424782753, -0.038686152547597885, 0.17203880846500397, 0.18500961363315582, 0.23920857906341553, 0.028950776904821396, 0.07196438312530518, 0.12000101804733276, 0.06378476321697235, 0.1028309315443039, 0.041278861463069916, 0.025719868019223213, 0.016715848818421364, -0.06302180141210556, -0.05754520744085312, 0.030517486855387688, -0.01313352957367897, -0.04069966450333595, -0.16936026513576508, 0.09610801190137863, 0.0034058187156915665, 0.07549026608467102, 0.16152046620845795, 0.060152530670166016, -0.10719622671604156, 0.154653400182724, 0.09787415713071823, 0.08995239436626434, -0.07149121165275574, 0.11695443838834763, 0.042611014097929, -0.0001982399116968736, 0.17321345210075378, 0.03346262127161026, 0.14425094425678253, -0.044629309326410294, -0.07287929207086563, -0.07177568972110748, -0.04053391143679619, 0.006203941069543362, 0.03128243237733841, -0.18755768239498138, 0.11796985566616058, 0.06607504934072495, 0.00736414548009634, -0.005991613958030939, -0.05340591073036194, 0.1675303727388382, 0.15960977971553802, 0.0766306146979332, 0.026288626715540886, -0.03269786015152931, -0.014615805819630623, -0.08978231996297836, 0.05488186329603195, 0.02854977734386921, 0.05824984237551689, -0.04007529467344284, -0.10551013797521591, -0.012431226670742035, -0.0058327545411884785, -0.022290624678134918, -0.09602153301239014, -0.11435811221599579, -0.04396599903702736, 0.24932387471199036, -0.06761185079813004, 0.04375799372792244, 0.05767390877008438, 0.040225520730018616, -0.013847432099282742, 0.0348946787416935, -0.028120698407292366, -0.019435739144682884, -0.019900428131222725, 0.01634974777698517, -0.0001519449142506346, -0.037566158920526505, -0.05821892246603966, -0.03118063695728779, -0.1024014800786972, -0.10642620921134949, -0.0000629091082373634, -0.042920731008052826, 0.013539633713662624, -0.016196271404623985, 0.023606229573488235, -0.08363696187734604, -0.04223988950252533, 0.030759556218981743, 0.04939231649041176, -0.08835294097661972, -0.1300876885652542, 0.00857609324157238, -0.03244468942284584, -0.03987769037485123, 0.06328389048576355, -0.0983601063489914, -0.12157551199197769, -0.0674690455198288, -0.03055870346724987, 0.14255374670028687, 0.24859224259853363, -0.03385566547513008, 0.0009191892459057271, 0.13929325342178345, -0.09237378090620041, -0.30570557713508606, -0.15372997522354126, -0.15550853312015533, -0.07920585572719574, 0.0555838942527771, -0.08843311667442322, 0.019696975126862526, 0.07531740516424179, -0.03949115052819252, 0.2128829061985016, -0.1619386076927185, -0.09442929923534393, 0.08830217272043228, 0.0784701406955719, 0.33048346638679504, -0.24181310832500458, 0.01671711541712284, -0.09992372244596481, -0.05125296115875244, 0.013136375695466995, -0.07670323550701141, 0.1279783844947815, 0.03544960170984268, 0.06222604215145111, 0.003136112354695797, -0.004357178695499897, 0.15444259345531464, -0.09326425194740295, 0.13063780963420868, -0.11791864782571793, -0.1019282191991806, 0.21169793605804443, -0.029597684741020203, -0.003326802747324109, -0.1995268613100052, -0.017521429806947708, -0.03530656546354294, 0.02967136725783348, -0.005812970921397209, 0.05825108662247658, 0.003385196439921856, -0.008759375661611557, -0.12908564507961273, -0.02869669906795025, -0.0574299655854702, 0.052134305238723755, 0.20922571420669556, -0.0747450590133667, -0.07319025695323944, 0.030683930963277817, -0.001153240562416613, 0.09211580455303192, -0.002254603197798133, -0.058743927627801895, -0.04292420297861099, 0.07824090123176575, -0.20564259588718414, 0.05905452370643616, 0.0016268292674794793, 0.001365993986837566, 0.007004925049841404, 0.00839656125754118, 0.003669883357360959, 0.12174265086650848, 0.19418860971927643, -0.019209641963243484, -0.02303892932832241, -0.028127769008278847, -0.0028570829890668392, 0.13735932111740112, -0.018221743404865265, 0.11337147653102875, 0.03393560275435448, 0.03866461664438248, 0.01162738911807537, 0.042879264801740646, -0.08367990702390671, -0.08900029957294464, 0.09130251407623291, -0.05182890594005585, -0.07942143827676773, 0.09662962704896927, 0.047298166900873184, 0.10171222686767578, -0.008204445242881775, 0.054147444665431976, 0.005197528284043074, -0.12353155761957169, 0.011655240319669247, 0.17922760546207428, -0.05131702870130539, -0.04654129594564438, -0.07225076854228973, 0.01499214768409729, -0.12483631074428558, 0.07362312078475952, 0.03770800679922104, -0.024236779659986496, 0.09669147431850433, -0.0385676734149456, -0.031101791188120842, -0.005492189899086952, -0.0495007298886776, 0.025531141087412834, -0.1719721108675003, -0.19697365164756775, 0.04899120703339577, 0.00207829219289124, -0.06540820002555847, -0.09526649862527847, -0.06816904246807098, 0.06589699536561966, -0.13164444267749786, 0.12171800434589386, -0.07512139528989792, 0.054380081593990326, -0.0410446971654892, -0.07090248912572861, -0.12174402177333832, -0.01492949016392231, -0.05124059319496155, -0.02058430016040802, 0.052757203578948975, 0.01220169011503458, -0.11042404174804688, -0.1163046807050705, 0.07117971032857895, -0.009870948269963264, 0.008701344020664692, 0.019864726811647415, -0.05638149008154869, 0.023305440321564674, -0.23047511279582977, -0.07378121465444565, 0.0946318730711937, 0.03228521719574928, -0.09850231558084488, 0.13308164477348328, 0.04396870359778404, -0.03363117575645447, 0.05234986171126366, 0.006673934869468212, 0.17355333268642426, -0.07289493829011917, 0.027046604081988335, -0.11137988418340683, -0.15629372000694275, -0.032900549471378326, 0.02565881796181202, 0.22271937131881714, 0.10286116600036621, 0.11414078623056412, -0.05193600058555603, 0.029475754126906395, -0.017892928794026375, 0.08014705032110214, 0.004994099028408527, -0.11435232311487198, -0.028704173862934113, -0.16224968433380127, -0.054482925683259964, -0.05127248913049698, 0.17437098920345306, 0.03629210963845253, -0.1490614265203476, 0.0058814059011638165, 0.12741699814796448, -0.14108552038669586, -0.014088333584368229, 0.1724669486284256, -0.05754481628537178, 0.0347503200173378, -0.1513909548521042, 0.03404717892408371, 0.07796824723482132, -0.03200056776404381, -0.007056177128106356, 0.1051093339920044, -0.018114887177944183, -0.012524203397333622, 0.03449494391679764, -0.0286547914147377, 0.08223294466733932, -0.09045584499835968, 0.04163426533341408, -0.005922887474298477, -0.048934947699308395, -0.10944133251905441, 0.20039622485637665, -0.017137546092271805, 0.009264372289180756, -0.03766164928674698, -0.007215815596282482, -0.10938391089439392, -0.08922018110752106, -0.06870098412036896, -0.1363399773836136, 0.07635719329118729, -0.05740777403116226, 0.01825106143951416, 0.0024078022688627243, 0.009800725616514683, -0.07543230801820755, -0.0025014847051352262, -0.15533974766731262, -0.060901083052158356, 0.03660057485103607, -0.020983805879950523, -0.034294240176677704, -0.04299808666110039, -0.037811193615198135, 0.009822389110922813, -0.06331761181354523, -0.06789285689592361, 0.06349790841341019, 0.06485453993082047, 0.056726545095443726, -0.1674349009990692, -0.10381916910409927, -0.08042624592781067, 0.04959199205040932, 0.06709856539964676, 0.15949812531471252, 0.030168330296874046, -0.01734294928610325, 0.046556998044252396, 0.1613289713859558, 0.023219671100378036, -0.07803602516651154, -0.04903748258948326, -0.14337027072906494, -0.1371868997812271, -0.013036858290433884, -0.065244659781456, -0.02945629693567753, 0.03646416589617729, 0.23981070518493652, 0.1991467922925949, -0.12367187440395355, 0.056554827839136124, -0.07050638645887375, 0.03418069705367088, -0.024946799501776695, 0.1547941267490387, 0.03366207331418991, 0.15994253754615784, -0.027248062193393707, -0.02186705730855465, -0.0638425424695015, 0.0145778339356184, -0.11511627584695816, 0.014990071766078472, -0.010053667239844799, -0.06651145964860916, -0.054290398955345154, 0.12110958993434906, -0.09741875529289246, 0.09239252656698227, 0.1887883096933365, -0.15641169250011444, -0.019867535680532455, -0.0331229642033577, 0.049198687076568604, 0.1066194698214531, 0.022639978677034378, -0.08330392837524414, -0.02482718788087368, -0.04008065536618233, 0.024031132459640503, -0.19774115085601807, -0.11984755098819733, -0.001931386417709291, -0.1317642480134964, 0.13679154217243195, -0.020647326484322548, 0.013462546281516552, 0.03472673520445824, -0.07298732548952103, -0.01247434876859188, 0.16746576130390167, 0.0021696244366467, -0.0379582941532135, -0.022980527952313423, -0.07085583359003067, -0.11607619374990463, 0.05030296742916107, 0.09624729305505753, 0.06499651819467545, -0.0006360329571180046, 0.15567849576473236, -0.015551101416349411, -0.036738816648721695, 0.1304047852754593, -0.1769493669271469, 0.07893262058496475, -0.004976040218025446, -0.028903160244226456, -0.07606187462806702, -0.041622746735811234, 0.03264767676591873, 0.09627795219421387, -0.14667533338069916, -0.04317459836602211, 0.041378892958164215, -0.1009782925248146, 0.10577627271413803, 0.036541033536195755, -0.08391183614730835, 0.00993050541728735, -0.1211213618516922, -0.0027017274405807257, -0.10660627484321594, 0.06506883352994919, 0.2113654911518097, -0.04049525409936905, 0.012998956255614758, -0.15399885177612305, 0.05650550499558449, -0.028856312856078148, -0.061665602028369904, -0.07961662113666534 ]
null
null
null
# Lora of isokaze/磯風/矶风 (Azur Lane) ## What Is This? This is the LoRA model of waifu isokaze/磯風/矶风 (Azur Lane). ## How Is It Trained? * This model is trained with [HCP-Diffusion](https://github.com/7eu7d7/HCP-Diffusion). * The [auto-training framework](https://github.com/deepghs/cyberharem) is maintained by [DeepGHS Team](https://huggingface.co/deepghs). * The base model used for training is [deepghs/animefull-latest](https://huggingface.co/deepghs/animefull-latest). * Dataset used for training is the `stage3-p480-800` in [CyberHarem/isokaze_azurlane](https://huggingface.co/datasets/CyberHarem/isokaze_azurlane), which contains 88 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 880 steps, 40 checkpoints were saved and evaluated. * **Trigger word is `isokaze_azurlane`.** * Pruned core tags for this waifu are `animal_ears, green_hair, animal_ear_fluff, hair_ornament, long_hair, green_eyes, fang, thick_eyebrows, bangs, tail, hair_between_eyes, hairband, black_hairband, very_long_hair, fox_ears`. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. ## How to Use It? ### If You Are Using A1111 WebUI v1.7+ **Just use it like the classic LoRA**. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 660, you need to download [`660/isokaze_azurlane.pt`](https://huggingface.co/CyberHarem/isokaze_azurlane/resolve/main/660/isokaze_azurlane.pt) as the embedding and [`660/isokaze_azurlane.safetensors`](https://huggingface.co/CyberHarem/isokaze_azurlane/resolve/main/660/isokaze_azurlane.safetensors) for loading Lora. By using both files together, you can generate images for the desired characters. ## Which Step Should I Use? We selected 5 good steps for you to choose. The best one is step 660. 1520 images (1.70 GiB) were generated for auto-testing. ![Metrics Plot](metrics_plot.png) The base model used for generating preview images is [Meina/MeinaMix_V11](https://huggingface.co/Meina/MeinaMix_V11). Here are the preview of the recommended steps: | Step | Epoch | CCIP | AI Corrupt | Bikini Plus | Score | Download | pattern_0 | pattern_1_0 | pattern_1_1 | portrait_0 | portrait_1 | portrait_2 | full_body_0 | full_body_1 | profile_0 | profile_1 | free_0 | free_1 | shorts | maid_0 | maid_1 | miko | yukata | suit | china | bikini_0 | bikini_1 | bikini_2 | sit | squat | kneel | jump | crossed_arms | angry | smile | cry | grin | n_lie_0 | n_lie_1 | n_stand_0 | n_stand_1 | n_stand_2 | n_sex_0 | n_sex_1 | |-------:|--------:|:----------|:-------------|:--------------|:----------|:-----------------------------------------------------------------------------------------------------|:-----------------------------------------|:---------------------------------------------|:---------------------------------------------|:-------------------------------------------|:-------------------------------------------|:-------------------------------------------|:---------------------------------------------|:---------------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-------------------------------|:-----------------------------------|:-------------------------------|:---------------------------------|:---------------------------------------|:---------------------------------------|:---------------------------------------|:-----------------------------|:---------------------------------|:---------------------------------|:-------------------------------|:-----------------------------------------------|:---------------------------------|:---------------------------------|:-----------------------------|:-------------------------------|:-------------------------------------|:-------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------------|:-------------------------------------|:-------------------------------------| | 660 | 30 | **0.994** | 0.975 | **0.854** | **0.797** | [Download](https://huggingface.co/CyberHarem/isokaze_azurlane/resolve/main/660/isokaze_azurlane.zip) | ![pattern_0](660/previews/pattern_0.png) | ![pattern_1_0](660/previews/pattern_1_0.png) | ![pattern_1_1](660/previews/pattern_1_1.png) | ![portrait_0](660/previews/portrait_0.png) | ![portrait_1](660/previews/portrait_1.png) | ![portrait_2](660/previews/portrait_2.png) | ![full_body_0](660/previews/full_body_0.png) | ![full_body_1](660/previews/full_body_1.png) | ![profile_0](660/previews/profile_0.png) | ![profile_1](660/previews/profile_1.png) | ![free_0](660/previews/free_0.png) | ![free_1](660/previews/free_1.png) | ![shorts](660/previews/shorts.png) | ![maid_0](660/previews/maid_0.png) | ![maid_1](660/previews/maid_1.png) | ![miko](660/previews/miko.png) | ![yukata](660/previews/yukata.png) | ![suit](660/previews/suit.png) | ![china](660/previews/china.png) | ![bikini_0](660/previews/bikini_0.png) | ![bikini_1](660/previews/bikini_1.png) | ![bikini_2](660/previews/bikini_2.png) | ![sit](660/previews/sit.png) | ![squat](660/previews/squat.png) | ![kneel](660/previews/kneel.png) | ![jump](660/previews/jump.png) | ![crossed_arms](660/previews/crossed_arms.png) | ![angry](660/previews/angry.png) | ![smile](660/previews/smile.png) | ![cry](660/previews/cry.png) | ![grin](660/previews/grin.png) | ![n_lie_0](660/previews/n_lie_0.png) | ![n_lie_1](660/previews/n_lie_1.png) | ![n_stand_0](660/previews/n_stand_0.png) | ![n_stand_1](660/previews/n_stand_1.png) | ![n_stand_2](660/previews/n_stand_2.png) | ![n_sex_0](660/previews/n_sex_0.png) | ![n_sex_1](660/previews/n_sex_1.png) | | 594 | 27 | 0.969 | 0.982 | 0.848 | 0.772 | [Download](https://huggingface.co/CyberHarem/isokaze_azurlane/resolve/main/594/isokaze_azurlane.zip) | ![pattern_0](594/previews/pattern_0.png) | ![pattern_1_0](594/previews/pattern_1_0.png) | ![pattern_1_1](594/previews/pattern_1_1.png) | ![portrait_0](594/previews/portrait_0.png) | ![portrait_1](594/previews/portrait_1.png) | ![portrait_2](594/previews/portrait_2.png) | ![full_body_0](594/previews/full_body_0.png) | ![full_body_1](594/previews/full_body_1.png) | ![profile_0](594/previews/profile_0.png) | ![profile_1](594/previews/profile_1.png) | ![free_0](594/previews/free_0.png) | ![free_1](594/previews/free_1.png) | ![shorts](594/previews/shorts.png) | ![maid_0](594/previews/maid_0.png) | ![maid_1](594/previews/maid_1.png) | ![miko](594/previews/miko.png) | ![yukata](594/previews/yukata.png) | ![suit](594/previews/suit.png) | ![china](594/previews/china.png) | ![bikini_0](594/previews/bikini_0.png) | ![bikini_1](594/previews/bikini_1.png) | ![bikini_2](594/previews/bikini_2.png) | ![sit](594/previews/sit.png) | ![squat](594/previews/squat.png) | ![kneel](594/previews/kneel.png) | ![jump](594/previews/jump.png) | ![crossed_arms](594/previews/crossed_arms.png) | ![angry](594/previews/angry.png) | ![smile](594/previews/smile.png) | ![cry](594/previews/cry.png) | ![grin](594/previews/grin.png) | ![n_lie_0](594/previews/n_lie_0.png) | ![n_lie_1](594/previews/n_lie_1.png) | ![n_stand_0](594/previews/n_stand_0.png) | ![n_stand_1](594/previews/n_stand_1.png) | ![n_stand_2](594/previews/n_stand_2.png) | ![n_sex_0](594/previews/n_sex_0.png) | ![n_sex_1](594/previews/n_sex_1.png) | | 880 | 40 | 0.968 | 0.979 | 0.843 | 0.760 | [Download](https://huggingface.co/CyberHarem/isokaze_azurlane/resolve/main/880/isokaze_azurlane.zip) | ![pattern_0](880/previews/pattern_0.png) | ![pattern_1_0](880/previews/pattern_1_0.png) | ![pattern_1_1](880/previews/pattern_1_1.png) | ![portrait_0](880/previews/portrait_0.png) | ![portrait_1](880/previews/portrait_1.png) | ![portrait_2](880/previews/portrait_2.png) | ![full_body_0](880/previews/full_body_0.png) | ![full_body_1](880/previews/full_body_1.png) | ![profile_0](880/previews/profile_0.png) | ![profile_1](880/previews/profile_1.png) | ![free_0](880/previews/free_0.png) | ![free_1](880/previews/free_1.png) | ![shorts](880/previews/shorts.png) | ![maid_0](880/previews/maid_0.png) | ![maid_1](880/previews/maid_1.png) | ![miko](880/previews/miko.png) | ![yukata](880/previews/yukata.png) | ![suit](880/previews/suit.png) | ![china](880/previews/china.png) | ![bikini_0](880/previews/bikini_0.png) | ![bikini_1](880/previews/bikini_1.png) | ![bikini_2](880/previews/bikini_2.png) | ![sit](880/previews/sit.png) | ![squat](880/previews/squat.png) | ![kneel](880/previews/kneel.png) | ![jump](880/previews/jump.png) | ![crossed_arms](880/previews/crossed_arms.png) | ![angry](880/previews/angry.png) | ![smile](880/previews/smile.png) | ![cry](880/previews/cry.png) | ![grin](880/previews/grin.png) | ![n_lie_0](880/previews/n_lie_0.png) | ![n_lie_1](880/previews/n_lie_1.png) | ![n_stand_0](880/previews/n_stand_0.png) | ![n_stand_1](880/previews/n_stand_1.png) | ![n_stand_2](880/previews/n_stand_2.png) | ![n_sex_0](880/previews/n_sex_0.png) | ![n_sex_1](880/previews/n_sex_1.png) | | 638 | 29 | 0.943 | **0.989** | 0.847 | 0.755 | [Download](https://huggingface.co/CyberHarem/isokaze_azurlane/resolve/main/638/isokaze_azurlane.zip) | ![pattern_0](638/previews/pattern_0.png) | ![pattern_1_0](638/previews/pattern_1_0.png) | ![pattern_1_1](638/previews/pattern_1_1.png) | ![portrait_0](638/previews/portrait_0.png) | ![portrait_1](638/previews/portrait_1.png) | ![portrait_2](638/previews/portrait_2.png) | ![full_body_0](638/previews/full_body_0.png) | ![full_body_1](638/previews/full_body_1.png) | ![profile_0](638/previews/profile_0.png) | ![profile_1](638/previews/profile_1.png) | ![free_0](638/previews/free_0.png) | ![free_1](638/previews/free_1.png) | ![shorts](638/previews/shorts.png) | ![maid_0](638/previews/maid_0.png) | ![maid_1](638/previews/maid_1.png) | ![miko](638/previews/miko.png) | ![yukata](638/previews/yukata.png) | ![suit](638/previews/suit.png) | ![china](638/previews/china.png) | ![bikini_0](638/previews/bikini_0.png) | ![bikini_1](638/previews/bikini_1.png) | ![bikini_2](638/previews/bikini_2.png) | ![sit](638/previews/sit.png) | ![squat](638/previews/squat.png) | ![kneel](638/previews/kneel.png) | ![jump](638/previews/jump.png) | ![crossed_arms](638/previews/crossed_arms.png) | ![angry](638/previews/angry.png) | ![smile](638/previews/smile.png) | ![cry](638/previews/cry.png) | ![grin](638/previews/grin.png) | ![n_lie_0](638/previews/n_lie_0.png) | ![n_lie_1](638/previews/n_lie_1.png) | ![n_stand_0](638/previews/n_stand_0.png) | ![n_stand_1](638/previews/n_stand_1.png) | ![n_stand_2](638/previews/n_stand_2.png) | ![n_sex_0](638/previews/n_sex_0.png) | ![n_sex_1](638/previews/n_sex_1.png) | | 572 | 26 | 0.914 | 0.984 | 0.849 | 0.742 | [Download](https://huggingface.co/CyberHarem/isokaze_azurlane/resolve/main/572/isokaze_azurlane.zip) | ![pattern_0](572/previews/pattern_0.png) | ![pattern_1_0](572/previews/pattern_1_0.png) | ![pattern_1_1](572/previews/pattern_1_1.png) | ![portrait_0](572/previews/portrait_0.png) | ![portrait_1](572/previews/portrait_1.png) | ![portrait_2](572/previews/portrait_2.png) | ![full_body_0](572/previews/full_body_0.png) | ![full_body_1](572/previews/full_body_1.png) | ![profile_0](572/previews/profile_0.png) | ![profile_1](572/previews/profile_1.png) | ![free_0](572/previews/free_0.png) | ![free_1](572/previews/free_1.png) | ![shorts](572/previews/shorts.png) | ![maid_0](572/previews/maid_0.png) | ![maid_1](572/previews/maid_1.png) | ![miko](572/previews/miko.png) | ![yukata](572/previews/yukata.png) | ![suit](572/previews/suit.png) | ![china](572/previews/china.png) | ![bikini_0](572/previews/bikini_0.png) | ![bikini_1](572/previews/bikini_1.png) | ![bikini_2](572/previews/bikini_2.png) | ![sit](572/previews/sit.png) | ![squat](572/previews/squat.png) | ![kneel](572/previews/kneel.png) | ![jump](572/previews/jump.png) | ![crossed_arms](572/previews/crossed_arms.png) | ![angry](572/previews/angry.png) | ![smile](572/previews/smile.png) | ![cry](572/previews/cry.png) | ![grin](572/previews/grin.png) | ![n_lie_0](572/previews/n_lie_0.png) | ![n_lie_1](572/previews/n_lie_1.png) | ![n_stand_0](572/previews/n_stand_0.png) | ![n_stand_1](572/previews/n_stand_1.png) | ![n_stand_2](572/previews/n_stand_2.png) | ![n_sex_0](572/previews/n_sex_0.png) | ![n_sex_1](572/previews/n_sex_1.png) | ## Anything Else? Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. ## All Steps We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * [Steps From 682 to 880](all/0.md) * [Steps From 462 to 660](all/1.md) * [Steps From 242 to 440](all/2.md) * [Steps From 22 to 220](all/3.md)
{"license": "mit", "tags": ["art", "not-for-all-audiences"], "datasets": ["CyberHarem/isokaze_azurlane"], "pipeline_tag": "text-to-image"}
text-to-image
CyberHarem/isokaze_azurlane
[ "art", "not-for-all-audiences", "text-to-image", "dataset:CyberHarem/isokaze_azurlane", "license:mit", "region:us" ]
2024-02-15T04:24:53+00:00
[]
[]
TAGS #art #not-for-all-audiences #text-to-image #dataset-CyberHarem/isokaze_azurlane #license-mit #region-us
Lora of isokaze/磯風/矶风 (Azur Lane) ================================= What Is This? ------------- This is the LoRA model of waifu isokaze/磯風/矶风 (Azur Lane). How Is It Trained? ------------------ * This model is trained with HCP-Diffusion. * The auto-training framework is maintained by DeepGHS Team. * The base model used for training is deepghs/animefull-latest. * Dataset used for training is the 'stage3-p480-800' in CyberHarem/isokaze\_azurlane, which contains 88 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 880 steps, 40 checkpoints were saved and evaluated. * Trigger word is 'isokaze\_azurlane'. * Pruned core tags for this waifu are 'animal\_ears, green\_hair, animal\_ear\_fluff, hair\_ornament, long\_hair, green\_eyes, fang, thick\_eyebrows, bangs, tail, hair\_between\_eyes, hairband, black\_hairband, very\_long\_hair, fox\_ears'. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. How to Use It? -------------- ### If You Are Using A1111 WebUI v1.7+ Just use it like the classic LoRA. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 660, you need to download '660/isokaze\_azurlane.pt' as the embedding and '660/isokaze\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters. Which Step Should I Use? ------------------------ We selected 5 good steps for you to choose. The best one is step 660. 1520 images (1.70 GiB) were generated for auto-testing. !Metrics Plot The base model used for generating preview images is Meina/MeinaMix\_V11. Here are the preview of the recommended steps: Anything Else? -------------- Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. All Steps --------- We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * Steps From 682 to 880 * Steps From 462 to 660 * Steps From 242 to 440 * Steps From 22 to 220
[ "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 660, you need to download '660/isokaze\\_azurlane.pt' as the embedding and '660/isokaze\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 660.\n\n\n1520 images (1.70 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 682 to 880\n* Steps From 462 to 660\n* Steps From 242 to 440\n* Steps From 22 to 220" ]
[ "TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/isokaze_azurlane #license-mit #region-us \n", "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 660, you need to download '660/isokaze\\_azurlane.pt' as the embedding and '660/isokaze\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 660.\n\n\n1520 images (1.70 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 682 to 880\n* Steps From 462 to 660\n* Steps From 242 to 440\n* Steps From 22 to 220" ]
[ 44, 38, 470 ]
[ "passage: TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/isokaze_azurlane #license-mit #region-us \n### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file." ]
[ 0.008174664340913296, -0.01723347418010235, -0.003882764372974634, 0.08050777018070221, 0.07979447394609451, 0.08488810062408447, 0.23552989959716797, 0.08054795116186142, 0.1312856525182724, -0.0714617595076561, 0.08515168726444244, 0.05449388176202774, -0.0042798565700650215, 0.03103134222328663, -0.028262551873922348, -0.1593867689371109, -0.060425352305173874, -0.029902884736657143, 0.010336157865822315, 0.014282076619565487, 0.0806766077876091, 0.007357516326010227, 0.10136282444000244, -0.04993438348174095, -0.03961480036377907, 0.05172464996576309, -0.03282592073082924, -0.04479197412729263, 0.030920125544071198, 0.07692146301269531, 0.12261015921831131, 0.011213045567274094, 0.06635279208421707, -0.16395717859268188, 0.06521381437778473, -0.009818315505981445, -0.10853328555822372, -0.008115396834909916, 0.022925330325961113, -0.04002470523118973, 0.12591546773910522, 0.02186606638133526, -0.11482048034667969, 0.038249582052230835, -0.13345655798912048, -0.02540610358119011, -0.051218386739492416, 0.046383146196603775, 0.1432833969593048, 0.05821685492992401, 0.02247052639722824, 0.058124225586652756, -0.0474240817129612, 0.08396259695291519, 0.12118282169103622, -0.13231229782104492, -0.06457342952489853, 0.09909205138683319, 0.008538096211850643, 0.14089985191822052, -0.08987965434789658, 0.09597480297088623, 0.07314647734165192, -0.05094106122851372, -0.15044981241226196, -0.09332538396120071, -0.20947128534317017, -0.013540387153625488, 0.008243043906986713, 0.02009052224457264, 0.40958860516548157, 0.0582582987844944, 0.03815731406211853, 0.0633963942527771, -0.06414933502674103, 0.02724859118461609, -0.09519116580486298, 0.14084157347679138, 0.03704095259308815, 0.09461377561092377, -0.0351937972009182, -0.09984545409679413, -0.11695746332406998, -0.06832903623580933, -0.08068092167377472, -0.0009727734723128378, 0.023725537583231926, 0.12083529680967331, -0.18962214887142181, 0.002597222803160548, -0.05856654793024063, -0.13294684886932373, 0.019567687064409256, -0.09979034215211868, 0.17180916666984558, 0.07170083373785019, -0.011698299087584019, 0.005341014359146357, 0.2522641718387604, 0.12556606531143188, 0.1846296340227127, 0.04609183594584465, -0.09053827822208405, 0.13140472769737244, 0.04547382891178131, -0.08049824833869934, -0.007760193664580584, -0.10326900333166122, 0.1378210335969925, -0.05214998871088028, 0.10683757066726685, -0.060974907130002975, -0.1163494661450386, 0.011813201010227203, -0.0986097902059555, 0.07246074825525284, 0.0388948954641819, 0.0059224506840109825, -0.047762420028448105, 0.04513678327202797, 0.04093191772699356, -0.03768687695264816, -0.008732017129659653, -0.013196044601500034, -0.0522049255669117, 0.05466911941766739, 0.10402298718690872, 0.030200276523828506, 0.05868513509631157, 0.002806790405884385, -0.014911901205778122, 0.0020308031234890223, -0.05089673772454262, 0.0012391263153403997, 0.04351312294602394, 0.03874833881855011, 0.09189853817224503, -0.15888641774654388, -0.07233235985040665, -0.013498570770025253, 0.055599045008420944, 0.008459106087684631, 0.10556136816740036, -0.006168937776237726, 0.056844428181648254, 0.008824477903544903, -0.022686224430799484, 0.021876245737075806, -0.10562628507614136, 0.08104997128248215, -0.017933903262019157, 0.09196116030216217, -0.1976550817489624, -0.004153440706431866, -0.047537170350551605, 0.014989829622209072, 0.06436241418123245, -0.010477319359779358, -0.11489759385585785, 0.12186172604560852, -0.010333165526390076, 0.07032572478055954, -0.08936698734760284, 0.04600771516561508, 0.027836309745907784, 0.08349096029996872, -0.0939137414097786, 0.012509976513683796, 0.1129651591181755, -0.14536890387535095, -0.16207271814346313, 0.09482430666685104, -0.023126039654016495, 0.029160859063267708, 0.03719688206911087, 0.16194696724414825, 0.16802209615707397, -0.18555207550525665, -0.021908901631832123, 0.05865800008177757, -0.020984169095754623, -0.08187083899974823, -0.01254358422011137, 0.10745445638895035, 0.03312426805496216, 0.03498127683997154, -0.03237565606832504, 0.12527666985988617, -0.03130587190389633, -0.08539402484893799, -0.0344109982252121, -0.07939352840185165, -0.0821995735168457, 0.05030425637960434, -0.005518514662981033, -0.05476905405521393, 0.016569286584854126, -0.14667510986328125, 0.16767896711826324, 0.01704929955303669, 0.019626734778285027, -0.07210642844438553, 0.11898703128099442, 0.006909723859280348, 0.00015580047329422086, 0.009131966158747673, -0.05710887908935547, -0.10408949106931686, 0.23558008670806885, 0.08365721255540848, 0.08072593063116074, 0.06157836690545082, -0.04968404769897461, -0.0686480924487114, 0.016020752489566803, 0.013740483671426773, -0.03853965923190117, 0.024579906836152077, -0.10403067618608475, 0.053025759756565094, -0.015238994732499123, 0.021239273250102997, 0.0037508648820221424, -0.029013626277446747, 0.06167512759566307, 0.010699831880629063, -0.021207647398114204, 0.09250663220882416, 0.04993478208780289, -0.020896373316645622, -0.07426891475915909, 0.0037250772584229708, 0.07167308032512665, -0.01292008999735117, -0.07870256900787354, 0.0244001392275095, -0.00414225785061717, 0.03639936447143555, 0.19997195899486542, -0.22559143602848053, 0.04118458926677704, 0.003732325043529272, 0.05278526991605759, 0.03848397359251976, 0.004860358778387308, -0.028330188244581223, 0.02842896431684494, -0.027455300092697144, 0.0646287128329277, -0.015067758969962597, 0.07247109711170197, -0.03250418230891228, -0.1407414972782135, -0.01914081536233425, -0.026713790372014046, 0.1636582762002945, -0.16719157993793488, 0.0663716122508049, 0.1966806948184967, -0.12131393700838089, 0.14512580633163452, -0.0043846010230481625, -0.014346965588629246, 0.01244099996984005, 0.03294359892606735, 0.0033848853781819344, 0.10577525943517685, -0.07946502417325974, -0.02951861545443535, 0.027850717306137085, -0.08296697586774826, 0.032462894916534424, -0.11925222724676132, -0.10990756005048752, -0.07000624388456345, -0.03735557198524475, -0.04179621487855911, 0.026894625276327133, -0.05323783680796623, 0.07385341823101044, -0.09457218647003174, -0.07864287495613098, -0.03070705384016037, -0.08331546932458878, 0.027435841038823128, 0.009745036251842976, -0.05830428749322891, -0.12464019656181335, -0.12665212154388428, -0.08615025877952576, -0.1476699858903885, -0.004768916871398687, 0.06917749345302582, -0.1129181906580925, -0.04329971224069595, 0.015403310768306255, -0.04157638177275658, 0.0927220806479454, -0.08174946904182434, 0.013383399695158005, 0.05817829817533493, -0.04389815777540207, -0.16745483875274658, -0.0021117532160133123, -0.06471925973892212, -0.05530910566449165, 0.16135235130786896, -0.16262060403823853, 0.1837686151266098, -0.02738274447619915, 0.05350710451602936, 0.06077747046947479, 0.032020989805459976, 0.1214303970336914, -0.11729803681373596, 0.0749296173453331, 0.19286592304706573, 0.036794938147068024, 0.07762771099805832, 0.11513466387987137, 0.0790087953209877, -0.10757458209991455, 0.03529461473226547, 0.06816049665212631, -0.1017083153128624, -0.08276136219501495, -0.0562901608645916, -0.11409611999988556, -0.054384175688028336, 0.05223513022065163, 0.05285094678401947, 0.046558231115341187, 0.13101406395435333, -0.05581056699156761, -0.015840444713830948, 0.09082797169685364, 0.05041193962097168, 0.07268358021974564, 0.01753848046064377, 0.05869894102215767, -0.1444307565689087, -0.046515241265296936, 0.16148465871810913, 0.21565629541873932, 0.23657888174057007, 0.019945252686738968, 0.07967548817396164, 0.11743231862783432, 0.09379364550113678, 0.09914100170135498, 0.06064026430249214, 0.006889949087053537, 0.019899677485227585, -0.07052061706781387, -0.05222893878817558, 0.014578809961676598, 0.006287466734647751, -0.051809221506118774, -0.14795981347560883, 0.1007109209895134, 0.001457073143683374, 0.08327321708202362, 0.14494803547859192, 0.039815597236156464, -0.10209815204143524, 0.15178774297237396, 0.09890592843294144, 0.09500911831855774, -0.06789663434028625, 0.12596139311790466, 0.052909377962350845, -0.004803565330803394, 0.15841436386108398, 0.03534519299864769, 0.14995650947093964, -0.03579452261328697, -0.07772223651409149, -0.07858245819807053, -0.05386737361550331, 0.0040569305419921875, 0.03698791563510895, -0.21015264093875885, 0.10956369340419769, 0.05626066401600838, 0.009905751794576645, -0.015488127246499062, -0.05408907309174538, 0.18666091561317444, 0.16264589130878448, 0.07960668951272964, 0.028134062886238098, -0.04976065084338188, -0.015249562449753284, -0.08273006230592728, 0.0516691692173481, 0.023519540205597878, 0.07107768207788467, -0.03724795579910278, -0.09929682314395905, -0.018666010349988937, -0.004399380646646023, 0.018146472051739693, -0.07796942442655563, -0.11193723231554031, -0.044491469860076904, 0.2559182941913605, -0.06340454518795013, 0.04888801649212837, 0.05327143147587776, 0.022808339446783066, -0.030805351212620735, 0.03773479908704758, -0.03841324895620346, -0.019104447215795517, -0.04234134405851364, 0.014076504856348038, 0.00411409605294466, -0.04930225759744644, -0.058785323053598404, -0.02656976319849491, -0.09788783639669418, -0.10262484103441238, 0.004788264166563749, -0.051524028182029724, 0.012765798717737198, -0.021797405555844307, 0.01211711298674345, -0.09543278813362122, -0.03354322537779808, 0.027243176475167274, 0.03517885133624077, -0.07959690690040588, -0.13249355554580688, -0.006986142601817846, -0.016370924189686775, -0.058825913816690445, 0.04283430054783821, -0.122515968978405, -0.10396108031272888, -0.06143660843372345, -0.040193282067775726, 0.12677650153636932, 0.23657627403736115, -0.023059874773025513, -0.005680999252945185, 0.14769577980041504, -0.09907487034797668, -0.31756591796875, -0.16603749990463257, -0.16098827123641968, -0.09733526408672333, 0.03279843553900719, -0.06934265792369843, 0.02640543505549431, 0.08193154633045197, -0.03836173564195633, 0.2102893739938736, -0.18837934732437134, -0.09786123782396317, 0.08813178539276123, 0.09096939861774445, 0.30393722653388977, -0.2505943775177002, 0.015967750921845436, -0.11753750592470169, -0.033716000616550446, 0.0043256268836557865, -0.06246061623096466, 0.12173041701316833, 0.03293372690677643, 0.07733683288097382, -0.0018730071606114507, -0.006197356153279543, 0.150755375623703, -0.0687018632888794, 0.1360720545053482, -0.12347237020730972, -0.1015310287475586, 0.20203310251235962, -0.03611859679222107, 0.0022702959831804037, -0.20631563663482666, -0.03548235073685646, -0.04387471452355385, 0.03772008419036865, -0.008056805469095707, 0.05515962839126587, -0.00499738659709692, -0.01876547373831272, -0.13294482231140137, -0.01578555628657341, -0.03705773130059242, 0.06150524690747261, 0.2331756353378296, -0.06586088985204697, -0.0721021369099617, 0.039704468101263046, -0.012041538022458553, 0.10557438433170319, 0.02532694861292839, -0.05521436035633087, -0.04281732812523842, 0.09076082706451416, -0.20098872482776642, 0.05731306970119476, 0.008555660024285316, -0.003433782374486327, 0.011581378988921642, 0.011561728082597256, 0.01748495176434517, 0.11913636326789856, 0.1809835284948349, -0.0029459474608302116, -0.019030526280403137, -0.01684233546257019, 0.021897833794355392, 0.1295148879289627, -0.019018467515707016, 0.1110742911696434, 0.020771609619259834, 0.037239428609609604, 0.004068885464221239, 0.05561263486742973, -0.08321400731801987, -0.08478239178657532, 0.09799116849899292, -0.04877062886953354, -0.07880409061908722, 0.09015647321939468, 0.04617823660373688, 0.07618601620197296, -0.0007935534813441336, 0.04355867952108383, 0.018565280362963676, -0.1265353411436081, 0.0337393656373024, 0.20261982083320618, -0.08835645765066147, -0.06775395572185516, -0.0685616135597229, 0.013047376647591591, -0.12544307112693787, 0.07138556987047195, 0.0410640612244606, -0.039908044040203094, 0.11892788112163544, -0.04292947053909302, -0.03395439684391022, 0.0054268790408968925, -0.05764475464820862, 0.03595218434929848, -0.13957840204238892, -0.19654446840286255, 0.044729605317115784, -0.005899729207158089, -0.06538379937410355, -0.08898128569126129, -0.08742856234312057, 0.0648474171757698, -0.16081304848194122, 0.13775980472564697, -0.07287692278623581, 0.056646525859832764, -0.03608083352446556, -0.05393229052424431, -0.10999659448862076, -0.017464155331254005, -0.05692067742347717, -0.022480368614196777, 0.05522587522864342, 0.01715932972729206, -0.11614561080932617, -0.11911185085773468, 0.06354586035013199, -0.006771184504032135, -0.004539014771580696, 0.020535748451948166, -0.06431552767753601, 0.02195383980870247, -0.23482371866703033, -0.06052367016673088, 0.08502736687660217, 0.04363827034831047, -0.08713269978761673, 0.12607786059379578, 0.04707450792193413, -0.022237984463572502, 0.03733453154563904, 0.00398034555837512, 0.173696830868721, -0.0732402503490448, 0.02722550556063652, -0.11837346851825714, -0.1670459657907486, -0.025463519617915154, 0.032998234033584595, 0.2294582575559616, 0.08198468387126923, 0.12355393916368484, -0.048728860914707184, 0.027309129014611244, -0.009487039409577847, 0.07438914477825165, 0.01622341386973858, -0.10597941279411316, -0.03832636773586273, -0.17209084331989288, -0.06467258930206299, -0.060829583555459976, 0.16683687269687653, 0.036079827696084976, -0.13336017727851868, -0.004919905215501785, 0.11114942282438278, -0.16502587497234344, -0.015092816203832626, 0.178762286901474, -0.0440511628985405, 0.02644670009613037, -0.15867838263511658, 0.029595280066132545, 0.07936849445104599, -0.0334727019071579, -0.006378277204930782, 0.12936930358409882, -0.007003136910498142, 0.0021945748012512922, 0.037481751292943954, -0.033945370465517044, 0.087489552795887, -0.0898590087890625, 0.06358136981725693, 0.0010311446385458112, -0.041924621909856796, -0.1092442125082016, 0.18529950082302094, -0.018307359889149666, 0.013674856163561344, -0.05889537185430527, -0.000998235773295164, -0.09824872761964798, -0.0919324979186058, -0.06860208511352539, -0.13272309303283691, 0.07323317974805832, -0.06034477427601814, 0.01539726834744215, 0.004707331769168377, 0.018614375963807106, -0.07317843288183212, 0.015443979762494564, -0.1900084912776947, -0.047387395054101944, 0.018738405779004097, -0.01681480184197426, -0.02376924827694893, -0.040274783968925476, -0.04028487950563431, 0.019111812114715576, -0.0651819258928299, -0.06684146076440811, 0.05844176188111305, 0.07978487014770508, 0.05591300502419472, -0.162907674908638, -0.10779419541358948, -0.07466478645801544, 0.03508090600371361, 0.07824011892080307, 0.185195192694664, 0.03563772514462471, -0.0021925652399659157, 0.04512832686305046, 0.13368161022663116, 0.015228793025016785, -0.08757013827562332, -0.06190520524978638, -0.1328660398721695, -0.13562825322151184, -0.017592724412679672, -0.06756168603897095, -0.025226429104804993, 0.022842051461338997, 0.23729421198368073, 0.1906329244375229, -0.14849406480789185, 0.037373993545770645, -0.07977430522441864, 0.04027867317199707, -0.03260333463549614, 0.1617308259010315, 0.05009113997220993, 0.14370197057724, -0.028861137107014656, -0.030770381912589073, -0.07024721056222916, 0.020819084718823433, -0.10571014136075974, 0.032384973019361496, -0.008495026268064976, -0.07167457044124603, -0.061668142676353455, 0.10759292542934418, -0.11221876740455627, 0.07133330404758453, 0.18962997198104858, -0.14596931636333466, -0.01992487721145153, -0.03633495792746544, 0.05460182577371597, 0.11226038634777069, 0.016047419980168343, -0.0795837789773941, -0.022300804033875465, 0.0002781648363452405, 0.02880363166332245, -0.17536260187625885, -0.11103503406047821, -0.004313481505960226, -0.12843653559684753, 0.13679543137550354, -0.0022126047406345606, 0.0021128677763044834, 0.036759331822395325, -0.06898591667413712, -0.006371630355715752, 0.17339377105236053, 0.017466185614466667, -0.03312324360013008, -0.023705024272203445, -0.05616933852434158, -0.1064591333270073, 0.06364718079566956, 0.08869180083274841, 0.05081494525074959, 0.003207941073924303, 0.1676337718963623, -0.019809188321232796, -0.03955687955021858, 0.13663293421268463, -0.17809268832206726, 0.09420081973075867, 0.0014448333531618118, -0.015498453751206398, -0.07541504502296448, -0.04271147400140762, 0.046824101358652115, 0.0813942477107048, -0.16360414028167725, -0.04967694357037544, 0.06508710235357285, -0.1013864129781723, 0.06490688771009445, 0.04420315846800804, -0.0912843719124794, 0.017558105289936066, -0.12844975292682648, -0.003641613991931081, -0.10295691341161728, 0.04742048680782318, 0.19853655993938446, -0.032101843506097794, 0.014768398366868496, -0.1455906331539154, 0.05884592607617378, -0.03229157254099846, -0.04315381124615669, -0.07556731253862381 ]
null
null
ml-agents
# **poca** Agent playing **SoccerTwos** This is a trained model of a **poca** agent playing **SoccerTwos** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: Wajid333/poca-SoccerTwos 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
{"library_name": "ml-agents", "tags": ["SoccerTwos", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-SoccerTwos"]}
reinforcement-learning
Wajid333/poca-SoccerTwos
[ "ml-agents", "tensorboard", "onnx", "SoccerTwos", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-SoccerTwos", "region:us" ]
2024-02-15T04:25:28+00:00
[]
[]
TAGS #ml-agents #tensorboard #onnx #SoccerTwos #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SoccerTwos #region-us
# poca Agent playing SoccerTwos This is a trained model of a poca agent playing SoccerTwos using the Unity ML-Agents Library. ## Usage (with ML-Agents) The Documentation: URL We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your browser: URL - A *longer tutorial* to understand how works ML-Agents: URL ### Resume the training ### Watch your Agent play You can watch your agent playing directly in your browser 1. If the environment is part of ML-Agents official environments, go to URL 2. Step 1: Find your model_id: Wajid333/poca-SoccerTwos 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play
[ "# poca Agent playing SoccerTwos\n This is a trained model of a poca agent playing SoccerTwos\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: Wajid333/poca-SoccerTwos\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play" ]
[ "TAGS\n#ml-agents #tensorboard #onnx #SoccerTwos #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SoccerTwos #region-us \n", "# poca Agent playing SoccerTwos\n This is a trained model of a poca agent playing SoccerTwos\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: Wajid333/poca-SoccerTwos\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play" ]
[ 52, 206 ]
[ "passage: TAGS\n#ml-agents #tensorboard #onnx #SoccerTwos #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SoccerTwos #region-us \n# poca Agent playing SoccerTwos\n This is a trained model of a poca agent playing SoccerTwos\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: Wajid333/poca-SoccerTwos\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play" ]
[ -0.001977975247427821, -0.02186603471636772, -0.004597488325089216, 0.0521719828248024, 0.16398963332176208, -0.020471230149269104, 0.11455918848514557, 0.10563007742166519, 0.12684589624404907, 0.09259509295225143, 0.07223780453205109, 0.045688971877098083, 0.08815697580575943, 0.12815824151039124, 0.07087939232587814, -0.13742046058177948, -0.020941784605383873, -0.10395856201648712, 0.045096397399902344, 0.05343535169959068, 0.08097416907548904, -0.042318593710660934, 0.05963391438126564, 0.04479517042636871, -0.0878981426358223, 0.0034559110645204782, -0.06875423341989517, -0.04911606386303902, 0.013151089660823345, 0.0219606664031744, 0.025738561525940895, -0.07264016568660736, 0.0962647795677185, -0.18050901591777802, 0.024113252758979797, 0.03264502063393593, -0.013338138349354267, -0.06947547942399979, 0.1376718133687973, 0.05823852866888046, 0.11064580827951431, -0.06216523051261902, 0.08775021880865097, 0.052698493003845215, -0.08611306548118591, 0.08214566111564636, -0.09665846079587936, 0.01588043011724949, 0.2143324464559555, 0.14468911290168762, 0.005920429714024067, 0.07405312359333038, -0.0365891307592392, 0.015550660900771618, 0.13713973760604858, -0.2729189693927765, -0.07193359732627869, 0.13932819664478302, -0.016986476257443428, 0.08152880519628525, -0.047675225883722305, 0.04154885187745094, -0.008824377320706844, 0.02996431104838848, -0.0335860438644886, 0.019110150635242462, 0.2053803652524948, -0.015376587398350239, -0.032718732953071594, -0.12078513205051422, 0.010491326451301575, 0.07102382183074951, -0.06471288949251175, -0.17013905942440033, 0.03904350474476814, 0.09532300382852554, -0.05251163989305496, 0.014151322655379772, 0.06402391195297241, 0.007091742008924484, -0.02041042223572731, -0.0915832445025444, -0.03820933774113655, -0.06057709828019142, 0.032962482422590256, 0.12067950516939163, -0.027921730652451515, -0.040370386093854904, 0.04220558702945709, 0.07094033062458038, 0.06770279258489609, -0.0356954000890255, -0.028628120198845863, 0.003178764833137393, -0.1762024611234665, -0.08268161863088608, -0.02158687822520733, -0.03576221317052841, 0.04786133021116257, 0.11636266857385635, 0.11365697532892227, 0.005681675858795643, 0.013757161796092987, 0.05871602147817612, -0.025473900139331818, 0.06753039360046387, -0.0074858637526631355, 0.00939076580107212, 0.020611898973584175, 0.012830265797674656, 0.017215846106410027, -0.09894608706235886, -0.10304700583219528, 0.08742023259401321, -0.13658791780471802, 0.10653151571750641, 0.1150747761130333, -0.015652606263756752, -0.030376633629202843, -0.057104095816612244, 0.02909141778945923, -0.12848040461540222, 0.0834396481513977, 0.040485162287950516, -0.040995385497808456, -0.0999947264790535, -0.04326523467898369, 0.037703949958086014, -0.07732658088207245, 0.021597445011138916, -0.013059310615062714, 0.055239930748939514, -0.005719564389437437, -0.031701091676950455, 0.08413992077112198, -0.10251295566558838, -0.001684089656919241, -0.16651061177253723, -0.08256209641695023, -0.08090696483850479, 0.03866267204284668, -0.08965655416250229, -0.08945120871067047, -0.08443599194288254, 0.00924282893538475, -0.09114676713943481, 0.048334281891584396, -0.04286068677902222, -0.06930801272392273, -0.01872703619301319, -0.06340540945529938, 0.0868741050362587, 0.07810377329587936, 0.047031037509441376, -0.022885629907250404, 0.030865615233778954, -0.1476559042930603, 0.14331236481666565, -0.10993220657110214, 0.14528578519821167, -0.06070716306567192, 0.09561962634325027, 0.01234882790595293, 0.025358442217111588, 0.06587372720241547, 0.13112229108810425, -0.05934220179915428, -0.08928577601909637, 0.15075789391994476, -0.05095735564827919, -0.18637831509113312, 0.06778480857610703, 0.039610423147678375, 0.07105673849582672, 0.039380330592393875, 0.228054016828537, 0.18103937804698944, -0.3129420876502991, 0.10714686661958694, 0.002859151689335704, -0.12108071148395538, -0.0026424915995448828, 0.11758571863174438, -0.0988846942782402, 0.07264887541532516, -0.03952224180102348, -0.1600237786769867, 0.14746065437793732, -0.03350047767162323, -0.07324998080730438, 0.04700760915875435, -0.08141255378723145, -0.070281483232975, -0.0009681222727522254, 0.04191844165325165, -0.04774525761604309, -0.0035830659326165915, -0.026522133499383926, 0.04835136607289314, -0.008104370906949043, 0.04386989027261734, -0.08390520513057709, 0.1389743983745575, -0.034686796367168427, 0.02602350525557995, -0.11995057016611099, -0.13884449005126953, 0.003300404641777277, 0.06743017584085464, 0.08561799675226212, -0.09055080264806747, 0.04012071341276169, 0.10188595205545425, 0.034703146666288376, -0.05683739855885506, -0.11367764323949814, 0.0219026617705822, -0.06358598172664642, -0.10068365186452866, -0.04078780859708786, -0.05955440551042557, 0.0851324275135994, -0.13093991577625275, 0.04603235796093941, -0.09935460239648819, 0.08548728376626968, -0.0008078018436208367, -0.06168093904852867, 0.004807842429727316, 0.039108239114284515, 0.043195247650146484, -0.07115153223276138, 0.10637715458869934, 0.02783932536840439, -0.08327239006757736, 0.043285101652145386, 0.02556685544550419, -0.041718196123838425, 0.11594671756029129, 0.01946951076388359, -0.017207587137818336, 0.040961287915706635, -0.03966550529003143, -0.00934730377048254, -0.11245451122522354, -0.034852802753448486, 0.15780167281627655, 0.08718281984329224, 0.1104734018445015, -0.08703786134719849, -0.015009134076535702, 0.028305713087320328, -0.06829970329999924, -0.05024784803390503, 0.07324904948472977, 0.03353340923786163, -0.021268390119075775, 0.04337006434798241, 0.054075565189123154, 0.1254461109638214, 0.14356179535388947, 0.004646499641239643, -0.1188759058713913, 0.032417185604572296, 0.1295175403356552, 0.03764332830905914, 0.0167801883071661, 0.0070562660694122314, -0.045291051268577576, -0.011795593425631523, -0.01743951439857483, -0.04007287696003914, -0.09183023124933243, -0.0740257203578949, 0.06300339102745056, -0.030199971050024033, -0.0035490787122398615, -0.03312317281961441, -0.017192943021655083, 0.07688646018505096, 0.08811144530773163, 0.011128269135951996, 0.012330097146332264, -0.05396393686532974, -0.12550006806850433, 0.05626673251390457, -0.08024559915065765, -0.214172825217247, -0.12615786492824554, -0.08327994495630264, -0.08042309433221817, 0.05094468966126442, 0.07095251977443695, -0.12234245240688324, 0.012529183179140091, -0.08145859837532043, -0.043579041957855225, 0.040707461535930634, -0.06963800638914108, 0.20036837458610535, 0.10747409611940384, -0.009834797121584415, -0.07079098373651505, -0.007409647572785616, 0.013167022727429867, -0.0880294218659401, -0.01875670813024044, 0.001137247309088707, 0.15096734464168549, 0.0933813527226448, 0.007427227217704058, 0.04841101914644241, -0.025516806170344353, 0.10504692047834396, -0.09071895480155945, 0.020094946026802063, 0.0668063759803772, -0.028445085510611534, 0.07930707931518555, 0.029096165671944618, 0.03120400942862034, -0.03558320924639702, 0.034346941858530045, 0.020611818879842758, -0.06564931571483612, -0.18935662508010864, -0.11004789173603058, -0.029038891196250916, 0.11111441254615784, 0.1189379170536995, 0.08636591583490372, -0.06111893430352211, 0.004002819769084454, 0.010378090664744377, -0.05160067602992058, 0.14173246920108795, 0.1254977136850357, -0.12443998456001282, -0.010157084092497826, 0.00907852966338396, -0.06488090008497238, 0.024584852159023285, 0.08331415802240372, -0.0246871430426836, 0.06110269948840141, 0.09088769555091858, 0.04688525199890137, 0.035114504396915436, -0.0760437399148941, -0.06849434971809387, 0.11434271931648254, 0.05747408792376518, -0.007964856922626495, -0.028174513950943947, -0.0660303384065628, -0.06591792404651642, 0.07480820268392563, 0.12381841987371445, -0.055948641151189804, -0.1316508948802948, 0.05353685840964317, 0.10061179101467133, 0.1583375632762909, 0.002816410269588232, -0.136219322681427, -0.05553795397281647, -0.013066429644823074, -0.11050475388765335, -0.0015473593957722187, 0.00012667954433709383, 0.05445944145321846, -0.15805132687091827, 0.04035469517111778, 0.07345199584960938, 0.15044942498207092, 0.048392023891210556, -0.0067175617441535, 0.02645150013267994, 0.019870685413479805, -0.014772932976484299, 0.05138172209262848, -0.1655939817428589, 0.04336345195770264, -0.012646009214222431, 0.10455439239740372, -0.05836670845746994, -0.00025512021966278553, 0.05851540341973305, -0.027321694418787956, 0.16339567303657532, 0.0719931498169899, -0.034604039043188095, -0.16394293308258057, -0.10757651925086975, -0.08318678289651871, -0.007536308374255896, -0.06657911092042923, 0.08566914498806, 0.023454980924725533, -0.011864429339766502, -0.1024932861328125, 0.08874686807394028, -0.03670342639088631, -0.07673268020153046, -0.028064612299203873, -0.044901199638843536, 0.06728492677211761, -0.038479194045066833, 0.012705064378678799, -0.08668772876262665, 0.15418539941310883, 0.08979515731334686, -0.032570842653512955, -0.0939277783036232, -0.002332062693312764, -0.10405395179986954, -0.02292710915207863, 0.03683847561478615, 0.0047226580791175365, 0.10153196007013321, -0.11199019849300385, 0.008837578818202019, -0.006303204223513603, -0.12196581065654755, -0.051388729363679886, -0.009417155757546425, 0.16742807626724243, 0.04741358011960983, 0.0318586602807045, 0.029703103005886078, 0.02576729841530323, 0.015531547367572784, -0.09313995391130447, 0.17613501846790314, 0.17165397107601166, -0.06657464802265167, 0.033771008253097534, -0.02801896259188652, 0.026218803599476814, -0.07921291887760162, -0.036223042756319046, 0.17940381169319153, 0.26783487200737, -0.05587837100028992, 0.2132738083600998, 0.01210265327244997, -0.09936724603176117, -0.19332194328308105, -0.05949610844254494, 0.052646737545728683, -0.031055817380547523, 0.150687113404274, -0.12039870023727417, 0.07639284431934357, 0.011438393965363503, -0.0032455516047775745, 0.02790083736181259, -0.16134455800056458, -0.09128429740667343, 0.006392448674887419, 0.07845314592123032, -0.000234358521993272, -0.08157481998205185, -0.054992690682411194, -0.01372462697327137, -0.23895111680030823, 0.04619792476296425, -0.12065518647432327, 0.05270949378609657, 0.024847639724612236, 0.05732692778110504, 0.05480575934052467, -0.003301378805190325, 0.1677713394165039, 0.007408088073134422, -0.044939544051885605, -0.07210525125265121, 0.016201116144657135, 0.08061444014310837, -0.06949795037508011, 0.056983642280101776, 0.06510951370000839, -0.03929603844881058, -0.22190667688846588, -0.008051654323935509, -0.005164995323866606, 0.02921057865023613, -0.03934803605079651, 0.011998923495411873, 0.003252339782193303, 0.06371238827705383, 0.08086985349655151, 0.04554975777864456, 0.11139810085296631, -0.02221781387925148, -0.001815487165004015, 0.0518978051841259, 0.05442829057574272, 0.0443543866276741, -0.0836058184504509, -0.06481249630451202, -0.06968918442726135, 0.02252127230167389, -0.04202992469072342, 0.013006595894694328, 0.03591494634747505, 0.028820395469665527, -0.030995788052678108, 0.04396000877022743, -0.10672979801893234, 0.02112920954823494, 0.06480731070041656, -0.021358348429203033, -0.050103358924388885, -0.05564210191369057, -0.06568542122840881, 0.039567481726408005, -0.126539945602417, 0.05040104314684868, -0.03150033950805664, -0.006824245676398277, 0.04465855285525322, -0.007304768078029156, -0.07160244137048721, 0.017111578956246376, -0.013685770332813263, 0.03504341468214989, -0.05003555864095688, 0.15731048583984375, 0.019666973501443863, -0.059517670422792435, 0.018177516758441925, 0.14088311791419983, -0.10750515013933182, -0.0799887552857399, -0.031066371127963066, 0.09604759514331818, 0.0690813660621643, -0.04535314068198204, 0.020377950742840767, -0.08459163457155228, 0.09959423542022705, -0.07958676666021347, -0.022588882595300674, -0.11175885796546936, 0.060887500643730164, 0.0586186908185482, -0.04879765957593918, 0.08503095805644989, -0.008735095150768757, -0.0476202517747879, -0.0813845694065094, 0.009960105642676353, 0.0391981303691864, 0.12150410562753677, -0.006826474331319332, -0.0277510117739439, -0.17256516218185425, 0.026611266657710075, -0.05199401453137398, -0.014800424687564373, -0.1856328696012497, -0.002534259343519807, -0.026019295677542686, 0.015849847346544266, 0.0391460657119751, 0.03635846823453903, -0.05282532423734665, -0.076359324157238, -0.04103345796465874, 0.12389951944351196, -0.053580280393362045, -0.01581515558063984, -0.023228848353028297, -0.04617864266037941, 0.05127095803618431, 0.08021081984043121, -0.001739910221658647, -0.021222542971372604, -0.11324189603328705, 0.02127097360789776, -0.049999628216028214, -0.05661851540207863, 0.0770464837551117, -0.14767901599407196, 0.040352798998355865, -0.02748228795826435, -0.1045621857047081, 0.025089893490076065, 0.12918245792388916, -0.04590894281864166, 0.07668185234069824, 0.0186065174639225, -0.10533666610717773, -0.08220318704843521, 0.02953350730240345, 0.09348729252815247, 0.03390783444046974, 0.0501854382455349, -0.09583393484354019, 0.1770295798778534, -0.12471824884414673, -0.01643492840230465, 0.0035411177668720484, 0.058029092848300934, -0.005255019757896662, -0.1526361107826233, 0.028937991708517075, -0.018233012408018112, 0.08122190088033676, 0.07048927992582321, 0.05008124187588692, 0.021898416802287102, -0.011411873623728752, 0.1180967316031456, 0.02718423306941986, 0.07752348482608795, -0.036777738481760025, 0.016445036977529526, 0.05632483959197998, 0.0026962507981806993, 0.03618912398815155, -0.10977183282375336, 0.07524891942739487, 0.0977306067943573, 0.08713138848543167, 0.05064034089446068, 0.05722387507557869, -0.08915203809738159, -0.14945314824581146, -0.060108572244644165, 0.060536351054906845, -0.04264168441295624, -0.06567799299955368, 0.09934891760349274, 0.17192089557647705, -0.2449711561203003, 0.0523257739841938, -0.0088726207613945, 0.056279003620147705, -0.05353932827711105, -0.0809951052069664, 0.01926923729479313, -0.18310488760471344, 0.06980645656585693, -0.047987472265958786, 0.006311447359621525, -0.09315139055252075, -0.012581907212734222, 0.013098275288939476, 0.09222442656755447, -0.08202137798070908, -0.0693579912185669, 0.08129197359085083, -0.04771001264452934, 0.05276501178741455, -0.06213199719786644, -0.025753185153007507, -0.02752630226314068, -0.057816531509160995, -0.009277233853936195, 0.08388223499059677, 0.0069085718132555485, 0.05958355963230133, -0.060620710253715515, -0.0736684650182724, 0.08781343698501587, -0.014598358422517776, 0.011937343515455723, 0.11233769357204437, 0.08046351373195648, -0.09314218908548355, -0.03166792541742325, 0.14814773201942444, -0.040232762694358826, -0.06370425969362259, -0.07920711487531662, 0.12740902602672577, -0.00903682503849268, -0.011888126842677593, -0.03248540312051773, -0.12851394712924957, -0.03534495458006859, 0.24809041619300842, 0.08640635013580322, -0.027283992618322372, 0.017293056473135948, -0.06164398789405823, 0.008611548691987991, 0.033084623515605927, 0.11379624158143997, 0.027784988284111023, 0.09039577841758728, -0.0649828091263771, -0.005452395882457495, -0.06436135619878769, -0.06483297049999237, -0.1571880728006363, 0.05160662159323692, 0.04856085032224655, -0.016134222969412804, -0.04564933478832245, 0.12278895825147629, -0.11964233964681625, -0.08351190388202667, 0.1829250305891037, -0.0682426318526268, -0.047147370874881744, -0.025567753240466118, -0.049887675791978836, 0.04996369406580925, 0.08632166683673859, 0.05173683166503906, 0.0471971370279789, 0.07838275283575058, -0.014377528801560402, -0.05183294042944908, -0.01902106963098049, 0.013793517835438251, -0.11194845288991928, 0.18547745048999786, -0.03391994163393974, 0.034558143466711044, 0.05604919046163559, 0.07707247883081436, -0.13280026614665985, 0.0358271524310112, 0.020331719890236855, -0.07773077487945557, 0.03107094205915928, 0.029827160760760307, -0.06653320044279099, 0.02388378418982029, 0.07752751559019089, -0.07278642803430557, 0.014324412681162357, 0.082077756524086, -0.006568188779056072, -0.04485446214675903, 0.10443738102912903, -0.1264781802892685, 0.10661973059177399, 0.10460308939218521, -0.06251361966133118, 0.023263974115252495, -0.01133277639746666, 0.0724470391869545, 0.033656101673841476, 0.06387655436992645, -0.04353664070367813, -0.14152862131595612, 0.013242029584944248, 0.03884240984916687, 0.02863350696861744, -0.24577279388904572, -0.0838109627366066, -0.01984814926981926, -0.05502938851714134, -0.008210626430809498, 0.10551893711090088, 0.11266690492630005, -0.050888676196336746, -0.02009948156774044, -0.21230070292949677, 0.05041462555527687, 0.20308174192905426, -0.03643973916769028, -0.02717341110110283 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
B2111797/recipe_gener_v7
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T04:28:10+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 57, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.05622259899973869, 0.16002345085144043, -0.004987028427422047, 0.023115945979952812, 0.0962471067905426, 0.011845538392663002, 0.06785304099321365, 0.11496778577566147, -0.020396295934915543, 0.11142492294311523, 0.03292480856180191, 0.0972127765417099, 0.11474913358688354, 0.16215258836746216, 0.004439093638211489, -0.23455148935317993, 0.04782992601394653, -0.12695099413394928, -0.033447545021772385, 0.11785799264907837, 0.14491069316864014, -0.10402194410562515, 0.07766910642385483, -0.030544815585017204, -0.009361269883811474, -0.03290390968322754, -0.06365230679512024, -0.05152205005288124, 0.05037128925323486, 0.06932847946882248, 0.06591591984033585, 0.007509593386203051, 0.09122733771800995, -0.2655104100704193, 0.02280162274837494, 0.07630051672458649, -0.0015554219717159867, 0.07497020810842514, 0.048351652920246124, -0.08209776133298874, 0.0788840726017952, -0.05696587264537811, 0.14718368649482727, 0.08216129243373871, -0.08924587815999985, -0.1965435892343521, -0.08464295417070389, 0.10284840315580368, 0.18357418477535248, 0.05158785358071327, -0.024141347035765648, 0.10476154088973999, -0.08419200032949448, 0.008797040209174156, 0.06024181470274925, -0.06443428993225098, -0.05412506312131882, 0.06934051215648651, 0.07975570857524872, 0.07967228442430496, -0.13025140762329102, -0.014651902951300144, 0.011243549175560474, 0.007594773545861244, 0.08504551649093628, 0.022028017789125443, 0.14595499634742737, 0.04393624886870384, -0.13030564785003662, -0.044304780662059784, 0.09771761298179626, 0.04345165938138962, -0.053857799619436264, -0.2537047266960144, -0.024983759969472885, -0.03927002474665642, -0.03094942681491375, -0.038562554866075516, 0.04431856796145439, -0.011080716736614704, 0.08032315224409103, -0.01118796318769455, -0.08149448037147522, -0.041395120322704315, 0.06544242054224014, 0.062143467366695404, 0.026896316558122635, -0.01158317644149065, 0.00973866879940033, 0.1224486380815506, 0.10907839238643646, -0.12763150036334991, -0.05768941715359688, -0.06755511462688446, -0.08307720720767975, -0.04300352931022644, 0.03337155282497406, 0.044020529836416245, 0.04436098039150238, 0.2466370165348053, 0.01108562108129263, 0.05453123152256012, 0.045806169509887695, 0.010608446784317493, 0.06787561625242233, 0.11606968939304352, -0.062306761741638184, -0.09178462624549866, -0.029058339074254036, 0.09215214103460312, 0.006741520017385483, -0.042814407497644424, -0.060904473066329956, 0.06479041278362274, 0.012608112767338753, 0.12110785394906998, 0.08444269746541977, 0.0026690615341067314, -0.07305197417736053, -0.06963318586349487, 0.18848419189453125, -0.1598394364118576, 0.047875016927719116, 0.031182926148176193, -0.038971830159425735, -0.0014042917173355818, 0.008752269670367241, 0.02394084818661213, -0.020246321335434914, 0.08923295140266418, -0.05574449151754379, -0.03784004598855972, -0.11079790443181992, -0.03252100944519043, 0.030985163524746895, 0.0051483530551195145, -0.027043871581554413, -0.033837489783763885, -0.09040277451276779, -0.059588029980659485, 0.0922931432723999, -0.07471107691526413, -0.04984431713819504, -0.013726521283388138, -0.07691634446382523, 0.023329194635152817, 0.016799474135041237, 0.08357251435518265, -0.02157396264374256, 0.0384126678109169, -0.0560205839574337, 0.0631464347243309, 0.11269522458314896, 0.029363946989178658, -0.053069718182086945, 0.05750001594424248, -0.24315528571605682, 0.10326608270406723, -0.07320205867290497, 0.050549428910017014, -0.15059062838554382, -0.026000602170825005, 0.044471126049757004, 0.00805877335369587, -0.013138634152710438, 0.14088952541351318, -0.21621745824813843, -0.0323486253619194, 0.16741067171096802, -0.0939871072769165, -0.07602590322494507, 0.059108685702085495, -0.05233629792928696, 0.10869261622428894, 0.04351044446229935, -0.02232111617922783, 0.060673557221889496, -0.14475463330745697, -0.01067100279033184, -0.04139741137623787, -0.02402937039732933, 0.16397778689861298, 0.07567544281482697, -0.06286642700433731, 0.08052356541156769, 0.024165838956832886, -0.017831770703196526, -0.04484899342060089, -0.023361295461654663, -0.10819391161203384, 0.009856974706053734, -0.06032416597008705, 0.02424289658665657, -0.025761527940630913, -0.09367526322603226, -0.02868773601949215, -0.1802000105381012, -0.009223134256899357, 0.0881323292851448, -0.011722641065716743, -0.021903391927480698, -0.12039245665073395, 0.011948852799832821, 0.031212422996759415, 0.002984174294397235, -0.13029038906097412, -0.05838731303811073, 0.027675874531269073, -0.16422230005264282, 0.03272955119609833, -0.05597274377942085, 0.05056252330541611, 0.03445037454366684, -0.03187771514058113, -0.033117350190877914, 0.009550533257424831, 0.006354342680424452, -0.010578392073512077, -0.2502359449863434, -0.02440580166876316, -0.0219739843159914, 0.17386503517627716, -0.21793730556964874, 0.04213962331414223, 0.07686693966388702, 0.14929872751235962, 0.006240781396627426, -0.038500864058732986, 0.010139784775674343, -0.08222103863954544, -0.030560437589883804, -0.0643099993467331, -0.012082485482096672, -0.03717579320073128, -0.05608142167329788, 0.05165567249059677, -0.16133594512939453, -0.028727244585752487, 0.1057019829750061, 0.06860516220331192, -0.14001330733299255, -0.019125886261463165, -0.04171464592218399, -0.043496038764715195, -0.05877087265253067, -0.0552728995680809, 0.1185101792216301, 0.05596614256501198, 0.04696191847324371, -0.06956122815608978, -0.07775315642356873, 0.007865429855883121, -0.017090093344449997, -0.017978519201278687, 0.08920905739068985, 0.07311701774597168, -0.12023317068815231, 0.09247473627328873, 0.10194233059883118, 0.09365488588809967, 0.108615942299366, -0.017981963232159615, -0.08929306268692017, -0.04584396257996559, 0.02045595459640026, 0.013332244008779526, 0.14797501266002655, -0.01403066236525774, 0.056954506784677505, 0.03922648727893829, -0.01123172789812088, 0.012020308524370193, -0.09384570270776749, 0.027314940467476845, 0.034342724829912186, -0.020308034494519234, 0.03796098753809929, -0.04001156985759735, 0.019826533272862434, 0.08712323755025864, 0.04676510766148567, 0.04415108636021614, 0.011758276261389256, -0.04233846068382263, -0.10904491692781448, 0.173858180642128, -0.12615609169006348, -0.24583272635936737, -0.14115718007087708, 0.0015609683468937874, 0.04152948409318924, -0.009671499952673912, 0.003867273684591055, -0.07054664939641953, -0.11710625886917114, -0.0934595838189125, 0.018713686615228653, 0.04491026699542999, -0.07426843047142029, -0.0596279613673687, 0.059872306883335114, 0.03894329443573952, -0.14430272579193115, 0.022237464785575867, 0.047419775277376175, -0.09032250195741653, -0.006925572175532579, 0.08398029953241348, 0.06729988008737564, 0.17764869332313538, 0.009659109637141228, -0.021044570952653885, 0.03080335259437561, 0.21258224546909332, -0.14283664524555206, 0.11252175271511078, 0.14021345973014832, -0.09024007618427277, 0.08099348843097687, 0.1948828399181366, 0.039186809211969376, -0.10478170961141586, 0.03259138762950897, 0.02489176020026207, -0.028939135372638702, -0.25018003582954407, -0.0680207833647728, 0.002590036718174815, -0.04892077296972275, 0.07092583924531937, 0.0918794497847557, 0.09946957975625992, 0.015428726561367512, -0.09732488542795181, -0.08017807453870773, 0.0468163788318634, 0.10640767961740494, 0.0070237633772194386, -0.01532268337905407, 0.08905128389596939, -0.03260866180062294, 0.018378758803009987, 0.0954233929514885, 0.00412675691768527, 0.17459604144096375, 0.05586163327097893, 0.17767499387264252, 0.07751350849866867, 0.06634163856506348, 0.019167855381965637, 0.0069374511949718, 0.02067388966679573, 0.017508454620838165, -0.004214957356452942, -0.08522020280361176, -0.00457410141825676, 0.12029227614402771, 0.06321834027767181, 0.024303704500198364, 0.0137604009360075, -0.03941800817847252, 0.08438141644001007, 0.17332784831523895, 0.0020201504230499268, -0.18486954271793365, -0.07240456342697144, 0.07921045273542404, -0.0910051167011261, -0.10552998632192612, -0.03353073075413704, 0.03346012532711029, -0.1747758537530899, 0.02097497321665287, -0.017018353566527367, 0.10809773951768875, -0.13855572044849396, -0.018670624122023582, 0.06328251957893372, 0.07232730835676193, -0.0028869258239865303, 0.06308864802122116, -0.153975248336792, 0.1050168052315712, 0.016289174556732178, 0.06754438579082489, -0.09747608006000519, 0.10138221830129623, -0.006303760688751936, -0.007241528946906328, 0.13875643908977509, 0.010596190579235554, -0.05694379657506943, -0.08987913280725479, -0.10555228590965271, -0.008462639525532722, 0.12933635711669922, -0.15157614648342133, 0.0847775787115097, -0.028662750497460365, -0.043171048164367676, 0.0024383023846894503, -0.1199452206492424, -0.1302652359008789, -0.1875755488872528, 0.058235347270965576, -0.1366453617811203, 0.039557021111249924, -0.10582595318555832, -0.04340389743447304, -0.028466427698731422, 0.2041483372449875, -0.2317875325679779, -0.0682469978928566, -0.1541893482208252, -0.08429346233606339, 0.14446710050106049, -0.04730919376015663, 0.08914490789175034, -0.0013825427740812302, 0.19013537466526031, 0.024473950266838074, -0.02387205697596073, 0.10308998823165894, -0.09543927758932114, -0.19450686872005463, -0.08603953570127487, 0.15582145750522614, 0.13931062817573547, 0.03702725097537041, -0.004593946039676666, 0.029260434210300446, -0.020000332966446877, -0.12535293400287628, 0.025526588782668114, 0.1793687790632248, 0.07859015464782715, 0.023437971249222755, -0.025896867737174034, -0.10993997752666473, -0.06524094194173813, -0.0335373692214489, 0.02718053013086319, 0.18264614045619965, -0.07421271502971649, 0.1900695115327835, 0.13626199960708618, -0.05445687845349312, -0.1955246478319168, 0.018216576427221298, 0.040417760610580444, 0.010847307741641998, 0.03138056397438049, -0.2078717201948166, 0.09027513861656189, 0.0014845491386950016, -0.05172133818268776, 0.141556978225708, -0.174949511885643, -0.1512570083141327, 0.06491631269454956, 0.0364508256316185, -0.19348180294036865, -0.117862768471241, -0.08817066252231598, -0.046907443553209305, -0.17498233914375305, 0.10519181191921234, 0.016932250931859016, 0.009516867808997631, 0.03492651879787445, 0.02640140987932682, 0.011080757714807987, -0.03873949125409126, 0.19461296498775482, -0.02505207620561123, 0.029532426968216896, -0.08079101145267487, -0.06136554479598999, 0.0607450045645237, -0.05577658861875534, 0.07896649837493896, -0.020188091322779655, 0.012835816480219364, -0.1100873053073883, -0.0468425452709198, -0.027396185323596, 0.017321845516562462, -0.09195652604103088, -0.09473495930433273, -0.05146971344947815, 0.09373841434717178, 0.08845265954732895, -0.036603908985853195, -0.04043547809123993, -0.07348548620939255, 0.0325477197766304, 0.17183002829551697, 0.17659065127372742, 0.038550034165382385, -0.08084331452846527, -0.005880105309188366, -0.01188716571778059, 0.04436201974749565, -0.22519725561141968, 0.06208868324756622, 0.04557957127690315, 0.015879612416028976, 0.11362850666046143, -0.018783990293741226, -0.16298477351665497, -0.06594224274158478, 0.06143777072429657, -0.06664001196622849, -0.18599680066108704, 0.0032026967965066433, 0.058006007224321365, -0.1646854728460312, -0.037671029567718506, 0.042260222136974335, -0.0045668939128518105, -0.04300284758210182, 0.01627597212791443, 0.08071378618478775, 0.005054219625890255, 0.07112491130828857, 0.05733523517847061, 0.0842885971069336, -0.10417009145021439, 0.07519911974668503, 0.08007751405239105, -0.08229218423366547, 0.031453702598810196, 0.08910130709409714, -0.061817802488803864, -0.03069761022925377, 0.032593827694654465, 0.07753410935401917, 0.019773589447140694, -0.041717879474163055, 0.008655321784317493, -0.09745000302791595, 0.06339588761329651, 0.09504765272140503, 0.03549657016992569, 0.014742289669811726, 0.034356739372015, 0.04988397657871246, -0.07460241764783859, 0.11766603589057922, 0.022336218506097794, 0.01780087500810623, -0.044981084764003754, -0.05459042266011238, 0.032110098749399185, -0.022974027320742607, -0.010163158178329468, -0.03885438293218613, -0.07015778869390488, -0.018130742013454437, -0.15929651260375977, -0.014899281784892082, -0.04085385054349899, 0.007158880587667227, 0.02551902085542679, -0.03834335505962372, 0.007963370531797409, 0.012195355258882046, -0.07085035741329193, -0.061454467475414276, -0.022903166711330414, 0.09224231541156769, -0.16436699032783508, 0.025155464187264442, 0.08285263180732727, -0.12099926173686981, 0.09775067120790482, 0.021939631551504135, 0.0031351554207503796, 0.028338242322206497, -0.1542527824640274, 0.04096807911992073, -0.024365095421671867, 0.01272035762667656, 0.04409142583608627, -0.22033950686454773, 0.001463581225834787, -0.03818526118993759, -0.05954346805810928, -0.010227864608168602, -0.033079732209444046, -0.11291328817605972, 0.09883669763803482, 0.008058897219598293, -0.08219768106937408, -0.030809206888079643, 0.03451729565858841, 0.08243680745363235, -0.02608415111899376, 0.15152283012866974, 0.0016822130419313908, 0.07172226905822754, -0.17519205808639526, -0.021702464669942856, -0.011611736379563808, 0.02207101881504059, -0.014536668546497822, -0.015496513806283474, 0.042471300810575485, -0.02421419881284237, 0.19108575582504272, -0.026401294395327568, 0.038726791739463806, 0.06405707448720932, 0.01593620702624321, -0.014801506884396076, 0.10957890748977661, 0.05975057929754257, 0.02399693801999092, 0.022115202620625496, 0.007329683285206556, -0.039842452853918076, -0.014149460941553116, -0.19538825750350952, 0.06474217027425766, 0.1377464383840561, 0.08781574666500092, -0.01322576031088829, 0.07683692127466202, -0.10024392604827881, -0.12397097796201706, 0.11215250939130783, -0.06283260136842728, -0.007701667957007885, -0.06531554460525513, 0.13346771895885468, 0.14944057166576385, -0.18992236256599426, 0.06835456937551498, -0.06228158622980118, -0.05332518368959427, -0.11744599789381027, -0.1957325041294098, -0.055616896599531174, -0.056456826627254486, -0.014700124971568584, -0.048795297741889954, 0.07307228446006775, 0.05693497136235237, 0.012962869368493557, 0.003600025549530983, 0.0766802653670311, -0.015357231721282005, 0.0008028073934838176, 0.03077360987663269, 0.06600049883127213, 0.013312965631484985, -0.02929985709488392, 0.020537450909614563, -0.007275243755429983, 0.04005419462919235, 0.06378308683633804, 0.038119763135910034, -0.02801438421010971, 0.01591232419013977, -0.03770609200000763, -0.10940317064523697, 0.0409080907702446, -0.028551526367664337, -0.08112191408872604, 0.13721226155757904, 0.02428387477993965, 0.005870606284588575, -0.02180131897330284, 0.24582624435424805, -0.07231455296278, -0.09001907706260681, -0.1473579704761505, 0.10211005061864853, -0.04095151647925377, 0.06560079753398895, 0.04110138490796089, -0.10732010751962662, 0.013498948886990547, 0.12688814103603363, 0.15896959602832794, -0.044884394854307175, 0.020156091079115868, 0.03252736106514931, 0.003683826420456171, -0.04006262496113777, 0.05253688618540764, 0.0694650411605835, 0.14883354306221008, -0.04907030612230301, 0.08928520232439041, 0.005485867150127888, -0.10256236046552658, -0.03822692111134529, 0.11808354407548904, -0.017866896465420723, 0.018703164532780647, -0.057248231023550034, 0.11889533698558807, -0.059861693531274796, -0.23005777597427368, 0.06317704170942307, -0.0720362737774849, -0.14286935329437256, -0.021647587418556213, 0.07456772774457932, -0.017636949196457863, 0.02658887766301632, 0.07326807081699371, -0.07681973278522491, 0.19899281859397888, 0.038975972682237625, -0.05729197710752487, -0.05658522993326187, 0.0789351835846901, -0.114089734852314, 0.2792985737323761, 0.01164181251078844, 0.04984506592154503, 0.10365619510412216, -0.016686614602804184, -0.13768579065799713, 0.015234606340527534, 0.09244892746210098, -0.09004336595535278, 0.03869183734059334, 0.2132277488708496, -0.002569539239630103, 0.1152428612112999, 0.07714667171239853, -0.07265080511569977, 0.04592108353972435, -0.1130065843462944, -0.0718315914273262, -0.086885966360569, 0.09441597014665604, -0.07240451127290726, 0.14123490452766418, 0.12318195402622223, -0.053516924381256104, 0.010368985123932362, -0.031209774315357208, 0.04651070013642311, 0.007842876948416233, 0.10365527868270874, 0.010769560933113098, -0.18099099397659302, 0.022656621411442757, 0.018202748149633408, 0.10856854915618896, -0.17241089046001434, -0.09672945737838745, 0.04725200682878494, 0.001958663808181882, -0.059874359518289566, 0.1282012164592743, 0.057909298688173294, 0.04923510178923607, -0.043742597103118896, -0.017267800867557526, -0.009560109116137028, 0.13584671914577484, -0.10737434774637222, -0.0021453071385622025 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.0` ```yaml base_model: mistralai/Mistral-7B-v0.1 model_type: MistralForCausalLM tokenizer_type: LlamaTokenizer is_mistral_derived_model: true load_in_8bit: false load_in_4bit: false strict: false datasets: - path: mhenrichsen/alpaca_2k_test type: alpaca dataset_prepared_path: last_run_prepared val_set_size: 0.05 output_dir: ./Mistral_FFT sequence_len: 4096 sample_packing: true pad_to_sequence_len: true adapter: lora_model_dir: lora_r: lora_alpha: lora_dropout: lora_target_linear: lora_fan_in_fan_out: wandb_project: wandb_entity: wandb_watch: wandb_name: wandb_log_model: gradient_accumulation_steps: 1 micro_batch_size: 1 num_epochs: 1 optimizer: adamw_bnb_8bit lr_scheduler: cosine learning_rate: 0.0002 train_on_inputs: false group_by_length: false bf16: auto fp16: tf32: false gradient_checkpointing: true early_stopping_patience: resume_from_checkpoint: local_rank: logging_steps: 1 xformers_attention: flash_attention: true flash_attn_cross_entropy: false flash_attn_rms_norm: true flash_attn_fuse_qkv: false flash_attn_fuse_mlp: true warmup_steps: 100 evals_per_epoch: 4 eval_table_size: eval_sample_packing: False saves_per_epoch: 1 debug: deepspeed: deepspeed_configs/zero2.json weight_decay: 0.1 fsdp: fsdp_config: special_tokens: ``` </details><br> # Mistral_FFT This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.2369 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 1 - eval_batch_size: 1 - seed: 42 - distributed_type: multi-GPU - num_devices: 3 - total_train_batch_size: 3 - total_eval_batch_size: 3 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 100 - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.9016 | 0.03 | 1 | 1.1080 | | 0.8288 | 0.25 | 8 | 0.8722 | | 1.0797 | 0.5 | 16 | 0.9858 | | 1.036 | 0.75 | 24 | 1.1281 | | 1.4318 | 1.0 | 32 | 1.2369 | ### Framework versions - Transformers 4.38.0.dev0 - Pytorch 2.1.2+cu118 - Datasets 2.17.0 - Tokenizers 0.15.0
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "mistralai/Mistral-7B-v0.1", "model-index": [{"name": "Mistral_FFT", "results": []}]}
text-generation
joseagmz/Mistral_FFT
[ "transformers", "pytorch", "tensorboard", "safetensors", "mistral", "text-generation", "generated_from_trainer", "base_model:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T04:29:03+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #safetensors #mistral #text-generation #generated_from_trainer #base_model-mistralai/Mistral-7B-v0.1 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
<img src="URL alt="Built with Axolotl" width="200" height="32"/> See axolotl config axolotl version: '0.4.0' Mistral\_FFT ============ This model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on the None dataset. It achieves the following results on the evaluation set: * Loss: 1.2369 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.0002 * train\_batch\_size: 1 * eval\_batch\_size: 1 * seed: 42 * distributed\_type: multi-GPU * num\_devices: 3 * total\_train\_batch\_size: 3 * total\_eval\_batch\_size: 3 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: cosine * lr\_scheduler\_warmup\_steps: 100 * num\_epochs: 1 ### Training results ### Framework versions * Transformers 4.38.0.dev0 * Pytorch 2.1.2+cu118 * Datasets 2.17.0 * Tokenizers 0.15.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 3\n* total\\_train\\_batch\\_size: 3\n* total\\_eval\\_batch\\_size: 3\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 100\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.0" ]
[ "TAGS\n#transformers #pytorch #tensorboard #safetensors #mistral #text-generation #generated_from_trainer #base_model-mistralai/Mistral-7B-v0.1 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 3\n* total\\_train\\_batch\\_size: 3\n* total\\_eval\\_batch\\_size: 3\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 100\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.0" ]
[ 86, 165, 4, 38 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #safetensors #mistral #text-generation #generated_from_trainer #base_model-mistralai/Mistral-7B-v0.1 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 1\n* eval\\_batch\\_size: 1\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 3\n* total\\_train\\_batch\\_size: 3\n* total\\_eval\\_batch\\_size: 3\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 100\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.0" ]
[ -0.10205147415399551, 0.11104954779148102, -0.005103578791022301, 0.07587040215730667, 0.09228968620300293, 0.05022608861327171, 0.13960163295269012, 0.1392371654510498, -0.04063837230205536, 0.13996362686157227, 0.11349740624427795, 0.07068509608507156, 0.07862435281276703, 0.16190506517887115, -0.013364739716053009, -0.23395077884197235, 0.0284760519862175, -0.04201376065611839, -0.11062177270650864, 0.10562101751565933, 0.07911345362663269, -0.11216519773006439, 0.08539211750030518, -0.02120226062834263, -0.10879954695701599, -0.05408269166946411, -0.039131686091423035, -0.021849118173122406, 0.09209329634904861, 0.04078130051493645, 0.09414611011743546, 0.03673205152153969, 0.09989326447248459, -0.22384509444236755, 0.003984707407653332, 0.0767521858215332, 0.01190943457186222, 0.08251094073057175, 0.10622477531433105, 0.020986948162317276, 0.108693927526474, -0.10340917110443115, 0.05484730377793312, 0.03423544391989708, -0.10194722563028336, -0.18054042756557465, -0.09238584339618683, 0.08125151693820953, 0.111581951379776, 0.06358436495065689, -0.0012773004127666354, 0.07067886739969254, -0.04015582799911499, 0.07882550358772278, 0.23846621811389923, -0.2322188764810562, -0.07112189382314682, 0.03733127564191818, 0.05366478115320206, 0.08869029581546783, -0.08457241207361221, -0.025490207597613335, 0.016542954370379448, 0.024229321628808975, 0.09080037474632263, 0.0050606438890099525, 0.03891090676188469, -0.003217353718355298, -0.1407020390033722, -0.07684147357940674, 0.12274319678544998, 0.04981933906674385, -0.008978858590126038, -0.09442498534917831, -0.0668049082159996, -0.1996190994977951, -0.03634101524949074, -0.004114001058042049, 0.023772235959768295, -0.03691738098859787, -0.018971525132656097, 0.02961837872862816, -0.07662142068147659, -0.08266773074865341, 0.019334498792886734, 0.06398151814937592, 0.05563626065850258, 0.00031278884853236377, 0.021366437897086143, 0.1116562932729721, 0.04220530390739441, -0.15772411227226257, -0.006628209259361029, 0.005547582171857357, -0.06784101575613022, -0.013283567503094673, 0.006465314421802759, 0.04128652811050415, 0.07176810503005981, 0.1453552544116974, -0.08076079934835434, 0.06710261106491089, 0.05453978851437569, 0.020707983523607254, -0.06686288863420486, 0.12956318259239197, -0.07844235748052597, -0.06839589774608612, -0.026447344571352005, 0.11439760774374008, 0.0297316312789917, -0.01616210862994194, -0.07562742382287979, 0.03082180954515934, 0.11014078557491302, 0.06390631943941116, 0.020844701677560806, 0.032868560403585434, -0.0630245953798294, -0.025894619524478912, 0.13077464699745178, -0.10269510000944138, 0.03334518522024155, 0.0511801652610302, -0.049929339438676834, -0.01161257829517126, 0.009727970696985722, -0.014516577124595642, -0.025480175390839577, 0.07477706670761108, -0.0848497524857521, -0.03174283355474472, -0.0703268051147461, -0.10551817715167999, 0.04486449435353279, -0.06758584827184677, -0.024157507345080376, -0.08611206710338593, -0.11312054097652435, -0.04428308084607124, 0.036909058690071106, -0.06344007700681686, -0.07180524617433548, -0.03984714671969414, -0.09548372030258179, 0.04570630565285683, -0.002862736349925399, 0.11406044661998749, -0.06364801526069641, 0.0751533955335617, -0.009197449311614037, 0.051293447613716125, 0.07149115949869156, 0.039214324206113815, -0.042207445949316025, 0.0824892595410347, -0.14589381217956543, 0.04556167870759964, -0.09710683673620224, 0.062347736209630966, -0.13148672878742218, -0.102984718978405, 0.025412730872631073, -0.0319475494325161, 0.06892167776823044, 0.11686769872903824, -0.1451936513185501, -0.05349984019994736, 0.16983871161937714, -0.08531217277050018, -0.10555848479270935, 0.11609324812889099, 0.008279827423393726, -0.08469782024621964, 0.011255837045609951, 0.11699224263429642, 0.1492622345685959, -0.11824619024991989, -0.022087037563323975, 0.0041585564613342285, 0.09527303278446198, 0.017750242725014687, 0.10086776316165924, -0.019043073058128357, 0.05116068944334984, 0.011867841705679893, -0.05908355861902237, 0.03428628668189049, -0.08807048946619034, -0.0848429724574089, -0.0388515442609787, -0.07554961740970612, 0.0008594731916673481, 0.0397198423743248, 0.00976471696048975, -0.0653216764330864, -0.12046371400356293, -0.017745180055499077, 0.1177644357085228, -0.07981568574905396, 0.004800752736628056, -0.05530889332294464, 0.09495566785335541, -0.00750574329867959, 0.004238449037075043, -0.149492084980011, -0.11435690522193909, 0.06358359009027481, -0.08469713479280472, -0.013193161226809025, 0.005532289855182171, 0.05566927418112755, 0.10018762201070786, -0.03296889364719391, -0.0478549525141716, -0.017575562000274658, -0.003409880679100752, -0.07663580775260925, -0.2367965131998062, -0.06461818516254425, -0.022196203470230103, 0.14028610289096832, -0.19139660894870758, 0.026799729093909264, 0.036836300045251846, 0.12031581997871399, 0.007937183603644371, -0.038060497492551804, -0.005328606814146042, 0.05183000490069389, -0.04101534187793732, -0.08514610677957535, 0.036795806139707565, -0.013913866132497787, -0.07819896191358566, -0.0003521908656693995, -0.17381489276885986, 0.07366662472486496, 0.08785136044025421, 0.04631226882338524, -0.09048265218734741, -0.030411824584007263, -0.06541575491428375, -0.06446274369955063, -0.0029899992514401674, -0.011129345744848251, 0.09648105502128601, 0.016235923394560814, 0.09924343973398209, -0.06821438670158386, -0.062062203884124756, 0.03607175126671791, 0.01591695286333561, -0.015455816872417927, 0.1419801414012909, 0.07463735342025757, -0.08814491331577301, 0.1326836794614792, 0.10568290203809738, -0.06050722673535347, 0.09608185291290283, -0.07632023096084595, -0.07424241304397583, -0.048327311873435974, 0.05187135934829712, 0.03784717619419098, 0.09276550263166428, -0.06883338838815689, 0.013890586793422699, 0.03445201739668846, 0.009753676131367683, 0.004632663447409868, -0.1714213490486145, 0.00823256466537714, 0.021385831758379936, -0.08983331173658371, 0.037888359278440475, -0.018131602555513382, -0.0024717412889003754, 0.08813696354627609, 0.0062223803251981735, -0.042740289121866226, -0.020178189501166344, -0.019162306562066078, -0.08347038924694061, 0.2050354778766632, -0.1151605099439621, -0.12280405312776566, -0.13796234130859375, 0.036125414073467255, -0.022782335057854652, -0.004937958437949419, 0.02438107691705227, -0.051554687321186066, -0.04757517948746681, -0.09489108622074127, -0.011163000017404556, 0.00001829074608394876, 0.029784219339489937, 0.017060890793800354, 0.004469578620046377, 0.05518035963177681, -0.10255907475948334, -0.003594939596951008, 0.02032715640962124, -0.056915316730737686, 0.02456817962229252, 0.03622644022107124, 0.09864721447229385, 0.15530310571193695, 0.039163943380117416, -0.004458875861018896, -0.01403838861733675, 0.17895206809043884, -0.07958564162254333, 0.02799282781779766, 0.09971015900373459, 0.011649614199995995, 0.07243563234806061, 0.16081105172634125, 0.03954193741083145, -0.04839462414383888, -0.005726081319153309, 0.02198171429336071, -0.020750055089592934, -0.22128647565841675, -0.04618680849671364, -0.043391164392232895, 0.05327976495027542, 0.10778553783893585, 0.052171796560287476, 0.014116179198026657, 0.041929394006729126, -0.044151753187179565, 0.01782136783003807, 0.028832633048295975, 0.06686892360448837, 0.052582401782274246, 0.04514636844396591, 0.11308789998292923, -0.028248365968465805, -0.003663119161501527, 0.051833197474479675, 0.008960695937275887, 0.225430428981781, -0.030728919431567192, 0.22756005823612213, 0.04988579824566841, 0.15870332717895508, 0.0038864859379827976, 0.052671000361442566, 0.021279683336615562, 0.014407290145754814, 0.010143506340682507, -0.06318136304616928, -0.008247983641922474, 0.03886168450117111, 0.036566413938999176, 0.018773026764392853, -0.07265113294124603, 0.04581007733941078, 0.04811064898967743, 0.2856237590312958, 0.050912272185087204, -0.31204402446746826, -0.09217233955860138, 0.03207968547940254, -0.03176344186067581, -0.0249190516769886, 0.008921177126467228, 0.14781567454338074, -0.08915189653635025, 0.06060705706477165, -0.059408415108919144, 0.07739534229040146, -0.06384971737861633, 0.0012181580532342196, 0.11149589717388153, 0.11563421040773392, 0.01875612884759903, 0.06149118021130562, -0.21723733842372894, 0.2431492656469345, -0.004452757071703672, 0.02742261067032814, -0.05082029104232788, 0.05330187827348709, 0.004749925807118416, 0.040212418884038925, 0.08393983542919159, -0.006979281082749367, -0.11864744126796722, -0.18694503605365753, -0.1358848363161087, 0.0022215568460524082, 0.11267425864934921, -0.07580214738845825, 0.1081199049949646, -0.032368555665016174, -0.03904346749186516, 0.0341046042740345, -0.05226081237196922, -0.0784204825758934, -0.10764268785715103, 0.05268080532550812, -0.02961786277592182, -0.0029193793889135122, -0.08401530981063843, -0.08730956166982651, -0.08890248090028763, 0.16155016422271729, -0.17019477486610413, -0.0564071461558342, -0.11976627260446548, 0.047904565930366516, 0.16695468127727509, -0.09833614528179169, 0.045504674315452576, -0.025842372328042984, 0.1027868390083313, 0.021222343668341637, -0.04568665847182274, 0.08556189388036728, -0.08804615586996078, -0.2459443211555481, -0.054173924028873444, 0.12720152735710144, 0.029272831976413727, 0.0588049478828907, -0.03904341906309128, 0.029460696503520012, -0.014556379988789558, -0.10739642381668091, 0.024976998567581177, 0.10503972321748734, 0.06382006406784058, 0.02238212153315544, -0.04435287043452263, -0.01308432500809431, -0.03887290135025978, -0.03511917218565941, 0.080237977206707, 0.28449514508247375, -0.10577777773141861, 0.05268498882651329, 0.05896349251270294, -0.06871555745601654, -0.1782340556383133, -0.055770691484212875, 0.09973971545696259, 0.020647989585995674, -0.009423399344086647, -0.16004200279712677, 0.04917868971824646, 0.11220576614141464, -0.025734659284353256, 0.11044054478406906, -0.3390224277973175, -0.13987405598163605, 0.06651093810796738, 0.08503680676221848, -0.05923612415790558, -0.20073430240154266, -0.07384516298770905, -0.002968677319586277, -0.12153307348489761, 0.08083470910787582, -0.018226729705929756, 0.10138954222202301, -0.02878907136619091, 0.0002312206634087488, 0.009698199108242989, -0.06585696339607239, 0.17453426122665405, -0.0021879100240767, 0.05228699743747711, -0.04972142353653908, -0.009936424903571606, 0.061804670840501785, -0.0717400386929512, 0.029627857729792595, -0.11733773350715637, 0.048566341400146484, -0.07490265369415283, -0.014233868569135666, -0.06992752850055695, 0.014800194650888443, -0.05805031582713127, -0.02561766654253006, -0.044997524470090866, 0.046739362180233, 0.06458362936973572, -0.01817421428859234, 0.10585500299930573, 0.015524695627391338, 0.12827421724796295, 0.15691877901554108, 0.06447616964578629, 0.03273140266537666, -0.09633719176054001, -0.016055431216955185, -0.01482329424470663, 0.031430844217538834, -0.11228447407484055, 0.015234380029141903, 0.14445379376411438, 0.020063528791069984, 0.1067691296339035, 0.045133087784051895, -0.06633971631526947, -0.013026577420532703, 0.06734073907136917, -0.12396708130836487, -0.1499142348766327, 0.005976437591016293, 0.010745447129011154, -0.14999857544898987, -0.0006290266173891723, 0.11335789412260056, -0.042020320892333984, -0.008449737913906574, 0.007529626600444317, 0.07051725685596466, -0.011199025437235832, 0.2200963944196701, 0.03120182827115059, 0.09573440998792648, -0.08604836463928223, 0.07360705733299255, 0.07373198866844177, -0.11149702966213226, 0.015551237389445305, 0.12238749861717224, -0.09223097562789917, -0.03210105001926422, 0.08560796082019806, 0.07276103645563126, -0.004097891040146351, -0.03629552945494652, -0.09887084364891052, -0.13703382015228271, 0.08367998152971268, 0.10169609636068344, 0.03980432450771332, 0.06442722678184509, 0.010051253251731396, 0.020157987251877785, -0.0767652615904808, 0.13252542912960052, 0.06346318125724792, 0.08010025322437286, -0.13632480800151825, 0.10932895541191101, -0.006037987768650055, 0.020037773996591568, -0.001721333828754723, 0.048433706164360046, -0.11906783282756805, -0.028823405504226685, -0.1278727799654007, 0.03431564196944237, -0.06818655133247375, -0.008639350533485413, 0.01042361930012703, -0.0385793000459671, -0.025693589821457863, 0.020174618810415268, -0.07333938777446747, -0.06379896402359009, -0.0543021634221077, 0.07977690547704697, -0.12454552203416824, -0.026884635910391808, 0.0302604828029871, -0.1051601693034172, 0.09120769053697586, 0.02297547087073326, 0.03936685621738434, 0.003244943916797638, -0.08916986733675003, 0.03612101450562477, 0.02892244979739189, 0.021022485569119453, 0.029357101768255234, -0.14709928631782532, -0.009105256758630276, -0.03762134164571762, -0.02062767557799816, -0.00009074130502995104, 0.03510432690382004, -0.11657016724348068, 0.025933662429451942, -0.027346324175596237, -0.05048943683505058, -0.07244198024272919, 0.040541596710681915, 0.09650032222270966, -0.01747920736670494, 0.1318662017583847, -0.07236181944608688, 0.06530889123678207, -0.2220498025417328, -0.008251804858446121, 0.009110924787819386, -0.06874970346689224, -0.07342319190502167, -0.024374308064579964, 0.09251390397548676, -0.05006644129753113, 0.09459884464740753, -0.052744485437870026, 0.027664391323924065, 0.019233502447605133, -0.03046373277902603, 0.039626702666282654, 0.07208621501922607, 0.13382159173488617, 0.032300446182489395, -0.03181714937090874, 0.031700704246759415, -0.013788570649921894, 0.049141064286231995, 0.025421587750315666, 0.18122828006744385, 0.1291033923625946, 0.01655600033700466, 0.06469593197107315, 0.07191230356693268, -0.1486256718635559, -0.12672659754753113, 0.09721167385578156, -0.0959649533033371, 0.12858524918556213, -0.03805823624134064, 0.16499552130699158, 0.09069075435400009, -0.2075779139995575, 0.02883223257958889, -0.04315060377120972, -0.09605875611305237, -0.09726133942604065, -0.09290264546871185, -0.08147204667329788, -0.13580335676670074, -0.010117446072399616, -0.10782245546579361, 0.037490636110305786, 0.07753649353981018, 0.03560889884829521, 0.02546846494078636, 0.11262664943933487, 0.07142575085163116, 0.022533254697918892, 0.033071309328079224, 0.05087420716881752, -0.004418181721121073, -0.018885411322116852, -0.09462473541498184, 0.020344460383057594, -0.023478657007217407, 0.05044659227132797, -0.02341393008828163, -0.029812702909111977, 0.0833415761590004, 0.004986943211406469, -0.08991537988185883, 0.02338995784521103, -0.01840285211801529, 0.010003599338233471, 0.07878965139389038, 0.012386632151901722, 0.0022176741622388363, -0.013584943488240242, 0.15407373011112213, -0.06905080378055573, -0.07739628851413727, -0.10749512910842896, 0.21948622167110443, -0.028812645003199577, -0.014751888811588287, 0.04992855712771416, -0.04962417855858803, -0.029142476618289948, 0.1425287276506424, 0.20049427449703217, -0.02845402993261814, -0.014822565950453281, 0.030471524223685265, -0.013883516192436218, -0.019329678267240524, 0.09069506824016571, 0.10255855321884155, 0.09252645075321198, -0.0681166872382164, -0.008735118433833122, -0.011483324691653252, -0.027094725519418716, -0.07163097709417343, 0.03691234439611435, 0.007827822118997574, 0.009965849108994007, -0.01771985925734043, 0.05887484923005104, -0.05299029499292374, -0.09344898909330368, 0.07786358892917633, -0.17552515864372253, -0.16460274159908295, -0.025985345244407654, 0.09393315017223358, 0.0005982358707115054, 0.0429937019944191, -0.0013052013237029314, -0.02645113877952099, 0.128783717751503, -0.021477164700627327, -0.08392365276813507, -0.0674297884106636, 0.05313778296113014, -0.07247769832611084, 0.17996175587177277, -0.03669577091932297, 0.057184115052223206, 0.1289466768503189, 0.018885888159275055, -0.12644648551940918, 0.022105395793914795, 0.0992271900177002, -0.10728719085454941, 0.0366622656583786, 0.1264578253030777, -0.03701504319906235, 0.08738075196743011, 0.04616988077759743, -0.06701906770467758, -0.01737711951136589, -0.044175803661346436, -0.01723981276154518, -0.05086539313197136, -0.018484242260456085, -0.042450200766325, 0.15891270339488983, 0.19599628448486328, -0.052533652633428574, -0.022044358775019646, -0.027375217527151108, 0.022153163328766823, 0.026043426245450974, 0.12551437318325043, -0.01816023513674736, -0.2501623034477234, 0.039540939033031464, -0.013949686661362648, 0.060062434524297714, -0.16598466038703918, -0.09752063453197479, 0.02104497328400612, -0.03152112290263176, -0.09673244506120682, 0.11420969665050507, 0.05381792411208153, 0.037267331033945084, -0.049947138875722885, -0.08537225425243378, -0.054636623710393906, 0.14928212761878967, -0.1641344428062439, -0.08220068365335464 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
token-classification
SKNahin/NER_Deberta2
[ "transformers", "safetensors", "deberta", "token-classification", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-15T04:40:08+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #deberta #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #deberta #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 49, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #deberta #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.07582134753465652, 0.1588554084300995, -0.0037710394244641066, 0.02595895528793335, 0.11814748495817184, 0.009787439368665218, 0.07563809305429459, 0.10580293834209442, -0.01635763794183731, 0.12473590672016144, 0.039848506450653076, 0.10315580666065216, 0.10913156718015671, 0.19314853847026825, -0.0022752105724066496, -0.2091170996427536, 0.0622878298163414, -0.11678310483694077, 0.010382972657680511, 0.12100998312234879, 0.14090564846992493, -0.11161413788795471, 0.06912626326084137, -0.042492516338825226, -0.018259624019265175, -0.031717754900455475, -0.06345462054014206, -0.05417080596089363, 0.06679418683052063, 0.055831894278526306, 0.0655626431107521, 0.02029980719089508, 0.08335743844509125, -0.2845470905303955, 0.019045550376176834, 0.07766879349946976, 0.002123442944139242, 0.0609569177031517, 0.07383845746517181, -0.07475782185792923, 0.09538372606039047, -0.0625537782907486, 0.15552008152008057, 0.07236507534980774, -0.09746827185153961, -0.1846913844347, -0.08630084246397018, 0.10301622748374939, 0.17936977744102478, 0.055135540664196014, -0.03497261554002762, 0.14088936150074005, -0.07017937302589417, 0.016605326905846596, 0.06468700617551804, -0.07290935516357422, -0.05348815396428108, 0.06037161126732826, 0.0742330476641655, 0.09727266430854797, -0.13022415339946747, -0.010263546369969845, 0.04170728102326393, 0.018862606957554817, 0.11035116761922836, 0.01817622222006321, 0.1315252184867859, 0.02978314645588398, -0.14483527839183807, -0.06033659726381302, 0.10560949891805649, 0.03417903184890747, -0.05956808477640152, -0.24999532103538513, -0.007466548588126898, -0.03454916179180145, -0.028890257701277733, -0.04949915036559105, 0.04322859272360802, -0.02760316990315914, 0.09133703261613846, 0.0029181726276874542, -0.06773355603218079, -0.05198833718895912, 0.09289798140525818, 0.06621047854423523, 0.028308380395174026, -0.028138238936662674, 0.01582835055887699, 0.12160526216030121, 0.10461324453353882, -0.1136874184012413, -0.062054362148046494, -0.06239290162920952, -0.08454962074756622, -0.047138769179582596, 0.0364193357527256, 0.06810244172811508, 0.0531260222196579, 0.2071358561515808, -0.006076968740671873, 0.04858258739113808, 0.03387023136019707, 0.012895005755126476, 0.07188420742750168, 0.07482324540615082, -0.05877743288874626, -0.13807903230190277, -0.02900126948952675, 0.1174110546708107, 0.007540363352745771, -0.030544515699148178, -0.035413552075624466, 0.058053579181432724, 0.051524870097637177, 0.12658938765525818, 0.06639471650123596, 0.015262601897120476, -0.07834838330745697, -0.0524817518889904, 0.18440617620944977, -0.15845707058906555, 0.02396208792924881, 0.016546091064810753, -0.051321811974048615, -0.031176572665572166, 0.016762053593993187, 0.010351779870688915, -0.027110083028674126, 0.09371508657932281, -0.06443753838539124, -0.0473901741206646, -0.10924410820007324, -0.053515657782554626, 0.031546659767627716, -0.020617445930838585, -0.024359513074159622, -0.04260227829217911, -0.12834255397319794, -0.07842771708965302, 0.0685746818780899, -0.0641399547457695, -0.06342477351427078, -0.0371922142803669, -0.06433878093957901, 0.012278573587536812, -0.003310238244011998, 0.1162533387541771, -0.030190767720341682, 0.05034793168306351, -0.05589374527335167, 0.06739762425422668, 0.136927992105484, 0.030936243012547493, -0.06643268465995789, 0.06668223440647125, -0.2147911936044693, 0.10741851478815079, -0.08720431476831436, 0.02960197627544403, -0.16360381245613098, -0.019591867923736572, 0.03655751794576645, 0.035667140036821365, -0.009463734924793243, 0.14313486218452454, -0.18016517162322998, -0.03503832593560219, 0.1872098445892334, -0.12664759159088135, -0.09166831523180008, 0.05657235532999039, -0.06124594062566757, 0.13424597680568695, 0.058792535215616226, -0.021857086569070816, 0.053700074553489685, -0.13882024586200714, -0.023953566327691078, -0.06349705904722214, -0.01692943274974823, 0.1536400020122528, 0.0591365285217762, -0.04715650901198387, 0.029974887147545815, 0.01766093075275421, -0.026321327313780785, -0.049422211945056915, -0.03356151655316353, -0.09505396336317062, 0.010490937158465385, -0.08083124458789825, 0.019366491585969925, -0.024703530594706535, -0.09222285449504852, -0.03895638883113861, -0.15414710342884064, 0.013223775662481785, 0.10018230229616165, -0.003258682554587722, -0.031090516597032547, -0.10073129087686539, -0.0023207797203212976, 0.015256554819643497, -0.005745676811784506, -0.14969810843467712, -0.058194927871227264, 0.02466205134987831, -0.17005768418312073, 0.02577390894293785, -0.044527240097522736, 0.03756798058748245, 0.04352610558271408, -0.04543379694223404, -0.03406783938407898, 0.016558455303311348, 0.021964261308312416, -0.023266246542334557, -0.2578973174095154, -0.01380295492708683, -0.05108672380447388, 0.173564150929451, -0.24888335168361664, 0.04812554270029068, 0.06300827115774155, 0.12234839797019958, 0.008925281465053558, -0.041419293731451035, 0.03762400150299072, -0.05390049144625664, -0.03631659969687462, -0.06965415179729462, -0.008510426618158817, -0.03528457134962082, -0.045454367995262146, 0.03824324160814285, -0.18064084649085999, -0.026336830109357834, 0.11385028064250946, 0.07398278266191483, -0.16777318716049194, -0.07124035060405731, -0.03505503386259079, -0.06051163747906685, -0.07906899601221085, -0.05589747801423073, 0.0883159190416336, 0.04612462595105171, 0.05180882290005684, -0.0675927922129631, -0.060058094561100006, 0.012471389956772327, -0.011969506740570068, -0.0303481537848711, 0.08634323626756668, 0.06659138202667236, -0.12965452671051025, 0.10808708518743515, 0.07467851042747498, 0.07148940861225128, 0.1054377406835556, 0.007283661514520645, -0.0933234840631485, -0.019234750419855118, 0.0287589393556118, 0.014921999536454678, 0.15269333124160767, -0.06041413173079491, 0.03667184337973595, 0.0400078184902668, -0.0238299872726202, 0.008937633596360683, -0.0961742103099823, 0.020881228148937225, 0.0288530420511961, -0.011045076884329319, 0.024098359048366547, -0.05350760370492935, 0.015204832889139652, 0.10605532675981522, 0.033052194863557816, 0.029378650709986687, 0.016013238579034805, -0.04219504073262215, -0.12545357644557953, 0.1795472502708435, -0.09741300344467163, -0.24795091152191162, -0.1238713338971138, -0.004490654915571213, 0.03899503871798515, -0.010224607773125172, 0.022714829072356224, -0.05680026113986969, -0.10976943373680115, -0.10007549822330475, 0.033845312893390656, 0.06473647058010101, -0.08570653945207596, -0.06949411332607269, 0.05287871137261391, 0.042630840092897415, -0.12978382408618927, 0.020360087975859642, 0.04170483723282814, -0.07134698331356049, 0.006720427889376879, 0.05840947851538658, 0.07955096662044525, 0.1792677789926529, 0.010206159204244614, -0.02045300044119358, 0.01256562676280737, 0.21880127489566803, -0.1466495245695114, 0.09328395873308182, 0.141775980591774, -0.06401344388723373, 0.08160442858934402, 0.20209866762161255, 0.030180789530277252, -0.10358195006847382, 0.03875577077269554, 0.03624651953577995, -0.034544941037893295, -0.24263033270835876, -0.07476656138896942, 0.005603810306638479, -0.06450692564249039, 0.10198681056499481, 0.08651577681303024, 0.10705448687076569, 0.04512255638837814, -0.11235596984624863, -0.06125471368432045, 0.05285392701625824, 0.11886214464902878, -0.02292727865278721, -0.002029747236520052, 0.09566584974527359, -0.023103266954421997, 0.022581107914447784, 0.09030848741531372, 0.02752724476158619, 0.18029075860977173, 0.04536215215921402, 0.13428856432437897, 0.08979280292987823, 0.05752971023321152, 0.01326517853885889, 0.018752997741103172, 0.018830908462405205, 0.03053511306643486, -0.020200124010443687, -0.08164151757955551, -0.009718762710690498, 0.13676214218139648, 0.023310136049985886, 0.03856613487005234, 0.0037885159254074097, -0.04802961274981499, 0.0742470771074295, 0.17449264228343964, 0.016129810363054276, -0.22968627512454987, -0.06662201136350632, 0.07476157695055008, -0.07414892315864563, -0.11967059224843979, -0.018175605684518814, 0.028066670522093773, -0.18243610858917236, 0.03899575024843216, -0.024058742448687553, 0.1009836345911026, -0.12115946412086487, -0.02023388259112835, 0.03995189815759659, 0.059546250849962234, -0.031342994421720505, 0.07075176388025284, -0.19329877197742462, 0.1349882036447525, 0.008857686072587967, 0.06893958151340485, -0.10071657598018646, 0.0806516632437706, 0.01673419401049614, 0.0031026999931782484, 0.1628592163324356, -0.004056477919220924, -0.06180410459637642, -0.10156970471143723, -0.085969477891922, -0.01386621780693531, 0.0969386100769043, -0.12537075579166412, 0.09289414435625076, -0.005195782519876957, -0.03387987241148949, -0.002357541350647807, -0.13537321984767914, -0.13680528104305267, -0.17812253534793854, 0.045711059123277664, -0.12590152025222778, 0.04698599874973297, -0.10713580250740051, -0.05455242469906807, -0.039548274129629135, 0.19008749723434448, -0.21812191605567932, -0.08449868857860565, -0.14965391159057617, -0.0647575631737709, 0.11447085440158844, -0.04327482357621193, 0.0829252079129219, 0.011816042475402355, 0.19854897260665894, -0.0006443834863603115, -0.003518373239785433, 0.09804367274045944, -0.09864502400159836, -0.20901942253112793, -0.10028961300849915, 0.1335003674030304, 0.13788676261901855, 0.039611831307411194, 0.004660561680793762, 0.023397788405418396, -0.0015464697498828173, -0.11214857548475266, 0.03287322819232941, 0.15810757875442505, 0.10996973514556885, 0.0365285724401474, -0.025456419214606285, -0.134415864944458, -0.10064711421728134, -0.04832622781395912, 0.007826434448361397, 0.1952485889196396, -0.06921055167913437, 0.16170531511306763, 0.1607891321182251, -0.06080915778875351, -0.20997200906276703, 0.031182322651147842, 0.03326820209622383, -0.0014625083422288299, 0.047811076045036316, -0.20409850776195526, 0.07823505252599716, 0.017151953652501106, -0.057960692793130875, 0.13218380510807037, -0.18051816523075104, -0.14781798422336578, 0.09219598770141602, 0.07739659398794174, -0.19650183618068695, -0.1307976245880127, -0.09491356462240219, -0.048903729766607285, -0.0986228659749031, 0.09035423398017883, -0.008246852084994316, 0.005571926478296518, 0.03216136619448662, 0.01862735114991665, 0.015546685084700584, -0.0485796183347702, 0.19612248241901398, -0.0004862592031713575, 0.051562365144491196, -0.07412069290876389, -0.07256710529327393, 0.03558439388871193, -0.07121629267930984, 0.08863470703363419, -0.01741286925971508, 0.006189014296978712, -0.11463242024183273, -0.06591381877660751, -0.04193699732422829, 0.03264250233769417, -0.08610426634550095, -0.09879012405872345, -0.04721202328801155, 0.10410426557064056, 0.09013193845748901, -0.03732464835047722, -0.06628967076539993, -0.08886818587779999, 0.04777882248163223, 0.21941284835338593, 0.18009066581726074, 0.07462029904127121, -0.07408244162797928, -0.006450060289353132, -0.02119174227118492, 0.06165141984820366, -0.20728842914104462, 0.04867798089981079, 0.0371476411819458, 0.03147519752383232, 0.1294110119342804, -0.026139631867408752, -0.16417503356933594, -0.05014671012759209, 0.05455317720770836, -0.07321128249168396, -0.15782763063907623, 0.007312591653317213, 0.08040127158164978, -0.15300150215625763, -0.0399666391313076, 0.04038683697581291, -0.03046327270567417, -0.031874220818281174, 0.002157751005142927, 0.08237545937299728, 0.020008308812975883, 0.10707804560661316, 0.06407194584608078, 0.10628766566514969, -0.10405796021223068, 0.07105699181556702, 0.08430515974760056, -0.10969457775354385, 0.03989127278327942, 0.057201582938432693, -0.06448391079902649, -0.035053741186857224, 0.03124030865728855, 0.08985407650470734, 0.025547293946146965, -0.07282499223947525, 0.004657800309360027, -0.11015290766954422, 0.06599664688110352, 0.1322651207447052, 0.04004150256514549, 0.0137094771489501, 0.042923275381326675, 0.033320680260658264, -0.10062410682439804, 0.11902371048927307, 0.05028797686100006, 0.03921814635396004, -0.055055614560842514, -0.02014131471514702, 0.040924906730651855, -0.020380757749080658, -0.0170382559299469, -0.038348399102687836, -0.07091455906629562, -0.012284004129469395, -0.1709912270307541, 0.021539563313126564, -0.0643090084195137, 0.01184446457773447, 0.016018250957131386, -0.028759492561221123, 0.0066799623891711235, 0.013610909692943096, -0.07205694168806076, -0.04707391932606697, -0.0035674909595400095, 0.10853823274374008, -0.1683083474636078, 0.006899724714457989, 0.0807538852095604, -0.12974224984645844, 0.08663125336170197, 0.0028446076903492212, -0.0010952386073768139, 0.01968505047261715, -0.1273455023765564, 0.06220780685544014, -0.005134644918143749, 0.0054603819735348225, 0.03317052498459816, -0.21593356132507324, 0.0022072684951126575, -0.0508023202419281, -0.06484345346689224, -0.0030691337306052446, -0.03311242535710335, -0.11482230573892593, 0.10371197760105133, 0.01689654216170311, -0.0779203251004219, -0.021994326263666153, 0.04691578447818756, 0.10877161473035812, -0.04842940717935562, 0.14637391269207, -0.017828896641731262, 0.059009574353694916, -0.18030768632888794, -0.021076207980513573, -0.017547423020005226, 0.01863904669880867, -0.03460176661610603, -0.004361789207905531, 0.05265168845653534, -0.018918506801128387, 0.22060424089431763, -0.02269667573273182, 0.028595924377441406, 0.06291016191244125, -0.00589276198297739, -0.013385145924985409, 0.09999861568212509, 0.045825641602277756, 0.01133752427995205, 0.023678036406636238, 0.006810458842664957, -0.04093582555651665, -0.006517220288515091, -0.1356368511915207, 0.07505546510219574, 0.16590814292430878, 0.08395548164844513, -0.005548030138015747, 0.048898518085479736, -0.11000822484493256, -0.10603458434343338, 0.09506268054246902, -0.034125830978155136, -0.017605112865567207, -0.0482698529958725, 0.13643765449523926, 0.1578799933195114, -0.19135646522045135, 0.06331014633178711, -0.06746772676706314, -0.056378982961177826, -0.10576022416353226, -0.17683377861976624, -0.05985268950462341, -0.03730166703462601, -0.013985122554004192, -0.06049531325697899, 0.058749984949827194, 0.10166352242231369, 0.014109558425843716, 0.006751217879354954, 0.08392417430877686, -0.024542953819036484, 0.0066293105483055115, 0.039478328078985214, 0.06338965147733688, 0.015577137470245361, -0.05866408720612526, 0.007817541249096394, 0.0013342619640752673, 0.03782322257757187, 0.05102457478642464, 0.030786391347646713, -0.013074606657028198, 0.0072399284690618515, -0.022229334339499474, -0.10115569829940796, 0.04019651189446449, -0.023845037445425987, -0.053124140948057175, 0.1545589715242386, 0.024088291451334953, -0.007400582078844309, -0.020909536629915237, 0.23240384459495544, -0.06599795818328857, -0.07759324461221695, -0.1394353210926056, 0.15041467547416687, -0.03572135791182518, 0.05887357145547867, 0.04635220393538475, -0.1078629195690155, 0.03809856250882149, 0.13536633551120758, 0.1431995928287506, -0.04122826084494591, 0.011672002263367176, 0.009141024202108383, 0.003373268526047468, -0.028277069330215454, 0.05200572684407234, 0.05153002589941025, 0.1268187165260315, -0.06070486456155777, 0.09818972647190094, -0.006758245639503002, -0.09154469519853592, -0.024163398891687393, 0.13355377316474915, 0.002042521024122834, 0.023738635703921318, -0.08125443011522293, 0.12700043618679047, -0.05307674780488014, -0.2563377320766449, 0.07183146476745605, -0.06298090517520905, -0.15155023336410522, -0.018484145402908325, 0.018409613519906998, -0.004543652758002281, 0.024991953745484352, 0.06298631429672241, -0.0635342076420784, 0.16134211421012878, 0.03548634424805641, -0.06381282955408096, -0.07436368614435196, 0.07651722431182861, -0.07090579718351364, 0.3068118989467621, 0.006135785020887852, 0.05688599869608879, 0.09663521498441696, -0.036371853202581406, -0.1401294320821762, 0.03042936511337757, 0.09035595506429672, -0.04897838085889816, 0.061034440994262695, 0.2049657553434372, -0.012143735773861408, 0.11445674300193787, 0.07419917732477188, -0.08735579252243042, 0.04469425231218338, -0.0924081802368164, -0.09559591859579086, -0.09037184715270996, 0.09341852366924286, -0.055895816534757614, 0.14971210062503815, 0.1230359748005867, -0.04733745753765106, 0.009310171008110046, -0.018910879269242287, 0.05801442265510559, 0.003590804524719715, 0.11238192021846771, 0.02891220524907112, -0.19292861223220825, 0.030544010922312737, -0.0002170102234231308, 0.10042719542980194, -0.2395145148038864, -0.0887673869729042, 0.042938072234392166, -0.0011809789575636387, -0.05844983085989952, 0.12348521500825882, 0.047044940292835236, 0.04414747655391693, -0.05313320830464363, -0.043901123106479645, -0.011250684037804604, 0.16178865730762482, -0.10214143991470337, -0.0059404755011200905 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # whisper-small-hi This model is a fine-tuned version of [vasista22/whisper-hindi-small](https://huggingface.co/vasista22/whisper-hindi-small) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0000 - Wer: 0.0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - training_steps: 4000 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:------:|:----:|:---------------:|:---:| | 0.0001 | 1000.0 | 1000 | 0.0001 | 0.0 | | 0.0 | 2000.0 | 2000 | 0.0000 | 0.0 | | 0.0 | 3000.0 | 3000 | 0.0000 | 0.0 | | 0.0 | 4000.0 | 4000 | 0.0000 | 0.0 | ### Framework versions - Transformers 4.38.0.dev0 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "vasista22/whisper-hindi-small", "model-index": [{"name": "whisper-small-hi", "results": []}]}
automatic-speech-recognition
Aditya757864/whisper-small-hi
[ "transformers", "tensorboard", "safetensors", "whisper", "automatic-speech-recognition", "generated_from_trainer", "base_model:vasista22/whisper-hindi-small", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2024-02-15T04:48:04+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #base_model-vasista22/whisper-hindi-small #license-apache-2.0 #endpoints_compatible #region-us
whisper-small-hi ================ This model is a fine-tuned version of vasista22/whisper-hindi-small on the None dataset. It achieves the following results on the evaluation set: * Loss: 0.0000 * Wer: 0.0 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 1e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 500 * training\_steps: 4000 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.38.0.dev0 * Pytorch 2.1.0+cu121 * Datasets 2.17.0 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 4000\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #base_model-vasista22/whisper-hindi-small #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 4000\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ 72, 130, 4, 38 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #base_model-vasista22/whisper-hindi-small #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 4000\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ -0.13280795514583588, 0.12275300174951553, -0.003591773798689246, 0.07313892245292664, 0.10650574415922165, -0.0032112705521285534, 0.1263386458158493, 0.13842539489269257, -0.055297620594501495, 0.08970507234334946, 0.0983264222741127, 0.09303717315196991, 0.06187349930405617, 0.15507325530052185, -0.040786005556583405, -0.2565773129463196, 0.02946130558848381, 0.006196336820721626, -0.04906604811549187, 0.11995530873537064, 0.09135676920413971, -0.12611374258995056, 0.030605174601078033, -0.001582689699716866, -0.09466569870710373, -0.025266584008932114, 0.010482246987521648, -0.09743169695138931, 0.10740852355957031, -0.008346575312316418, 0.10032612085342407, 0.05581187084317207, 0.0863095223903656, -0.20002439618110657, 0.01692219078540802, 0.056682031601667404, 0.04062271863222122, 0.06898599863052368, 0.06503108888864517, -0.003055352484807372, 0.07826648652553558, -0.07084313780069351, 0.08754641562700272, 0.027048680931329727, -0.09595687687397003, -0.29117920994758606, -0.08312512189149857, 0.06658807396888733, 0.10126521438360214, 0.08603623509407043, -0.010777683928608894, 0.10087238997220993, -0.042468488216400146, 0.09646792709827423, 0.2230679839849472, -0.2655259370803833, -0.0480327345430851, -0.042714521288871765, 0.03707534447312355, 0.08574539422988892, -0.08487386256456375, -0.030977029353380203, 0.020374707877635956, 0.047166887670755386, 0.10713790357112885, -0.004992593079805374, -0.0586697980761528, -0.024151571094989777, -0.13860541582107544, -0.0615551620721817, 0.10777515172958374, 0.013251874595880508, -0.04554695263504982, -0.0821966901421547, -0.0609157420694828, -0.1713588386774063, -0.05854904651641846, -0.0033738536294549704, 0.022655487060546875, -0.04024964198470116, -0.05886746197938919, -0.008473617024719715, -0.07982105761766434, -0.09134157747030258, 0.010176518000662327, 0.1764010787010193, 0.06431115418672562, 0.007330296561121941, -0.02374173142015934, 0.08002512902021408, 0.033333733677864075, -0.1613309383392334, -0.0050949826836586, 0.030525272712111473, -0.01207397598773241, 0.010521444492042065, -0.030383335426449776, -0.051562804728746414, 0.03156966716051102, 0.12507440149784088, -0.10422607511281967, 0.07529597729444504, 0.002059251768514514, 0.033056966960430145, -0.11991430819034576, 0.17771653831005096, -0.040029048919677734, -0.007778857834637165, 0.022449200972914696, 0.13517417013645172, 0.06738560646772385, -0.027143031358718872, -0.07256902754306793, 0.010223476216197014, 0.10621906816959381, 0.06469609588384628, -0.026719143614172935, 0.03667827323079109, -0.054510075598955154, -0.011441675946116447, 0.026027627289295197, -0.13044416904449463, 0.006910684984177351, 0.023931337520480156, -0.0556352436542511, -0.04440974071621895, 0.033253978937864304, -0.006508121266961098, -0.04521818831562996, 0.07771217077970505, -0.05455808714032173, -0.008371195755898952, -0.07452714443206787, -0.1044129729270935, 0.01811010017991066, -0.0700879618525505, -0.009349332191050053, -0.10233572125434875, -0.14097467064857483, -0.02471686154603958, 0.025673968717455864, -0.030102506279945374, -0.051768507808446884, -0.05744723975658417, -0.09591037780046463, 0.03740255907177925, -0.03043263591825962, 0.065883107483387, -0.06004450097680092, 0.11962413787841797, 0.03265107795596123, 0.07634400576353073, 0.02619076333940029, 0.05236000195145607, -0.0671815350651741, 0.049384262412786484, -0.16880667209625244, 0.06789235025644302, -0.09805115312337875, 0.04475102201104164, -0.10326284170150757, -0.10694839805364609, 0.03339105471968651, -0.013150055892765522, 0.11422445625066757, 0.11464595794677734, -0.16167280077934265, -0.07649850100278854, 0.20697282254695892, -0.1317894160747528, -0.11169017106294632, 0.13517336547374725, -0.0012362998677417636, -0.027582544833421707, 0.053962867707014084, 0.21363481879234314, 0.10848847776651382, -0.11125880479812622, 0.006069973576813936, -0.015352408401668072, 0.07290475815534592, -0.025684352964162827, 0.08214318752288818, -0.016319040209054947, 0.018906734883785248, 0.01964619569480419, -0.03746132180094719, 0.05831894651055336, -0.08813630044460297, -0.08784470707178116, -0.037651609629392624, -0.09146410971879959, 0.040089111775159836, 0.04539396986365318, 0.029930943623185158, -0.10649021714925766, -0.10427625477313995, -0.0018682396039366722, 0.1285993456840515, -0.08868489414453506, 0.030603114515542984, -0.10854149609804153, 0.11166691035032272, -0.03137754276394844, -0.01145889237523079, -0.16541504859924316, 0.007845335640013218, 0.04350753873586655, -0.05914483591914177, 0.0035847993567585945, -0.08275347948074341, 0.07438350468873978, 0.07668822258710861, -0.037126798182725906, -0.07233158499002457, -0.05028744041919708, 0.000039717771869618446, -0.08560556173324585, -0.20302565395832062, -0.04616285860538483, -0.039394788444042206, 0.11444620043039322, -0.17149539291858673, 0.03174600750207901, 0.019391469657421112, 0.0990455225110054, 0.036906614899635315, -0.0283811092376709, 0.011930323205888271, 0.08365374058485031, 0.0010507393162697554, -0.0645454078912735, 0.05744054540991783, 0.02141558937728405, -0.09402615576982498, 0.03954727575182915, -0.16197176277637482, 0.10448116809129715, 0.11500594764947891, -0.007312926463782787, -0.054361287504434586, 0.01179423090070486, -0.05272288993000984, -0.03803476318717003, -0.008153589442372322, 0.016379551962018013, 0.16605882346630096, 0.012844969518482685, 0.140509694814682, -0.09363873302936554, -0.02427653595805168, 0.03455020859837532, -0.023887164890766144, 0.003738716011866927, 0.12438451498746872, 0.022792445495724678, -0.012331821024417877, 0.1135745495557785, 0.11029326170682907, -0.08822081238031387, 0.1334298551082611, -0.07795537263154984, -0.08323682099580765, -0.010955600999295712, 0.027589092031121254, 0.025163156911730766, 0.10271328687667847, -0.1221994087100029, -0.014149974100291729, 0.028762497007846832, 0.005734769627451897, 0.016074396669864655, -0.21704410016536713, -0.004556865431368351, 0.025881286710500717, -0.08087756484746933, -0.03343962877988815, 0.00030018624966032803, 0.016049155965447426, 0.10143949091434479, -0.0032394651789218187, -0.06668721139431, 0.005402504000812769, -0.023181477561593056, -0.07452801614999771, 0.18704932928085327, -0.1142706647515297, -0.19294694066047668, -0.1301659196615219, -0.006167907267808914, -0.008268388919532299, 0.0014753196155652404, 0.062118157744407654, -0.10290947556495667, -0.04358730465173721, -0.08788453787565231, 0.011162523180246353, 0.015698067843914032, 0.030486220493912697, 0.02618682198226452, 0.002265924820676446, 0.08617054671049118, -0.10882585495710373, -0.003763417014852166, -0.027690628543496132, -0.035230137407779694, 0.03346555307507515, 0.05597285181283951, 0.07654742151498795, 0.16139617562294006, -0.006548761390149593, 0.012852441519498825, -0.030173083767294884, 0.1717441976070404, -0.07963256537914276, -0.02479369193315506, 0.18072882294654846, -0.01307281106710434, 0.06140971928834915, 0.14270690083503723, 0.033690303564071655, -0.08277714997529984, -0.0014317813329398632, -0.007579516619443893, -0.03666619211435318, -0.23597921431064606, -0.056786637753248215, -0.03581112623214722, 0.02220684476196766, 0.08818702399730682, 0.03530705347657204, -0.003299236297607422, 0.04010049253702164, -0.012597729451954365, -0.010217205621302128, 0.010990439914166927, 0.07192245125770569, 0.10109663754701614, 0.023622721433639526, 0.123548723757267, -0.03179153427481651, -0.01966671459376812, 0.02698771096765995, 0.0011155239772051573, 0.22359363734722137, -0.03354862332344055, 0.16789397597312927, 0.05424480140209198, 0.16956888139247894, 0.041290294378995895, 0.050525594502687454, -0.007550188340246677, -0.004313078708946705, 0.01826503500342369, -0.055443763732910156, -0.07315146178007126, 0.01120955403894186, -0.014115379191935062, 0.0540626123547554, -0.11780223250389099, 0.03284686803817749, 0.0230417400598526, 0.3229511082172394, 0.04868699610233307, -0.3102239966392517, -0.11538200080394745, 0.005687727127224207, -0.06367471069097519, -0.03993631526827812, 0.051200900226831436, 0.14812856912612915, -0.06057857722043991, 0.035574838519096375, -0.06486035883426666, 0.0759659931063652, -0.06840506941080093, 0.01612464338541031, 0.03804792836308479, 0.09438072890043259, 0.017602281644940376, 0.043236784636974335, -0.24451744556427002, 0.281030535697937, -0.006163068115711212, 0.08888593316078186, -0.03630603477358818, 0.014536608010530472, 0.026997817680239677, 0.014999045990407467, 0.08406278491020203, -0.014688022434711456, -0.08207882195711136, -0.18938089907169342, -0.10879279673099518, 0.01874689944088459, 0.11596211045980453, 0.01602599397301674, 0.10269150882959366, -0.03512376919388771, -0.014654584228992462, 0.03835304453969002, -0.11897998303174973, -0.06351790577173233, -0.068222776055336, 0.02024889923632145, 0.0588856004178524, 0.023646248504519463, -0.11110623180866241, -0.10692483931779861, -0.06001431122422218, 0.1002359464764595, -0.09877852350473404, -0.07034172117710114, -0.10747223347425461, 0.03646448254585266, 0.12929047644138336, -0.08647051453590393, 0.058382291346788406, 0.007447331678122282, 0.10602083802223206, 0.014104041270911694, -0.039468880742788315, 0.08019442111253738, -0.08302264660596848, -0.21836283802986145, -0.03656614571809769, 0.18004418909549713, 0.017752448096871376, 0.05289355292916298, -0.019260283559560776, 0.023854468017816544, -0.00919449981302023, -0.07229788601398468, 0.03323063254356384, 0.04909438639879227, 0.025770079344511032, 0.029165271669626236, -0.03040073625743389, -0.03434615582227707, -0.07190832495689392, -0.03511020913720131, 0.14917203783988953, 0.26589879393577576, -0.0787154957652092, 0.09140753000974655, 0.10703346878290176, -0.04684619978070259, -0.21061399579048157, -0.03339243307709694, 0.09626132994890213, 0.01834542490541935, -0.021515265107154846, -0.17246992886066437, 0.04512125253677368, 0.07107926905155182, -0.041793666779994965, 0.09058459848165512, -0.32266199588775635, -0.14341548085212708, 0.11621024459600449, 0.1376161277294159, 0.06592688709497452, -0.15808123350143433, -0.06473373621702194, -0.004543788731098175, -0.086994469165802, 0.057177186012268066, -0.080694280564785, 0.11552119255065918, -0.0209124106913805, 0.06042541563510895, 0.024850184097886086, -0.06567330658435822, 0.12698425352573395, -0.04483689367771149, 0.057449568063020706, -0.045366644859313965, 0.025514602661132812, 0.04674794152379036, -0.056113652884960175, 0.03705745190382004, -0.07257080078125, 0.06193196401000023, -0.06352865695953369, -0.028746018186211586, -0.09411994367837906, 0.010612696409225464, -0.022778920829296112, -0.03376098349690437, -0.004217394627630711, 0.03185300901532173, 0.05604666471481323, -0.007339233532547951, 0.12094743549823761, -0.023476164788007736, 0.15063777565956116, 0.13565479218959808, 0.10459139198064804, -0.0447491817176342, -0.051203805953264236, -0.006027937866747379, -0.03583288565278053, 0.06494147330522537, -0.13161630928516388, 0.03813037648797035, 0.12083014845848083, 0.035211823880672455, 0.1214217096567154, 0.05272230878472328, -0.05573531985282898, 0.02978011965751648, 0.06698105484247208, -0.11609625816345215, -0.15117810666561127, 0.000605731678660959, 0.049825917929410934, -0.11267033964395523, 0.045869432389736176, 0.11878717690706253, -0.07212536036968231, -0.01641053333878517, -0.013650092296302319, 0.03147418051958084, -0.03160158544778824, 0.21320928633213043, 0.05614088103175163, 0.07431910932064056, -0.10549090802669525, 0.10468760877847672, 0.04695204645395279, -0.1434924602508545, 0.04411010071635246, 0.11523792147636414, -0.0985371321439743, -0.038453251123428345, 0.00757264532148838, 0.08010583370923996, 0.031960126012563705, -0.0713135227560997, -0.1402072012424469, -0.13622869551181793, 0.08084413409233093, 0.1595039963722229, 0.040800128132104874, 0.02232293225824833, -0.03444031625986099, 0.02955821342766285, -0.10934453457593918, 0.11840089410543442, 0.05678422003984451, 0.058571912348270416, -0.13937847316265106, 0.16307328641414642, 0.020146820694208145, 0.04203985258936882, -0.019549956545233727, 0.011120508424937725, -0.09379398077726364, 0.025932537391781807, -0.15927204489707947, 0.029772499576210976, -0.03900522366166115, -0.0029249796643853188, -0.006771255750209093, -0.048102959990501404, -0.046197958290576935, 0.04882419854402542, -0.10280817747116089, -0.03141652047634125, -0.0114411860704422, 0.03802593797445297, -0.1061471551656723, -0.03521022945642471, 0.026125164702534676, -0.09439755976200104, 0.09685306251049042, 0.06346943229436874, -0.0020762248896062374, 0.053609948605298996, -0.08852118998765945, -0.004577721003443003, 0.048870690166950226, 0.01661870628595352, 0.02113797888159752, -0.13352231681346893, -0.01273853238672018, -0.006576987449079752, -0.00866010319441557, 0.0007077474729157984, 0.09698093682527542, -0.12431333214044571, 0.013953402638435364, -0.0077881100587546825, -0.03026825189590454, -0.07237183302640915, 0.03292965516448021, 0.07361600548028946, 0.03401252627372742, 0.1455567181110382, -0.11243987083435059, 0.039257682859897614, -0.19807536900043488, 0.0012592601124197245, -0.0286362636834383, -0.09551293402910233, -0.10645837336778641, -0.007819975726306438, 0.09242306649684906, -0.06698409467935562, 0.0741555392742157, -0.04128547012805939, 0.04949670284986496, 0.03197258338332176, -0.07134762406349182, 0.0038119666278362274, 0.04433198645710945, 0.19542773067951202, 0.0409247986972332, -0.040060125291347504, 0.05199332907795906, -0.015196104533970356, 0.0620882622897625, 0.09451691806316376, 0.13296937942504883, 0.18341775238513947, 0.08479399979114532, 0.08514903485774994, 0.06803908199071884, -0.061028700321912766, -0.1760627031326294, 0.039805617183446884, -0.05957252159714699, 0.10739372670650482, -0.010185322724282742, 0.22555455565452576, 0.13191762566566467, -0.14180554449558258, 0.06372110545635223, -0.03129258006811142, -0.08109280467033386, -0.10633177310228348, -0.07400495558977127, -0.082253098487854, -0.14379197359085083, 0.008728865534067154, -0.1081719845533371, 0.03714407980442047, 0.07884510606527328, 0.02958393096923828, 0.009543310850858688, 0.14761507511138916, 0.03971964120864868, 0.01719674840569496, 0.08746778964996338, 0.016580641269683838, -0.013917258940637112, -0.020902365446090698, -0.10846119374036789, 0.044241245836019516, -0.003961574751883745, 0.04743807017803192, -0.024787751957774162, -0.08345706760883331, 0.04572936147451401, -0.0060513755306601524, -0.10843399167060852, 0.03291013836860657, 0.006649099290370941, 0.07137906551361084, 0.08199276775121689, 0.03001512959599495, -0.006287498865276575, -0.0056850784458220005, 0.23682618141174316, -0.08800079673528671, -0.11070779711008072, -0.11319095641374588, 0.2527534067630768, 0.014557878486812115, -0.02375718578696251, 0.03318263962864876, -0.07951780408620834, -0.04075352102518082, 0.17565779387950897, 0.1654212325811386, -0.03467663377523422, 0.001573087996803224, -0.012424668297171593, -0.008716845884919167, -0.07762549072504044, 0.11273422092199326, 0.14334002137184143, 0.0782570168375969, -0.06555217504501343, -0.03596263751387596, -0.04876701533794403, -0.019172143191099167, -0.06880801916122437, 0.06838680803775787, -0.015827197581529617, -0.009689697064459324, -0.04078768938779831, 0.06024894118309021, -0.06196719408035278, -0.11287284642457962, 0.0197414793074131, -0.2015303373336792, -0.17413103580474854, -0.022047756239771843, 0.07397007197141647, 0.03134312853217125, 0.0319422222673893, 0.009761244989931583, 0.005684920120984316, 0.07961694151163101, -0.013777617365121841, -0.04134686291217804, -0.07341387122869492, 0.08310122042894363, -0.10615681856870651, 0.19768604636192322, -0.028543587774038315, 0.05097917839884758, 0.1096542477607727, 0.06522595137357712, -0.09603617340326309, 0.08924222737550735, 0.07043648511171341, -0.11863909661769867, 0.026395907625555992, 0.1811792254447937, -0.04348872974514961, 0.1015729084610939, 0.034096308052539825, -0.11380232870578766, -0.018197841942310333, -0.05073988437652588, -0.0632539689540863, -0.04944323003292084, -0.019647806882858276, -0.04395345225930214, 0.12700745463371277, 0.1791222244501114, -0.06033537536859512, -0.0156268160790205, -0.05108011141419411, 0.006072756834328175, 0.0431843027472496, 0.050909314304590225, -0.019746560603380203, -0.2551959753036499, 0.013408822007477283, 0.015845758840441704, 0.022221392020583153, -0.2557966113090515, -0.07165069878101349, 0.0062532625161111355, -0.049183014780282974, -0.09445945918560028, 0.09205612540245056, 0.0719868615269661, 0.04811985418200493, -0.04278358444571495, -0.055222038179636, -0.032494012266397476, 0.1825321912765503, -0.17686937749385834, -0.07977034896612167 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
null
Junmai/sample_tokenizer
[ "transformers", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
2024-02-15T04:49:01+00:00
[ "1910.09700" ]
[]
TAGS #transformers #arxiv-1910.09700 #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #arxiv-1910.09700 #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 26, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.08389580249786377, 0.19830818474292755, -0.0013316317927092314, 0.02313883788883686, 0.11396584659814835, 0.01961737498641014, 0.053626976907253265, 0.14538456499576569, 0.0060051376931369305, 0.10656800121068954, 0.066679947078228, 0.09131570905447006, 0.09678101539611816, 0.20042605698108673, 0.04371999576687813, -0.17659740149974823, 0.010636410675942898, -0.06930278241634369, -0.010073255747556686, 0.11651819199323654, 0.141214057803154, -0.10151198506355286, 0.07627976685762405, -0.03319970890879631, -0.02870541252195835, -0.0070160143077373505, -0.07769215852022171, -0.05755697935819626, 0.07573003321886063, 0.054863471537828445, 0.04207949340343475, -0.0008347301045432687, 0.08447454124689102, -0.2674994468688965, 0.013753628358244896, 0.07452993094921112, 0.010659529827535152, 0.05990942195057869, 0.07833302766084671, -0.04036625102162361, 0.12881849706172943, -0.06320446729660034, 0.13035163283348083, 0.0906217098236084, -0.0681561604142189, -0.24378153681755066, -0.08239314705133438, 0.06505522131919861, 0.12533815205097198, 0.07694927603006363, -0.02823091857135296, 0.16422191262245178, -0.07247646898031235, 0.019290022552013397, 0.09481704235076904, -0.1151006743311882, -0.060644298791885376, 0.08318385481834412, 0.14101974666118622, 0.10340547561645508, -0.1255619376897812, -0.012289565056562424, 0.04275871813297272, 0.045979104936122894, 0.07389909774065018, 0.011339850723743439, 0.1143413558602333, 0.05629947781562805, -0.13526225090026855, -0.05700986459851265, 0.14547574520111084, 0.023872992023825645, -0.057064127177000046, -0.2138909548521042, -0.002902575535699725, -0.07730814069509506, -0.011685127392411232, -0.06846728920936584, 0.0291305985301733, -0.01194276288151741, 0.060226380825042725, -0.0496203787624836, -0.09797755628824234, -0.046314824372529984, 0.1015089675784111, 0.054820988327264786, 0.011354796588420868, -0.01489334274083376, 0.03576440364122391, 0.13432876765727997, 0.04213530570268631, -0.10012737661600113, -0.07065672427415848, -0.0701170489192009, -0.09620913118124008, -0.03947552293539047, 0.04272124543786049, 0.020167991518974304, 0.042202774435281754, 0.2283228635787964, 0.024096308276057243, 0.05459817871451378, 0.029667891561985016, 0.0026177873369306326, 0.03211980313062668, 0.1073630079627037, -0.041210614144802094, -0.188126802444458, -0.03292805701494217, 0.0931866466999054, -0.009821015410125256, -0.028658604249358177, -0.033444397151470184, 0.035014089196920395, 0.08379437029361725, 0.11821532249450684, 0.08875755965709686, -0.012828069739043713, -0.037612639367580414, -0.03493109717965126, 0.2115669697523117, -0.14141373336315155, 0.045799970626831055, -0.022097334265708923, -0.018195297569036484, -0.06905751675367355, 0.030103791505098343, 0.01831657998263836, -0.003142025787383318, 0.06966056674718857, -0.061253178864717484, -0.05794486775994301, -0.11518853157758713, -0.045523155480623245, 0.04711875319480896, -0.024105608463287354, -0.024469668045639992, -0.07765042781829834, -0.11219723522663116, -0.06417357176542282, 0.06612563133239746, -0.04156653955578804, -0.03974827378988266, 0.005308232270181179, -0.07131324708461761, 0.008387917652726173, 0.008993842639029026, 0.12122467905282974, -0.030063031241297722, 0.05833350867033005, -0.002476902212947607, 0.05916252359747887, 0.10643328726291656, 0.03227818012237549, -0.08492200076580048, 0.057466037571430206, -0.20633617043495178, 0.08371785283088684, -0.11420095711946487, 0.034276340156793594, -0.17048145830631256, -0.024183684960007668, 0.008447963744401932, 0.023597201332449913, 0.023726604878902435, 0.1338067352771759, -0.2097422182559967, -0.016196569427847862, 0.14133213460445404, -0.09649793803691864, -0.12422871589660645, 0.07990546524524689, -0.03459475561976433, 0.1747698187828064, 0.038475677371025085, -0.019652999937534332, 0.09909367561340332, -0.15559963881969452, -0.05852397903800011, -0.026064254343509674, -0.008927824907004833, 0.08823978155851364, 0.07542291283607483, -0.05844951793551445, 0.02285866066813469, 0.02562655322253704, -0.04727208614349365, -0.0268824752420187, -0.05256075784564018, -0.10127434879541397, -0.023140445351600647, -0.09642518311738968, 0.026515161618590355, 0.000058677000197349116, -0.07310442626476288, -0.028560271486639977, -0.17347893118858337, -0.02563360333442688, 0.10103316605091095, 0.004820956848561764, -0.007559072691947222, -0.08540112525224686, 0.022149885073304176, -0.05362366884946823, -0.006164622958749533, -0.16996455192565918, -0.03558015450835228, 0.051895126700401306, -0.14917676150798798, 0.015460150316357613, -0.07327745854854584, 0.07047311216592789, 0.02098717913031578, -0.05859505757689476, -0.03108096309006214, 0.0007694467785768211, 0.004292082041501999, -0.06229274719953537, -0.1903683841228485, -0.058886781334877014, -0.041500482708215714, 0.15720732510089874, -0.24841000139713287, 0.0300158578902483, 0.03247617185115814, 0.13185922801494598, 0.007058668415993452, -0.06344027817249298, 0.02096918225288391, -0.04676475748419762, -0.050621338188648224, -0.06898977607488632, -0.009901339188218117, -0.014539826661348343, -0.031393732875585556, 0.012980648316442966, -0.14970256388187408, -0.060514215379953384, 0.09452559798955917, 0.11224991828203201, -0.14555825293064117, 0.00204002158716321, -0.0460561066865921, -0.07002599537372589, -0.07487804442644119, -0.0761631652712822, 0.07739497721195221, 0.044650159776210785, 0.049250341951847076, -0.06317461282014847, -0.06234706938266754, 0.023210179060697556, 0.005524294450879097, -0.019023682922124863, 0.0948529988527298, 0.074309803545475, -0.09122881293296814, 0.07973480224609375, 0.08461450785398483, 0.04414684325456619, 0.086973637342453, 0.005991141777485609, -0.11396963149309158, -0.03062884695827961, 0.037754856050014496, 0.024159027263522148, 0.15351562201976776, -0.08692087233066559, 0.030462130904197693, 0.052177220582962036, -0.03854219615459442, 0.03157065063714981, -0.0923321321606636, 0.025362705811858177, 0.021495236083865166, -0.006555700208991766, 0.05864228308200836, -0.018769768998026848, -0.01403577346354723, 0.06336429715156555, 0.05677810311317444, 0.044270504266023636, 0.02595379762351513, -0.02093072421848774, -0.1278371512889862, 0.16537296772003174, -0.09028079360723495, -0.2540280222892761, -0.17074446380138397, 0.015454737469553947, 0.03706491366028786, -0.021728800609707832, 0.039588842540979385, -0.06286025792360306, -0.10237989574670792, -0.09417891502380371, 0.0029635571409016848, 0.023925531655550003, -0.058347854763269424, -0.0817074254155159, 0.060779985040426254, 0.04047083482146263, -0.13689260184764862, 0.0349188968539238, 0.06170675903558731, -0.03042641654610634, 0.0018567070364952087, 0.07321398705244064, 0.12743599712848663, 0.14838241040706635, -0.006730219814926386, -0.012446845881640911, 0.035035960376262665, 0.229813352227211, -0.1490442156791687, 0.10630457103252411, 0.14053207635879517, -0.021705523133277893, 0.06635113060474396, 0.1461038440465927, 0.023231739178299904, -0.07546708732843399, 0.04147516191005707, 0.04027445614337921, -0.04228919371962547, -0.2589097023010254, -0.05694316700100899, -0.00946022942662239, -0.07043391466140747, 0.09718906134366989, 0.09238530695438385, 0.11972260475158691, 0.0337289460003376, -0.05568677559494972, -0.025771914049983025, -0.003401360474526882, 0.114128477871418, -0.027640055865049362, -0.004564122296869755, 0.07965842634439468, -0.05878787487745285, 0.011684526689350605, 0.09941446036100388, 0.019347423687577248, 0.17601320147514343, 0.02533329278230667, 0.10681075602769852, 0.06725578010082245, 0.09347675740718842, -0.0015635732561349869, 0.034774236381053925, 0.05337131395936012, 0.022044572979211807, 0.010453542694449425, -0.09408048540353775, -0.012431944720447063, 0.13713060319423676, 0.019816776737570763, 0.009031654335558414, 0.008926562033593655, -0.01010479498654604, 0.03131420537829399, 0.20501568913459778, 0.0009575071162544191, -0.22537250816822052, -0.09500737488269806, 0.059459153562784195, -0.06931101530790329, -0.143676295876503, -0.02094252221286297, 0.030270220711827278, -0.17292405664920807, 0.016790566965937614, -0.0316389761865139, 0.09112390875816345, -0.07145322859287262, -0.028050832450389862, 0.06891903281211853, 0.07569212466478348, -0.012108199298381805, 0.07973295450210571, -0.19069278240203857, 0.12254468351602554, 0.03037673607468605, 0.08605273067951202, -0.11708726733922958, 0.07849059253931046, -0.0019813794642686844, -0.014807495288550854, 0.17999744415283203, -0.014062200672924519, -0.0586031936109066, -0.08878950774669647, -0.08704045414924622, -0.011727320961654186, 0.10361312329769135, -0.09322915226221085, 0.09586969763040543, -0.02775636687874794, -0.03705112263560295, 0.012418309226632118, -0.10469507426023483, -0.1636953055858612, -0.18679304420948029, 0.06244563311338425, -0.07802703976631165, 0.012347841635346413, -0.11227322369813919, -0.06334327906370163, -0.01575082167983055, 0.23160123825073242, -0.16648635268211365, -0.07049825042486191, -0.1498587429523468, -0.03997112438082695, 0.17463743686676025, -0.042160745710134506, 0.06849376112222672, -0.021383514627814293, 0.1873992383480072, -0.008081548847258091, -0.013158116489648819, 0.06569221615791321, -0.09637628495693207, -0.16879262030124664, -0.05748843029141426, 0.14160962402820587, 0.10863390564918518, 0.05731578543782234, -0.0038195757661014795, 0.013171887956559658, -0.03383830562233925, -0.09896382689476013, 0.013824623078107834, 0.13817466795444489, 0.0034514935687184334, 0.00682973163202405, -0.03995988517999649, -0.07027145475149155, -0.05825701728463173, -0.07912654429674149, 0.057147104293107986, 0.187900573015213, -0.09512355923652649, 0.1602867990732193, 0.12431421875953674, -0.06468851119279861, -0.2306901067495346, 0.03996593505144119, 0.04701630026102066, 0.007666614837944508, 0.022401191294193268, -0.19138796627521515, 0.09788824617862701, 0.0009011493530124426, -0.06807263940572739, 0.14616990089416504, -0.16564498841762543, -0.1461436152458191, 0.08002161979675293, 0.025075770914554596, -0.22560662031173706, -0.14821304380893707, -0.1037549376487732, -0.03735695406794548, -0.13707835972309113, 0.048581719398498535, 0.02614329755306244, 0.019834673032164574, 0.025222565978765488, 0.005338077899068594, 0.029657263308763504, -0.07272187620401382, 0.1870686560869217, -0.020297454670071602, 0.0072362530045211315, -0.050640691071748734, -0.04617878794670105, 0.09227550774812698, -0.06150037795305252, 0.11741586774587631, 0.018679620698094368, 0.018796883523464203, -0.1431548148393631, -0.049209367483854294, -0.060803934931755066, 0.04456847906112671, -0.07284719496965408, -0.09393193572759628, -0.04137463867664337, 0.08888561278581619, 0.07211937010288239, -0.032792408019304276, -0.0027768779546022415, -0.07569456845521927, 0.09405932575464249, 0.184477761387825, 0.17357055842876434, 0.009977072477340698, -0.07020942866802216, 0.024555526673793793, -0.042279548943042755, 0.03349342197179794, -0.24652716517448425, 0.03456863760948181, 0.066053606569767, 0.03803660348057747, 0.08509242534637451, -0.016836483031511307, -0.1781480610370636, -0.04086102172732353, 0.08498652279376984, -0.06206206604838371, -0.19876568019390106, -0.02703288197517395, 0.08424776047468185, -0.20383712649345398, -0.032998621463775635, 0.041543323546648026, -0.03834589570760727, -0.02396267279982567, -0.002415500348433852, 0.06396626681089401, -0.008327016606926918, 0.12156640738248825, 0.06747189164161682, 0.10266115516424179, -0.09284433722496033, 0.08920657634735107, 0.10416955500841141, -0.09140542894601822, 0.03545991703867912, 0.10264154523611069, -0.05670900270342827, -0.04460543021559715, 0.033935222774744034, 0.05925208330154419, -0.028357384726405144, -0.06409841030836105, -0.000502707262057811, -0.0359574519097805, 0.04993389546871185, 0.08058220148086548, 0.036113787442445755, -0.01202210783958435, 0.06544706225395203, 0.028145326301455498, -0.11693570017814636, 0.10949387401342392, 0.04405685141682625, 0.04509059712290764, -0.07182393968105316, -0.012280966155230999, 0.015999672934412956, 0.032540347427129745, -0.019734015688300133, -0.014576527290046215, -0.03146412968635559, -0.007561005651950836, -0.1553635597229004, -0.02064543403685093, -0.06516171246767044, 0.006067827809602022, 0.022207623347640038, -0.03830232471227646, -0.012014663778245449, 0.01381110493093729, -0.07979435473680496, -0.07571027427911758, -0.01700955256819725, 0.08539021760225296, -0.1381402313709259, 0.006627439055591822, 0.07182712107896805, -0.10980239510536194, 0.07347989827394485, -0.0048679932951927185, 0.017079560086131096, 0.010923396795988083, -0.11654401570558548, 0.04386281594634056, -0.005810429807752371, 0.01551580335944891, 0.022556742653250694, -0.171111062169075, 0.011553828604519367, -0.038553636521101, -0.03114982508122921, 0.011926400475203991, -0.025060230866074562, -0.11875922232866287, 0.08676479011774063, -0.028097305446863174, -0.037512701004743576, -0.03292486071586609, 0.06296087801456451, 0.08736220002174377, -0.011740099638700485, 0.09667140990495682, -0.025766119360923767, 0.04818311333656311, -0.1756584197282791, -0.01910574547946453, -0.050167568027973175, 0.02537350542843342, -0.01759655587375164, -0.0070639788173139095, 0.055272240191698074, -0.004191063344478607, 0.20991376042366028, -0.03921036794781685, 0.1548677533864975, 0.05199402943253517, -0.009925156831741333, 0.010884369723498821, 0.05032730847597122, 0.06423956155776978, 0.031145188957452774, 0.00853167474269867, 0.04660189896821976, -0.004552975296974182, -0.020357951521873474, -0.13699717819690704, 0.02791593410074711, 0.16117429733276367, 0.061918217688798904, 0.0392887257039547, 0.03704594820737839, -0.1422400325536728, -0.09538721293210983, 0.10306388139724731, -0.0331864058971405, 0.014331420883536339, -0.08317886292934418, 0.17621558904647827, 0.12328410148620605, -0.1574767529964447, 0.0577850341796875, -0.07234696298837662, -0.05066767707467079, -0.1024852767586708, -0.11832084506750107, -0.06293155997991562, -0.06027044355869293, -0.004747506696730852, -0.042489297688007355, 0.05734556168317795, 0.026751231402158737, -0.003270963439717889, -0.006759525276720524, 0.12665949761867523, -0.0249644722789526, -0.004145825747400522, 0.04152364656329155, 0.0326087586581707, 0.019319625571370125, -0.05872373282909393, 0.017997145652770996, 0.018602589145302773, 0.022180357947945595, 0.06835069507360458, 0.0260987039655447, -0.059317342936992645, 0.044286735355854034, 0.00319746439345181, -0.11313364654779434, 0.018146557733416557, -0.00002245741598017048, -0.05020225793123245, 0.13557326793670654, 0.04076748713850975, 0.01548024732619524, -0.029270920902490616, 0.24342355132102966, -0.07199113070964813, -0.08681939542293549, -0.13965600728988647, 0.11511493474245071, -0.023563209921121597, 0.03755274787545204, 0.016542524099349976, -0.12659503519535065, 0.011511262506246567, 0.18531471490859985, 0.12824349105358124, 0.012459068559110165, -0.007656481582671404, 0.05736639350652695, -0.0007639875984750688, -0.05985576659440994, 0.05051197111606598, 0.0664999932050705, 0.16097788512706757, -0.09069112688302994, 0.0652846097946167, -0.008405503816902637, -0.0831485390663147, -0.027498632669448853, 0.11705785244703293, -0.022675158455967903, 0.02148384228348732, -0.03778035193681717, 0.11204422265291214, -0.052532415837049484, -0.2719486355781555, 0.02952493168413639, -0.09503202140331268, -0.13993041217327118, -0.02591860294342041, 0.041448429226875305, -0.03349510580301285, 0.01577647216618061, 0.06254769116640091, -0.045389387756586075, 0.18837277591228485, 0.025987716391682625, -0.08679025620222092, -0.07755549252033234, 0.05874146893620491, -0.08695939928293228, 0.2789687216281891, 0.003863075515255332, 0.04782010242342949, 0.12108923494815826, -0.03053574077785015, -0.18664880096912384, 0.014769754372537136, 0.11989909410476685, -0.09114406257867813, 0.07780203968286514, 0.18139931559562683, -0.005561648402363062, 0.12649618089199066, 0.04705416411161423, -0.03877115994691849, 0.03976387158036232, -0.02721380814909935, -0.03821522742509842, -0.12209630757570267, 0.05661242455244064, -0.0612691193819046, 0.15957388281822205, 0.1158948540687561, -0.05964287370443344, 0.001120698289014399, -0.06126941740512848, 0.06300627440214157, 0.014774397015571594, 0.12115653604269028, 0.018452486023306847, -0.2023056596517563, 0.05087360367178917, -0.03283824771642685, 0.08166342973709106, -0.254973828792572, -0.08186668157577515, 0.07622263580560684, -0.019022729247808456, -0.04275642707943916, 0.12311509251594543, 0.06101066991686821, 0.03676839917898178, -0.03853875398635864, -0.08537755906581879, -0.01412904355674982, 0.15376435220241547, -0.14123432338237762, -0.029574336484074593 ]
null
null
stable-baselines3
# **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
{"library_name": "stable-baselines3", "tags": ["LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "PPO", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "LunarLander-v2", "type": "LunarLander-v2"}, "metrics": [{"type": "mean_reward", "value": "283.31 +/- 20.51", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
Lounarisnia/ppo-LunarLander-v2-3e6
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
2024-02-15T04:49:43+00:00
[]
[]
TAGS #stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
# PPO Agent playing LunarLander-v2 This is a trained model of a PPO agent playing LunarLander-v2 using the stable-baselines3 library. ## Usage (with Stable-baselines3) TODO: Add your code
[ "# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.", "## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ "TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n", "# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.", "## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ 39, 41, 17 ]
[ "passage: TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ 0.03942384943366051, 0.04900386184453964, -0.005304091144353151, 0.026427261531352997, 0.107408307492733, -0.026511888951063156, 0.11188238859176636, 0.0814051404595375, 0.10722193866968155, 0.04762078449130058, 0.08338645845651627, 0.06030960753560066, 0.05080918222665787, 0.2571701407432556, 0.04754156619310379, -0.22987541556358337, 0.036159250885248184, -0.04869936779141426, 0.12395193427801132, 0.07178173214197159, -0.0038484656251966953, -0.06485428661108017, 0.020415637642145157, -0.013290755450725555, 0.05367108806967735, 0.04282612353563309, -0.01716216839849949, -0.08207534998655319, 0.07169748842716217, -0.06345846503973007, 0.06986866891384125, 0.07677983492612839, 0.13218913972377777, -0.17832116782665253, 0.029566360637545586, 0.02571309357881546, -0.07189024239778519, 0.01342033501714468, 0.008019951172173023, 0.05120139941573143, 0.17303818464279175, 0.019879888743162155, 0.07844575494527817, -0.0025605305563658476, -0.15412317216396332, -0.018950799480080605, 0.0436202734708786, 0.12546207010746002, 0.08808347582817078, 0.04605821147561073, 0.01970590092241764, 0.17503218352794647, -0.054352790117263794, -0.028833400458097458, 0.21759237349033356, -0.2881564497947693, -0.031460098922252655, 0.321048766374588, 0.06997483223676682, 0.09725230932235718, -0.07540661096572876, -0.03619609400629997, 0.007783263456076384, -0.013137873262166977, -0.028666524216532707, -0.07447073608636856, 0.17313385009765625, 0.05152064561843872, -0.05057951435446739, -0.09541505575180054, 0.16948209702968597, 0.006921638268977404, 0.0018855923553928733, -0.019282981753349304, 0.009060598909854889, 0.07402525842189789, -0.016097044572234154, -0.07255112379789352, 0.057438433170318604, 0.05330665782094002, 0.019649166613817215, -0.1435653269290924, -0.10762494057416916, -0.022740179672837257, -0.008012006990611553, 0.17786912620067596, -0.009255532175302505, 0.042902372777462006, 0.003065188182517886, 0.10384012013673782, -0.12480384111404419, -0.03354184702038765, -0.0454259067773819, -0.07565800100564957, -0.0223417766392231, -0.02058211714029312, -0.03580251708626747, 0.07184842973947525, 0.11971849203109741, 0.027368178591132164, 0.09350208193063736, 0.047715865075588226, -0.03206788748502731, 0.06343851238489151, 0.05555703118443489, 0.14222665131092072, 0.05807621404528618, 0.012854371219873428, 0.13179877400398254, 0.055213116109371185, 0.033023182302713394, -0.0613492950797081, -0.18252409994602203, 0.07489913702011108, -0.07031869143247604, 0.007941240444779396, 0.12051256000995636, -0.04480670019984245, -0.1183447614312172, -0.037500523030757904, -0.017392054200172424, -0.06224250793457031, -0.025395862758159637, 0.0547584593296051, -0.02883218228816986, -0.03973718360066414, 0.0011496668448671699, 0.09384800493717194, 0.00953749567270279, -0.1752052903175354, 0.03303423151373863, -0.025042934343218803, -0.10782608389854431, 0.009975161403417587, 0.0022444494534283876, 0.03394931182265282, 0.04408763721585274, -0.11822668462991714, -0.30899152159690857, -0.07652641832828522, 0.05490870401263237, -0.06516939401626587, -0.18425025045871735, -0.13193942606449127, 0.02454492449760437, -0.09037084132432938, -0.044885024428367615, -0.12759265303611755, -0.028549788519740105, 0.01743689924478531, 0.011519349180161953, 0.10758619755506516, -0.0106219332665205, -0.012188062071800232, -0.1571401208639145, 0.008273907005786896, -0.20951123535633087, 0.0890483483672142, -0.019150104373693466, 0.037884220480918884, -0.032381169497966766, -0.07404014468193054, 0.030707746744155884, 0.052499737590551376, -0.01474119070917368, 0.13510210812091827, -0.15592676401138306, -0.03691192343831062, -0.007996266707777977, -0.13611900806427002, -0.04786273464560509, -0.10358831286430359, -0.04357128217816353, 0.13354332745075226, 0.018664736300706863, 0.15356586873531342, -0.08709818124771118, -0.0722038671374321, 0.20489206910133362, -0.010411538183689117, -0.12820468842983246, -0.076752208173275, 0.10165707021951675, 0.021510310471057892, -0.056606587022542953, -0.02523270808160305, -0.1839766949415207, -0.0152357779443264, -0.04550420492887497, -0.047039128839969635, 0.01796751655638218, -0.010888241231441498, 0.13837894797325134, 0.08494598418474197, 0.05018039792776108, -0.06086122244596481, -0.006730288732796907, 0.10779471695423126, 0.08823856711387634, 0.008680110797286034, 0.023406028747558594, -0.05774238705635071, 0.09552932530641556, -0.04003755748271942, -0.0142367510125041, -0.08283266425132751, -0.036246106028556824, -0.026256313547492027, 0.17507147789001465, 0.09440762549638748, 0.2257927656173706, 0.09567736834287643, 0.039160262793302536, 0.031270865350961685, -0.13181598484516144, -0.1425403207540512, -0.0017254541162401438, 0.09020978957414627, -0.14270411431789398, -0.04119925573468208, -0.08974775671958923, -0.17768175899982452, -0.12202505767345428, 0.0006432619411498308, -0.17960017919540405, 0.06390921026468277, 0.05408334732055664, -0.035177867859601974, 0.03272094577550888, 0.13032332062721252, -0.011533179320394993, -0.03967514634132385, 0.0831870287656784, 0.0379033200442791, -0.041234664618968964, -0.021742934361100197, 0.11885567009449005, 0.15673065185546875, 0.13124459981918335, -0.03511447086930275, 0.004914294462651014, 0.07076404243707657, -0.02309088408946991, 0.06539414077997208, 0.0558244064450264, 0.20973342657089233, 0.188301220536232, 0.038996949791908264, 0.008822928182780743, -0.07048165798187256, 0.0855446457862854, -0.0742373839020729, -0.14302679896354675, -0.05579735338687897, 0.08729292452335358, 0.016605578362941742, 0.023469142615795135, 0.08711627870798111, 0.024545932188630104, 0.09132762253284454, 0.15968108177185059, 0.01990218088030815, -0.09659269452095032, -0.050218869000673294, 0.01175848301500082, 0.027713103219866753, 0.04794301092624664, -0.04514073207974434, -0.00937939714640379, 0.017020760104060173, -0.10303554683923721, 0.031789086759090424, -0.1413339376449585, -0.1358717679977417, 0.044326696544885635, 0.003906996920704842, 0.010907664895057678, 0.02786896750330925, -0.0038291432429105043, 0.019039705395698547, 0.04351753741502762, -0.06975466758012772, 0.047416772693395615, -0.024745507165789604, -0.020031947642564774, 0.03340689837932587, -0.057257164269685745, -0.205775648355484, -0.17696654796600342, 0.00013708483311347663, -0.09910997003316879, 0.10194740444421768, 0.018308809027075768, -0.12373185902833939, 0.047737859189510345, -0.05822649225592613, 0.027574289590120316, -0.01875593699514866, -0.049130141735076904, 0.10507171601057053, 0.1525275856256485, -0.016146350651979446, 0.018018173053860664, -0.04865182936191559, -0.10157987475395203, -0.19632206857204437, 0.0691583976149559, 0.04680244252085686, 0.014610917307436466, 0.10669491440057755, 0.018072687089443207, 0.02367905154824257, -0.007674071006476879, -0.016521066427230835, -0.011659215204417706, -0.08781040459871292, 0.31909599900245667, 0.04510033503174782, -0.025173069909214973, 0.02041010931134224, -0.0043001663871109486, -0.028083480894565582, 0.03263787180185318, -0.0985708013176918, -0.07548979669809341, -0.08774089068174362, -0.04367410019040108, -0.09784720093011856, 0.053299110382795334, 0.05916472524404526, 0.003188040340319276, -0.07727594673633575, 0.04221395403146744, 0.11369874328374863, -0.0923808291554451, -0.07137343287467957, 0.07477962225675583, 0.0972946360707283, -0.07331304252147675, 0.00012658814375754446, 0.00874367356300354, 0.023951783776283264, 0.037102166563272476, 0.06778035312891006, -0.03966575115919113, 0.08589404821395874, -0.19917890429496765, 0.0372927263379097, 0.106058269739151, 0.023754918947815895, 0.0638108178973198, 0.07643651217222214, -0.1058402881026268, -0.008500572293996811, -0.032518330961465836, -0.21341575682163239, 0.1668180525302887, 0.1355515867471695, 0.06788124144077301, -0.025637222453951836, -0.00461410591378808, -0.0649740919470787, 0.05773647129535675, 0.02723747305572033, -0.14758841693401337, 0.004883295856416225, 0.06064270809292793, 0.026899009943008423, 0.01614922471344471, 0.07971042394638062, 0.014697225764393806, -0.1801026314496994, -0.014406266622245312, 0.10730406641960144, 0.002390873385593295, 0.0053148469887673855, -0.03175045922398567, -0.1755964607000351, 0.0751047357916832, 0.004285442177206278, 0.07233936339616776, -0.1676585078239441, 0.14297930896282196, -0.10089799761772156, 0.07726949453353882, -0.004285062663257122, -0.021311495453119278, 0.02507244050502777, -0.0541163794696331, 0.15163759887218475, 0.01058570109307766, -0.021810131147503853, -0.1200498715043068, -0.1717042326927185, -0.019227758049964905, -0.11788936704397202, -0.11679866164922714, 0.050424277782440186, 0.062185097485780716, 0.04923136904835701, -0.061147067695856094, 0.1518532931804657, -0.047422297298908234, 0.060713399201631546, -0.06893875449895859, -0.06755045056343079, 0.03764858841896057, -0.12588608264923096, -0.08176055550575256, 0.05573027580976486, 0.19166934490203857, 0.15833087265491486, -0.02816431224346161, -0.03472423925995827, -0.047419581562280655, -0.006212298292666674, -0.007802055217325687, 0.0275666993111372, 0.023223137483000755, 0.07315318286418915, -0.07681374251842499, -0.11649256944656372, 0.033787861466407776, -0.06713802367448807, -0.055589709430933, -0.015439179725944996, 0.1513158082962036, 0.04671623185276985, 0.07720734924077988, -0.018946662545204163, 0.03887668624520302, -0.001724981120787561, -0.056474871933460236, 0.16197094321250916, 0.03885216265916824, -0.05193585529923439, 0.06837689876556396, 0.053174007683992386, 0.043745119124650955, 0.03011113777756691, -0.026783017441630363, 0.206032395362854, 0.1980147808790207, 0.014206883497536182, 0.2175983190536499, 0.03177616000175476, -0.03772832080721855, -0.1300560086965561, -0.065880686044693, -0.006372632458806038, 0.03559038043022156, 0.08070417493581772, -0.18207235634326935, -0.015011128038167953, -0.05689644813537598, -0.034518610686063766, -0.15059494972229004, -0.28553900122642517, -0.05957856774330139, 0.20075850188732147, 0.14706264436244965, 0.27519428730010986, -0.10432573407888412, 0.035197313874959946, 0.02663275972008705, -0.04912831634283066, -0.006501141935586929, 0.00018665487004909664, 0.10268618166446686, -0.15421873331069946, 0.1176437959074974, 0.08486983180046082, -0.019002694636583328, 0.01058861706405878, -0.1619086116552353, 0.00936629343777895, -0.12191236019134521, 0.05354422330856323, 0.1400289237499237, -0.048128653317689896, -0.054873593151569366, 0.14033560454845428, -0.024562934413552284, -0.22685599327087402, -0.04648222774267197, -0.043600670993328094, -0.010640020482242107, 0.026607351377606392, -0.1013401448726654, 0.04101909324526787, 0.1330099105834961, 0.009380043484270573, 0.1147187277674675, 0.11749245226383209, -0.052566803991794586, 0.10792597383260727, 0.2257719188928604, -0.018785694614052773, 0.04689010605216026, -0.12743118405342102, -0.0012336712097749114, -0.028270328417420387, 0.013657891191542149, -0.09504974633455276, -0.09938385337591171, 0.02366873063147068, 0.02872389927506447, 0.009118586778640747, 0.0921793207526207, -0.029922157526016235, 0.0759170651435852, 0.06817561388015747, -0.13014446198940277, -0.16288450360298157, 0.015828335657715797, -0.007344507612287998, 0.08354310691356659, 0.00027861111448146403, 0.08878035843372345, -0.11932205408811569, -0.018093237653374672, -0.03153328225016594, -0.03319635987281799, -0.130486860871315, -0.07138993591070175, 0.06156524643301964, 0.028095467016100883, -0.06602972000837326, 0.1398407518863678, 0.026440169662237167, 0.15942534804344177, 0.049197953194379807, 0.012499804608523846, 0.07227300107479095, -0.05345509201288223, 0.1283530443906784, 0.13818155229091644, -0.00868943240493536, -0.05460423603653908, -0.1013643890619278, -0.10236792266368866, 0.08925779908895493, -0.05773641914129257, 0.07476430386304855, -0.14885357022285461, -0.06675903499126434, 0.015772046521306038, 0.016141414642333984, -0.09562095999717712, 0.02571965754032135, -0.01625603251159191, -0.18119946122169495, 0.056570518761873245, -0.048285093158483505, 0.0440407395362854, -0.06347788125276566, -0.1110161691904068, -0.17226378619670868, 0.06091433763504028, 0.08593481779098511, -0.053876690566539764, -0.12229149043560028, 0.011023230850696564, -0.00012518465518951416, -0.06341652572154999, -0.05023367330431938, 0.09722746908664703, -0.11020902544260025, 0.031452205032110214, -0.012567701749503613, 0.08853451162576675, -0.03510405123233795, -0.011538895778357983, 0.044220831245183945, -0.08039166033267975, -0.009481523185968399, 0.03534642979502678, -0.026372017338871956, -0.04127239063382149, -0.2689029574394226, 0.0036654395516961813, 0.0341104120016098, 0.02497158572077751, 0.07856601476669312, 0.011906822212040424, 0.021174922585487366, 0.03993808850646019, -0.15396519005298615, -0.013395369984209538, 0.14574195444583893, -0.07689505815505981, -0.022186370566487312, 0.05703273415565491, -0.09054436534643173, 0.013882770203053951, -0.030287226662039757, 0.1345842480659485, 0.023923413828015327, 0.06404478847980499, -0.0851147472858429, 0.10106813907623291, -0.1451139897108078, -0.04998219385743141, -0.01244612317532301, 0.09761348366737366, 0.07019034773111343, -0.10272270441055298, 0.014697125181555748, 0.04210108891129494, 0.19416837394237518, 0.016384804621338844, -0.0356343574821949, -0.03396720811724663, 0.004015897400677204, 0.22076453268527985, 0.03044266067445278, 0.10457023978233337, 0.07281364500522614, -0.026583973318338394, 0.12624378502368927, 0.09929762035608292, 0.11280370503664017, -0.055645186454057693, 0.13904185593128204, 0.04667386785149574, 0.038641396909952164, 0.0614289753139019, 0.06836545467376709, 0.09098632633686066, -0.0008288522367365658, 0.1138714924454689, 0.013811973854899406, -0.02422109805047512, -0.021335409954190254, 0.17759373784065247, 0.10501719266176224, -0.14769648015499115, 0.029047364369034767, -0.01258957851678133, 0.039933037012815475, -0.014194529503583908, -0.15634691715240479, -0.07240267097949982, -0.3315149247646332, 0.1226184144616127, -0.07119352370500565, 0.019930170848965645, 0.007913772016763687, -0.037425633519887924, -0.03296699747443199, -0.04477746784687042, 0.13151589035987854, -0.013641550205647945, -0.006079165264964104, -0.04815853759646416, -0.015360191464424133, -0.11607866734266281, -0.11200575530529022, -0.013207737356424332, -0.13671602308750153, -0.010119039565324783, 0.05595948174595833, 0.003977729007601738, 0.01821410097181797, -0.03142618387937546, 0.0024383175186812878, 0.06541839241981506, -0.05751744285225868, 0.056182678788900375, 0.12097269296646118, 0.08766137808561325, -0.1058853268623352, 0.031048951670527458, 0.2011747509241104, 0.04359564557671547, -0.12483977526426315, 0.01449228823184967, 0.1819491684436798, 0.004885740112513304, 0.017068125307559967, -0.006097703706473112, -0.0540788508951664, -0.07554277032613754, 0.1251034289598465, 0.08296554535627365, -0.09985227137804031, 0.015833314508199692, -0.0726347416639328, -0.01594804972410202, -0.06374675035476685, 0.10130585730075836, 0.09538925439119339, 0.04440245032310486, -0.10621760785579681, -0.08487539738416672, -0.10891728103160858, 0.040588874369859695, -0.08629853278398514, -0.07311757653951645, 0.09629398584365845, -0.07057105004787445, -0.07029950618743896, 0.025521177798509598, -0.17978744208812714, -0.009467960335314274, 0.1711762249469757, -0.24654000997543335, -0.0916430801153183, -0.10857923328876495, 0.14477859437465668, 0.016497576609253883, 0.1013975441455841, -0.006207061931490898, -0.007889035157859325, -0.20577777922153473, 0.024890204891562462, -0.05293011665344238, -0.02073732763528824, 0.07814782857894897, -0.09476397186517715, 0.22629831731319427, -0.08276885002851486, 0.020940175279974937, 0.012659613974392414, 0.0870661810040474, -0.030675338581204414, 0.09283176809549332, -0.03660329803824425, -0.12576518952846527, -0.03620953485369682, 0.03001813031733036, 0.013904244638979435, 0.10071761906147003, 0.09772487729787827, -0.03414725139737129, 0.03389119729399681, 0.09747414290904999, 0.04172342270612717, -0.023843804374337196, 0.0360250361263752, -0.17077107727527618, 0.02182629331946373, -0.018498148769140244, -0.06935930997133255, 0.03687669709324837, -0.06603235751390457, 0.1639697551727295, 0.04022442549467087, 0.0670473501086235, -0.036152735352516174, 0.0073931049555540085, -0.014454689808189869, -0.013775371946394444, -0.026180334389209747, -0.17259705066680908, -0.10422050207853317, -0.1347656100988388, -0.012701659463346004, -0.034971047192811966, 0.04591470584273338, 0.023234914988279343, -0.0003200018545612693, -0.014577031135559082, -0.12090865522623062, 0.04360328987240791, 0.11146783083677292, -0.04631396010518074, -0.026193076744675636 ]
null
null
null
# Lora of murasaki/紫/紫 (Azur Lane) ## What Is This? This is the LoRA model of waifu murasaki/紫/紫 (Azur Lane). ## How Is It Trained? * This model is trained with [HCP-Diffusion](https://github.com/7eu7d7/HCP-Diffusion). * The [auto-training framework](https://github.com/deepghs/cyberharem) is maintained by [DeepGHS Team](https://huggingface.co/deepghs). * The base model used for training is [deepghs/animefull-latest](https://huggingface.co/deepghs/animefull-latest). * Dataset used for training is the `stage3-p480-800` in [CyberHarem/murasaki_azurlane](https://huggingface.co/datasets/CyberHarem/murasaki_azurlane), which contains 77 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 800 steps, 40 checkpoints were saved and evaluated. * **Trigger word is `murasaki_azurlane`.** * Pruned core tags for this waifu are `long_hair, purple_hair, breasts, purple_eyes, hair_ribbon, ribbon, large_breasts, very_long_hair, black_ribbon`. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. ## How to Use It? ### If You Are Using A1111 WebUI v1.7+ **Just use it like the classic LoRA**. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 720, you need to download [`720/murasaki_azurlane.pt`](https://huggingface.co/CyberHarem/murasaki_azurlane/resolve/main/720/murasaki_azurlane.pt) as the embedding and [`720/murasaki_azurlane.safetensors`](https://huggingface.co/CyberHarem/murasaki_azurlane/resolve/main/720/murasaki_azurlane.safetensors) for loading Lora. By using both files together, you can generate images for the desired characters. ## Which Step Should I Use? We selected 5 good steps for you to choose. The best one is step 720. 1520 images (1.52 GiB) were generated for auto-testing. ![Metrics Plot](metrics_plot.png) The base model used for generating preview images is [Meina/MeinaMix_V11](https://huggingface.co/Meina/MeinaMix_V11). Here are the preview of the recommended steps: | Step | Epoch | CCIP | AI Corrupt | Bikini Plus | Score | Download | pattern_0_0 | pattern_0_1 | pattern_1 | portrait_0 | portrait_1 | portrait_2 | full_body_0 | full_body_1 | profile_0 | profile_1 | free_0 | free_1 | shorts | maid_0 | maid_1 | miko | yukata | suit | china | bikini_0 | bikini_1 | bikini_2 | sit | squat | kneel | jump | crossed_arms | angry | smile | cry | grin | n_lie_0 | n_lie_1 | n_stand_0 | n_stand_1 | n_stand_2 | n_sex_0 | n_sex_1 | |-------:|--------:|:----------|:-------------|:--------------|:----------|:-------------------------------------------------------------------------------------------------------|:---------------------------------------------|:---------------------------------------------|:-----------------------------------------|:-------------------------------------------|:-------------------------------------------|:-------------------------------------------|:---------------------------------------------|:---------------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-----------------------------------|:-------------------------------|:-----------------------------------|:-------------------------------|:---------------------------------|:---------------------------------------|:---------------------------------------|:---------------------------------------|:-----------------------------|:---------------------------------|:---------------------------------|:-------------------------------|:-----------------------------------------------|:---------------------------------|:---------------------------------|:-----------------------------|:-------------------------------|:-------------------------------------|:-------------------------------------|:-----------------------------------------|:-----------------------------------------|:-----------------------------------------|:-------------------------------------|:-------------------------------------| | 720 | 38 | **0.983** | **0.967** | **0.860** | **0.808** | [Download](https://huggingface.co/CyberHarem/murasaki_azurlane/resolve/main/720/murasaki_azurlane.zip) | ![pattern_0_0](720/previews/pattern_0_0.png) | ![pattern_0_1](720/previews/pattern_0_1.png) | ![pattern_1](720/previews/pattern_1.png) | ![portrait_0](720/previews/portrait_0.png) | ![portrait_1](720/previews/portrait_1.png) | ![portrait_2](720/previews/portrait_2.png) | ![full_body_0](720/previews/full_body_0.png) | ![full_body_1](720/previews/full_body_1.png) | ![profile_0](720/previews/profile_0.png) | ![profile_1](720/previews/profile_1.png) | ![free_0](720/previews/free_0.png) | ![free_1](720/previews/free_1.png) | ![shorts](720/previews/shorts.png) | ![maid_0](720/previews/maid_0.png) | ![maid_1](720/previews/maid_1.png) | ![miko](720/previews/miko.png) | ![yukata](720/previews/yukata.png) | ![suit](720/previews/suit.png) | ![china](720/previews/china.png) | ![bikini_0](720/previews/bikini_0.png) | ![bikini_1](720/previews/bikini_1.png) | ![bikini_2](720/previews/bikini_2.png) | ![sit](720/previews/sit.png) | ![squat](720/previews/squat.png) | ![kneel](720/previews/kneel.png) | ![jump](720/previews/jump.png) | ![crossed_arms](720/previews/crossed_arms.png) | ![angry](720/previews/angry.png) | ![smile](720/previews/smile.png) | ![cry](720/previews/cry.png) | ![grin](720/previews/grin.png) | ![n_lie_0](720/previews/n_lie_0.png) | ![n_lie_1](720/previews/n_lie_1.png) | ![n_stand_0](720/previews/n_stand_0.png) | ![n_stand_1](720/previews/n_stand_1.png) | ![n_stand_2](720/previews/n_stand_2.png) | ![n_sex_0](720/previews/n_sex_0.png) | ![n_sex_1](720/previews/n_sex_1.png) | | 640 | 34 | 0.948 | 0.959 | 0.856 | 0.780 | [Download](https://huggingface.co/CyberHarem/murasaki_azurlane/resolve/main/640/murasaki_azurlane.zip) | ![pattern_0_0](640/previews/pattern_0_0.png) | ![pattern_0_1](640/previews/pattern_0_1.png) | ![pattern_1](640/previews/pattern_1.png) | ![portrait_0](640/previews/portrait_0.png) | ![portrait_1](640/previews/portrait_1.png) | ![portrait_2](640/previews/portrait_2.png) | ![full_body_0](640/previews/full_body_0.png) | ![full_body_1](640/previews/full_body_1.png) | ![profile_0](640/previews/profile_0.png) | ![profile_1](640/previews/profile_1.png) | ![free_0](640/previews/free_0.png) | ![free_1](640/previews/free_1.png) | ![shorts](640/previews/shorts.png) | ![maid_0](640/previews/maid_0.png) | ![maid_1](640/previews/maid_1.png) | ![miko](640/previews/miko.png) | ![yukata](640/previews/yukata.png) | ![suit](640/previews/suit.png) | ![china](640/previews/china.png) | ![bikini_0](640/previews/bikini_0.png) | ![bikini_1](640/previews/bikini_1.png) | ![bikini_2](640/previews/bikini_2.png) | ![sit](640/previews/sit.png) | ![squat](640/previews/squat.png) | ![kneel](640/previews/kneel.png) | ![jump](640/previews/jump.png) | ![crossed_arms](640/previews/crossed_arms.png) | ![angry](640/previews/angry.png) | ![smile](640/previews/smile.png) | ![cry](640/previews/cry.png) | ![grin](640/previews/grin.png) | ![n_lie_0](640/previews/n_lie_0.png) | ![n_lie_1](640/previews/n_lie_1.png) | ![n_stand_0](640/previews/n_stand_0.png) | ![n_stand_1](640/previews/n_stand_1.png) | ![n_stand_2](640/previews/n_stand_2.png) | ![n_sex_0](640/previews/n_sex_0.png) | ![n_sex_1](640/previews/n_sex_1.png) | | 460 | 24 | 0.938 | 0.954 | 0.854 | 0.771 | [Download](https://huggingface.co/CyberHarem/murasaki_azurlane/resolve/main/460/murasaki_azurlane.zip) | ![pattern_0_0](460/previews/pattern_0_0.png) | ![pattern_0_1](460/previews/pattern_0_1.png) | ![pattern_1](460/previews/pattern_1.png) | ![portrait_0](460/previews/portrait_0.png) | ![portrait_1](460/previews/portrait_1.png) | ![portrait_2](460/previews/portrait_2.png) | ![full_body_0](460/previews/full_body_0.png) | ![full_body_1](460/previews/full_body_1.png) | ![profile_0](460/previews/profile_0.png) | ![profile_1](460/previews/profile_1.png) | ![free_0](460/previews/free_0.png) | ![free_1](460/previews/free_1.png) | ![shorts](460/previews/shorts.png) | ![maid_0](460/previews/maid_0.png) | ![maid_1](460/previews/maid_1.png) | ![miko](460/previews/miko.png) | ![yukata](460/previews/yukata.png) | ![suit](460/previews/suit.png) | ![china](460/previews/china.png) | ![bikini_0](460/previews/bikini_0.png) | ![bikini_1](460/previews/bikini_1.png) | ![bikini_2](460/previews/bikini_2.png) | ![sit](460/previews/sit.png) | ![squat](460/previews/squat.png) | ![kneel](460/previews/kneel.png) | ![jump](460/previews/jump.png) | ![crossed_arms](460/previews/crossed_arms.png) | ![angry](460/previews/angry.png) | ![smile](460/previews/smile.png) | ![cry](460/previews/cry.png) | ![grin](460/previews/grin.png) | ![n_lie_0](460/previews/n_lie_0.png) | ![n_lie_1](460/previews/n_lie_1.png) | ![n_stand_0](460/previews/n_stand_0.png) | ![n_stand_1](460/previews/n_stand_1.png) | ![n_stand_2](460/previews/n_stand_2.png) | ![n_sex_0](460/previews/n_sex_0.png) | ![n_sex_1](460/previews/n_sex_1.png) | | 520 | 28 | 0.937 | 0.958 | 0.853 | 0.769 | [Download](https://huggingface.co/CyberHarem/murasaki_azurlane/resolve/main/520/murasaki_azurlane.zip) | ![pattern_0_0](520/previews/pattern_0_0.png) | ![pattern_0_1](520/previews/pattern_0_1.png) | ![pattern_1](520/previews/pattern_1.png) | ![portrait_0](520/previews/portrait_0.png) | ![portrait_1](520/previews/portrait_1.png) | ![portrait_2](520/previews/portrait_2.png) | ![full_body_0](520/previews/full_body_0.png) | ![full_body_1](520/previews/full_body_1.png) | ![profile_0](520/previews/profile_0.png) | ![profile_1](520/previews/profile_1.png) | ![free_0](520/previews/free_0.png) | ![free_1](520/previews/free_1.png) | ![shorts](520/previews/shorts.png) | ![maid_0](520/previews/maid_0.png) | ![maid_1](520/previews/maid_1.png) | ![miko](520/previews/miko.png) | ![yukata](520/previews/yukata.png) | ![suit](520/previews/suit.png) | ![china](520/previews/china.png) | ![bikini_0](520/previews/bikini_0.png) | ![bikini_1](520/previews/bikini_1.png) | ![bikini_2](520/previews/bikini_2.png) | ![sit](520/previews/sit.png) | ![squat](520/previews/squat.png) | ![kneel](520/previews/kneel.png) | ![jump](520/previews/jump.png) | ![crossed_arms](520/previews/crossed_arms.png) | ![angry](520/previews/angry.png) | ![smile](520/previews/smile.png) | ![cry](520/previews/cry.png) | ![grin](520/previews/grin.png) | ![n_lie_0](520/previews/n_lie_0.png) | ![n_lie_1](520/previews/n_lie_1.png) | ![n_stand_0](520/previews/n_stand_0.png) | ![n_stand_1](520/previews/n_stand_1.png) | ![n_stand_2](520/previews/n_stand_2.png) | ![n_sex_0](520/previews/n_sex_0.png) | ![n_sex_1](520/previews/n_sex_1.png) | | 680 | 36 | 0.941 | 0.962 | 0.839 | 0.740 | [Download](https://huggingface.co/CyberHarem/murasaki_azurlane/resolve/main/680/murasaki_azurlane.zip) | ![pattern_0_0](680/previews/pattern_0_0.png) | ![pattern_0_1](680/previews/pattern_0_1.png) | ![pattern_1](680/previews/pattern_1.png) | ![portrait_0](680/previews/portrait_0.png) | ![portrait_1](680/previews/portrait_1.png) | ![portrait_2](680/previews/portrait_2.png) | ![full_body_0](680/previews/full_body_0.png) | ![full_body_1](680/previews/full_body_1.png) | ![profile_0](680/previews/profile_0.png) | ![profile_1](680/previews/profile_1.png) | ![free_0](680/previews/free_0.png) | ![free_1](680/previews/free_1.png) | ![shorts](680/previews/shorts.png) | ![maid_0](680/previews/maid_0.png) | ![maid_1](680/previews/maid_1.png) | ![miko](680/previews/miko.png) | ![yukata](680/previews/yukata.png) | ![suit](680/previews/suit.png) | ![china](680/previews/china.png) | ![bikini_0](680/previews/bikini_0.png) | ![bikini_1](680/previews/bikini_1.png) | ![bikini_2](680/previews/bikini_2.png) | ![sit](680/previews/sit.png) | ![squat](680/previews/squat.png) | ![kneel](680/previews/kneel.png) | ![jump](680/previews/jump.png) | ![crossed_arms](680/previews/crossed_arms.png) | ![angry](680/previews/angry.png) | ![smile](680/previews/smile.png) | ![cry](680/previews/cry.png) | ![grin](680/previews/grin.png) | ![n_lie_0](680/previews/n_lie_0.png) | ![n_lie_1](680/previews/n_lie_1.png) | ![n_stand_0](680/previews/n_stand_0.png) | ![n_stand_1](680/previews/n_stand_1.png) | ![n_stand_2](680/previews/n_stand_2.png) | ![n_sex_0](680/previews/n_sex_0.png) | ![n_sex_1](680/previews/n_sex_1.png) | ## Anything Else? Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. ## All Steps We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * [Steps From 620 to 800](all/0.md) * [Steps From 420 to 600](all/1.md) * [Steps From 220 to 400](all/2.md) * [Steps From 20 to 200](all/3.md)
{"license": "mit", "tags": ["art", "not-for-all-audiences"], "datasets": ["CyberHarem/murasaki_azurlane"], "pipeline_tag": "text-to-image"}
text-to-image
CyberHarem/murasaki_azurlane
[ "art", "not-for-all-audiences", "text-to-image", "dataset:CyberHarem/murasaki_azurlane", "license:mit", "region:us" ]
2024-02-15T04:50:22+00:00
[]
[]
TAGS #art #not-for-all-audiences #text-to-image #dataset-CyberHarem/murasaki_azurlane #license-mit #region-us
Lora of murasaki/紫/紫 (Azur Lane) ================================ What Is This? ------------- This is the LoRA model of waifu murasaki/紫/紫 (Azur Lane). How Is It Trained? ------------------ * This model is trained with HCP-Diffusion. * The auto-training framework is maintained by DeepGHS Team. * The base model used for training is deepghs/animefull-latest. * Dataset used for training is the 'stage3-p480-800' in CyberHarem/murasaki\_azurlane, which contains 77 images. * Batch size is 4, resolution is 720x720, clustering into 5 buckets. * Batch size for regularization dataset is 16, resolution is 720x720, clustering into 20 buckets. * Trained for 800 steps, 40 checkpoints were saved and evaluated. * Trigger word is 'murasaki\_azurlane'. * Pruned core tags for this waifu are 'long\_hair, purple\_hair, breasts, purple\_eyes, hair\_ribbon, ribbon, large\_breasts, very\_long\_hair, black\_ribbon'. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. How to Use It? -------------- ### If You Are Using A1111 WebUI v1.7+ Just use it like the classic LoRA. The LoRA we provided are bundled with the embedding file. ### If You Are Using A1111 WebUI v1.6 or Lower After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 720, you need to download '720/murasaki\_azurlane.pt' as the embedding and '720/murasaki\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters. Which Step Should I Use? ------------------------ We selected 5 good steps for you to choose. The best one is step 720. 1520 images (1.52 GiB) were generated for auto-testing. !Metrics Plot The base model used for generating preview images is Meina/MeinaMix\_V11. Here are the preview of the recommended steps: Anything Else? -------------- Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. All Steps --------- We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * Steps From 620 to 800 * Steps From 420 to 600 * Steps From 220 to 400 * Steps From 20 to 200
[ "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 720, you need to download '720/murasaki\\_azurlane.pt' as the embedding and '720/murasaki\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 720.\n\n\n1520 images (1.52 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 620 to 800\n* Steps From 420 to 600\n* Steps From 220 to 400\n* Steps From 20 to 200" ]
[ "TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/murasaki_azurlane #license-mit #region-us \n", "### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file.", "### If You Are Using A1111 WebUI v1.6 or Lower\n\n\nAfter downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora.\n\n\nFor example, if you want to use the model from step 720, you need to download '720/murasaki\\_azurlane.pt' as the embedding and '720/murasaki\\_azurlane.safetensors' for loading Lora. By using both files together, you can generate images for the desired characters.\n\n\nWhich Step Should I Use?\n------------------------\n\n\nWe selected 5 good steps for you to choose. The best one is step 720.\n\n\n1520 images (1.52 GiB) were generated for auto-testing.\n\n\n!Metrics Plot\n\n\nThe base model used for generating preview images is Meina/MeinaMix\\_V11.\n\n\nHere are the preview of the recommended steps:\n\n\n\nAnything Else?\n--------------\n\n\nBecause the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret:\n\n\n1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail.\n2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits.\n3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm.\n4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters.\n5. Individuals who finds the generated image content offensive to their values.\n\n\nAll Steps\n---------\n\n\nWe uploaded the files in all steps. you can check the images, metrics and download them in the following links:\n\n\n* Steps From 620 to 800\n* Steps From 420 to 600\n* Steps From 220 to 400\n* Steps From 20 to 200" ]
[ 45, 38, 467 ]
[ "passage: TAGS\n#art #not-for-all-audiences #text-to-image #dataset-CyberHarem/murasaki_azurlane #license-mit #region-us \n### If You Are Using A1111 WebUI v1.7+\n\n\nJust use it like the classic LoRA. The LoRA we provided are bundled with the embedding file." ]
[ 0.013026767410337925, -0.007562323473393917, -0.003739025676622987, 0.07970427721738815, 0.08803433179855347, 0.08568010479211807, 0.23701341450214386, 0.07847987115383148, 0.11838795989751816, -0.06459178030490875, 0.08918838948011398, 0.05709420144557953, 0.004876198247075081, 0.04488932341337204, -0.02419363521039486, -0.1750452220439911, -0.06266576796770096, -0.03325406834483147, 0.014773810282349586, 0.02001582831144333, 0.08231084048748016, 0.004891762975603342, 0.10130881518125534, -0.04732998460531235, -0.03596479445695877, 0.04471703991293907, -0.037260908633470535, -0.046911463141441345, 0.02204066514968872, 0.08842775970697403, 0.12010505050420761, 0.0023407454136759043, 0.0643898993730545, -0.15112192928791046, 0.06791350990533829, -0.013876172713935375, -0.10871061682701111, -0.01043756678700447, 0.021000448614358902, -0.026974568143486977, 0.12636567652225494, 0.026717333123087883, -0.12007282674312592, 0.043051790446043015, -0.13049016892910004, -0.017399460077285767, -0.04862510785460472, 0.05125856399536133, 0.15307201445102692, 0.056966401636600494, 0.01997033692896366, 0.05343441665172577, -0.04600996896624565, 0.08423494547605515, 0.11016510426998138, -0.12943010032176971, -0.07647231966257095, 0.10111143440008163, 0.008347061462700367, 0.14123967289924622, -0.09515006840229034, 0.09675159305334091, 0.06861062347888947, -0.05229979008436203, -0.16188766062259674, -0.09740585088729858, -0.19761091470718384, -0.010701227001845837, 0.02044025808572769, 0.03174493461847305, 0.40337055921554565, 0.06489557772874832, 0.038416288793087006, 0.060843538492918015, -0.06627149879932404, 0.03414734825491905, -0.1014251708984375, 0.1410479098558426, 0.0378836914896965, 0.09065048396587372, -0.042212408035993576, -0.09688612073659897, -0.11466166377067566, -0.07136257737874985, -0.08117663115262985, -0.02797756716609001, 0.021478069946169853, 0.11558881402015686, -0.19460813701152802, -0.010348598472774029, -0.05685773864388466, -0.12806446850299835, 0.013221833854913712, -0.098284050822258, 0.1639181226491928, 0.06614714115858078, -0.017259448766708374, -0.012461596168577671, 0.24979516863822937, 0.10377014428377151, 0.17890316247940063, 0.059554439038038254, -0.09104953706264496, 0.13109923899173737, 0.0460159033536911, -0.08737725764513016, -0.02705383114516735, -0.09271810203790665, 0.13594500720500946, -0.045176900923252106, 0.11027444899082184, -0.06292295455932617, -0.10910525172948837, 0.017888620495796204, -0.11088890582323074, 0.06117694079875946, 0.026003332808613777, 0.02024342678487301, -0.039772484451532364, 0.04156586155295372, 0.02220804989337921, -0.04679690673947334, -0.012417278252542019, -0.009050569497048855, -0.06589491665363312, 0.04024820774793625, 0.10491978377103806, 0.034986186772584915, 0.06542474031448364, -0.0053440285846591, -0.01794077455997467, 0.0077524916268885136, -0.048349685966968536, 0.010121042840182781, 0.048075027763843536, 0.04440760985016823, 0.08520214259624481, -0.16331873834133148, -0.08279652893543243, -0.013446642085909843, 0.05358605459332466, 0.012029040604829788, 0.09738034009933472, -0.0011749041732400656, 0.06524105370044708, 0.010823112912476063, -0.026991194114089012, 0.047782838344573975, -0.10279476642608643, 0.0893290638923645, -0.006601711735129356, 0.0960366502404213, -0.1824628859758377, -0.0023231615778058767, -0.03773796558380127, 0.015516508370637894, 0.05485468730330467, -0.0075123365968465805, -0.10711504518985748, 0.12969627976417542, -0.012708425521850586, 0.08275707811117172, -0.09777195006608963, 0.040892742574214935, 0.02267080917954445, 0.0828995406627655, -0.09138520807027817, 0.005103326868265867, 0.10812842845916748, -0.1388830840587616, -0.16463327407836914, 0.09262189269065857, -0.024146709591150284, 0.04174447059631348, 0.037952765822410583, 0.15489380061626434, 0.1751111000776291, -0.1816587597131729, -0.02006913721561432, 0.06113008037209511, -0.01569337584078312, -0.07532428950071335, -0.0047536264173686504, 0.1125023365020752, 0.020045368000864983, 0.04326675832271576, -0.026897238567471504, 0.12514881789684296, -0.03390365466475487, -0.08696594834327698, -0.02730787731707096, -0.08938638865947723, -0.07499715685844421, 0.05794275179505348, 0.003242113161832094, -0.05004361644387245, 0.017166541889309883, -0.14609181880950928, 0.16389138996601105, 0.018616922199726105, 0.01910393126308918, -0.08409728109836578, 0.10385308414697647, 0.00628611259162426, 0.003572766436263919, 0.0191201139241457, -0.0538480207324028, -0.11053936928510666, 0.2372369021177292, 0.08148732036352158, 0.08182648569345474, 0.06197001412510872, -0.04611215367913246, -0.0748995989561081, 0.022122330963611603, 0.016761217266321182, -0.036968328058719635, 0.02513561025261879, -0.09939522296190262, 0.04989476501941681, -0.014625283889472485, 0.03169732540845871, 0.002189639024436474, -0.028954457491636276, 0.05933646112680435, 0.014927499927580357, -0.02684570476412773, 0.0923490971326828, 0.06311790645122528, -0.01824377104640007, -0.06464351713657379, 0.00803230982273817, 0.07760901004076004, -0.0077133807353675365, -0.0930289551615715, 0.041021592915058136, -0.021816639229655266, 0.06046343967318535, 0.1970541924238205, -0.218502476811409, 0.04611575976014137, 0.014295408502221107, 0.04925490915775299, 0.03239617124199867, 0.006366894114762545, -0.027879852801561356, 0.02252461388707161, -0.03234133496880531, 0.07444038987159729, -0.015046633780002594, 0.0715152770280838, -0.029586369171738625, -0.13658837974071503, -0.022138170897960663, -0.025983601808547974, 0.16934117674827576, -0.18552166223526, 0.0699010118842125, 0.17087221145629883, -0.11822905391454697, 0.1519831120967865, 0.00021746719721704721, -0.013698418624699116, 0.005675546359270811, 0.03780049458146095, -0.002801888855174184, 0.09672058373689651, -0.08645369112491608, -0.02926025725901127, 0.02201896905899048, -0.08498319238424301, 0.03371753916144371, -0.11242473125457764, -0.11958957463502884, -0.0699458196759224, -0.033071815967559814, -0.04205494746565819, 0.03265739232301712, -0.04924427717924118, 0.07701261341571808, -0.09250124543905258, -0.062375374138355255, -0.023873237892985344, -0.08722291141748428, 0.015148735605180264, 0.011541252955794334, -0.05765252932906151, -0.15943875908851624, -0.11263208091259003, -0.10609307885169983, -0.14338603615760803, -0.009004314430058002, 0.06254543364048004, -0.1256828010082245, -0.041951846331357956, 0.008568373508751392, -0.04615076631307602, 0.08512871712446213, -0.07665080577135086, 0.02438226155936718, 0.05235583335161209, -0.033015649765729904, -0.1674128770828247, 0.006109215784817934, -0.06340905278921127, -0.06494861096143723, 0.159310445189476, -0.15431465208530426, 0.1818457543849945, -0.040174007415771484, 0.05442525073885918, 0.0655440241098404, 0.030762793496251106, 0.13360457122325897, -0.11267745494842529, 0.08281969279050827, 0.18579614162445068, 0.042681917548179626, 0.0769117921590805, 0.1317943036556244, 0.08133655041456223, -0.10714726895093918, 0.0417051799595356, 0.07930883020162582, -0.09971408545970917, -0.08149358630180359, -0.042686935514211655, -0.11835816502571106, -0.05499987304210663, 0.04330987110733986, 0.059732623398303986, 0.05720693618059158, 0.12338270992040634, -0.055996883660554886, -0.0016011609695851803, 0.10071559250354767, 0.0424453429877758, 0.06537508964538574, 0.01812180131673813, 0.05797940865159035, -0.151513934135437, -0.03798665851354599, 0.15856806933879852, 0.21609091758728027, 0.2242986112833023, 0.02033565565943718, 0.07650819420814514, 0.12681207060813904, 0.08542542159557343, 0.10626895725727081, 0.0423312671482563, 0.006014388054609299, 0.012574741616845131, -0.07227714359760284, -0.051804788410663605, 0.013200628571212292, 0.011147155426442623, -0.03476068750023842, -0.14159899950027466, 0.10551559180021286, 0.007154181133955717, 0.07971682399511337, 0.13928700983524323, 0.028220780193805695, -0.10144460946321487, 0.1541370451450348, 0.09728571027517319, 0.08193369954824448, -0.05540802702307701, 0.12944923341274261, 0.055390745401382446, -0.008984275162220001, 0.1677747219800949, 0.0298271756619215, 0.15396998822689056, -0.025932898744940758, -0.08113177120685577, -0.07403047382831573, -0.062083303928375244, 0.003279043361544609, 0.030136363580822945, -0.22424301505088806, 0.10980966687202454, 0.06548650562763214, 0.022217513993382454, -0.002643251558765769, -0.056200530380010605, 0.19739331305027008, 0.1589106321334839, 0.07469391822814941, 0.023671936243772507, -0.054583024233579636, -0.027448374778032303, -0.07951861619949341, 0.05455766245722771, 0.022745080292224884, 0.07905461639165878, -0.042170073837041855, -0.09559973329305649, -0.01727345585823059, -0.00700397277250886, 0.049957554787397385, -0.09634371101856232, -0.11104481667280197, -0.0419858917593956, 0.25329259037971497, -0.060838956385850906, 0.05075208470225334, 0.05483807623386383, 0.007478783838450909, -0.03630410507321358, 0.04422314465045929, -0.03177504613995552, -0.018244605511426926, -0.0317978709936142, 0.002027528127655387, 0.0008041720138862729, -0.05538520961999893, -0.06366303563117981, -0.03440028056502342, -0.09316612035036087, -0.10334205627441406, -0.008028445765376091, -0.03966424614191055, 0.01787806861102581, -0.020626883953809738, 0.006097805220633745, -0.09547159820795059, -0.034509964287281036, 0.01797221414744854, 0.02983693964779377, -0.06800559163093567, -0.1291925460100174, -0.004992731846868992, -0.002143785823136568, -0.055038485676050186, 0.025895189493894577, -0.11506348848342896, -0.08902840316295624, -0.05763055011630058, -0.046604275703430176, 0.12178884446620941, 0.22536733746528625, -0.018435098230838776, 0.0020454723853617907, 0.15572546422481537, -0.09342719614505768, -0.313409686088562, -0.17291942238807678, -0.16264788806438446, -0.10020279884338379, 0.03040791116654873, -0.07721591740846634, 0.02411762811243534, 0.08585338294506073, -0.03470494970679283, 0.20205126702785492, -0.20597440004348755, -0.10582274198532104, 0.0728377252817154, 0.10151691734790802, 0.31910574436187744, -0.2624204456806183, 0.01907777041196823, -0.11290736496448517, -0.03995766490697861, -0.008379657752811909, -0.08004666864871979, 0.12375303357839584, 0.03850951045751572, 0.07922405004501343, -0.012068371288478374, -0.004930188413709402, 0.14835099875926971, -0.08570975065231323, 0.12528930604457855, -0.12943211197853088, -0.0918096974492073, 0.20431649684906006, -0.03193909674882889, 0.0030670668929815292, -0.20796111226081848, -0.038851264864206314, -0.04878853261470795, 0.03399510309100151, -0.004125331528484821, 0.04794273525476456, -0.0060296435840427876, -0.014863459393382072, -0.12961941957473755, -0.00607253098860383, -0.026834893971681595, 0.062206756323575974, 0.21170593798160553, -0.06833510100841522, -0.05114913731813431, 0.022108402103185654, -0.010238687507808208, 0.10782884061336517, 0.016008678823709488, -0.06234949082136154, -0.043851058930158615, 0.0952180027961731, -0.2043023556470871, 0.05294916406273842, -0.005841487552970648, -0.005673081614077091, 0.01693277433514595, 0.00785826612263918, 0.023892633616924286, 0.13216182589530945, 0.1808609962463379, 0.007285060361027718, -0.04713507369160652, -0.015037656761705875, 0.02301948331296444, 0.13481755554676056, -0.025067579001188278, 0.10357043147087097, 0.02256532572209835, 0.04153842851519585, 0.007691334933042526, 0.059632860124111176, -0.07918737083673477, -0.09173990041017532, 0.09995826333761215, -0.042990874499082565, -0.08023561537265778, 0.09727633744478226, 0.049108125269412994, 0.07131274789571762, 0.0022757877595722675, 0.05649459734559059, 0.019125204533338547, -0.12543143332004547, 0.0072751776315271854, 0.20107021927833557, -0.04914954677224159, -0.07148387283086777, -0.0692133978009224, 0.019193777814507484, -0.11742905527353287, 0.08799365907907486, 0.02916301041841507, -0.033224206417798996, 0.12371402233839035, -0.042106419801712036, -0.027539577335119247, 0.00914523284882307, -0.057449933141469955, 0.028751932084560394, -0.14592300355434418, -0.19769753515720367, 0.05620778724551201, 0.004571276716887951, -0.06587675958871841, -0.09644706547260284, -0.092591293156147, 0.06868492066860199, -0.15211084485054016, 0.14579567313194275, -0.08287985622882843, 0.05659882724285126, -0.04222286120057106, -0.05210746079683304, -0.12014317512512207, -0.022194715216755867, -0.04659072682261467, -0.01898675598204136, 0.056548796594142914, 0.014400600455701351, -0.12068374454975128, -0.11505278199911118, 0.05967089161276817, -0.002018850762397051, -0.008497701026499271, 0.01908198371529579, -0.07262430340051651, 0.020431961864233017, -0.23708386719226837, -0.05991276353597641, 0.08123636990785599, 0.0354122631251812, -0.09314844757318497, 0.12169944494962692, 0.04888284206390381, -0.025450678542256355, 0.048065975308418274, 0.005709515418857336, 0.18631795048713684, -0.07177245616912842, 0.028831256553530693, -0.13222987949848175, -0.1587904393672943, -0.04089801385998726, 0.029888290911912918, 0.23085427284240723, 0.07886390388011932, 0.11645402759313583, -0.05573349818587303, 0.024490447714924812, -0.014110677875578403, 0.07453909516334534, 0.019784318283200264, -0.10002069175243378, -0.04683883860707283, -0.1722240298986435, -0.06777095049619675, -0.06619145721197128, 0.1686968356370926, 0.030445702373981476, -0.1539190262556076, 0.0026126087177544832, 0.10004615783691406, -0.16982628405094147, -0.010848530568182468, 0.1877400428056717, -0.04288702830672264, 0.026094838976860046, -0.14587950706481934, 0.030022095888853073, 0.07092604041099548, -0.029173187911510468, -0.008188641630113125, 0.13492417335510254, 0.009638035669922829, 0.006906747817993164, 0.025905927643179893, -0.03052508644759655, 0.08842901140451431, -0.06780105829238892, 0.044680774211883545, -0.007880161516368389, -0.045106589794158936, -0.12043027579784393, 0.21767351031303406, -0.014898164197802544, 0.013906550593674183, -0.06449268758296967, 0.001173527562059462, -0.09772270172834396, -0.09932434558868408, -0.0784570723772049, -0.1301494836807251, 0.08290880918502808, -0.057078272104263306, 0.008291185833513737, 0.0019239645916968584, 0.018788795918226242, -0.07216601073741913, 0.020393919199705124, -0.17540444433689117, -0.044320981949567795, 0.012736436910927296, -0.0197144765406847, -0.021591126918792725, -0.04366343468427658, -0.04387117922306061, 0.026234956458210945, -0.06935053318738937, -0.0657866820693016, 0.05766259506344795, 0.07270865887403488, 0.06597478687763214, -0.16228951513767242, -0.11147043853998184, -0.07132118940353394, 0.03634883835911751, 0.07855279743671417, 0.18165458738803864, 0.039241351187229156, -0.009584135375916958, 0.033849555999040604, 0.12407470494508743, 0.02284465916454792, -0.07813289761543274, -0.07462552934885025, -0.14595992863178253, -0.13948774337768555, -0.028066769242286682, -0.0621083527803421, -0.025582313537597656, 0.01867021434009075, 0.23777411878108978, 0.19138801097869873, -0.15210510790348053, 0.03755912184715271, -0.07920726388692856, 0.04123891517519951, -0.03645355626940727, 0.1770937740802765, 0.050931382924318314, 0.15057048201560974, -0.03379904478788376, -0.039562519639730453, -0.069175586104393, 0.01292650680989027, -0.1002921313047409, 0.041722372174263, -0.006777461152523756, -0.06383277475833893, -0.06307639181613922, 0.10785272717475891, -0.1247771605849266, 0.06662343442440033, 0.1837296038866043, -0.13652652502059937, -0.012055957689881325, -0.040316611528396606, 0.05371761694550514, 0.11423759162425995, 0.011796320788562298, -0.07520892471075058, -0.012514122761785984, -0.0045129903592169285, 0.037176623940467834, -0.1757640242576599, -0.11733264476060867, -0.0004282773588784039, -0.10733392089605331, 0.1406038999557495, -0.007499363739043474, -0.0036998731084167957, 0.03480591997504234, -0.06952330470085144, 0.004121360834687948, 0.18553639948368073, 0.026520973071455956, -0.029349427670240402, -0.03073844686150551, -0.05272417515516281, -0.0995614156126976, 0.07846953719854355, 0.08589665591716766, 0.05562399700284004, 0.0015277499333024025, 0.15854448080062866, -0.011656023561954498, -0.0401025228202343, 0.14720255136489868, -0.17590230703353882, 0.09711183607578278, 0.006745760329067707, -0.015211387537419796, -0.07118199020624161, -0.041605524718761444, 0.04867985099554062, 0.06932266801595688, -0.1827884316444397, -0.05423454940319061, 0.0576418861746788, -0.10212443023920059, 0.04955611377954483, 0.046302903443574905, -0.0998016893863678, 0.012012151069939137, -0.12462915480136871, -0.008357822895050049, -0.09701721370220184, 0.04985027760267258, 0.19329488277435303, -0.033785898238420486, 0.014376556500792503, -0.1586226522922516, 0.05741400644183159, -0.034309543669223785, -0.047869812697172165, -0.08198888599872589 ]
null
null
transformers
--- library_name: transformers license: bigscience-openrail-m --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"language": ["en"], "license": "apache-2.0", "datasets": ["tyson0420/stackexchange-4dpo-filby-clang-keywords", "tyson0420/valid_stack_exchange_ai_fil"]}
text-generation
tyson0420/mixtral_stack_llama
[ "transformers", "safetensors", "mixtral", "text-generation", "en", "dataset:tyson0420/stackexchange-4dpo-filby-clang-keywords", "dataset:tyson0420/valid_stack_exchange_ai_fil", "arxiv:1910.09700", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-15T04:52:39+00:00
[ "1910.09700" ]
[ "en" ]
TAGS #transformers #safetensors #mixtral #text-generation #en #dataset-tyson0420/stackexchange-4dpo-filby-clang-keywords #dataset-tyson0420/valid_stack_exchange_ai_fil #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
--- library_name: transformers license: bigscience-openrail-m --- # Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #mixtral #text-generation #en #dataset-tyson0420/stackexchange-4dpo-filby-clang-keywords #dataset-tyson0420/valid_stack_exchange_ai_fil #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 113, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #mixtral #text-generation #en #dataset-tyson0420/stackexchange-4dpo-filby-clang-keywords #dataset-tyson0420/valid_stack_exchange_ai_fil #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware" ]
[ -0.04917329549789429, 0.159444198012352, -0.005255247000604868, 0.00272384868003428, 0.07941984385251999, -0.013671926222741604, 0.08847717940807343, 0.1011357456445694, 0.023035015910863876, 0.12323157489299774, 0.01975179836153984, 0.11816491931676865, 0.10905490070581436, 0.13553780317306519, -0.007312314584851265, -0.22050537168979645, 0.0858142077922821, -0.11567307263612747, 0.006561290007084608, 0.11594890803098679, 0.14055022597312927, -0.10590668022632599, 0.07996485382318497, -0.01691150851547718, -0.004137367941439152, -0.02722829580307007, -0.0674920380115509, -0.05023694410920143, 0.05035029351711273, 0.08334821462631226, 0.06607183814048767, 0.02448614500463009, 0.048611175268888474, -0.2675463855266571, 0.021170448511838913, 0.06675003468990326, -0.008031340315937996, 0.07004541158676147, 0.07113886624574661, -0.06971383094787598, 0.09581557661294937, -0.03576907888054848, 0.11617699265480042, 0.07672020047903061, -0.08604032546281815, -0.15191005170345306, -0.0811733677983284, 0.051491815596818924, 0.16698451340198517, 0.05710611119866371, -0.043269895017147064, 0.14378847181797028, -0.08508328348398209, 0.023194696754217148, 0.11221089214086533, -0.048878151923418045, -0.045437563210725784, 0.0287158340215683, 0.10467290878295898, 0.06894008815288544, -0.09504052996635437, -0.014134470373392105, 0.006371748633682728, 0.02223430946469307, 0.05507232993841171, 0.0017595833633095026, 0.13569743931293488, 0.018362626433372498, -0.1573374718427658, -0.044254936277866364, 0.1222454234957695, 0.02716434746980667, -0.053142204880714417, -0.23417137563228607, -0.040049970149993896, 0.013218180276453495, -0.02606169693171978, -0.0392722561955452, 0.03353700041770935, -0.016699910163879395, 0.09205532819032669, -0.0036446228623390198, -0.06581644713878632, -0.00793917290866375, 0.04897766932845116, 0.0870412141084671, 0.018565725535154343, -0.02774174138903618, -0.001774336094968021, 0.1270972490310669, 0.11206378042697906, -0.13004475831985474, -0.05731557682156563, -0.0654267892241478, -0.07157726585865021, -0.039042938500642776, 0.03471720218658447, 0.02824552357196808, 0.07023438811302185, 0.23174849152565002, 0.014079303480684757, 0.03133433684706688, 0.04224804788827896, 0.01564931310713291, 0.0683014988899231, 0.12377067655324936, -0.06349463015794754, -0.10143774002790451, -0.02099798060953617, 0.09425445646047592, 0.02380591258406639, -0.03836876153945923, -0.06843895465135574, 0.05436325818300247, 0.03692470118403435, 0.13959693908691406, 0.1244005486369133, -0.01414475031197071, -0.09831419587135315, -0.06195094436407089, 0.2028370052576065, -0.15643231570720673, 0.029813576489686966, 0.001225842977873981, -0.026028158143162727, -0.009351106360554695, 0.006496645975857973, 0.026696451008319855, -0.04888121038675308, 0.08623113483190536, -0.07222171127796173, -0.0197281576693058, -0.08609470725059509, -0.05720553174614906, 0.03522742912173271, 0.0004336716956458986, -0.021620163694024086, -0.04257795959711075, -0.07294006645679474, -0.06771755963563919, 0.0725022703409195, -0.07391168177127838, -0.06593453139066696, -0.0035161739215254784, -0.07193607836961746, 0.01533143874257803, 0.004292632453143597, 0.05588972568511963, -0.03106873668730259, 0.011993634514510632, -0.04946980997920036, 0.048195671290159225, 0.09783918410539627, 0.02850295789539814, -0.08329110592603683, 0.07394484430551529, -0.262620210647583, 0.11420039087533951, -0.06326963752508163, 0.05658520385622978, -0.14510010182857513, -0.00976523756980896, 0.06550855189561844, 0.01585584506392479, -0.0056763035245239735, 0.160483255982399, -0.23366419970989227, -0.03629021719098091, 0.1561349332332611, -0.11767983436584473, -0.09025364369153976, 0.055383000522851944, -0.03572366386651993, 0.10446309298276901, 0.04513826593756676, -0.02966165728867054, 0.053354695439338684, -0.13566941022872925, -0.0429818369448185, -0.03939862176775932, 0.005505614448338747, 0.154044970870018, 0.07329685240983963, -0.07664402574300766, 0.07922793924808502, 0.012477763928472996, -0.02345927432179451, -0.03231925144791603, -0.013191151432693005, -0.09416377544403076, 0.00912422314286232, -0.0339631661772728, 0.009044947102665901, -0.03804418817162514, -0.09895049035549164, -0.02070617489516735, -0.1773558109998703, 0.014069675467908382, 0.07334288954734802, -0.010382955893874168, -0.0077637662179768085, -0.10300272703170776, 0.045083750039339066, -0.0026198800187557936, -0.006322176661342382, -0.16765892505645752, -0.09180165082216263, 0.04309355095028877, -0.202597975730896, 0.027368798851966858, -0.037530191242694855, 0.018101010471582413, 0.032425716519355774, -0.01977786421775818, -0.030618827790021896, 0.016136767342686653, 0.010343847796320915, -0.01928294263780117, -0.2272108942270279, -0.024160247296094894, -0.038415949791669846, 0.1435362845659256, -0.18818937242031097, 0.0308246910572052, 0.1092105358839035, 0.15341828763484955, 0.008982861414551735, -0.03927591070532799, 0.02155085653066635, -0.062260955572128296, -0.049124930053949356, -0.06212574616074562, -0.010670637711882591, -0.030419394373893738, -0.08814933896064758, 0.06874540448188782, -0.19054609537124634, -0.052808426320552826, 0.10300670564174652, 0.0888579711318016, -0.12672488391399384, -0.046989571303129196, -0.03831856697797775, -0.054915059357881546, -0.07601006329059601, -0.07396500557661057, 0.10788487643003464, 0.06233633682131767, 0.02349206991493702, -0.08345288783311844, -0.08471456170082092, -0.00023762646014802158, -0.01809031516313553, -0.037169430404901505, 0.10005117952823639, 0.06205682083964348, -0.12086454778909683, 0.08978922665119171, 0.11698704957962036, 0.08434737473726273, 0.07913234829902649, -0.015440837480127811, -0.08128631860017776, -0.04636470973491669, 0.018984666094183922, 0.018216727301478386, 0.11244434118270874, -0.027994442731142044, 0.05846548452973366, 0.04741149768233299, -0.030760014429688454, 0.011282337829470634, -0.09006369113922119, 0.04246736690402031, 0.01903342828154564, -0.01903204806149006, 0.04246202111244202, -0.023676788434386253, 0.026997510343790054, 0.08900199830532074, 0.03050501085817814, 0.031204575672745705, 0.01172877661883831, -0.0429353304207325, -0.11571165919303894, 0.17631638050079346, -0.10959918797016144, -0.26493191719055176, -0.12395641207695007, 0.0018378393724560738, 0.02549203485250473, -0.01667417399585247, 0.02046087384223938, -0.05285373330116272, -0.13057276606559753, -0.10933247953653336, 0.0355575792491436, 0.07869265228509903, -0.060962975025177, -0.09545407444238663, 0.0443233922123909, 0.041797954589128494, -0.1310717612504959, 0.007485806941986084, 0.0495121143758297, -0.02829156257212162, -0.012775895185768604, 0.089695505797863, 0.08395125716924667, 0.16661621630191803, 0.04143492877483368, -0.02852776274085045, 0.03413091599941254, 0.2670392096042633, -0.13602055609226227, 0.1153063103556633, 0.16006074845790863, -0.10539029538631439, 0.09307090938091278, 0.2310069501399994, 0.024150412529706955, -0.10214875638484955, 0.04240044951438904, 0.01604459621012211, -0.03835557401180267, -0.24305155873298645, -0.08106372505426407, -0.0072272117249667645, -0.07264931499958038, 0.0699998140335083, 0.08802943676710129, 0.06270089000463486, 0.0316031239926815, -0.11317365616559982, -0.09068504720926285, 0.03385813534259796, 0.11195014417171478, 0.016765806823968887, -0.004566340707242489, 0.0918010026216507, -0.030456645414233208, -0.007800812367349863, 0.0943313017487526, -0.012939698994159698, 0.18524353206157684, 0.05579017102718353, 0.17454735934734344, 0.08192053437232971, 0.03731748089194298, 0.03714594617486, 0.021213535219430923, 0.03494301065802574, 0.00888274796307087, -0.004674122203141451, -0.09535159170627594, 0.011408960446715355, 0.134618878364563, 0.08385623246431351, 0.012290491722524166, 0.03485238179564476, -0.04035727679729462, 0.08749092370271683, 0.15520527958869934, 0.025805484503507614, -0.20718753337860107, -0.03723223879933357, 0.09219832718372345, -0.0863165408372879, -0.11104071140289307, -0.028063347563147545, 0.03209707513451576, -0.17207089066505432, 0.03663593530654907, -0.023061970248818398, 0.11009027808904648, -0.1646798849105835, -0.0281680840998888, 0.03630552813410759, 0.08046577125787735, -0.014647131785750389, 0.05828634649515152, -0.15414175391197205, 0.09882957488298416, 0.021083751693367958, 0.03688739240169525, -0.10323180258274078, 0.11412379890680313, 0.0036410309839993715, -0.014667574316263199, 0.1646251678466797, 0.019604241475462914, -0.07594810426235199, -0.0455283597111702, -0.11713066697120667, -0.014258998446166515, 0.11228493601083755, -0.14758288860321045, 0.09493040293455124, -0.0227464959025383, -0.044307831674814224, -0.010357828810811043, -0.1205720603466034, -0.12738938629627228, -0.19279815256595612, 0.08292695879936218, -0.11021353304386139, 0.04350744187831879, -0.1026453971862793, -0.04952351003885269, 0.0029436093755066395, 0.21076816320419312, -0.20075222849845886, -0.0941859781742096, -0.1508646011352539, -0.0984816923737526, 0.16845019161701202, -0.05967541038990021, 0.08194226771593094, -0.017277253791689873, 0.19751043617725372, 0.002675285330042243, -0.00936389621347189, 0.11160832643508911, -0.10421882569789886, -0.18569251894950867, -0.0760996863245964, 0.13611198961734772, 0.1450709104537964, 0.03501749038696289, -0.0041993665508925915, 0.04038218781352043, -0.026031265035271645, -0.12430023401975632, -0.00015944361803121865, 0.14894284307956696, 0.07060328871011734, 0.029040906578302383, -0.008158832788467407, -0.0989009216427803, -0.09300340712070465, -0.04114557057619095, 0.029524190351366997, 0.20595069229602814, -0.05700298398733139, 0.1417115330696106, 0.12267964333295822, -0.04838309809565544, -0.1873467117547989, 0.008191348053514957, 0.04083095118403435, 0.0012412446085363626, 0.04040246829390526, -0.17719316482543945, 0.11014509201049805, -0.007253906689584255, -0.04714829847216606, 0.15003906190395355, -0.15700890123844147, -0.1627928763628006, 0.10009519755840302, 0.0380186103284359, -0.1810317486524582, -0.11068997532129288, -0.11382531374692917, -0.058580365031957626, -0.10149820894002914, 0.10621960461139679, 0.02670869790017605, -0.005181923042982817, 0.055108729749917984, 0.021866051480174065, 0.023763973265886307, -0.03303416073322296, 0.21680311858654022, 0.0012072831159457564, 0.008521946147084236, -0.07476281374692917, -0.0866866484284401, 0.032643307000398636, -0.05820051580667496, 0.06671221554279327, -0.008007671684026718, 0.001291082939133048, -0.055465731769800186, -0.047840606421232224, -0.035647351294755936, 0.04995804652571678, -0.08734151721000671, -0.09178612381219864, -0.05922382324934006, 0.09727267920970917, 0.07483026385307312, -0.024814052507281303, 0.011252245865762234, -0.09253175556659698, 0.016474122181534767, 0.20175965130329132, 0.17717300355434418, 0.04304102808237076, -0.05751610919833183, 0.015994958579540253, -0.0066413357853889465, 0.03202632814645767, -0.15358084440231323, 0.055669330060482025, 0.04654789716005325, -0.002825300209224224, 0.12281286716461182, -0.02292597107589245, -0.17589612305164337, -0.0630887970328331, 0.05932660028338432, -0.07516144216060638, -0.1818586140871048, -0.01902373507618904, 0.050299473106861115, -0.1758131980895996, -0.057503364980220795, 0.0631212592124939, 0.020295172929763794, -0.038396310061216354, 0.023236358538269997, 0.0802735760807991, 0.019963014870882034, 0.07709354907274246, 0.04949900880455971, 0.09394130855798721, -0.1055506095290184, 0.06748341768980026, 0.07733830064535141, -0.09777271747589111, 0.0052496446296572685, 0.07383108139038086, -0.060843296349048615, -0.029178114607930183, -0.017640048637986183, 0.06399904191493988, -0.03336961939930916, -0.049789901822805405, -0.03615652769804001, -0.10920335352420807, 0.0642518401145935, 0.13237246870994568, 0.05255913734436035, 0.01635412871837616, 0.042759813368320465, 0.013675248250365257, -0.047372523695230484, 0.11832531541585922, 0.04093621298670769, 0.03404923900961876, -0.058507490903139114, -0.07676422595977783, 0.044274356216192245, -0.004205274395644665, -0.012198632583022118, -0.036054447293281555, -0.07889270782470703, -0.019864995032548904, -0.16194452345371246, 0.0030951343942433596, -0.0747697576880455, 0.003196167293936014, 0.01583745703101158, -0.04572472721338272, -0.00995628722012043, 0.003634963184595108, -0.07900094240903854, -0.06115785241127014, -0.025349145755171776, 0.11435715109109879, -0.14934752881526947, 0.021754277870059013, 0.09390439093112946, -0.13465166091918945, 0.10048136115074158, -0.001065612304955721, -0.01586104743182659, 0.010994807817041874, -0.11689449101686478, 0.04623278230428696, 0.0068559846840798855, 0.020054573193192482, 0.05829663574695587, -0.21973557770252228, -0.014623850584030151, -0.020447812974452972, -0.054956622421741486, -0.018384497612714767, -0.04640485718846321, -0.11966411769390106, 0.09986605495214462, 0.021449171006679535, -0.10038334131240845, -0.0299672931432724, 0.035489488393068314, 0.08403090387582779, -0.024778468534350395, 0.15400782227516174, 0.011038771830499172, 0.08125106245279312, -0.15866495668888092, -0.012412501499056816, -0.006921084132045507, 0.010924470610916615, 0.04398532211780548, -0.019811324775218964, 0.039930686354637146, -0.016890449449419975, 0.21464373171329498, -0.039104804396629333, 0.059651926159858704, 0.06774275749921799, 0.0719514861702919, -0.008844029158353806, 0.09232623130083084, 0.024353457614779472, 0.022477174177765846, 0.017234286293387413, -0.0013523765373975039, -0.04189547896385193, -0.009671194478869438, -0.16587483882904053, 0.09826433658599854, 0.11899146437644958, 0.10636626929044724, 0.016207469627261162, 0.07207745313644409, -0.0821867287158966, -0.09722921252250671, 0.0878533273935318, -0.020608721300959587, 0.00605057505890727, -0.04194945469498634, 0.14923056960105896, 0.15366053581237793, -0.17779384553432465, 0.05941151827573776, -0.018067825585603714, -0.04955171048641205, -0.13730598986148834, -0.21582551300525665, -0.05871215835213661, -0.03122333623468876, -0.02549716643989086, -0.06215694919228554, 0.05611354857683182, 0.08035208284854889, -0.0032608609180897474, 0.004231151659041643, 0.08772693574428558, -0.00678171357139945, -0.041337043046951294, 0.032485947012901306, 0.05526771768927574, 0.02232791669666767, -0.01740451343357563, 0.01924903132021427, -0.010540341027081013, 0.06606951355934143, 0.04348386824131012, 0.05225410312414169, -0.034254275262355804, 0.021849999204277992, -0.04683902487158775, -0.11716337502002716, 0.022005949169397354, -0.02931932918727398, -0.08443448692560196, 0.12722229957580566, 0.02170259691774845, 0.013286080211400986, -0.015018871985375881, 0.21373336017131805, -0.08422058075666428, -0.08403007686138153, -0.12689438462257385, 0.0702289417386055, -0.0525101013481617, 0.07042352110147476, 0.03626779466867447, -0.10914495587348938, 0.0025060574989765882, 0.10397161543369293, 0.12982620298862457, -0.03973928466439247, 0.027616076171398163, 0.024570032954216003, 0.010188291780650616, -0.0461280457675457, 0.02660255879163742, 0.05464548617601395, 0.18471843004226685, -0.059707801789045334, 0.09605742245912552, -0.022321701049804688, -0.10229262709617615, -0.038026463240385056, 0.10862389951944351, 0.0031248799059540033, 0.025702111423015594, -0.06073428690433502, 0.12349840253591537, -0.08574005216360092, -0.19040195643901825, 0.03967317193746567, -0.05359802395105362, -0.14679642021656036, -0.021509340032935143, 0.04358862712979317, -0.0237599927932024, 0.03542368486523628, 0.04898432269692421, -0.08895615488290787, 0.1777046173810959, 0.036549173295497894, -0.07048788666725159, -0.05768734961748123, 0.07877302169799805, -0.10435562580823898, 0.2505630552768707, 0.014521227218210697, 0.054792292416095734, 0.11268874257802963, -0.018451273441314697, -0.14787261188030243, 0.04407281428575516, 0.09746965020895004, -0.0947326049208641, 0.0669516772031784, 0.18028433620929718, -0.000962135789450258, 0.12459766119718552, 0.0867563784122467, -0.05047536641359329, 0.06532526016235352, -0.08311869204044342, -0.07833926379680634, -0.09611684083938599, 0.07082532346248627, -0.09262320399284363, 0.1398409605026245, 0.10962647199630737, -0.05700530484318733, 0.0023807426914572716, -0.021320415660738945, 0.04793087765574455, 0.016487855464220047, 0.08963468670845032, 0.01065150462090969, -0.1717887967824936, 0.028010763227939606, 0.06377994269132614, 0.10537014156579971, -0.1614166498184204, -0.08918020129203796, 0.035186249762773514, -0.008172756060957909, -0.07408014684915543, 0.11609157174825668, 0.03241659700870514, 0.03957958519458771, -0.045669570565223694, -0.03112339973449707, -0.004236721433699131, 0.1289798468351364, -0.11685369908809662, -0.020306479185819626 ]
null
null
transformers
# Model Trained Using AutoTrain This model was trained using AutoTrain. For more information, please visit [AutoTrain](https://hf.co/docs/autotrain). # Architecture This model finetuned version of llama-2-7b # Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_path = "OmlyCheeini/Llama-Discore" tokenizer = AutoTokenizer.from_pretrained(model_path) model = AutoModelForCausalLM.from_pretrained( model_path, device_map="auto", torch_dtype='auto' ).eval() # Prompt content: "hi" messages = [ {"role": "user", "content": "hi"} ] input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors='pt') output_ids = model.generate(input_ids.to('cuda')) response = tokenizer.decode(output_ids[0][input_ids.shape[1]:], skip_special_tokens=True) # Model response: "Hello! How can I assist you today?" print(response) ```
{"license": "other", "library_name": "transformers", "tags": ["autotrain", "text-generation"], "widget": [{"text": "Hello "}], "pipeline_tag": "text2text-generation"}
text2text-generation
OnlyCheeini/Llama-Discore
[ "transformers", "safetensors", "autotrain", "text-generation", "text2text-generation", "license:other", "endpoints_compatible", "has_space", "region:us" ]
2024-02-15T04:54:33+00:00
[]
[]
TAGS #transformers #safetensors #autotrain #text-generation #text2text-generation #license-other #endpoints_compatible #has_space #region-us
# Model Trained Using AutoTrain This model was trained using AutoTrain. For more information, please visit AutoTrain. # Architecture This model finetuned version of llama-2-7b # Usage
[ "# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.", "# Architecture \n\nThis model finetuned version of llama-2-7b", "# Usage" ]
[ "TAGS\n#transformers #safetensors #autotrain #text-generation #text2text-generation #license-other #endpoints_compatible #has_space #region-us \n", "# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.", "# Architecture \n\nThis model finetuned version of llama-2-7b", "# Usage" ]
[ 47, 29, 14, 3 ]
[ "passage: TAGS\n#transformers #safetensors #autotrain #text-generation #text2text-generation #license-other #endpoints_compatible #has_space #region-us \n# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.# Architecture \n\nThis model finetuned version of llama-2-7b# Usage" ]
[ -0.008456175215542316, 0.10585593432188034, 0.00010751133959274739, 0.0625758096575737, 0.09050112962722778, -0.04680807515978813, 0.22228413820266724, 0.07465317100286484, -0.07591325789690018, -0.02226267382502556, 0.12589320540428162, 0.16557243466377258, -0.038947079330682755, 0.1880701184272766, 0.0035622292198240757, -0.24351045489311218, 0.044247038662433624, -0.01850030943751335, 0.004072871990501881, 0.09191162139177322, 0.12365809828042984, -0.07924796640872955, 0.082940474152565, 0.043493643403053284, -0.18188130855560303, 0.06430009007453918, 0.022179175168275833, -0.10202371329069138, 0.14349474012851715, 0.045437704771757126, 0.1152179092168808, 0.03472519293427467, 0.11344697326421738, -0.11731590330600739, 0.016111239790916443, 0.012263883836567402, -0.030644956976175308, 0.06108202040195465, 0.08157791197299957, -0.02201654016971588, 0.16526451706886292, 0.11028936505317688, 0.06215750798583031, 0.004729487933218479, -0.07700981944799423, -0.04770230874419212, -0.00533448439091444, 0.11605779826641083, 0.11606231331825256, 0.08739468455314636, 0.015370829962193966, 0.0872826799750328, -0.12660284340381622, 0.07853034883737564, -0.010432635433971882, -0.2771224081516266, -0.04994563013315201, 0.14675717055797577, 0.10585753619670868, 0.06224067509174347, -0.09966444969177246, 0.08170628547668457, 0.10360492765903473, -0.022677190601825714, 0.0826651081442833, -0.03519333899021149, -0.12435544282197952, -0.003361908020451665, -0.09228374063968658, -0.005802831146866083, 0.1764335185289383, -0.06245522201061249, -0.011041617020964622, -0.10561051219701767, -0.05329984426498413, -0.02087464928627014, 0.015895919874310493, -0.09374391287565231, 0.0009064450860023499, 0.07472287118434906, 0.014229083433747292, -0.08856289833784103, -0.14632685482501984, -0.04697178676724434, -0.08364979922771454, 0.09515907615423203, -0.004119524732232094, 0.016571378335356712, -0.1010129451751709, 0.12427310645580292, 0.052411943674087524, -0.12243807315826416, 0.047156594693660736, -0.05529917776584625, 0.040697094053030014, -0.06145742908120155, -0.03383062034845352, -0.04726928472518921, 0.047570183873176575, 0.1799950897693634, 0.15384036302566528, 0.005799374543130398, -0.056010082364082336, 0.03582978621125221, -0.026820968836545944, 0.13121040165424347, 0.0023151529021561146, -0.04260507971048355, 0.07087017595767975, 0.025913825258612633, 0.03089630790054798, -0.022674567997455597, -0.22465436160564423, -0.02131473645567894, 0.03818926960229874, 0.08356963843107224, -0.019331850111484528, 0.07323478162288666, 0.004781695082783699, 0.020425383001565933, 0.0055899485014379025, -0.06508044898509979, 0.056847505271434784, -0.06629077345132828, 0.010158460587263107, -0.09371267259120941, 0.04514196887612343, 0.07932744920253754, 0.02916055917739868, 0.018505049869418144, -0.1037210002541542, -0.04449882730841637, -0.11046157777309418, -0.03736937791109085, 0.03139909729361534, -0.028159581124782562, 0.04736492782831192, -0.20312674343585968, -0.2472597360610962, -0.007712002377957106, 0.06830956041812897, -0.029523208737373352, -0.07540467381477356, -0.030776672065258026, -0.017048025503754616, 0.053577929735183716, -0.05014987289905548, 0.10684430599212646, -0.009561102837324142, 0.028178470209240913, -0.04012781381607056, 0.03782278299331665, -0.11736344546079636, 0.02450106292963028, -0.12633778154850006, 0.03411455824971199, -0.07065683603286743, 0.011148488149046898, -0.017467420548200607, 0.12066460400819778, -0.023984409868717194, 0.04706978425383568, -0.05740833282470703, 0.04720832780003548, 0.007870455272495747, 0.11096004396677017, -0.13237033784389496, 0.008065249770879745, 0.10196208953857422, -0.11953302472829819, -0.11684060096740723, 0.07966487109661102, -0.06109170243144035, 0.2386387437582016, 0.0784616693854332, 0.16958120465278625, 0.12369177490472794, -0.09296610951423645, 0.0550692155957222, 0.032894644886255264, -0.01699707843363285, -0.08227679878473282, 0.017080193385481834, 0.03010755218565464, -0.13906726241111755, 0.042048174887895584, 0.07133155316114426, 0.06131388619542122, 0.011871768161654472, -0.09718719869852066, -0.03488842025399208, -0.09382674843072891, 0.011609002947807312, -0.004110036883503199, 0.12210992723703384, -0.035666514188051224, -0.04021228477358818, 0.060181453824043274, 0.09368313103914261, 0.02655816823244095, -0.030670437961816788, -0.05457537621259689, 0.0760032907128334, -0.0715000107884407, -0.0007268101326189935, -0.10395964235067368, -0.05145685374736786, -0.034474391490221024, 0.05015837773680687, 0.02977924980223179, 0.14244818687438965, 0.053808290511369705, 0.01873781345784664, -0.027930989861488342, 0.00011324237129883841, 0.10982435941696167, 0.051101747900247574, -0.14045602083206177, -0.10655674338340759, 0.04041222110390663, -0.08582409471273422, 0.16173535585403442, -0.21378690004348755, 0.0316009595990181, -0.09669668972492218, 0.13926434516906738, -0.0075653595849871635, 0.06763533502817154, -0.07567699998617172, 0.012216990813612938, -0.09974660724401474, -0.020710188895463943, 0.06111341342329979, 0.02845185063779354, -0.024551711976528168, 0.11796440184116364, -0.20739765465259552, 0.2432900220155716, 0.11564561724662781, -0.03515440225601196, -0.05555100739002228, -0.06936144828796387, 0.008800476789474487, -0.003310115309432149, -0.05603376030921936, -0.04311018064618111, 0.040926866233348846, -0.08981920033693314, 0.1733148843050003, -0.06631173938512802, -0.020401161164045334, -0.025847578421235085, -0.09637074917554855, -0.03951418399810791, 0.02452990412712097, 0.06602675467729568, -0.20625261962413788, 0.13833318650722504, 0.2555360794067383, -0.08551308512687683, 0.21818682551383972, 0.04040350764989853, 0.006124396342784166, -0.022272717207670212, -0.023885104805231094, -0.009356766939163208, 0.017893461510539055, 0.0019711353816092014, -0.02368718571960926, 0.01997310109436512, 0.016953200101852417, 0.07874009758234024, -0.14815320074558258, -0.02587234042584896, 0.018943147733807564, 0.028308922424912453, 0.08243346214294434, 0.06867768615484238, -0.06769303232431412, 0.09762588143348694, -0.02675093151628971, -0.1316162496805191, 0.09758155047893524, 0.023511484265327454, -0.07263559848070145, 0.15127605199813843, -0.09296984225511551, -0.22423583269119263, -0.2120324820280075, -0.11025621742010117, -0.001241446239873767, 0.0718262568116188, 0.062528096139431, -0.050594527274370193, -0.05193489044904709, -0.04808253049850464, -0.05242551118135452, -0.017491377890110016, -0.013582905754446983, -0.054618559777736664, 0.04680690914392471, 0.03218105062842369, -0.11085409671068192, -0.027364928275346756, 0.023502811789512634, -0.07245532423257828, 0.014239664189517498, -0.0445554293692112, 0.08201810717582703, 0.14303681254386902, -0.04664752259850502, 0.007469487376511097, -0.003666508477181196, 0.1296858936548233, -0.05444878712296486, 0.023901190608739853, 0.19861875474452972, -0.014033188112080097, 0.043907299637794495, 0.22862429916858673, -0.0008426204440183938, -0.06027241796255112, 0.057169653475284576, -0.021768685430288315, -0.09933101385831833, -0.1635131537914276, -0.07452105730772018, -0.04815930128097534, -0.038343947380781174, 0.10579679161310196, 0.07823363691568375, 0.21597683429718018, 0.14306136965751648, 0.04471883922815323, 0.06983812898397446, -0.009723023511469364, 0.07821168005466461, 0.16350550949573517, -0.03519980609416962, 0.15251140296459198, -0.09826429188251495, -0.166761577129364, 0.050975047051906586, 0.05586386099457741, 0.03540343418717384, 0.11146733909845352, -0.06914149969816208, 0.04105185344815254, 0.1043444573879242, 0.09478423744440079, 0.09376069903373718, 0.06439505517482758, -0.055490534752607346, -0.022795025259256363, -0.03620203584432602, -0.0373389795422554, 0.1357925683259964, -0.015633998438715935, -0.044494181871414185, -0.032442908734083176, 0.0069008176214993, 0.012200473807752132, 0.06683759391307831, 0.0243578739464283, -0.2826240658760071, 0.06588519364595413, 0.07866764813661575, -0.05705796182155609, -0.11218485236167908, 0.05122154951095581, 0.04916895553469658, -0.13667075335979462, 0.020031701773405075, -0.004086927976459265, 0.08442991971969604, 0.041312944144010544, 0.04148883372545242, -0.09439370781183243, -0.021205218508839607, -0.003994745202362537, 0.15117022395133972, -0.28919464349746704, 0.16733495891094208, 0.006196172442287207, 0.03331175819039345, -0.10541355609893799, 0.028858356177806854, 0.08714873343706131, 0.1669645458459854, 0.07764139026403427, -0.05258377641439438, -0.13626384735107422, -0.11057744920253754, -0.07093647122383118, -0.006237450521439314, 0.011370408348739147, -0.04937458038330078, 0.0492626316845417, -0.1148831769824028, 0.004104220308363438, 0.033558301627635956, 0.094245046377182, -0.09702886641025543, -0.17627067863941193, -0.005827234126627445, 0.11280304938554764, 0.08279778063297272, -0.0624089241027832, -0.06309574842453003, -0.1105550229549408, 0.17481562495231628, 0.09726582467556, 0.0009277306962758303, -0.1275724172592163, -0.07380161434412003, 0.029928335919976234, -0.0369882769882679, 0.06842340528964996, -0.04420296102762222, 0.11065031588077545, -0.04388027638196945, -0.10158532857894897, 0.07149497419595718, -0.10352080315351486, -0.05456863343715668, -0.08916515856981277, 0.022814517840743065, -0.07319948822259903, -0.021635200828313828, 0.10816018283367157, 0.03024011291563511, -0.09608498960733414, -0.06849499046802521, -0.05798263102769852, 0.025933805853128433, 0.003789391601458192, -0.08907753974199295, -0.14261721074581146, -0.05635278299450874, 0.017194408923387527, -0.08434072136878967, 0.14507360756397247, 0.14727096259593964, -0.10362134873867035, 0.12531337141990662, 0.17857541143894196, -0.11480683088302612, -0.3073054254055023, -0.10606743395328522, -0.04912916198372841, 0.05067092925310135, 0.06879612803459167, -0.08542971312999725, 0.10924302041530609, 0.02435920573771, -0.09993449598550797, -0.016270235180854797, -0.14825227856636047, -0.14831134676933289, 0.2172226458787918, 0.013908647932112217, 0.20132257044315338, -0.12112051993608475, -0.03879290819168091, -0.11646027117967606, 0.014611806720495224, 0.05128025263547897, -0.14736546576023102, 0.09071208536624908, 0.006256596185266972, 0.05997338891029358, 0.03733256086707115, -0.017277095466852188, 0.111129030585289, -0.06952036172151566, 0.08252651989459991, -0.1597813069820404, -0.014283456839621067, 0.08282611519098282, -0.08440543711185455, 0.07729235291481018, -0.13749049603939056, 0.04146907106041908, -0.10310821235179901, -0.04997536912560463, -0.0013526033144444227, 0.0746508464217186, -0.017568405717611313, -0.08314380794763565, -0.027607984840869904, 0.005681721959263086, 0.004058546386659145, -0.03347446396946907, 0.09005533158779144, -0.030542168766260147, 0.15139630436897278, 0.25377997756004333, 0.1586880087852478, -0.019669676199555397, 0.10246199369430542, -0.02734304405748844, -0.09079936146736145, 0.09354861080646515, -0.09609495848417282, 0.04035671427845955, 0.05993932858109474, -0.058447469025850296, 0.09491200000047684, 0.04963095113635063, -0.00012921456072945148, -0.008716845884919167, 0.17915929853916168, -0.12796077132225037, 0.021667271852493286, -0.049099717289209366, 0.09980373829603195, -0.027184201404452324, 0.013079309836030006, 0.13396495580673218, -0.06522107124328613, -0.01964338682591915, -0.0001408313401043415, -0.024633893743157387, -0.05216601490974426, 0.08790542930364609, 0.04127603396773338, 0.02647962048649788, -0.040749117732048035, 0.0201095100492239, 0.08595885336399078, 0.0331011600792408, 0.04181922972202301, 0.017022710293531418, -0.057938102632761, -0.093525730073452, 0.043566130101680756, 0.16959325969219208, -0.17412248253822327, -0.09285514056682587, -0.03027958981692791, -0.10508663207292557, 0.010543271899223328, 0.11864814907312393, 0.08012627810239792, 0.018438728526234627, -0.046331603080034256, -0.02970566414296627, -0.08734731376171112, 0.10414554178714752, -0.018251333385705948, 0.07611528784036636, -0.20509880781173706, 0.08299397677183151, -0.035079989582300186, 0.018272021785378456, -0.08677832782268524, 0.0008368991548195481, -0.1345815807580948, 0.006570211611688137, -0.197976216673851, -0.031546127051115036, -0.0462162010371685, -0.010760772973299026, 0.02012508548796177, -0.045044880360364914, -0.06653449684381485, -0.021742314100265503, -0.11552363634109497, 0.028766879811882973, -0.021896053105592728, 0.06315126270055771, -0.05061149224638939, -0.049370914697647095, 0.0229240320622921, -0.012347209267318249, 0.042701445519924164, -0.0274477768689394, 0.010119456797838211, 0.04520189017057419, -0.11725261807441711, 0.0227664764970541, 0.06469155848026276, 0.02352737821638584, 0.011645968072116375, -0.04635288193821907, -0.013616237789392471, 0.08537551760673523, 0.045284729450941086, 0.028942659497261047, -0.015106463804841042, -0.10528063774108887, 0.005124611780047417, -0.0017156205140054226, -0.13157415390014648, -0.018972959369421005, 0.015611114911735058, 0.02949143387377262, -0.04163437709212303, 0.20060060918331146, -0.08593445271253586, -0.007889271713793278, -0.06164011359214783, 0.03716413304209709, -0.05340370535850525, -0.08613167703151703, -0.07323911041021347, -0.1278122067451477, -0.04393896088004112, 0.001960249152034521, 0.23821358382701874, 0.13266269862651825, -0.050017017871141434, 0.01870202086865902, 0.07264783978462219, 0.09844136238098145, -0.00021126159117557108, 0.24470263719558716, 0.07363349199295044, 0.021707819774746895, -0.0760994479060173, 0.04250824823975563, 0.029697049409151077, 0.02454003132879734, 0.005425279960036278, 0.04823360592126846, -0.11016377806663513, 0.10115982592105865, 0.06181210279464722, -0.03155885636806488, -0.11644445359706879, -0.11920303851366043, -0.10523749887943268, 0.014801539480686188, -0.11502154171466827, 0.014708204194903374, 0.20624960958957672, -0.053836219012737274, -0.02158239483833313, -0.023368913680315018, -0.021732797846198082, -0.19136641919612885, -0.17167720198631287, -0.10411783307790756, -0.07705360651016235, -0.009213494136929512, -0.025462772697210312, 0.0728018581867218, 0.05278646945953369, 0.017255054786801338, -0.04454243183135986, 0.11202879250049591, -0.07766169309616089, -0.03639144077897072, -0.014491420239210129, -0.025740021839737892, 0.05748014897108078, -0.1286650449037552, -0.013095716014504433, -0.10632780194282532, -0.006913208868354559, -0.03719548508524895, -0.033442914485931396, 0.024720044806599617, -0.014261671341955662, -0.056751005351543427, -0.004899314604699612, -0.02558334544301033, 0.025044407695531845, 0.07621218264102936, 0.05741266533732414, -0.002307205693796277, -0.06316999346017838, 0.02450580894947052, 0.22026990354061127, -0.07595822960138321, -0.13288025557994843, -0.09300410002470016, 0.2424864023923874, -0.016485067084431648, 0.08115395903587341, -0.06379278749227524, -0.008473794907331467, 0.023489126935601234, 0.3112630546092987, 0.22716160118579865, -0.07498447597026825, 0.02197171188890934, -0.03721361234784126, -0.000491452869027853, -0.02827967330813408, 0.16349922120571136, 0.021572867408394814, 0.1599520891904831, -0.04819977656006813, 0.05447735637426376, -0.02603466436266899, -0.055025096982717514, -0.10549348592758179, 0.11309078335762024, -0.061517104506492615, -0.03535357117652893, -0.02888774313032627, 0.02135276608169079, -0.06440738588571548, 0.11233370006084442, -0.05290057137608528, -0.045609693974256516, -0.09572772681713104, 0.029106058180332184, 0.13520827889442444, -0.04422350972890854, 0.01656281389296055, -0.053662192076444626, -0.012162054888904095, 0.0022562441881746054, -0.04962541162967682, -0.06872659921646118, -0.031927429139614105, 0.0946340560913086, 0.019100414589047432, 0.20069733262062073, -0.01781184785068035, 0.020583661273121834, 0.10974788665771484, -0.027751529589295387, -0.10831490904092789, 0.09666634351015091, 0.006736826617270708, -0.08326106518507004, 0.11839373409748077, -0.01521723810583353, -0.011360702104866505, 0.015929095447063446, 0.032579049468040466, -0.06065254285931587, 0.07825785875320435, -0.08522350341081619, -0.040708672255277634, -0.059551190584897995, 0.06608381122350693, -0.03972524031996727, 0.1676206737756729, 0.07639437913894653, -0.040826600044965744, -0.014097432605922222, -0.033654265105724335, 0.07146131992340088, 0.032715339213609695, -0.07638584822416306, -0.026727471500635147, -0.15407581627368927, -0.051300033926963806, 0.0710080936551094, -0.022615617141127586, -0.2598153352737427, -0.08552069216966629, -0.07792970538139343, -0.04642710089683533, -0.12668578326702118, 0.08828074485063553, 0.2271517515182495, 0.02103402465581894, -0.009808441624045372, -0.152297705411911, -0.0465136393904686, 0.03290353715419769, -0.0846087783575058, -0.08241627365350723 ]