diff --git "a/LoRa_Merge_Script.ipynb" "b/LoRa_Merge_Script.ipynb" --- "a/LoRa_Merge_Script.ipynb" +++ "b/LoRa_Merge_Script.ipynb" @@ -3,8 +3,7 @@ "nbformat_minor": 0, "metadata": { "colab": { - "provenance": [], - "gpuType": "T4" + "provenance": [] }, "kernelspec": { "name": "python3", @@ -12,8 +11,7 @@ }, "language_info": { "name": "python" - }, - "accelerator": "GPU" + } }, "cells": [ { @@ -32,44 +30,2315 @@ "source": [ "#initialize\n", "import torch\n", - "from safetensors.torch import load_file\n", + "from safetensors.torch import load_file, save_file\n", "from google.colab import drive\n", "drive.mount('/content/drive')" ], "metadata": { - "id": "CBVTifA_ZwdC" + "id": "CBVTifA_ZwdC", + "outputId": "8ce58389-8263-4016-8ebe-f61708ffef95", + "colab": { + "base_uri": "https://localhost:8080/" + } }, - "execution_count": null, + "execution_count": 1, + "outputs": [ + { + "output_type": "stream", + "name": "stdout", + "text": [ + "Mounted at /content/drive\n" + ] + } + ] + }, + { + "cell_type": "code", + "source": [ + "from safetensors.torch import load_file, save_file\n", + "_puff = load_file('/content/drive/MyDrive/Saved from Chrome/pffy3FLUX.safetensors')\n", + "puff = {}\n", + "\n", + "#alpha = 64\n", + "#rank = 64\n", + "\n", + "# = > so scale = 1\n", + "#desired scale = 0.5\n", + "# so multiply matrices by 2 and set alpha to 32\n", + "device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\n", + "for key in _puff:\n", + " if f'{key}'.find('alpha')>-1:\n", + " puff[f'{key}'] = torch.tensor(32).to(device=device , dtype = torch.float16)\n", + " #print(puff[f'{key}'])\n", + " continue\n", + " puff[f'{key}'] = 2*_puff[f'{key}'].to(device=device , dtype = torch.float16)\n", + "\n", + " #print(puff[f'{key}'].shape)\n", + "\n", + "save_file(puff, 'buff.safetensors')" + ], + "metadata": { + "id": "SKYzFxehkfG8" + }, + "execution_count": 32, "outputs": [] }, { "cell_type": "code", "source": [ + "from safetensors.torch import load_file, save_file\n", + "_puff = load_file('/content/drive/MyDrive/Saved from Chrome/pfbkFLUX.safetensors')\n", + "puff = {}\n", "\n", + "#alpha = 64\n", + "#rank = 64\n", "\n", + "# = > so scale = 1\n", + "#desired scale = 0.5\n", + "# so multiply matrices by 2 and set alpha to 32\n", "device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\n", + "for key in _puff:\n", + " if f'{key}'.find('alpha')>-1:\n", + " puff[f'{key}'] = torch.tensor(32).to(device=device , dtype = torch.float16)\n", + " #print(puff[f'{key}'])\n", + " continue\n", + " puff[f'{key}'] = 2*_puff[f'{key}'].to(device=device , dtype = torch.float16)\n", "\n", - "doll = load_file('/content/drive/MyDrive/Saved from Chrome/dolls.safetensors')\n", - "euro = load_file('/content/drive/MyDrive/Saved from Chrome/euro.safetensors')\n", - "scale = load_file('/content/drive/MyDrive/Saved from Chrome/scale.safetensors')\n", - "cgi = load_file('/content/drive/MyDrive/Saved from Chrome/cgi.safetensors')\n", - "guns = load_file('/content/drive/MyDrive/Saved from Chrome/guns.safetensors')\n", - "iris = load_file('/content/drive/MyDrive/Saved from Chrome/iris.safetensors')\n", + " #print(puff[f'{key}'].shape)\n", "\n", - "for key in doll:\n", - " doll[f'{key}'] = doll[f'{key}'].to(device = device , dtype=torch.float16)\n", - " euro[f'{key}'] = euro[f'{key}'].to(device = device , dtype=torch.float16)\n", - " scale[f'{key}'] = scale[f'{key}'].to(device = device , dtype=torch.float16)\n", - " iris[f'{key}'] = iris[f'{key}'].to(device = device , dtype=torch.float16)\n", - " cgi[f'{key}'] = cgi[f'{key}'].to(device = device , dtype=torch.float16)\n", - " guns[f'{key}'] = guns[f'{key}'].to(device = device , dtype=torch.float16)" + "save_file(puff, 'puff.safetensors')" ], "metadata": { - "id": "1oxeJYHRqxQC" + "id": "U8fCk78GimS8" }, "execution_count": 28, "outputs": [] }, + { + "cell_type": "code", + "source": [ + "from safetensors.torch import load_file, save_file\n", + "_tongue = load_file('/content/drive/MyDrive/Saved from Chrome/tongue-flux-v2.1.safetensors')\n", + "tongue = {}\n", + "# Scale = 32/16 = 2\n", + "# Desired scale = 0.5 => multiply all matrices by 4 and set alpha to 8\n", + "device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\n", + "for key in _tongue:\n", + " if f'{key}'.find('alpha')>-1:\n", + " tongue[f'{key}'] = torch.tensor(8).to(device=device , dtype = torch.float16)\n", + " continue\n", + " #-------#\n", + " tongue[f'{key}'] = 4*_tongue[f'{key}'].to(device=device , dtype = torch.float16)\n", + "#-------#\n", + "save_file(tongue, 'tongue.safetensors')" + ], + "metadata": { + "id": "lFNa6vgrgdSA" + }, + "execution_count": 23, + "outputs": [] + }, + { + "cell_type": "code", + "source": [ + "\n", + "\n", + "device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\n", + "\n", + "_oily = load_file('/content/drive/MyDrive/Saved from Chrome/OiledSkin_FluxDev.safetensors')\n", + "\n", + "star = load_file('/content/drive/MyDrive/Saved from Chrome/star_100_r32_16alpha.safetensors')\n", + "#A = vs , B = u\n", + "#lora_down = A , lora_up = B\n", + "\n", + "oily = {}\n", + "for key in _oily:\n", + " if not f'{key}'.find('_A.')>-1:continue\n", + " A = f'{key}'\n", + " B = f'{key}'.replace('_A.','_B.')\n", + " down = f'{key}'.replace('_A.','_down.')\n", + " up = f'{key}'.replace('_A.','_up.')\n", + " #-----#\n", + " oily[f'{up}'] = _oily[f'{B}'].to(device = device , dtype=torch.float16)\n", + " oily[f'{down}'] = _oily[f'{A}'].to(device = device , dtype=torch.float16)\n", + " #------#\n", + " if not f'{key}'.find('to_k.')>-1:continue\n", + " k = f'{key}'\n", + " q = k.replace('to_k.','to_q.')\n", + " v = k.replace('to_k.','to_v.')\n", + "\n", + "print(\"---------OILY---------\")\n", + "for key in oily:\n", + " print(key)\n", + " #if f'{key}'.find('alpha')>-1:print(key)\n", + "\n", + "print(\"---------STAR---------\")\n", + "for key in star:\n", + " break\n", + " print(key)" + ], + "metadata": { + "id": "1oxeJYHRqxQC", + "collapsed": true, + "outputId": "12e3a407-f9d1-403e-949b-31330be59577", + "colab": { + "base_uri": "https://localhost:8080/" + } + }, + "execution_count": 12, + "outputs": [ + { + "output_type": "stream", + "name": "stdout", + "text": [ + "---------OILY---------\n", + "transformer.single_transformer_blocks.0.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.0.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.0.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.0.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.0.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.0.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.0.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.0.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.0.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.0.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.0.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.0.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.1.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.1.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.1.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.1.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.1.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.1.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.1.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.1.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.1.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.1.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.1.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.1.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.10.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.10.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.10.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.10.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.10.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.10.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.10.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.10.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.10.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.10.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.10.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.10.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.11.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.11.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.11.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.11.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.11.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.11.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.11.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.11.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.11.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.11.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.11.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.11.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.12.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.12.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.12.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.12.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.12.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.12.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.12.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.12.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.12.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.12.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.12.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.12.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.13.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.13.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.13.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.13.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.13.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.13.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.13.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.13.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.13.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.13.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.13.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.13.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.14.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.14.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.14.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.14.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.14.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.14.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.14.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.14.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.14.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.14.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.14.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.14.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.15.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.15.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.15.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.15.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.15.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.15.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.15.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.15.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.15.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.15.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.15.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.15.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.16.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.16.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.16.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.16.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.16.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.16.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.16.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.16.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.16.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.16.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.16.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.16.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.17.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.17.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.17.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.17.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.17.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.17.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.17.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.17.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.17.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.17.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.17.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.17.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.18.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.18.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.18.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.18.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.18.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.18.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.18.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.18.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.18.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.18.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.18.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.18.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.19.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.19.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.19.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.19.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.19.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.19.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.19.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.19.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.19.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.19.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.19.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.19.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.2.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.2.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.2.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.2.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.2.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.2.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.2.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.2.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.2.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.2.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.2.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.2.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.20.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.20.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.20.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.20.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.20.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.20.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.20.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.20.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.20.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.20.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.20.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.20.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.21.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.21.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.21.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.21.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.21.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.21.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.21.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.21.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.21.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.21.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.21.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.21.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.22.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.22.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.22.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.22.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.22.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.22.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.22.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.22.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.22.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.22.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.22.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.22.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.23.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.23.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.23.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.23.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.23.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.23.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.23.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.23.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.23.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.23.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.23.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.23.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.24.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.24.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.24.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.24.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.24.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.24.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.24.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.24.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.24.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.24.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.24.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.24.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.25.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.25.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.25.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.25.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.25.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.25.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.25.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.25.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.25.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.25.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.25.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.25.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.26.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.26.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.26.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.26.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.26.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.26.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.26.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.26.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.26.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.26.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.26.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.26.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.27.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.27.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.27.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.27.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.27.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.27.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.27.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.27.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.27.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.27.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.27.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.27.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.28.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.28.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.28.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.28.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.28.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.28.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.28.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.28.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.28.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.28.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.28.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.28.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.29.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.29.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.29.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.29.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.29.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.29.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.29.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.29.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.29.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.29.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.29.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.29.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.3.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.3.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.3.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.3.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.3.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.3.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.3.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.3.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.3.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.3.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.3.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.3.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.30.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.30.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.30.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.30.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.30.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.30.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.30.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.30.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.30.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.30.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.30.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.30.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.31.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.31.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.31.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.31.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.31.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.31.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.31.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.31.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.31.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.31.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.31.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.31.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.32.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.32.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.32.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.32.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.32.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.32.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.32.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.32.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.32.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.32.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.32.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.32.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.33.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.33.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.33.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.33.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.33.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.33.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.33.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.33.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.33.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.33.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.33.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.33.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.34.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.34.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.34.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.34.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.34.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.34.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.34.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.34.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.34.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.34.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.34.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.34.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.35.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.35.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.35.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.35.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.35.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.35.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.35.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.35.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.35.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.35.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.35.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.35.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.36.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.36.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.36.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.36.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.36.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.36.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.36.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.36.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.36.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.36.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.36.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.36.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.37.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.37.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.37.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.37.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.37.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.37.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.37.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.37.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.37.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.37.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.37.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.37.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.4.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.4.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.4.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.4.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.4.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.4.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.4.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.4.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.4.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.4.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.4.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.4.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.5.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.5.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.5.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.5.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.5.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.5.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.5.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.5.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.5.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.5.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.5.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.5.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.6.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.6.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.6.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.6.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.6.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.6.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.6.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.6.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.6.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.6.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.6.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.6.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.7.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.7.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.7.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.7.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.7.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.7.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.7.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.7.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.7.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.7.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.7.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.7.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.8.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.8.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.8.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.8.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.8.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.8.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.8.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.8.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.8.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.8.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.8.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.8.proj_out.lora_down.weight\n", + "transformer.single_transformer_blocks.9.attn.to_k.lora_up.weight\n", + "transformer.single_transformer_blocks.9.attn.to_k.lora_down.weight\n", + "transformer.single_transformer_blocks.9.attn.to_q.lora_up.weight\n", + "transformer.single_transformer_blocks.9.attn.to_q.lora_down.weight\n", + "transformer.single_transformer_blocks.9.attn.to_v.lora_up.weight\n", + "transformer.single_transformer_blocks.9.attn.to_v.lora_down.weight\n", + "transformer.single_transformer_blocks.9.norm.linear.lora_up.weight\n", + "transformer.single_transformer_blocks.9.norm.linear.lora_down.weight\n", + "transformer.single_transformer_blocks.9.proj_mlp.lora_up.weight\n", + "transformer.single_transformer_blocks.9.proj_mlp.lora_down.weight\n", + "transformer.single_transformer_blocks.9.proj_out.lora_up.weight\n", + "transformer.single_transformer_blocks.9.proj_out.lora_down.weight\n", + "transformer.transformer_blocks.0.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.0.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.0.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.0.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.0.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.0.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.0.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.0.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.0.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.0.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.0.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.0.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.0.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.0.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.0.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.0.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.0.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.0.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.0.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.0.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.0.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.0.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.0.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.0.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.0.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.0.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.0.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.0.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.1.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.1.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.1.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.1.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.1.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.1.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.1.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.1.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.1.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.1.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.1.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.1.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.1.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.1.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.1.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.1.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.1.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.1.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.1.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.1.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.1.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.1.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.1.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.1.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.1.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.1.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.1.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.1.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.10.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.10.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.10.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.10.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.10.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.10.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.10.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.10.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.10.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.10.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.10.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.10.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.10.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.10.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.10.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.10.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.10.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.10.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.10.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.10.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.10.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.10.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.10.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.10.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.10.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.10.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.10.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.10.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.11.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.11.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.11.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.11.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.11.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.11.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.11.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.11.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.11.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.11.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.11.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.11.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.11.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.11.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.11.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.11.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.11.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.11.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.11.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.11.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.11.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.11.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.11.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.11.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.11.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.11.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.11.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.11.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.12.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.12.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.12.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.12.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.12.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.12.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.12.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.12.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.12.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.12.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.12.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.12.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.12.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.12.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.12.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.12.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.12.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.12.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.12.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.12.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.12.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.12.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.12.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.12.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.12.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.12.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.12.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.12.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.13.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.13.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.13.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.13.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.13.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.13.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.13.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.13.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.13.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.13.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.13.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.13.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.13.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.13.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.13.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.13.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.13.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.13.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.13.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.13.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.13.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.13.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.13.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.13.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.13.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.13.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.13.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.13.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.14.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.14.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.14.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.14.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.14.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.14.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.14.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.14.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.14.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.14.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.14.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.14.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.14.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.14.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.14.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.14.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.14.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.14.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.14.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.14.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.14.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.14.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.14.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.14.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.14.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.14.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.14.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.14.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.15.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.15.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.15.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.15.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.15.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.15.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.15.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.15.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.15.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.15.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.15.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.15.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.15.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.15.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.15.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.15.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.15.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.15.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.15.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.15.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.15.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.15.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.15.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.15.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.15.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.15.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.15.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.15.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.16.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.16.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.16.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.16.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.16.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.16.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.16.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.16.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.16.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.16.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.16.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.16.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.16.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.16.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.16.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.16.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.16.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.16.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.16.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.16.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.16.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.16.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.16.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.16.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.16.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.16.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.16.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.16.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.17.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.17.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.17.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.17.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.17.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.17.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.17.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.17.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.17.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.17.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.17.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.17.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.17.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.17.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.17.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.17.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.17.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.17.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.17.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.17.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.17.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.17.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.17.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.17.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.17.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.17.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.17.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.17.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.18.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.18.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.18.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.18.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.18.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.18.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.18.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.18.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.18.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.18.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.18.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.18.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.18.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.18.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.18.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.18.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.18.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.18.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.18.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.18.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.18.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.18.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.18.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.18.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.18.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.18.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.18.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.18.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.2.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.2.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.2.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.2.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.2.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.2.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.2.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.2.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.2.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.2.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.2.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.2.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.2.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.2.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.2.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.2.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.2.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.2.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.2.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.2.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.2.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.2.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.2.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.2.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.2.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.2.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.2.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.2.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.3.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.3.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.3.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.3.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.3.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.3.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.3.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.3.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.3.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.3.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.3.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.3.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.3.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.3.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.3.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.3.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.3.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.3.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.3.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.3.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.3.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.3.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.3.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.3.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.3.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.3.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.3.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.3.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.4.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.4.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.4.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.4.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.4.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.4.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.4.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.4.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.4.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.4.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.4.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.4.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.4.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.4.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.4.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.4.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.4.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.4.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.4.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.4.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.4.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.4.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.4.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.4.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.4.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.4.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.4.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.4.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.5.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.5.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.5.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.5.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.5.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.5.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.5.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.5.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.5.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.5.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.5.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.5.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.5.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.5.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.5.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.5.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.5.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.5.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.5.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.5.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.5.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.5.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.5.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.5.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.5.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.5.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.5.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.5.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.6.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.6.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.6.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.6.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.6.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.6.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.6.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.6.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.6.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.6.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.6.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.6.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.6.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.6.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.6.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.6.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.6.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.6.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.6.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.6.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.6.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.6.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.6.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.6.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.6.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.6.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.6.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.6.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.7.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.7.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.7.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.7.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.7.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.7.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.7.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.7.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.7.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.7.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.7.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.7.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.7.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.7.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.7.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.7.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.7.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.7.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.7.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.7.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.7.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.7.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.7.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.7.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.7.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.7.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.7.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.7.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.8.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.8.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.8.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.8.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.8.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.8.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.8.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.8.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.8.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.8.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.8.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.8.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.8.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.8.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.8.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.8.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.8.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.8.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.8.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.8.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.8.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.8.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.8.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.8.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.8.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.8.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.8.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.8.norm1_context.linear.lora_down.weight\n", + "transformer.transformer_blocks.9.attn.add_k_proj.lora_up.weight\n", + "transformer.transformer_blocks.9.attn.add_k_proj.lora_down.weight\n", + "transformer.transformer_blocks.9.attn.add_q_proj.lora_up.weight\n", + "transformer.transformer_blocks.9.attn.add_q_proj.lora_down.weight\n", + "transformer.transformer_blocks.9.attn.add_v_proj.lora_up.weight\n", + "transformer.transformer_blocks.9.attn.add_v_proj.lora_down.weight\n", + "transformer.transformer_blocks.9.attn.to_add_out.lora_up.weight\n", + "transformer.transformer_blocks.9.attn.to_add_out.lora_down.weight\n", + "transformer.transformer_blocks.9.attn.to_k.lora_up.weight\n", + "transformer.transformer_blocks.9.attn.to_k.lora_down.weight\n", + "transformer.transformer_blocks.9.attn.to_out.0.lora_up.weight\n", + "transformer.transformer_blocks.9.attn.to_out.0.lora_down.weight\n", + "transformer.transformer_blocks.9.attn.to_q.lora_up.weight\n", + "transformer.transformer_blocks.9.attn.to_q.lora_down.weight\n", + "transformer.transformer_blocks.9.attn.to_v.lora_up.weight\n", + "transformer.transformer_blocks.9.attn.to_v.lora_down.weight\n", + "transformer.transformer_blocks.9.ff.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.9.ff.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.9.ff.net.2.lora_up.weight\n", + "transformer.transformer_blocks.9.ff.net.2.lora_down.weight\n", + "transformer.transformer_blocks.9.ff_context.net.0.proj.lora_up.weight\n", + "transformer.transformer_blocks.9.ff_context.net.0.proj.lora_down.weight\n", + "transformer.transformer_blocks.9.ff_context.net.2.lora_up.weight\n", + "transformer.transformer_blocks.9.ff_context.net.2.lora_down.weight\n", + "transformer.transformer_blocks.9.norm1.linear.lora_up.weight\n", + "transformer.transformer_blocks.9.norm1.linear.lora_down.weight\n", + "transformer.transformer_blocks.9.norm1_context.linear.lora_up.weight\n", + "transformer.transformer_blocks.9.norm1_context.linear.lora_down.weight\n", + "---------STAR---------\n" + ] + } + ] + }, + { + "cell_type": "code", + "source": [ + "down = 'lora_unet_double_blocks_0_img_attn_qkv.lora_down.weight'\n", + "up = 'lora_unet_double_blocks_0_img_attn_qkv.lora_up.weight'\n", + "tgt = star\n", + "print(\"STAR\")\n", + "print(tgt[f'{up}'].shape)\n", + "#print(torch.matmul(tgt[f'{up}'],tgt[f'{down}']).shape)\n", + "\n", + "down = 'transformer.transformer_blocks.0.attn.to_k.lora_down.weight'\n", + "up = 'transformer.transformer_blocks.0.attn.to_k.lora_up.weight'\n", + "tgt = oily\n", + "print(\"VS. OILY\")\n", + "print(tgt[f'{up}'].shape)\n", + "#print(torch.matmul(tgt[f'{up}'],tgt[f'{down}']).shape)\n" + ], + "metadata": { + "id": "GoDfgENYaWD7", + "outputId": "9336ae1a-6244-4e76-f291-82cda4482831", + "colab": { + "base_uri": "https://localhost:8080/" + } + }, + "execution_count": 17, + "outputs": [ + { + "output_type": "stream", + "name": "stdout", + "text": [ + "STAR\n", + "torch.Size([9216, 32])\n", + "VS. OILY\n", + "torch.Size([3072, 32])\n" + ] + } + ] + }, + { + "cell_type": "code", + "source": [ + "for key in oily:\n", + " print(oily[f'{key}'].shape)" + ], + "metadata": { + "id": "xQhVLouEfmGE", + "outputId": "662176b3-480d-48eb-f5db-97ec71b5e970", + "colab": { + "base_uri": "https://localhost:8080/" + } + }, + "execution_count": 18, + "outputs": [ + { + "output_type": "stream", + "name": "stdout", + "text": [ + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([9216, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 15360])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([12288, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([3072, 32])\n", + "torch.Size([32, 12288])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n", + "torch.Size([18432, 32])\n", + "torch.Size([32, 3072])\n" + ] + } + ] + }, + { + "cell_type": "code", + "source": [ + "\n", + "import torch\n", + "from safetensors.torch import load_file, save_file\n", + "import torch.nn as nn\n", + "from torch import linalg as LA\n", + "import os\n", + "import math\n", + "import random\n", + "import numpy as np\n", + "device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\n", + "def _filter(tgt , pcnt, largest):\n", + " num_topk = math.floor(tgt.numel()*(pcnt/100))\n", + " y = tgt.flatten().to(device = device , dtype=torch.float32)\n", + " values,indices = torch.topk( y , num_topk , largest=largest)\n", + " _values,_indices = torch.topk( -y , num_topk , largest=largest)\n", + " y = y*0\n", + " y[indices] = 1\n", + " y[_indices] = 1\n", + " y = y.unflatten(0,tgt.shape).to(device = device , dtype=torch.float32)\n", + " return torch.mul(tgt,y)\n", + "\n", + "#----#\n", + "\n", + "# For pcnt = 30 , 'filter_and_save' will keep all top 30 % values\n", + "#, and the lowest (negative) 30% values for each layer delta_W in this lora\n", + "# Then save the new filtered lora as a .safetensor file\n", + "def filter_and_save(_lora , savefile_name, new_rank , new_alpha, thresh):\n", + " lora = {}\n", + " count = 0\n", + " for key in _lora:count = count + 1\n", + " NUM_ITEMS = count\n", + " count = 0\n", + " thresh = resolution*0.000001 # 1e-6\n", + " #-------#\n", + " for key in _lora:\n", + " if f'{key}'.find('alpha') > -1:\n", + " lora[f'{key}'] = torch.tensor(new_alpha).to(device = device , dtype = torch.float32)\n", + " count = count + 1\n", + " print(f'{count} / {NUM_ITEMS}')\n", + " continue\n", + " #------#\n", + " if not f'{key}'.find('lora_down') > -1: continue\n", + " up = f'{key}'.replace('lora_down' , 'lora_up')\n", + " down = f'{key}'\n", + " #-------#\n", + " delta_W = torch.matmul(_lora[up],_lora[down]).to(device = device , dtype=torch.float32)\n", + " #---#\n", + " N = delta_W.numel()\n", + " y = delta_W.flatten().to(device = device , dtype=torch.float32)\n", + " values,indices = torch.sort(y, descending = False) # smallest -> largest elements\n", + " y = torch.zeros(y.shape).to(device = device , dtype=torch.float32)\n", + " y[indices[values>thresh]] = 1\n", + " y[indices[values<-thresh]] = 1\n", + " y = y.unflatten(0,delta_W.shape).to(device = device , dtype=torch.float32)\n", + " delta_W = torch.mul(delta_W,y).to(device = device , dtype=torch.float32)\n", + " #------#\n", + " tmp={}\n", + " tmp['u'], tmp['s'], tmp['Vh'] = torch.svd(delta_W)\n", + " tmp['u'] = tmp['u'][:,: new_rank]\n", + " tmp['s'] = tmp['s'][: new_rank]\n", + " #-------#\n", + " tmp['u'] = torch.round(torch.matmul(tmp['u'], torch.diag(tmp['s'])),decimals=6)\n", + " tmp['Vh'] = torch.round(tmp['Vh'].t()[: new_rank,:],decimals=6)\n", + " #-------#\n", + " for key in tmp:tmp[f'{key}'] = tmp[f'{key}'].contiguous()\n", + " lora[up] = tmp['u'].to(device = device , dtype=torch.float32)\n", + " lora[down] = tmp['Vh'].to(device = device , dtype=torch.float32)\n", + " #-------#\n", + " count = count + 2\n", + " print(f'{count} / {NUM_ITEMS}')\n", + " #-------#\n", + " print(f'done!')\n", + " print(f'casting params to fp16....')\n", + " for key in _lora: lora[f'{key}'] = lora[f'{key}'].to(device = device , dtype=torch.float16)\n", + " #-------#\n", + " print(f'done!')\n", + " print(f'saving {savefile_name}...')\n", + " save_file(lora , f'{savefile_name}')\n", + "#--------#\n", + "\n", + "\n", + "new_rank = 32\n", + "new_alpha = new_rank/2\n", + "resolution = 100\n", + "star = load_file('/kaggle/input/flux-loras/yeero.safetensors')\n", + "for key in star:\n", + " star[f'{key}'] = star[f'{key}'].to(device = device , dtype = torch.float32)\n", + "\n", + "filter_and_save(star , f'yeero_{resolution}_r{new_rank}_{new_alpha}alpha.safetensors' , new_rank , new_alpha, resolution)\n", + "\n", + "#pcnt = 30\n", + "#new_rank = 6\n", + "#filter_and_save(yeero , f'yeero_topk{pcnt}_r{new_rank}.safetensors' , pcnt , new_rank)\n", + "#filter_and_save(euro , f'euro_topk{pcnt}_r{new_rank}.safetensors' , pcnt , new_rank)\n", + "#filter_and_save(star , f'star_topk{pcnt}_r{new_rank}.safetensors' , pcnt , new_rank)\n" + ], + "metadata": { + "id": "f46xbSVkUlDl" + }, + "execution_count": null, + "outputs": [] + }, { "cell_type": "code", "source": [ @@ -78,7 +2347,7 @@ "metadata": { "id": "JuGDCX5272Bh" }, - "execution_count": 10, + "execution_count": null, "outputs": [] }, { @@ -95,7 +2364,7 @@ "metadata": { "id": "FftDdBRG7su6" }, - "execution_count": 57, + "execution_count": null, "outputs": [] }, { @@ -109,7 +2378,7 @@ "metadata": { "id": "RII9SEqh8KH2" }, - "execution_count": 60, + "execution_count": null, "outputs": [] }, { @@ -157,7 +2426,7 @@ "metadata": { "id": "hJL6QEclHdHn" }, - "execution_count": 104, + "execution_count": null, "outputs": [] }, { @@ -174,7 +2443,7 @@ "id": "ckyBSQi5Ll4F", "outputId": "341f7192-083d-4423-f61f-4f49d5756e79" }, - "execution_count": 106, + "execution_count": null, "outputs": [ { "output_type": "stream", @@ -199,7 +2468,7 @@ "id": "XLwslN61hiIJ", "outputId": "9e3cbba6-3727-4772-f453-fecf8a408790" }, - "execution_count": 16, + "execution_count": null, "outputs": [ { "output_type": "execute_result", @@ -225,7 +2494,7 @@ "id": "AKwh0lZ1f8dJ", "outputId": "59186526-bd73-4efe-925a-3e7a9c738e53" }, - "execution_count": 13, + "execution_count": null, "outputs": [ { "output_type": "execute_result", @@ -264,7 +2533,7 @@ "id": "SNCvvkb2h3Zb", "outputId": "725fabd1-3fe2-4ac2-f24c-5f9309d45e4a" }, - "execution_count": 37, + "execution_count": null, "outputs": [ { "output_type": "execute_result", @@ -391,7 +2660,7 @@ "id": "9L_g5Zp9Du2E", "outputId": "a3aa2bde-061e-43f5-ca35-96bdc470be80" }, - "execution_count": 33, + "execution_count": null, "outputs": [ { "output_type": "stream", @@ -535,7 +2804,7 @@ "metadata": { "id": "k7Pq-kDbuNnQ" }, - "execution_count": 64, + "execution_count": null, "outputs": [] }, { @@ -550,7 +2819,7 @@ "id": "ANBPfP7tuOoa", "outputId": "24300487-f874-4f1b-beb7-0f441ec7df4a" }, - "execution_count": 65, + "execution_count": null, "outputs": [ { "output_type": "execute_result", @@ -576,7 +2845,7 @@ "id": "zN92j8JJuQ6G", "outputId": "b810f4e6-a8f3-426a-ae52-ffbd44fb3f00" }, - "execution_count": 66, + "execution_count": null, "outputs": [ { "output_type": "execute_result", @@ -597,8 +2866,7 @@ "\n", "\n", "\n", - "\n", - "" + "\n" ], "metadata": { "colab": { @@ -607,7 +2875,7 @@ "id": "py-JMJzhsAI4", "outputId": "207cd809-031c-48e3-af0a-98bc114d910e" }, - "execution_count": 85, + "execution_count": null, "outputs": [ { "output_type": "stream", @@ -648,7 +2916,7 @@ "metadata": { "id": "DHdy4DptowYG" }, - "execution_count": 47, + "execution_count": null, "outputs": [] }, { @@ -663,7 +2931,7 @@ "id": "WeNJ0bquphtx", "outputId": "442bfb2e-c1ab-4549-a4ea-ca80d3cc9a7d" }, - "execution_count": 46, + "execution_count": null, "outputs": [ { "output_type": "execute_result", @@ -689,7 +2957,7 @@ "id": "xqZp3Xo8pQuW", "outputId": "68e5c25e-3391-45e7-9c73-45e0174ddbc1" }, - "execution_count": 48, + "execution_count": null, "outputs": [ { "output_type": "execute_result", @@ -722,7 +2990,7 @@ "id": "VFNw0Nck8V6Q", "outputId": "e48bab98-18f7-43bb-d1cf-89f3e00f7ccf" }, - "execution_count": 39, + "execution_count": null, "outputs": [ { "output_type": "stream",