Spaces:
				
			
			
	
			
			
		Runtime error
		
	
	
	
			
			
	
	
	
	
		
		
		Runtime error
		
	FEAT: New version of model trained as sequence classification
Browse filesThis view is limited to 50 files because it contains too many changes. Β 
							See raw diff
- app.py +18 -14
 - data/example_data.py +0 -0
 - inference_tokenizer.py +0 -1
 - model/3c090bb39725194fae09a603caac8c3d9014df49/awscliv2.zip +3 -0
 - model/3c090bb39725194fae09a603caac8c3d9014df49/onnx/model.onnx +3 -0
 - model/{6a62f122a90e090b285f0344a1d79e753f2000bb β 4a70ad1033ceec48447d5319d0863d442e976823}/config.json +0 -0
 - model/4a70ad1033ceec48447d5319d0863d442e976823/info.json +4 -0
 - model/4a70ad1033ceec48447d5319d0863d442e976823/meta-info.json +56 -0
 - model/{6a62f122a90e090b285f0344a1d79e753f2000bb β 4a70ad1033ceec48447d5319d0863d442e976823}/pytorch_model.bin +1 -1
 - model/{e09d71f55f4b6fc20135f856bf029322a3265d8d β 4a70ad1033ceec48447d5319d0863d442e976823}/special_tokens_map.json +3 -0
 - model/4a70ad1033ceec48447d5319d0863d442e976823/tokenizer_config.json +15 -0
 - model/{6a62f122a90e090b285f0344a1d79e753f2000bb β 4a70ad1033ceec48447d5319d0863d442e976823}/training_args.bin +1 -1
 - model/{6a62f122a90e090b285f0344a1d79e753f2000bb β 4a70ad1033ceec48447d5319d0863d442e976823}/vocab.txt +0 -0
 - model/6a62f122a90e090b285f0344a1d79e753f2000bb/info.json +0 -4
 - model/6a62f122a90e090b285f0344a1d79e753f2000bb/meta-info.json +0 -1
 - model/6a62f122a90e090b285f0344a1d79e753f2000bb/special_tokens_map.json +0 -1
 - model/6a62f122a90e090b285f0344a1d79e753f2000bb/tokenizer_config.json +0 -1
 - model/berttokenizer.zip +3 -0
 - model/berttokenizer/special_tokens_map.json +8 -0
 - model/{e09d71f55f4b6fc20135f856bf029322a3265d8d β berttokenizer}/tokenizer.json +10 -10
 - model/{c3c3bdb7ad80396e69de171995e2038f900940c8 β berttokenizer}/tokenizer_config.json +1 -0
 - model/{c3c3bdb7ad80396e69de171995e2038f900940c8 β berttokenizer}/vocab.txt +0 -0
 - model/c3c3bdb7ad80396e69de171995e2038f900940c8/info.json +0 -4
 - model/c3c3bdb7ad80396e69de171995e2038f900940c8/meta-info.json +0 -1
 - model/d1dd8365cbf16ff423f537e2291c61a91c717ed1/onnx/model.onnx +3 -0
 - model/e09d71f55f4b6fc20135f856bf029322a3265d8d/info.json +0 -4
 - model/e09d71f55f4b6fc20135f856bf029322a3265d8d/meta-info.json +0 -1
 - model/e09d71f55f4b6fc20135f856bf029322a3265d8d/optimizer.pt +0 -3
 - model/f1f881389fb38108e623689999ceaaaf398c5e92/info.json +0 -4
 - model/f1f881389fb38108e623689999ceaaaf398c5e92/meta-info.json +0 -1
 - model/f1f881389fb38108e623689999ceaaaf398c5e92/special_tokens_map.json +0 -1
 - model/f1f881389fb38108e623689999ceaaaf398c5e92/tokenizer_config.json +0 -1
 - model/{e09d71f55f4b6fc20135f856bf029322a3265d8d β new_version_dummy}/config.json +3 -2
 - model/new_version_dummy/onnx/model.onnx +3 -0
 - model/{e09d71f55f4b6fc20135f856bf029322a3265d8d β new_version_dummy}/pytorch_model.bin +2 -2
 - model/{c3c3bdb7ad80396e69de171995e2038f900940c8 β new_version_dummy}/special_tokens_map.json +3 -0
 - model/{c3c3bdb7ad80396e69de171995e2038f900940c8 β new_version_dummy}/tokenizer.json +9 -0
 - model/{e09d71f55f4b6fc20135f856bf029322a3265d8d β new_version_dummy}/tokenizer_config.json +1 -1
 - model/{f1f881389fb38108e623689999ceaaaf398c5e92 β new_version_dummy}/training_args.bin +2 -2
 - model/{e09d71f55f4b6fc20135f856bf029322a3265d8d β new_version_dummy}/vocab.txt +0 -0
 - model/{f1f881389fb38108e623689999ceaaaf398c5e92 β new_version_v1}/config.json +2 -2
 - model/new_version_v1/onnx/model.onnx +3 -0
 - model/{c3c3bdb7ad80396e69de171995e2038f900940c8 β new_version_v1}/pytorch_model.bin +2 -2
 - model/new_version_v1/special_tokens_map.json +10 -0
 - model/new_version_v1/tokenizer_config.json +15 -0
 - model/{c3c3bdb7ad80396e69de171995e2038f900940c8 β new_version_v1}/training_args.bin +2 -2
 - model/{f1f881389fb38108e623689999ceaaaf398c5e92 β new_version_v1}/vocab.txt +0 -0
 - model/{c3c3bdb7ad80396e69de171995e2038f900940c8 β new_version_v2}/config.json +4 -3
 - model/new_version_v2/info.json +4 -0
 - model/new_version_v2/meta-info.json +28 -0
 
    	
        app.py
    CHANGED
    
    | 
         @@ -22,7 +22,7 @@ def get_model(_model_path): 
     | 
|
| 22 | 
         
             
                        _model_package = meta_info["kwargs"].get("model_package", "transformers")
         
     | 
| 23 | 
         
             
                        _model_class = meta_info["kwargs"].get("model_class", "BertForNextSentencePrediction")
         
     | 
| 24 | 
         
             
                else:
         
     | 
| 25 | 
         
            -
                    raise FileNotFoundError(" 
     | 
| 26 | 
         | 
| 27 | 
         
             
                model_class = get_class(_model_package, _model_class)
         
     | 
| 28 | 
         
             
                _model = model_class.from_pretrained(_model_path)
         
     | 
| 
         @@ -32,8 +32,8 @@ def get_model(_model_path): 
     | 
|
| 32 | 
         | 
| 33 | 
         
             
            def get_tokenizer(tokenizer_path):
         
     | 
| 34 | 
         
             
                print(f"Getting tokenizer at {tokenizer_path}")
         
     | 
| 35 | 
         
            -
                from transformers import  
     | 
| 36 | 
         
            -
                tokenizer =  
     | 
| 37 | 
         
             
                if os.path.isfile(os.path.join(tokenizer_path, "meta-info.json")):
         
     | 
| 38 | 
         
             
                    with open(os.path.join(tokenizer_path, "meta-info.json"), "r") as f:
         
     | 
| 39 | 
         
             
                        meta_info = json.load(f)
         
     | 
| 
         @@ -44,8 +44,7 @@ def get_tokenizer(tokenizer_path): 
     | 
|
| 44 | 
         | 
| 45 | 
         
             
                if special_token != " ":
         
     | 
| 46 | 
         
             
                    tokenizer.add_special_tokens({"additional_special_tokens": [special_token]})
         
     | 
| 47 | 
         
            -
             
     | 
| 48 | 
         
            -
                # print(tokenizer_args)
         
     | 
| 49 | 
         
             
                _inference_tokenizer = NextSentencePredictionTokenizer(tokenizer, **tokenizer_args)
         
     | 
| 50 | 
         
             
                return _inference_tokenizer
         
     | 
| 51 | 
         | 
| 
         @@ -108,7 +107,10 @@ option = st.selectbox("Choose type of input:", 
     | 
|
| 108 | 
         
             
                                   "02 - JSON (aggregated)",
         
     | 
| 109 | 
         
             
                                   "03 - JSON (example CA-OOD)",
         
     | 
| 110 | 
         
             
                                   "04 - JSON (example Elysai)",
         
     | 
| 111 | 
         
            -
                                   "05 - Diagnostic mode" 
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 112 | 
         | 
| 113 | 
         
             
            with st.form("input_text"):
         
     | 
| 114 | 
         
             
                if "01" in option:
         
     | 
| 
         @@ -129,10 +131,9 @@ with st.form("input_text"): 
     | 
|
| 129 | 
         
             
                        ax.pie([prop_follow, prop_not_follow], labels=["Probability - Follow", "Probability - Not Follow"],
         
     | 
| 130 | 
         
             
                               autopct='%1.1f%%')
         
     | 
| 131 | 
         
             
                        st.pyplot(fig)
         
     | 
| 132 | 
         
            -
                elif "02" in option or "03" in option or "04" in option:
         
     | 
| 133 | 
         
            -
                    from data.example_data import ca_ood, elysai
         
     | 
| 134 | 
         | 
| 135 | 
         
            -
                    choices = [ca_ood, elysai]
         
     | 
| 136 | 
         
             
                    option: str
         
     | 
| 137 | 
         
             
                    # > Python 3.10
         
     | 
| 138 | 
         
             
                    # match option.split("-")[0].strip():
         
     | 
| 
         @@ -143,12 +144,14 @@ with st.form("input_text"): 
     | 
|
| 143 | 
         
             
                    #     case _:
         
     | 
| 144 | 
         
             
                    #         text = ""
         
     | 
| 145 | 
         
             
                    option = option.split("-")[0].strip()
         
     | 
| 
         | 
|
| 146 | 
         
             
                    if option == "03":
         
     | 
| 147 | 
         
            -
                        text = json.dumps( 
     | 
| 148 | 
         
             
                    elif option == "04":
         
     | 
| 149 | 
         
            -
                        text = json.dumps( 
     | 
| 150 | 
         
            -
                     
     | 
| 151 | 
         
            -
                         
     | 
| 
         | 
|
| 152 | 
         
             
                    context = st.text_area("Insert JSON here:", value=str(text))
         
     | 
| 153 | 
         | 
| 154 | 
         
             
                    if "{" in context:
         
     | 
| 
         @@ -158,7 +161,8 @@ with st.form("input_text"): 
     | 
|
| 158 | 
         | 
| 159 | 
         
             
                    submitted = st.form_submit_button("Submit")
         
     | 
| 160 | 
         
             
                    if submitted:
         
     | 
| 161 | 
         
            -
                        for datapoint in data_for_evaluation:
         
     | 
| 
         | 
|
| 162 | 
         
             
                            c, s, human_label = datapoint
         
     | 
| 163 | 
         
             
                            input_tensor = inference_tokenizer.get_item(context=c, actual_sentence=s)
         
     | 
| 164 | 
         
             
                            output_model = model(**input_tensor.data).logits
         
     | 
| 
         | 
|
| 22 | 
         
             
                        _model_package = meta_info["kwargs"].get("model_package", "transformers")
         
     | 
| 23 | 
         
             
                        _model_class = meta_info["kwargs"].get("model_class", "BertForNextSentencePrediction")
         
     | 
| 24 | 
         
             
                else:
         
     | 
| 25 | 
         
            +
                    raise FileNotFoundError("Model is provided without meta-info.json. Cannot interfere proper configuration!")
         
     | 
| 26 | 
         | 
| 27 | 
         
             
                model_class = get_class(_model_package, _model_class)
         
     | 
| 28 | 
         
             
                _model = model_class.from_pretrained(_model_path)
         
     | 
| 
         | 
|
| 32 | 
         | 
| 33 | 
         
             
            def get_tokenizer(tokenizer_path):
         
     | 
| 34 | 
         
             
                print(f"Getting tokenizer at {tokenizer_path}")
         
     | 
| 35 | 
         
            +
                from transformers import AutoTokenizer
         
     | 
| 36 | 
         
            +
                tokenizer = AutoTokenizer.from_pretrained(tokenizer_path)
         
     | 
| 37 | 
         
             
                if os.path.isfile(os.path.join(tokenizer_path, "meta-info.json")):
         
     | 
| 38 | 
         
             
                    with open(os.path.join(tokenizer_path, "meta-info.json"), "r") as f:
         
     | 
| 39 | 
         
             
                        meta_info = json.load(f)
         
     | 
| 
         | 
|
| 44 | 
         | 
| 45 | 
         
             
                if special_token != " ":
         
     | 
| 46 | 
         
             
                    tokenizer.add_special_tokens({"additional_special_tokens": [special_token]})
         
     | 
| 47 | 
         
            +
             
     | 
| 
         | 
|
| 48 | 
         
             
                _inference_tokenizer = NextSentencePredictionTokenizer(tokenizer, **tokenizer_args)
         
     | 
| 49 | 
         
             
                return _inference_tokenizer
         
     | 
| 50 | 
         | 
| 
         | 
|
| 107 | 
         
             
                                   "02 - JSON (aggregated)",
         
     | 
| 108 | 
         
             
                                   "03 - JSON (example CA-OOD)",
         
     | 
| 109 | 
         
             
                                   "04 - JSON (example Elysai)",
         
     | 
| 110 | 
         
            +
                                   "05 - Diagnostic mode",
         
     | 
| 111 | 
         
            +
                                   "06 - JSON (example Elysai - large)"])
         
     | 
| 112 | 
         
            +
             
     | 
| 113 | 
         
            +
            progres_bar = st.progress(0.0, text="Inference")
         
     | 
| 114 | 
         | 
| 115 | 
         
             
            with st.form("input_text"):
         
     | 
| 116 | 
         
             
                if "01" in option:
         
     | 
| 
         | 
|
| 131 | 
         
             
                        ax.pie([prop_follow, prop_not_follow], labels=["Probability - Follow", "Probability - Not Follow"],
         
     | 
| 132 | 
         
             
                               autopct='%1.1f%%')
         
     | 
| 133 | 
         
             
                        st.pyplot(fig)
         
     | 
| 134 | 
         
            +
                elif "02" in option or "03" in option or "04" in option or "06" in option:
         
     | 
| 135 | 
         
            +
                    from data.example_data import ca_ood, elysai, elysai_large
         
     | 
| 136 | 
         | 
| 
         | 
|
| 137 | 
         
             
                    option: str
         
     | 
| 138 | 
         
             
                    # > Python 3.10
         
     | 
| 139 | 
         
             
                    # match option.split("-")[0].strip():
         
     | 
| 
         | 
|
| 144 | 
         
             
                    #     case _:
         
     | 
| 145 | 
         
             
                    #         text = ""
         
     | 
| 146 | 
         
             
                    option = option.split("-")[0].strip()
         
     | 
| 147 | 
         
            +
                    text = ""
         
     | 
| 148 | 
         
             
                    if option == "03":
         
     | 
| 149 | 
         
            +
                        text = json.dumps(ca_ood)
         
     | 
| 150 | 
         
             
                    elif option == "04":
         
     | 
| 151 | 
         
            +
                        text = json.dumps(elysai)
         
     | 
| 152 | 
         
            +
                    elif option == "06":
         
     | 
| 153 | 
         
            +
                        text = json.dumps(elysai_large)
         
     | 
| 154 | 
         
            +
             
     | 
| 155 | 
         
             
                    context = st.text_area("Insert JSON here:", value=str(text))
         
     | 
| 156 | 
         | 
| 157 | 
         
             
                    if "{" in context:
         
     | 
| 
         | 
|
| 161 | 
         | 
| 162 | 
         
             
                    submitted = st.form_submit_button("Submit")
         
     | 
| 163 | 
         
             
                    if submitted:
         
     | 
| 164 | 
         
            +
                        for idx, datapoint in enumerate(data_for_evaluation):
         
     | 
| 165 | 
         
            +
                            progres_bar.progress(idx/len(data_for_evaluation), text="Inference")
         
     | 
| 166 | 
         
             
                            c, s, human_label = datapoint
         
     | 
| 167 | 
         
             
                            input_tensor = inference_tokenizer.get_item(context=c, actual_sentence=s)
         
     | 
| 168 | 
         
             
                            output_model = model(**input_tensor.data).logits
         
     | 
    	
        data/example_data.py
    CHANGED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        inference_tokenizer.py
    CHANGED
    
    | 
         @@ -1,7 +1,6 @@ 
     | 
|
| 1 | 
         
             
            import torch
         
     | 
| 2 | 
         
             
            from typing import Dict, List
         
     | 
| 3 | 
         | 
| 4 | 
         
            -
             
     | 
| 5 | 
         
             
            class NextSentencePredictionTokenizer:
         
     | 
| 6 | 
         | 
| 7 | 
         
             
                def __init__(self, _tokenizer, **_tokenizer_args):
         
     | 
| 
         | 
|
| 1 | 
         
             
            import torch
         
     | 
| 2 | 
         
             
            from typing import Dict, List
         
     | 
| 3 | 
         | 
| 
         | 
|
| 4 | 
         
             
            class NextSentencePredictionTokenizer:
         
     | 
| 5 | 
         | 
| 6 | 
         
             
                def __init__(self, _tokenizer, **_tokenizer_args):
         
     | 
    	
        model/3c090bb39725194fae09a603caac8c3d9014df49/awscliv2.zip
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:e4962990f24634814b6d4834a5c105a524c8895fca478a8fc17f7cc7e6191fa4
         
     | 
| 3 | 
         
            +
            size 57717779
         
     | 
    	
        model/3c090bb39725194fae09a603caac8c3d9014df49/onnx/model.onnx
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:a1e7543fd13846a928c1309459cd3a068dc82221a4d22ca647e054d658ea1063
         
     | 
| 3 | 
         
            +
            size 433513952
         
     | 
    	
        model/{6a62f122a90e090b285f0344a1d79e753f2000bb β 4a70ad1033ceec48447d5319d0863d442e976823}/config.json
    RENAMED
    
    | 
         
            File without changes
         
     | 
    	
        model/4a70ad1033ceec48447d5319d0863d442e976823/info.json
    ADDED
    
    | 
         @@ -0,0 +1,4 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "model": "BERT-NSP-v7",
         
     | 
| 3 | 
         
            +
              "description": "Model trained on full version of DailyDialogue and CommonDialogues + down=sampled version of SODA and AirDialogue. Using [unused1] token to divide sentences in context. Improved training arguments (warmup, smaller learning rate). Using frozen test set to better compare models and therefore trained longer time (about 60 epochs). The model also have bigger classification head (from one layer liner as classical). Added method for **Data Augmentation**. More info can be found at https://wandb.ai/alquist/next-sentence-prediction/runs/b40zgvoo/overview?workspace=user-petr-lorenc"
         
     | 
| 4 | 
         
            +
            }
         
     | 
    	
        model/4a70ad1033ceec48447d5319d0863d442e976823/meta-info.json
    ADDED
    
    | 
         @@ -0,0 +1,56 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "args": [],
         
     | 
| 3 | 
         
            +
              "kwargs": {
         
     | 
| 4 | 
         
            +
                "model_package": "models",
         
     | 
| 5 | 
         
            +
                "model_class": "OwnBertForNextSentencePrediction",
         
     | 
| 6 | 
         
            +
                "data_root": "/home/lorenpe2/project/data",
         
     | 
| 7 | 
         
            +
                "data_sources": [
         
     | 
| 8 | 
         
            +
                  [
         
     | 
| 9 | 
         
            +
                    "COMMON_DIALOGUES",
         
     | 
| 10 | 
         
            +
                    "common_dialogues/train.json",
         
     | 
| 11 | 
         
            +
                    "common_dialogues/valid_frozen.json",
         
     | 
| 12 | 
         
            +
                    "common_dialogues/test_frozen.json"
         
     | 
| 13 | 
         
            +
                  ],
         
     | 
| 14 | 
         
            +
                  [
         
     | 
| 15 | 
         
            +
                    "DAILY_DIALOGUES",
         
     | 
| 16 | 
         
            +
                    "daily_dialogues/dialogues_text.train.txt",
         
     | 
| 17 | 
         
            +
                    "daily_dialogues/dev_frozen.json",
         
     | 
| 18 | 
         
            +
                    "daily_dialogues/test_frozen.json"
         
     | 
| 19 | 
         
            +
                  ],
         
     | 
| 20 | 
         
            +
                  [
         
     | 
| 21 | 
         
            +
                    "DAILY_DIALOGUES",
         
     | 
| 22 | 
         
            +
                    "air_dialogue/subsampled_train.txt",
         
     | 
| 23 | 
         
            +
                    "air_dialogue/subsampled_validation_frozen.json",
         
     | 
| 24 | 
         
            +
                    "air_dialogue/subsampled_test_frozen.json"
         
     | 
| 25 | 
         
            +
                  ],
         
     | 
| 26 | 
         
            +
                  [
         
     | 
| 27 | 
         
            +
                    "DAILY_DIALOGUES",
         
     | 
| 28 | 
         
            +
                    "soda/subsampled_train.txt",
         
     | 
| 29 | 
         
            +
                    "soda/subsampled_validation_frozen.json",
         
     | 
| 30 | 
         
            +
                    "soda/subsampled_test_frozen.json"
         
     | 
| 31 | 
         
            +
                  ]
         
     | 
| 32 | 
         
            +
                ],
         
     | 
| 33 | 
         
            +
                "pretrained_model": "bert-base-uncased",
         
     | 
| 34 | 
         
            +
                "tokenizer": "bert-base-uncased",
         
     | 
| 35 | 
         
            +
                "approach": "UNIQUE_RANDOM_CONTEXT",
         
     | 
| 36 | 
         
            +
                "data_augmentation": [
         
     | 
| 37 | 
         
            +
                  "ADD_PARTLY_BROKEN_CONTEXT",
         
     | 
| 38 | 
         
            +
                  "ADD_SMALLER_CONTEXT"
         
     | 
| 39 | 
         
            +
                ],
         
     | 
| 40 | 
         
            +
                "special_token": "[unused1]",
         
     | 
| 41 | 
         
            +
                "learning_rate": 5e-07,
         
     | 
| 42 | 
         
            +
                "warmup_ratio": 0.1,
         
     | 
| 43 | 
         
            +
                "freeze_prefinetuning": true,
         
     | 
| 44 | 
         
            +
                "prefinenuting_epoch": 10,
         
     | 
| 45 | 
         
            +
                "finetuning_epochs": 75
         
     | 
| 46 | 
         
            +
              },
         
     | 
| 47 | 
         
            +
              "tokenizer_args": {
         
     | 
| 48 | 
         
            +
                "padding": "max_length",
         
     | 
| 49 | 
         
            +
                "max_length_ctx": 32,
         
     | 
| 50 | 
         
            +
                "max_length_res": 8,
         
     | 
| 51 | 
         
            +
                "truncation": "only_first",
         
     | 
| 52 | 
         
            +
                "return_tensors": "np",
         
     | 
| 53 | 
         
            +
                "is_split_into_words": true,
         
     | 
| 54 | 
         
            +
                "special_token": "[unused1]"
         
     | 
| 55 | 
         
            +
              }
         
     | 
| 56 | 
         
            +
            }
         
     | 
    	
        model/{6a62f122a90e090b285f0344a1d79e753f2000bb β 4a70ad1033ceec48447d5319d0863d442e976823}/pytorch_model.bin
    RENAMED
    
    | 
         @@ -1,3 +1,3 @@ 
     | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256: 
     | 
| 3 | 
         
             
            size 438871109
         
     | 
| 
         | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:8a0200d7532286b0e8aae550933d8a083274c7d8bcba41bd1ac989f9efb1bb1d
         
     | 
| 3 | 
         
             
            size 438871109
         
     | 
    	
        model/{e09d71f55f4b6fc20135f856bf029322a3265d8d β 4a70ad1033ceec48447d5319d0863d442e976823}/special_tokens_map.json
    RENAMED
    
    | 
         @@ -1,4 +1,7 @@ 
     | 
|
| 1 | 
         
             
            {
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 2 | 
         
             
              "cls_token": "[CLS]",
         
     | 
| 3 | 
         
             
              "mask_token": "[MASK]",
         
     | 
| 4 | 
         
             
              "pad_token": "[PAD]",
         
     | 
| 
         | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
            +
              "additional_special_tokens": [
         
     | 
| 3 | 
         
            +
                "[unused1]"
         
     | 
| 4 | 
         
            +
              ],
         
     | 
| 5 | 
         
             
              "cls_token": "[CLS]",
         
     | 
| 6 | 
         
             
              "mask_token": "[MASK]",
         
     | 
| 7 | 
         
             
              "pad_token": "[PAD]",
         
     | 
    	
        model/4a70ad1033ceec48447d5319d0863d442e976823/tokenizer_config.json
    ADDED
    
    | 
         @@ -0,0 +1,15 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "cls_token": "[CLS]",
         
     | 
| 3 | 
         
            +
              "do_basic_tokenize": true,
         
     | 
| 4 | 
         
            +
              "do_lower_case": true,
         
     | 
| 5 | 
         
            +
              "mask_token": "[MASK]",
         
     | 
| 6 | 
         
            +
              "model_max_length": 512,
         
     | 
| 7 | 
         
            +
              "never_split": null,
         
     | 
| 8 | 
         
            +
              "pad_token": "[PAD]",
         
     | 
| 9 | 
         
            +
              "sep_token": "[SEP]",
         
     | 
| 10 | 
         
            +
              "special_tokens_map_file": null,
         
     | 
| 11 | 
         
            +
              "strip_accents": null,
         
     | 
| 12 | 
         
            +
              "tokenize_chinese_chars": true,
         
     | 
| 13 | 
         
            +
              "tokenizer_class": "BertTokenizer",
         
     | 
| 14 | 
         
            +
              "unk_token": "[UNK]"
         
     | 
| 15 | 
         
            +
            }
         
     | 
    	
        model/{6a62f122a90e090b285f0344a1d79e753f2000bb β 4a70ad1033ceec48447d5319d0863d442e976823}/training_args.bin
    RENAMED
    
    | 
         @@ -1,3 +1,3 @@ 
     | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256: 
     | 
| 3 | 
         
             
            size 3195
         
     | 
| 
         | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:fb391844e5d1b871e851254cc8388e4803682dd147af1fdb5067fa64fbe530aa
         
     | 
| 3 | 
         
             
            size 3195
         
     | 
    	
        model/{6a62f122a90e090b285f0344a1d79e753f2000bb β 4a70ad1033ceec48447d5319d0863d442e976823}/vocab.txt
    RENAMED
    
    | 
         
            File without changes
         
     | 
    	
        model/6a62f122a90e090b285f0344a1d79e753f2000bb/info.json
    DELETED
    
    | 
         @@ -1,4 +0,0 @@ 
     | 
|
| 1 | 
         
            -
            {
         
     | 
| 2 | 
         
            -
              "model": "BERT-NSP-v6",
         
     | 
| 3 | 
         
            -
              "description": "Model trained on DailyDialogue and CommonDialogues. Using [unused1] token to divide sentences in context. Improved training arguments (warmup, smaller learning rate). Using frozen test set to better compare models and therefore trained longer time (about 60 epochs). The model also have bigger classification head (from one layer liner as classical). Added method for **Data Augmentation**. More info can be found at https://wandb.ai/alquist/next-sentence-prediction/runs/xvzhhw2r/overview?workspace=user-petr-lorenc"
         
     | 
| 4 | 
         
            -
            }
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
    	
        model/6a62f122a90e090b285f0344a1d79e753f2000bb/meta-info.json
    DELETED
    
    | 
         @@ -1 +0,0 @@ 
     | 
|
| 1 | 
         
            -
            {"args": [], "kwargs": {"model_package": "models", "model_class": "OwnBertForNextSentencePrediction", "data_root": "/home/lorenpe2/project/data", "data_sources": [["COMMON_DIALOGUES", "common_dialogues/train.json", "common_dialogues/valid_frozen.json", "common_dialogues/test_frozen.json"], ["DAILY_DIALOGUES", "daily_dialogues/dialogues_text.train.txt", "daily_dialogues/dev_frozen.json", "daily_dialogues/test_frozen.json"]], "pretrained_model": "bert-base-uncased", "tokenizer": "bert-base-uncased", "approach": "IGNORE_DUPLICITIES", "data_augmentation": ["ADD_PARTLY_BROKEN_CONTEXT"], "special_token": "[unused1]", "learning_rate": 5e-07, "warmup_ratio": 0.1, "freeze_prefinetuning": true, "prefinenuting_epoch": 10, "finetuning_epochs": 75}, "tokenizer_args": {"padding": "max_length", "max_length_ctx": 256, "max_length_res": 40, "truncation": "only_first", "return_tensors": "np", "is_split_into_words": true, "approach": "IGNORE_DUPLICITIES", "special_token": "[unused1]"}}
         
     | 
| 
         | 
|
| 
         | 
    	
        model/6a62f122a90e090b285f0344a1d79e753f2000bb/special_tokens_map.json
    DELETED
    
    | 
         @@ -1 +0,0 @@ 
     | 
|
| 1 | 
         
            -
            {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "additional_special_tokens": ["[unused1]"]}
         
     | 
| 
         | 
|
| 
         | 
    	
        model/6a62f122a90e090b285f0344a1d79e753f2000bb/tokenizer_config.json
    DELETED
    
    | 
         @@ -1 +0,0 @@ 
     | 
|
| 1 | 
         
            -
            {"do_lower_case": true, "do_basic_tokenize": true, "never_split": null, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "bert-base-uncased", "tokenizer_class": "BertTokenizer"}
         
     | 
| 
         | 
|
| 
         | 
    	
        model/berttokenizer.zip
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:1d253965be810ac94eed5cf080d45c958e282f0450d64dff77803b272e145d8e
         
     | 
| 3 | 
         
            +
            size 320047
         
     | 
    	
        model/berttokenizer/special_tokens_map.json
    ADDED
    
    | 
         @@ -0,0 +1,8 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "cls_token": "[CLS]",
         
     | 
| 3 | 
         
            +
              "mask_token": "[MASK]",
         
     | 
| 4 | 
         
            +
              "pad_token": "[PAD]",
         
     | 
| 5 | 
         
            +
              "sep_token": "[SEP]",
         
     | 
| 6 | 
         
            +
              "unk_token": "[UNK]",
         
     | 
| 7 | 
         
            +
              "additional_special_tokens": ["[unused1]", "[unused2]", "[unused3]", "[unused4]", "[unused5]", "[unused6]", "[unused7]", "[unused8]", "[unused9]", "[unused10]"]
         
     | 
| 8 | 
         
            +
            }
         
     | 
    	
        model/{e09d71f55f4b6fc20135f856bf029322a3265d8d β berttokenizer}/tokenizer.json
    RENAMED
    
    | 
         @@ -5,48 +5,48 @@ 
     | 
|
| 5 | 
         
             
              "added_tokens": [
         
     | 
| 6 | 
         
             
                {
         
     | 
| 7 | 
         
             
                  "id": 0,
         
     | 
| 
         | 
|
| 8 | 
         
             
                  "content": "[PAD]",
         
     | 
| 9 | 
         
             
                  "single_word": false,
         
     | 
| 10 | 
         
             
                  "lstrip": false,
         
     | 
| 11 | 
         
             
                  "rstrip": false,
         
     | 
| 12 | 
         
            -
                  "normalized": false 
     | 
| 13 | 
         
            -
                  "special": true
         
     | 
| 14 | 
         
             
                },
         
     | 
| 15 | 
         
             
                {
         
     | 
| 16 | 
         
             
                  "id": 100,
         
     | 
| 
         | 
|
| 17 | 
         
             
                  "content": "[UNK]",
         
     | 
| 18 | 
         
             
                  "single_word": false,
         
     | 
| 19 | 
         
             
                  "lstrip": false,
         
     | 
| 20 | 
         
             
                  "rstrip": false,
         
     | 
| 21 | 
         
            -
                  "normalized": false 
     | 
| 22 | 
         
            -
                  "special": true
         
     | 
| 23 | 
         
             
                },
         
     | 
| 24 | 
         
             
                {
         
     | 
| 25 | 
         
             
                  "id": 101,
         
     | 
| 
         | 
|
| 26 | 
         
             
                  "content": "[CLS]",
         
     | 
| 27 | 
         
             
                  "single_word": false,
         
     | 
| 28 | 
         
             
                  "lstrip": false,
         
     | 
| 29 | 
         
             
                  "rstrip": false,
         
     | 
| 30 | 
         
            -
                  "normalized": false 
     | 
| 31 | 
         
            -
                  "special": true
         
     | 
| 32 | 
         
             
                },
         
     | 
| 33 | 
         
             
                {
         
     | 
| 34 | 
         
             
                  "id": 102,
         
     | 
| 
         | 
|
| 35 | 
         
             
                  "content": "[SEP]",
         
     | 
| 36 | 
         
             
                  "single_word": false,
         
     | 
| 37 | 
         
             
                  "lstrip": false,
         
     | 
| 38 | 
         
             
                  "rstrip": false,
         
     | 
| 39 | 
         
            -
                  "normalized": false 
     | 
| 40 | 
         
            -
                  "special": true
         
     | 
| 41 | 
         
             
                },
         
     | 
| 42 | 
         
             
                {
         
     | 
| 43 | 
         
             
                  "id": 103,
         
     | 
| 
         | 
|
| 44 | 
         
             
                  "content": "[MASK]",
         
     | 
| 45 | 
         
             
                  "single_word": false,
         
     | 
| 46 | 
         
             
                  "lstrip": false,
         
     | 
| 47 | 
         
             
                  "rstrip": false,
         
     | 
| 48 | 
         
            -
                  "normalized": false 
     | 
| 49 | 
         
            -
                  "special": true
         
     | 
| 50 | 
         
             
                }
         
     | 
| 51 | 
         
             
              ],
         
     | 
| 52 | 
         
             
              "normalizer": {
         
     | 
| 
         | 
|
| 5 | 
         
             
              "added_tokens": [
         
     | 
| 6 | 
         
             
                {
         
     | 
| 7 | 
         
             
                  "id": 0,
         
     | 
| 8 | 
         
            +
                  "special": true,
         
     | 
| 9 | 
         
             
                  "content": "[PAD]",
         
     | 
| 10 | 
         
             
                  "single_word": false,
         
     | 
| 11 | 
         
             
                  "lstrip": false,
         
     | 
| 12 | 
         
             
                  "rstrip": false,
         
     | 
| 13 | 
         
            +
                  "normalized": false
         
     | 
| 
         | 
|
| 14 | 
         
             
                },
         
     | 
| 15 | 
         
             
                {
         
     | 
| 16 | 
         
             
                  "id": 100,
         
     | 
| 17 | 
         
            +
                  "special": true,
         
     | 
| 18 | 
         
             
                  "content": "[UNK]",
         
     | 
| 19 | 
         
             
                  "single_word": false,
         
     | 
| 20 | 
         
             
                  "lstrip": false,
         
     | 
| 21 | 
         
             
                  "rstrip": false,
         
     | 
| 22 | 
         
            +
                  "normalized": false
         
     | 
| 
         | 
|
| 23 | 
         
             
                },
         
     | 
| 24 | 
         
             
                {
         
     | 
| 25 | 
         
             
                  "id": 101,
         
     | 
| 26 | 
         
            +
                  "special": true,
         
     | 
| 27 | 
         
             
                  "content": "[CLS]",
         
     | 
| 28 | 
         
             
                  "single_word": false,
         
     | 
| 29 | 
         
             
                  "lstrip": false,
         
     | 
| 30 | 
         
             
                  "rstrip": false,
         
     | 
| 31 | 
         
            +
                  "normalized": false
         
     | 
| 
         | 
|
| 32 | 
         
             
                },
         
     | 
| 33 | 
         
             
                {
         
     | 
| 34 | 
         
             
                  "id": 102,
         
     | 
| 35 | 
         
            +
                  "special": true,
         
     | 
| 36 | 
         
             
                  "content": "[SEP]",
         
     | 
| 37 | 
         
             
                  "single_word": false,
         
     | 
| 38 | 
         
             
                  "lstrip": false,
         
     | 
| 39 | 
         
             
                  "rstrip": false,
         
     | 
| 40 | 
         
            +
                  "normalized": false
         
     | 
| 
         | 
|
| 41 | 
         
             
                },
         
     | 
| 42 | 
         
             
                {
         
     | 
| 43 | 
         
             
                  "id": 103,
         
     | 
| 44 | 
         
            +
                  "special": true,
         
     | 
| 45 | 
         
             
                  "content": "[MASK]",
         
     | 
| 46 | 
         
             
                  "single_word": false,
         
     | 
| 47 | 
         
             
                  "lstrip": false,
         
     | 
| 48 | 
         
             
                  "rstrip": false,
         
     | 
| 49 | 
         
            +
                  "normalized": false
         
     | 
| 
         | 
|
| 50 | 
         
             
                }
         
     | 
| 51 | 
         
             
              ],
         
     | 
| 52 | 
         
             
              "normalizer": {
         
     | 
    	
        model/{c3c3bdb7ad80396e69de171995e2038f900940c8 β berttokenizer}/tokenizer_config.json
    RENAMED
    
    | 
         @@ -3,6 +3,7 @@ 
     | 
|
| 3 | 
         
             
              "do_lower_case": true,
         
     | 
| 4 | 
         
             
              "mask_token": "[MASK]",
         
     | 
| 5 | 
         
             
              "model_max_length": 512,
         
     | 
| 
         | 
|
| 6 | 
         
             
              "pad_token": "[PAD]",
         
     | 
| 7 | 
         
             
              "sep_token": "[SEP]",
         
     | 
| 8 | 
         
             
              "special_tokens_map_file": null,
         
     | 
| 
         | 
|
| 3 | 
         
             
              "do_lower_case": true,
         
     | 
| 4 | 
         
             
              "mask_token": "[MASK]",
         
     | 
| 5 | 
         
             
              "model_max_length": 512,
         
     | 
| 6 | 
         
            +
              "name_or_path": "bert-base-uncased",
         
     | 
| 7 | 
         
             
              "pad_token": "[PAD]",
         
     | 
| 8 | 
         
             
              "sep_token": "[SEP]",
         
     | 
| 9 | 
         
             
              "special_tokens_map_file": null,
         
     | 
    	
        model/{c3c3bdb7ad80396e69de171995e2038f900940c8 β berttokenizer}/vocab.txt
    RENAMED
    
    | 
         
            File without changes
         
     | 
    	
        model/c3c3bdb7ad80396e69de171995e2038f900940c8/info.json
    DELETED
    
    | 
         @@ -1,4 +0,0 @@ 
     | 
|
| 1 | 
         
            -
            {
         
     | 
| 2 | 
         
            -
              "model": "BERT-NSP-v1",
         
     | 
| 3 | 
         
            -
              "description": "Model trained on DailyDialogue. Context is taken as is - no separation of individual turns. More info can be found at https://wandb.ai/alquist/next-sentence-prediction/runs/dm6ab7ma/logs?workspace=user-petr-lorenc"
         
     | 
| 4 | 
         
            -
            }
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
    	
        model/c3c3bdb7ad80396e69de171995e2038f900940c8/meta-info.json
    DELETED
    
    | 
         @@ -1 +0,0 @@ 
     | 
|
| 1 | 
         
            -
            {"args": [], "kwargs": {"data_root": "/home/lorenpe2/project/data", "data_sources": [["COMMON_DIALOGUES", "common_dialogues/train.json", "common_dialogues/valid.json", "common_dialogues/test.json"]], "pretrained_model": "bert-base-uncased", "tokenizer": "bert-base-uncased", "naive_approach": true, "special_token": " ", "learning_rate": 5e-05}, "tokenizer_args": {"padding": "max_length", "max_length_ctx": 256, "max_length_res": 64, "truncation": "only_first", "return_tensors": "np", "is_split_into_words": true, "naive_approach": true, "special_token": " "}}
         
     | 
| 
         | 
|
| 
         | 
    	
        model/d1dd8365cbf16ff423f537e2291c61a91c717ed1/onnx/model.onnx
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:a65ee19216a6cfe592c1b5d7b35dabe9182a60d9bf7dbe415a7ad491b6c64733
         
     | 
| 3 | 
         
            +
            size 438044019
         
     | 
    	
        model/e09d71f55f4b6fc20135f856bf029322a3265d8d/info.json
    DELETED
    
    | 
         @@ -1,4 +0,0 @@ 
     | 
|
| 1 | 
         
            -
            {
         
     | 
| 2 | 
         
            -
              "model": "BERT-NSP-v2",
         
     | 
| 3 | 
         
            -
              "description": "Model trained on DailyDialogue and CommonDialogues. Using [unused1] token to divide sentences in context. More info can be found at https://wandb.ai/alquist/next-sentence-prediction/runs/ll69cliu/logs?workspace=user-petr-lorenc"
         
     | 
| 4 | 
         
            -
            }
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
    	
        model/e09d71f55f4b6fc20135f856bf029322a3265d8d/meta-info.json
    DELETED
    
    | 
         @@ -1 +0,0 @@ 
     | 
|
| 1 | 
         
            -
            {"args": [], "kwargs": {"data_root": "/home/lorenpe2/project/data", "data_sources": [["COMMON_DIALOGUES", "common_dialogues/train.json", "common_dialogues/valid.json", "common_dialogues/test.json"], ["DAILY_DIALOGUES", "daily_dialogues/dialogues_text.train.txt", "daily_dialogues/dialogues_text.dev.txt", "daily_dialogues/dialogues_text.test.txt"]], "pretrained_model": "bert-base-uncased", "tokenizer": "bert-base-uncased", "naive_approach": true, "special_token": "[unused1]", "learning_rate": 5e-05}, "tokenizer_args": {"padding": "max_length", "max_length_ctx": 256, "max_length_res": 64, "truncation": "only_first", "return_tensors": "np", "is_split_into_words": true, "naive_approach": true, "special_token": "[unused1]"}}
         
     | 
| 
         | 
|
| 
         | 
    	
        model/e09d71f55f4b6fc20135f856bf029322a3265d8d/optimizer.pt
    DELETED
    
    | 
         @@ -1,3 +0,0 @@ 
     | 
|
| 1 | 
         
            -
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256:ab19babe7ae39d1b6746d1dec58ab7758b0bacc33346b66e9c7da94419bebf96
         
     | 
| 3 | 
         
            -
            size 124944384
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
    	
        model/f1f881389fb38108e623689999ceaaaf398c5e92/info.json
    DELETED
    
    | 
         @@ -1,4 +0,0 @@ 
     | 
|
| 1 | 
         
            -
            {
         
     | 
| 2 | 
         
            -
              "model": "BERT-NSP-v5",
         
     | 
| 3 | 
         
            -
              "description": "Model trained on DailyDialogue and CommonDialogues. Using [unused1] token to divide sentences in context. Improved training arguments (warmup, smaller learning rate). Using frozen test set to better compare models and therefore trained longer time (about 60 epochs). The model also have bigger classification head (from one layer liner as classical). More info can be found at https://wandb.ai/alquist/next-sentence-prediction/runs/vzpwetvm/overview?workspace=user-petr-lorenc"
         
     | 
| 4 | 
         
            -
            }
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
    	
        model/f1f881389fb38108e623689999ceaaaf398c5e92/meta-info.json
    DELETED
    
    | 
         @@ -1 +0,0 @@ 
     | 
|
| 1 | 
         
            -
            {"args": [], "kwargs": {"model_package": "models", "model_class": "OwnBertForNextSentencePrediction", "data_root": "/home/lorenpe2/project/data", "data_sources": [["COMMON_DIALOGUES", "common_dialogues/train.json", "common_dialogues/valid_frozen.json", "common_dialogues/test_frozen.json"], ["DAILY_DIALOGUES", "daily_dialogues/dialogues_text.train.txt", "daily_dialogues/dev_frozen.json", "daily_dialogues/test_frozen.json"]], "pretrained_model": "bert-base-uncased", "tokenizer": "bert-base-uncased", "approach": "IGNORE_DUPLICITIES", "special_token": "[unused1]", "learning_rate": 5e-07, "warmup_ratio": 0.1, "freeze_prefinetuning": true, "prefinenuting_epoch": 10, "finetuning_epochs": 75}, "tokenizer_args": {"padding": "max_length", "max_length_ctx": 256, "max_length_res": 40, "truncation": "only_first", "return_tensors": "np", "is_split_into_words": true, "approach": "IGNORE_DUPLICITIES", "special_token": "[unused1]"}}
         
     | 
| 
         | 
|
| 
         | 
    	
        model/f1f881389fb38108e623689999ceaaaf398c5e92/special_tokens_map.json
    DELETED
    
    | 
         @@ -1 +0,0 @@ 
     | 
|
| 1 | 
         
            -
            {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "additional_special_tokens": ["[unused1]"]}
         
     | 
| 
         | 
|
| 
         | 
    	
        model/f1f881389fb38108e623689999ceaaaf398c5e92/tokenizer_config.json
    DELETED
    
    | 
         @@ -1 +0,0 @@ 
     | 
|
| 1 | 
         
            -
            {"do_lower_case": true, "do_basic_tokenize": true, "never_split": null, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "bert-base-uncased", "tokenizer_class": "BertTokenizer"}
         
     | 
| 
         | 
|
| 
         | 
    	
        model/{e09d71f55f4b6fc20135f856bf029322a3265d8d β new_version_dummy}/config.json
    RENAMED
    
    | 
         @@ -1,7 +1,7 @@ 
     | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
             
              "_name_or_path": "bert-base-uncased",
         
     | 
| 3 | 
         
             
              "architectures": [
         
     | 
| 4 | 
         
            -
                " 
     | 
| 5 | 
         
             
              ],
         
     | 
| 6 | 
         
             
              "attention_probs_dropout_prob": 0.1,
         
     | 
| 7 | 
         
             
              "classifier_dropout": null,
         
     | 
| 
         @@ -18,8 +18,9 @@ 
     | 
|
| 18 | 
         
             
              "num_hidden_layers": 12,
         
     | 
| 19 | 
         
             
              "pad_token_id": 0,
         
     | 
| 20 | 
         
             
              "position_embedding_type": "absolute",
         
     | 
| 
         | 
|
| 21 | 
         
             
              "torch_dtype": "float32",
         
     | 
| 22 | 
         
            -
              "transformers_version": "4. 
     | 
| 23 | 
         
             
              "type_vocab_size": 2,
         
     | 
| 24 | 
         
             
              "use_cache": true,
         
     | 
| 25 | 
         
             
              "vocab_size": 30522
         
     | 
| 
         | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
             
              "_name_or_path": "bert-base-uncased",
         
     | 
| 3 | 
         
             
              "architectures": [
         
     | 
| 4 | 
         
            +
                "BertForSequenceClassification"
         
     | 
| 5 | 
         
             
              ],
         
     | 
| 6 | 
         
             
              "attention_probs_dropout_prob": 0.1,
         
     | 
| 7 | 
         
             
              "classifier_dropout": null,
         
     | 
| 
         | 
|
| 18 | 
         
             
              "num_hidden_layers": 12,
         
     | 
| 19 | 
         
             
              "pad_token_id": 0,
         
     | 
| 20 | 
         
             
              "position_embedding_type": "absolute",
         
     | 
| 21 | 
         
            +
              "problem_type": "single_label_classification",
         
     | 
| 22 | 
         
             
              "torch_dtype": "float32",
         
     | 
| 23 | 
         
            +
              "transformers_version": "4.30.0.dev0",
         
     | 
| 24 | 
         
             
              "type_vocab_size": 2,
         
     | 
| 25 | 
         
             
              "use_cache": true,
         
     | 
| 26 | 
         
             
              "vocab_size": 30522
         
     | 
    	
        model/new_version_dummy/onnx/model.onnx
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:adf2c61611e66c3efcca5e6d866b354596fff14c0f5d49ef4ff73bcbc77a20bc
         
     | 
| 3 | 
         
            +
            size 438201824
         
     | 
    	
        model/{e09d71f55f4b6fc20135f856bf029322a3265d8d β new_version_dummy}/pytorch_model.bin
    RENAMED
    
    | 
         @@ -1,3 +1,3 @@ 
     | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256: 
     | 
| 3 | 
         
            -
            size  
     | 
| 
         | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:04a843b2644531a9f6ac15659f22d89bc288c17886e9c6d2561431a7a0add441
         
     | 
| 3 | 
         
            +
            size 438007925
         
     | 
    	
        model/{c3c3bdb7ad80396e69de171995e2038f900940c8 β new_version_dummy}/special_tokens_map.json
    RENAMED
    
    | 
         @@ -1,4 +1,7 @@ 
     | 
|
| 1 | 
         
             
            {
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 2 | 
         
             
              "cls_token": "[CLS]",
         
     | 
| 3 | 
         
             
              "mask_token": "[MASK]",
         
     | 
| 4 | 
         
             
              "pad_token": "[PAD]",
         
     | 
| 
         | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
            +
              "additional_special_tokens": [
         
     | 
| 3 | 
         
            +
                "[unused1]"
         
     | 
| 4 | 
         
            +
              ],
         
     | 
| 5 | 
         
             
              "cls_token": "[CLS]",
         
     | 
| 6 | 
         
             
              "mask_token": "[MASK]",
         
     | 
| 7 | 
         
             
              "pad_token": "[PAD]",
         
     | 
    	
        model/{c3c3bdb7ad80396e69de171995e2038f900940c8 β new_version_dummy}/tokenizer.json
    RENAMED
    
    | 
         @@ -12,6 +12,15 @@ 
     | 
|
| 12 | 
         
             
                  "normalized": false,
         
     | 
| 13 | 
         
             
                  "special": true
         
     | 
| 14 | 
         
             
                },
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 15 | 
         
             
                {
         
     | 
| 16 | 
         
             
                  "id": 100,
         
     | 
| 17 | 
         
             
                  "content": "[UNK]",
         
     | 
| 
         | 
|
| 12 | 
         
             
                  "normalized": false,
         
     | 
| 13 | 
         
             
                  "special": true
         
     | 
| 14 | 
         
             
                },
         
     | 
| 15 | 
         
            +
                {
         
     | 
| 16 | 
         
            +
                  "id": 2,
         
     | 
| 17 | 
         
            +
                  "content": "[unused1]",
         
     | 
| 18 | 
         
            +
                  "single_word": false,
         
     | 
| 19 | 
         
            +
                  "lstrip": false,
         
     | 
| 20 | 
         
            +
                  "rstrip": false,
         
     | 
| 21 | 
         
            +
                  "normalized": false,
         
     | 
| 22 | 
         
            +
                  "special": true
         
     | 
| 23 | 
         
            +
                },
         
     | 
| 24 | 
         
             
                {
         
     | 
| 25 | 
         
             
                  "id": 100,
         
     | 
| 26 | 
         
             
                  "content": "[UNK]",
         
     | 
    	
        model/{e09d71f55f4b6fc20135f856bf029322a3265d8d β new_version_dummy}/tokenizer_config.json
    RENAMED
    
    | 
         @@ -1,11 +1,11 @@ 
     | 
|
| 1 | 
         
             
            {
         
     | 
| 
         | 
|
| 2 | 
         
             
              "cls_token": "[CLS]",
         
     | 
| 3 | 
         
             
              "do_lower_case": true,
         
     | 
| 4 | 
         
             
              "mask_token": "[MASK]",
         
     | 
| 5 | 
         
             
              "model_max_length": 512,
         
     | 
| 6 | 
         
             
              "pad_token": "[PAD]",
         
     | 
| 7 | 
         
             
              "sep_token": "[SEP]",
         
     | 
| 8 | 
         
            -
              "special_tokens_map_file": null,
         
     | 
| 9 | 
         
             
              "strip_accents": null,
         
     | 
| 10 | 
         
             
              "tokenize_chinese_chars": true,
         
     | 
| 11 | 
         
             
              "tokenizer_class": "BertTokenizer",
         
     | 
| 
         | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
            +
              "clean_up_tokenization_spaces": true,
         
     | 
| 3 | 
         
             
              "cls_token": "[CLS]",
         
     | 
| 4 | 
         
             
              "do_lower_case": true,
         
     | 
| 5 | 
         
             
              "mask_token": "[MASK]",
         
     | 
| 6 | 
         
             
              "model_max_length": 512,
         
     | 
| 7 | 
         
             
              "pad_token": "[PAD]",
         
     | 
| 8 | 
         
             
              "sep_token": "[SEP]",
         
     | 
| 
         | 
|
| 9 | 
         
             
              "strip_accents": null,
         
     | 
| 10 | 
         
             
              "tokenize_chinese_chars": true,
         
     | 
| 11 | 
         
             
              "tokenizer_class": "BertTokenizer",
         
     | 
    	
        model/{f1f881389fb38108e623689999ceaaaf398c5e92 β new_version_dummy}/training_args.bin
    RENAMED
    
    | 
         @@ -1,3 +1,3 @@ 
     | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256: 
     | 
| 3 | 
         
            -
            size  
     | 
| 
         | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:a5e18fad4f08f10e21116128e5ffdbc3b9da56804fae548529abfcafe8814d94
         
     | 
| 3 | 
         
            +
            size 3899
         
     | 
    	
        model/{e09d71f55f4b6fc20135f856bf029322a3265d8d β new_version_dummy}/vocab.txt
    RENAMED
    
    | 
         
            File without changes
         
     | 
    	
        model/{f1f881389fb38108e623689999ceaaaf398c5e92 β new_version_v1}/config.json
    RENAMED
    
    | 
         @@ -1,5 +1,5 @@ 
     | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
            -
              "_name_or_path": "bert-base-uncased",
         
     | 
| 3 | 
         
             
              "architectures": [
         
     | 
| 4 | 
         
             
                "OwnBertForNextSentencePrediction"
         
     | 
| 5 | 
         
             
              ],
         
     | 
| 
         @@ -19,7 +19,7 @@ 
     | 
|
| 19 | 
         
             
              "pad_token_id": 0,
         
     | 
| 20 | 
         
             
              "position_embedding_type": "absolute",
         
     | 
| 21 | 
         
             
              "torch_dtype": "float32",
         
     | 
| 22 | 
         
            -
              "transformers_version": "4. 
     | 
| 23 | 
         
             
              "type_vocab_size": 2,
         
     | 
| 24 | 
         
             
              "use_cache": true,
         
     | 
| 25 | 
         
             
              "vocab_size": 30522
         
     | 
| 
         | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
            +
              "_name_or_path": "/home/lorenpe2/project/hf_models/bert-base-uncased",
         
     | 
| 3 | 
         
             
              "architectures": [
         
     | 
| 4 | 
         
             
                "OwnBertForNextSentencePrediction"
         
     | 
| 5 | 
         
             
              ],
         
     | 
| 
         | 
|
| 19 | 
         
             
              "pad_token_id": 0,
         
     | 
| 20 | 
         
             
              "position_embedding_type": "absolute",
         
     | 
| 21 | 
         
             
              "torch_dtype": "float32",
         
     | 
| 22 | 
         
            +
              "transformers_version": "4.30.0.dev0",
         
     | 
| 23 | 
         
             
              "type_vocab_size": 2,
         
     | 
| 24 | 
         
             
              "use_cache": true,
         
     | 
| 25 | 
         
             
              "vocab_size": 30522
         
     | 
    	
        model/new_version_v1/onnx/model.onnx
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:1a4908d96ecead256b5c4f921a6ae08fe156bdb3a11aceb5aa7b8677dac9322e
         
     | 
| 3 | 
         
            +
            size 438174352
         
     | 
    	
        model/{c3c3bdb7ad80396e69de171995e2038f900940c8 β new_version_v1}/pytorch_model.bin
    RENAMED
    
    | 
         @@ -1,3 +1,3 @@ 
     | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256: 
     | 
| 3 | 
         
            -
            size  
     | 
| 
         | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:7ed3691d1c07d77a0a6796fb4a5f61a2e774849cb7489c97a183097cb4f693aa
         
     | 
| 3 | 
         
            +
            size 438856837
         
     | 
    	
        model/new_version_v1/special_tokens_map.json
    ADDED
    
    | 
         @@ -0,0 +1,10 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "additional_special_tokens": [
         
     | 
| 3 | 
         
            +
                "[unused1]"
         
     | 
| 4 | 
         
            +
              ],
         
     | 
| 5 | 
         
            +
              "cls_token": "[CLS]",
         
     | 
| 6 | 
         
            +
              "mask_token": "[MASK]",
         
     | 
| 7 | 
         
            +
              "pad_token": "[PAD]",
         
     | 
| 8 | 
         
            +
              "sep_token": "[SEP]",
         
     | 
| 9 | 
         
            +
              "unk_token": "[UNK]"
         
     | 
| 10 | 
         
            +
            }
         
     | 
    	
        model/new_version_v1/tokenizer_config.json
    ADDED
    
    | 
         @@ -0,0 +1,15 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "cls_token": "[CLS]",
         
     | 
| 3 | 
         
            +
              "do_basic_tokenize": true,
         
     | 
| 4 | 
         
            +
              "do_lower_case": true,
         
     | 
| 5 | 
         
            +
              "mask_token": "[MASK]",
         
     | 
| 6 | 
         
            +
              "model_max_length": 512,
         
     | 
| 7 | 
         
            +
              "never_split": null,
         
     | 
| 8 | 
         
            +
              "pad_token": "[PAD]",
         
     | 
| 9 | 
         
            +
              "sep_token": "[SEP]",
         
     | 
| 10 | 
         
            +
              "special_tokens_map_file": null,
         
     | 
| 11 | 
         
            +
              "strip_accents": null,
         
     | 
| 12 | 
         
            +
              "tokenize_chinese_chars": true,
         
     | 
| 13 | 
         
            +
              "tokenizer_class": "BertTokenizer",
         
     | 
| 14 | 
         
            +
              "unk_token": "[UNK]"
         
     | 
| 15 | 
         
            +
            }
         
     | 
    	
        model/{c3c3bdb7ad80396e69de171995e2038f900940c8 β new_version_v1}/training_args.bin
    RENAMED
    
    | 
         @@ -1,3 +1,3 @@ 
     | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256: 
     | 
| 3 | 
         
            -
            size  
     | 
| 
         | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:bed99c1c5b736531dce1c0822e055bb8ad10a1634b9235d9a35e7d6aa4356d58
         
     | 
| 3 | 
         
            +
            size 4091
         
     | 
    	
        model/{f1f881389fb38108e623689999ceaaaf398c5e92 β new_version_v1}/vocab.txt
    RENAMED
    
    | 
         
            File without changes
         
     | 
    	
        model/{c3c3bdb7ad80396e69de171995e2038f900940c8 β new_version_v2}/config.json
    RENAMED
    
    | 
         @@ -1,7 +1,7 @@ 
     | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
            -
              "_name_or_path": "bert-base-uncased",
         
     | 
| 3 | 
         
             
              "architectures": [
         
     | 
| 4 | 
         
            -
                " 
     | 
| 5 | 
         
             
              ],
         
     | 
| 6 | 
         
             
              "attention_probs_dropout_prob": 0.1,
         
     | 
| 7 | 
         
             
              "classifier_dropout": null,
         
     | 
| 
         @@ -18,8 +18,9 @@ 
     | 
|
| 18 | 
         
             
              "num_hidden_layers": 12,
         
     | 
| 19 | 
         
             
              "pad_token_id": 0,
         
     | 
| 20 | 
         
             
              "position_embedding_type": "absolute",
         
     | 
| 
         | 
|
| 21 | 
         
             
              "torch_dtype": "float32",
         
     | 
| 22 | 
         
            -
              "transformers_version": "4. 
     | 
| 23 | 
         
             
              "type_vocab_size": 2,
         
     | 
| 24 | 
         
             
              "use_cache": true,
         
     | 
| 25 | 
         
             
              "vocab_size": 30522
         
     | 
| 
         | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
            +
              "_name_or_path": "/home/lorenpe2/project/hf_models/bert-base-uncased",
         
     | 
| 3 | 
         
             
              "architectures": [
         
     | 
| 4 | 
         
            +
                "BertForSequenceClassification"
         
     | 
| 5 | 
         
             
              ],
         
     | 
| 6 | 
         
             
              "attention_probs_dropout_prob": 0.1,
         
     | 
| 7 | 
         
             
              "classifier_dropout": null,
         
     | 
| 
         | 
|
| 18 | 
         
             
              "num_hidden_layers": 12,
         
     | 
| 19 | 
         
             
              "pad_token_id": 0,
         
     | 
| 20 | 
         
             
              "position_embedding_type": "absolute",
         
     | 
| 21 | 
         
            +
              "problem_type": "single_label_classification",
         
     | 
| 22 | 
         
             
              "torch_dtype": "float32",
         
     | 
| 23 | 
         
            +
              "transformers_version": "4.30.0.dev0",
         
     | 
| 24 | 
         
             
              "type_vocab_size": 2,
         
     | 
| 25 | 
         
             
              "use_cache": true,
         
     | 
| 26 | 
         
             
              "vocab_size": 30522
         
     | 
    	
        model/new_version_v2/info.json
    ADDED
    
    | 
         @@ -0,0 +1,4 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "model": "BERT-SEQUENCE-CLASSIFICATION",
         
     | 
| 3 | 
         
            +
              "description": "Model trained on subset of DailyDialogue, CommonDialogues, ChitChatDataset, AirDialogue and SODA. Using [unused1] token to divide sentences in context. More info can be found at https://wandb.ai/alquist/next-sentence-prediction/runs/wfsx1sga/overview?workspace=user-petr-lorenc"
         
     | 
| 4 | 
         
            +
            }
         
     | 
    	
        model/new_version_v2/meta-info.json
    ADDED
    
    | 
         @@ -0,0 +1,28 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "args": [],
         
     | 
| 3 | 
         
            +
              "kwargs": {
         
     | 
| 4 | 
         
            +
                "model_package": "transformers",
         
     | 
| 5 | 
         
            +
                "model_class": "AutoModelForSequenceClassification",
         
     | 
| 6 | 
         
            +
                "data_root": "/home/lorenpe2/project/data",
         
     | 
| 7 | 
         
            +
                "data_sources": [],
         
     | 
| 8 | 
         
            +
                "pretrained_model": "bert-base-uncased",
         
     | 
| 9 | 
         
            +
                "tokenizer": "bert-base-uncased",
         
     | 
| 10 | 
         
            +
                "approach": "IGNORE_DUPLICITIES",
         
     | 
| 11 | 
         
            +
                "special_token": "[unused1]",
         
     | 
| 12 | 
         
            +
                "learning_rate": 5e-07,
         
     | 
| 13 | 
         
            +
                "warmup_ratio": 0.1,
         
     | 
| 14 | 
         
            +
                "freeze_prefinetuning": true,
         
     | 
| 15 | 
         
            +
                "prefinenuting_epoch": 10,
         
     | 
| 16 | 
         
            +
                "finetuning_epochs": 75
         
     | 
| 17 | 
         
            +
              },
         
     | 
| 18 | 
         
            +
              "tokenizer_args": {
         
     | 
| 19 | 
         
            +
                "padding": "max_length",
         
     | 
| 20 | 
         
            +
                "max_length_ctx": 256,
         
     | 
| 21 | 
         
            +
                "max_length_res": 40,
         
     | 
| 22 | 
         
            +
                "truncation": "only_first",
         
     | 
| 23 | 
         
            +
                "return_tensors": "np",
         
     | 
| 24 | 
         
            +
                "is_split_into_words": true,
         
     | 
| 25 | 
         
            +
                "approach": "IGNORE_DUPLICITIES",
         
     | 
| 26 | 
         
            +
                "special_token": "[unused1]"
         
     | 
| 27 | 
         
            +
              }
         
     | 
| 28 | 
         
            +
            }
         
     |