Update README.md
Browse files
    	
        README.md
    CHANGED
    
    | 
         @@ -1,3 +1,24 @@ 
     | 
|
| 1 | 
         
            -
             
     | 
| 2 | 
         
            -
             
     | 
| 3 | 
         
            -
             
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
             
     | 
| 2 | 
         
            +
            # t5-small-nlu-tm2-context3
         
     | 
| 3 | 
         
            +
             
     | 
| 4 | 
         
            +
            This model is a fine-tuned version of [GPT2-medium](https://huggingface.co/gpt2-medium) on [Schema-Guided Dialogue](https://huggingface.co/datasets/ConvLab/sgd).
         
     | 
| 5 | 
         
            +
             
     | 
| 6 | 
         
            +
            Refer to [ConvLab-3](https://github.com/ConvLab/ConvLab-3) for model description and usage.
         
     | 
| 7 | 
         
            +
             
     | 
| 8 | 
         
            +
            ## Training procedure
         
     | 
| 9 | 
         
            +
             
     | 
| 10 | 
         
            +
            ### Training hyperparameters
         
     | 
| 11 | 
         
            +
             
     | 
| 12 | 
         
            +
            The following hyperparameters were used during training:
         
     | 
| 13 | 
         
            +
            - learning_rate: 5e-5
         
     | 
| 14 | 
         
            +
            - train_batch_size: 64
         
     | 
| 15 | 
         
            +
            - gradient_accumulation_steps: 2
         
     | 
| 16 | 
         
            +
            - total_train_batch_size: 128
         
     | 
| 17 | 
         
            +
            - optimizer: AdamW
         
     | 
| 18 | 
         
            +
            - lr_scheduler_type: linear
         
     | 
| 19 | 
         
            +
            - num_epochs: 20
         
     | 
| 20 | 
         
            +
             
     | 
| 21 | 
         
            +
            ### Framework versions
         
     | 
| 22 | 
         
            +
             
     | 
| 23 | 
         
            +
            - Transformers 4.23.1
         
     | 
| 24 | 
         
            +
            - Pytorch 1.10.1+cu111
         
     |