Spaces:
				
			
			
	
			
			
		Paused
		
	
	
	
			
			
	
	
	
	
		
		
		Paused
		
	:zap: [Enhance] Upgrade gemma-7b to 1.1
Browse files- constants/models.py +1 -1
- messagers/message_composer.py +5 -5
    	
        constants/models.py
    CHANGED
    
    | @@ -3,7 +3,7 @@ MODEL_MAP = { | |
| 3 | 
             
                "nous-mixtral-8x7b": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
         | 
| 4 | 
             
                "mistral-7b": "mistralai/Mistral-7B-Instruct-v0.2",
         | 
| 5 | 
             
                "openchat-3.5": "openchat/openchat-3.5-0106",
         | 
| 6 | 
            -
                "gemma-7b": "google/gemma-7b-it",
         | 
| 7 | 
             
                "command-r-plus": "CohereForAI/c4ai-command-r-plus",
         | 
| 8 | 
             
                "default": "mistralai/Mixtral-8x7B-Instruct-v0.1",
         | 
| 9 | 
             
            }
         | 
|  | |
| 3 | 
             
                "nous-mixtral-8x7b": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
         | 
| 4 | 
             
                "mistral-7b": "mistralai/Mistral-7B-Instruct-v0.2",
         | 
| 5 | 
             
                "openchat-3.5": "openchat/openchat-3.5-0106",
         | 
| 6 | 
            +
                "gemma-7b": "google/gemma-1.1-7b-it",
         | 
| 7 | 
             
                "command-r-plus": "CohereForAI/c4ai-command-r-plus",
         | 
| 8 | 
             
                "default": "mistralai/Mixtral-8x7B-Instruct-v0.1",
         | 
| 9 | 
             
            }
         | 
    	
        messagers/message_composer.py
    CHANGED
    
    | @@ -125,7 +125,7 @@ class MessageComposer: | |
| 125 | 
             
                                )
         | 
| 126 | 
             
                        self.merged_str_list.append(f"GPT4 Correct Assistant:\n")
         | 
| 127 | 
             
                        self.merged_str = "\n".join(self.merged_str_list)
         | 
| 128 | 
            -
                    # https://huggingface.co/google/gemma-7b-it#chat-template
         | 
| 129 | 
             
                    elif self.model in ["gemma-7b"]:
         | 
| 130 | 
             
                        self.messages = self.concat_messages_by_role(messages)
         | 
| 131 | 
             
                        self.merged_str_list = []
         | 
| @@ -147,11 +147,11 @@ class MessageComposer: | |
| 147 | 
             
                                    f"{self.start_of_turn}user\n{content}{self.end_of_turn}"
         | 
| 148 | 
             
                                )
         | 
| 149 | 
             
                        self.merged_str_list.append(f"{self.start_of_turn}model\n")
         | 
| 150 | 
            -
                        self.merged_str = "\n".join(self.merged_str_list)
         | 
| 151 | 
             
                    # https://huggingface.co/NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO#prompt-format
         | 
| 152 | 
             
                    # https://huggingface.co/openchat/openchat-3.5-0106
         | 
| 153 | 
             
                    # elif self.model in ["openchat-3.5", "nous-mixtral-8x7b"]:
         | 
| 154 | 
            -
                    elif self.model in ["openchat-3.5", "command-r-plus"]:
         | 
| 155 | 
             
                        tokenizer = AutoTokenizer.from_pretrained(self.model_fullname)
         | 
| 156 | 
             
                        self.merged_str = tokenizer.apply_chat_template(
         | 
| 157 | 
             
                            messages, tokenize=False, add_generation_prompt=True
         | 
| @@ -167,9 +167,9 @@ class MessageComposer: | |
| 167 | 
             
            if __name__ == "__main__":
         | 
| 168 | 
             
                # model = "mixtral-8x7b"
         | 
| 169 | 
             
                # model = "nous-mixtral-8x7b"
         | 
| 170 | 
            -
                 | 
| 171 | 
             
                # model = "openchat-3.5"
         | 
| 172 | 
            -
                model = "command-r-plus"
         | 
| 173 | 
             
                composer = MessageComposer(model)
         | 
| 174 | 
             
                messages = [
         | 
| 175 | 
             
                    {
         | 
|  | |
| 125 | 
             
                                )
         | 
| 126 | 
             
                        self.merged_str_list.append(f"GPT4 Correct Assistant:\n")
         | 
| 127 | 
             
                        self.merged_str = "\n".join(self.merged_str_list)
         | 
| 128 | 
            +
                    # https://huggingface.co/google/gemma-1.1-7b-it#chat-template
         | 
| 129 | 
             
                    elif self.model in ["gemma-7b"]:
         | 
| 130 | 
             
                        self.messages = self.concat_messages_by_role(messages)
         | 
| 131 | 
             
                        self.merged_str_list = []
         | 
|  | |
| 147 | 
             
                                    f"{self.start_of_turn}user\n{content}{self.end_of_turn}"
         | 
| 148 | 
             
                                )
         | 
| 149 | 
             
                        self.merged_str_list.append(f"{self.start_of_turn}model\n")
         | 
| 150 | 
            +
                        self.merged_str = "<bos>" + "\n".join(self.merged_str_list)
         | 
| 151 | 
             
                    # https://huggingface.co/NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO#prompt-format
         | 
| 152 | 
             
                    # https://huggingface.co/openchat/openchat-3.5-0106
         | 
| 153 | 
             
                    # elif self.model in ["openchat-3.5", "nous-mixtral-8x7b"]:
         | 
| 154 | 
            +
                    elif self.model in ["openchat-3.5", "command-r-plus", "gemma-7b"]:
         | 
| 155 | 
             
                        tokenizer = AutoTokenizer.from_pretrained(self.model_fullname)
         | 
| 156 | 
             
                        self.merged_str = tokenizer.apply_chat_template(
         | 
| 157 | 
             
                            messages, tokenize=False, add_generation_prompt=True
         | 
|  | |
| 167 | 
             
            if __name__ == "__main__":
         | 
| 168 | 
             
                # model = "mixtral-8x7b"
         | 
| 169 | 
             
                # model = "nous-mixtral-8x7b"
         | 
| 170 | 
            +
                model = "gemma-7b"
         | 
| 171 | 
             
                # model = "openchat-3.5"
         | 
| 172 | 
            +
                # model = "command-r-plus"
         | 
| 173 | 
             
                composer = MessageComposer(model)
         | 
| 174 | 
             
                messages = [
         | 
| 175 | 
             
                    {
         |