Spaces:
Sleeping
Sleeping
| import copy | |
| import json | |
| import re | |
| import requests | |
| import uuid | |
| # from curl_cffi import requests | |
| from tclogger import logger | |
| from transformers import AutoTokenizer | |
| from constants.models import ( | |
| MODEL_MAP, | |
| STOP_SEQUENCES_MAP, | |
| TOKEN_LIMIT_MAP, | |
| TOKEN_RESERVED, | |
| ) | |
| from constants.envs import PROXIES | |
| from constants.headers import ( | |
| REQUESTS_HEADERS, | |
| HUGGINGCHAT_POST_HEADERS, | |
| HUGGINGCHAT_SETTINGS_POST_DATA, | |
| ) | |
| from messagers.message_outputer import OpenaiStreamOutputer | |
| class HuggingchatStreamer: | |
| def __init__(self, model: str): | |
| if model in MODEL_MAP.keys(): | |
| self.model = model | |
| else: | |
| self.model = "mixtral-8x7b" | |
| self.model_fullname = MODEL_MAP[self.model] | |
| self.message_outputer = OpenaiStreamOutputer(model=self.model) | |
| # export HF_ENDPOINT=https://hf-mirror.com | |
| self.tokenizer = AutoTokenizer.from_pretrained(self.model_fullname) | |
| def count_tokens(self, text): | |
| tokens = self.tokenizer.encode(text) | |
| token_count = len(tokens) | |
| logger.note(f"Prompt Token Count: {token_count}") | |
| return token_count | |
| def get_hf_chat_id(self): | |
| request_url = "https://huggingface.co/chat/settings" | |
| request_body = copy.deepcopy(HUGGINGCHAT_SETTINGS_POST_DATA) | |
| extra_body = { | |
| "activeModel": self.model_fullname, | |
| } | |
| request_body.update(extra_body) | |
| logger.note(f"> hf-chat ID:", end=" ") | |
| res = requests.post( | |
| request_url, | |
| headers=HUGGINGCHAT_POST_HEADERS, | |
| json=request_body, | |
| proxies=PROXIES, | |
| timeout=10, | |
| ) | |
| self.hf_chat_id = res.cookies.get("hf-chat") | |
| if self.hf_chat_id: | |
| logger.success(f"[{self.hf_chat_id}]") | |
| else: | |
| logger.warn(f"[{res.status_code}]") | |
| logger.warn(res.text) | |
| raise ValueError("Failed to get hf-chat ID!") | |
| def get_conversation_id(self, preprompt: str = ""): | |
| request_url = "https://huggingface.co/chat/conversation" | |
| request_headers = HUGGINGCHAT_POST_HEADERS | |
| extra_headers = { | |
| "Cookie": f"hf-chat={self.hf_chat_id}", | |
| } | |
| request_headers.update(extra_headers) | |
| request_body = { | |
| "model": self.model_fullname, | |
| "preprompt": preprompt, | |
| } | |
| logger.note(f"> Conversation ID:", end=" ") | |
| res = requests.post( | |
| request_url, | |
| headers=request_headers, | |
| json=request_body, | |
| proxies=PROXIES, | |
| timeout=10, | |
| ) | |
| if res.status_code == 200: | |
| conversation_id = res.json()["conversationId"] | |
| logger.success(f"[{conversation_id}]") | |
| else: | |
| logger.warn(f"[{res.status_code}]") | |
| raise ValueError("Failed to get conversation ID!") | |
| self.conversation_id = conversation_id | |
| return conversation_id | |
| def log_request(self, url, method="GET"): | |
| logger.note(f"> {method}:", end=" ") | |
| logger.mesg(f"{url}", end=" ") | |
| def log_response( | |
| self, res: requests.Response, stream=False, iter_lines=False, verbose=False | |
| ): | |
| status_code = res.status_code | |
| status_code_str = f"[{status_code}]" | |
| if status_code == 200: | |
| logger_func = logger.success | |
| else: | |
| logger_func = logger.warn | |
| logger_func(status_code_str) | |
| logger.enter_quiet(not verbose) | |
| if status_code != 200: | |
| logger_func(res.text) | |
| if stream: | |
| if not iter_lines: | |
| return | |
| for line in res.iter_lines(): | |
| line = line.decode("utf-8") | |
| line = re.sub(r"^data:\s*", "", line) | |
| line = line.strip() | |
| if line: | |
| try: | |
| data = json.loads(line, strict=False) | |
| msg_type = data.get("type") | |
| if msg_type == "status": | |
| msg_status = data.get("status") | |
| elif msg_type == "stream": | |
| content = data.get("token", "") | |
| logger_func(content, end="") | |
| elif msg_type == "finalAnswer": | |
| full_content = data.get("text") | |
| logger.success("\n[Finished]") | |
| break | |
| else: | |
| pass | |
| except Exception as e: | |
| logger.warn(e) | |
| else: | |
| logger_func(res.json()) | |
| logger.exit_quiet(not verbose) | |
| def chat_response( | |
| self, | |
| prompt: str = None, | |
| temperature: float = 0.5, | |
| top_p: float = 0.95, | |
| max_new_tokens: int = None, | |
| api_key: str = None, | |
| use_cache: bool = False, | |
| ): | |
| pass | |
| def chat_return_dict(self, stream_response): | |
| pass | |
| def chat_return_generator(self, stream_response): | |
| pass | |
| if __name__ == "__main__": | |
| streamer = HuggingchatStreamer(model="mixtral-8x7b") | |
| conversation_id = streamer.get_conversation_id() | |
| # python -m networks.huggingchat_streamer | |