roxky commited on
Commit
410cc8f
·
1 Parent(s): eac6483

Commit changes in Space

Browse files
app.py CHANGED
@@ -5,6 +5,7 @@ from g4f.typing import AsyncResult, Messages
5
  import json
6
  import re
7
  import time
 
8
  from urllib.parse import quote_plus
9
  from fastapi import FastAPI, Response, Request
10
  from fastapi.responses import RedirectResponse
@@ -29,7 +30,7 @@ class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
29
 
30
  models = [
31
  *g4f.Provider.OpenaiAccount.get_models(),
32
- *g4f.Provider.HuggingChat.get_models(),
33
  "flux",
34
  "flux-pro",
35
  "MiniMax-01",
@@ -48,7 +49,7 @@ class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
48
  model = model.split(' ')[-1]
49
  elif model in g4f.Provider.OpenaiAccount.get_models():
50
  pass
51
- elif model in g4f.Provider.HuggingChat.get_models():
52
  pass
53
  else:
54
  raise ModelNotSupportedError(f"Model: {model}")
@@ -58,12 +59,12 @@ class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
58
  def get_provider(cls, model):
59
  if model.startswith("MiniMax"):
60
  return "HailuoAI"
61
- elif model == "Copilot":
62
  return "CopilotAccount"
63
  elif model in g4f.Provider.OpenaiAccount.get_models():
64
  return "OpenaiAccount"
65
- elif model in g4f.Provider.HuggingChat.get_models():
66
- return "HuggingChat"
67
  return None
68
 
69
  @classmethod
@@ -77,6 +78,9 @@ class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
77
  **kwargs
78
  ) -> AsyncResult:
79
  debug.log(f"{__name__}: {api_key}")
 
 
 
80
 
81
  async with StreamSession(
82
  proxy=proxy,
@@ -86,12 +90,14 @@ class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
86
  model = cls.get_model(model)
87
  provider = cls.get_provider(model)
88
  async with session.post(f"{cls.url}/backend-api/v2/conversation", json={
 
89
  "model": model,
90
  "messages": messages,
91
- "provider": provider,
92
- **kwargs
93
  }, ssl=cls.ssl) as response:
 
94
  async for line in response.iter_lines():
 
95
  data = json.loads(line)
96
  data_type = data.pop("type")
97
  if data_type == "provider":
@@ -112,7 +118,22 @@ class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
112
  filename = f"{int(time.time())}_{quote_plus(match.group(1)[:100], '')}{extension}"
113
  download_url = f"/download/{filename}?url={cls.url}{match.group(3)}"
114
  return f"[![{match.group(1)}]({download_url})](/images/{filename})"
115
- yield re.sub(r'\[\!\[(.+?)\]\(([^)]+?)\)\]\(([^)]+?)\)', on_image, data["content"])
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
116
  elif data_type =="synthesize":
117
  yield SynthesizeData(**data[data_type])
118
  elif data_type == "parameters":
@@ -134,10 +155,74 @@ class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
134
 
135
  g4f.Provider.__map__["Feature"] = BackendApi
136
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
137
  def create_app():
138
  g4f.debug.logging = True
139
  g4f.api.AppConfig.gui = True
140
- g4f.api.AppConfig.demo = True
 
 
 
 
 
 
 
 
 
141
 
142
  app = FastAPI()
143
 
@@ -156,6 +241,8 @@ def create_app():
156
  api.register_authorization()
157
  api.register_validation_exception_handler()
158
 
 
 
159
  @app.get("/download/{filename}", response_class=RedirectResponse)
160
  async def download(filename, request: Request):
161
  filename = os.path.basename(filename)
@@ -163,9 +250,14 @@ def create_app():
163
  if not os.path.exists(target):
164
  url = str(request.query_params).split("url=", 1)[1]
165
  if url:
166
- source_url = url.replace("%2F", "/").replace("%3A", ":").replace("%3F", "?").replace("%3D", "=")
167
- await copy_images([source_url], target=target, ssl=False)
 
 
 
 
168
  if not os.path.exists(target):
 
169
  return Response(status_code=404)
170
  return RedirectResponse(f"/images/{filename}")
171
 
 
5
  import json
6
  import re
7
  import time
8
+ import logging
9
  from urllib.parse import quote_plus
10
  from fastapi import FastAPI, Response, Request
11
  from fastapi.responses import RedirectResponse
 
30
 
31
  models = [
32
  *g4f.Provider.OpenaiAccount.get_models(),
33
+ *g4f.Provider.PerplexityLabs.get_models(),
34
  "flux",
35
  "flux-pro",
36
  "MiniMax-01",
 
49
  model = model.split(' ')[-1]
50
  elif model in g4f.Provider.OpenaiAccount.get_models():
51
  pass
52
+ elif model in g4f.Provider.PerplexityLabs.get_models():
53
  pass
54
  else:
55
  raise ModelNotSupportedError(f"Model: {model}")
 
59
  def get_provider(cls, model):
60
  if model.startswith("MiniMax"):
61
  return "HailuoAI"
62
+ elif model == "Copilot" or "dall-e" in model:
63
  return "CopilotAccount"
64
  elif model in g4f.Provider.OpenaiAccount.get_models():
65
  return "OpenaiAccount"
66
+ elif model in g4f.Provider.PerplexityLabs.get_models():
67
+ return "PerplexityLabs"
68
  return None
69
 
70
  @classmethod
 
78
  **kwargs
79
  ) -> AsyncResult:
80
  debug.log(f"{__name__}: {api_key}")
81
+ if "dall-e" in model and "prompt" not in kwargs:
82
+ kwargs["prompt"] = messages[-1]["content"]
83
+ messages[-1]["content"] = f"Generate a image: {kwargs['prompt']}"
84
 
85
  async with StreamSession(
86
  proxy=proxy,
 
90
  model = cls.get_model(model)
91
  provider = cls.get_provider(model)
92
  async with session.post(f"{cls.url}/backend-api/v2/conversation", json={
93
+ **kwargs,
94
  "model": model,
95
  "messages": messages,
96
+ "provider": provider
 
97
  }, ssl=cls.ssl) as response:
98
+ is_thinking = 0
99
  async for line in response.iter_lines():
100
+ response.raise_for_status()
101
  data = json.loads(line)
102
  data_type = data.pop("type")
103
  if data_type == "provider":
 
118
  filename = f"{int(time.time())}_{quote_plus(match.group(1)[:100], '')}{extension}"
119
  download_url = f"/download/{filename}?url={cls.url}{match.group(3)}"
120
  return f"[![{match.group(1)}]({download_url})](/images/{filename})"
121
+ if "<think>" in data[data_type]:
122
+ data[data_type] = data[data_type].split("<think>", 1)
123
+ yield data[data_type][0]
124
+ yield Reasoning(data[data_type][1])
125
+ yield Reasoning(None, "Is thinking...")
126
+ is_thinking = time.time()
127
+ if "</think>" in data[data_type][1]:
128
+ data[data_type][1] = data[data_type].split("</think>", 1)
129
+ yield Reasoning(data[data_type][0])
130
+ yield Reasoning(None, f"Finished in {round(time.time()-is_thinking, 2)} seconds")
131
+ yield data[data_type][1]
132
+ is_thinking = 0
133
+ elif is_thinking:
134
+ yield Reasoning(data[data_type])
135
+ else:
136
+ yield re.sub(r'\[\!\[(.+?)\]\(([^)]+?)\)\]\(([^)]+?)\)', on_image, data[data_type])
137
  elif data_type =="synthesize":
138
  yield SynthesizeData(**data[data_type])
139
  elif data_type == "parameters":
 
155
 
156
  g4f.Provider.__map__["Feature"] = BackendApi
157
 
158
+ import asyncio
159
+ import uuid
160
+ from aiohttp import ClientSession, ClientError
161
+ from g4f.typing import Optional, Cookies
162
+ from g4f.image import is_accepted_format
163
+
164
+ async def copy_images(
165
+ images: list[str],
166
+ cookies: Optional[Cookies] = None,
167
+ headers: dict = None,
168
+ proxy: Optional[str] = None,
169
+ add_url: bool = True,
170
+ target: str = None,
171
+ ssl: bool = None
172
+ ) -> list[str]:
173
+ if add_url:
174
+ add_url = not cookies
175
+ #ensure_images_dir()
176
+ async with ClientSession(
177
+ #connector=get_connector(proxy=proxy),
178
+ cookies=cookies,
179
+ headers=headers,
180
+ ) as session:
181
+ async def copy_image(image: str, target: str = None) -> str:
182
+ if target is None or len(images) > 1:
183
+ target = os.path.join(images_dir, f"{int(time.time())}_{str(uuid.uuid4())}")
184
+ try:
185
+ if image.startswith("data:"):
186
+ pass
187
+ #with open(target, "wb") as f:
188
+ # f.write(extract_data_uri(image))
189
+ else:
190
+ try:
191
+ async with session.get(image, proxy=proxy, ssl=ssl) as response:
192
+ response.raise_for_status()
193
+ with open(target, "wb") as f:
194
+ async for chunk in response.content.iter_chunked(4096):
195
+ f.write(chunk)
196
+ except ClientError as e:
197
+ debug.log(f"copy_images failed: {e.__class__.__name__}: {e}")
198
+ return image
199
+ if "." not in target:
200
+ with open(target, "rb") as f:
201
+ extension = is_accepted_format(f.read(12)).split("/")[-1]
202
+ extension = "jpg" if extension == "jpeg" else extension
203
+ new_target = f"{target}.{extension}"
204
+ os.rename(target, new_target)
205
+ target = new_target
206
+ finally:
207
+ if "." not in target and os.path.exists(target):
208
+ os.unlink(target)
209
+ return f"/images/{os.path.basename(target)}{'?url=' + image if add_url and not image.startswith('data:') else ''}"
210
+
211
+ return await asyncio.gather(*[copy_image(image, target) for image in images])
212
+
213
  def create_app():
214
  g4f.debug.logging = True
215
  g4f.api.AppConfig.gui = True
216
+ g4f.api.AppConfig.demo = False
217
+
218
+ class NoHomeFilter(logging.Filter):
219
+ def filter(self, record):
220
+ if '"GET / HTTP/1.1" 200 OK' in record.getMessage():
221
+ return False
222
+ if '"GET /static/' in record.getMessage():
223
+ return False
224
+ return True
225
+ logging.getLogger("uvicorn.access").addFilter(NoHomeFilter())
226
 
227
  app = FastAPI()
228
 
 
241
  api.register_authorization()
242
  api.register_validation_exception_handler()
243
 
244
+ headers = {}
245
+
246
  @app.get("/download/{filename}", response_class=RedirectResponse)
247
  async def download(filename, request: Request):
248
  filename = os.path.basename(filename)
 
250
  if not os.path.exists(target):
251
  url = str(request.query_params).split("url=", 1)[1]
252
  if url:
253
+ source_url = url.replace("%2F", "/").replace("%3A", ":").replace("%3F", "?")
254
+ await copy_images(
255
+ [source_url],
256
+ target=target,
257
+ ssl=False,
258
+ headers=headers if source_url.startswith(BackendApi.url) else None)
259
  if not os.path.exists(target):
260
+
261
  return Response(status_code=404)
262
  return RedirectResponse(f"/images/{filename}")
263
 
deepseek4free-setup.sh ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ git clone https://github.com/xtekky/deepseek4free.git
2
+
3
+ echo "recursive-include dsk/wasm *" >> ./deepseek4free/MANIFEST.in
4
+
5
+ echo "from setuptools import find_packages, setup
6
+
7
+ INSTALL_REQUIRE = [
8
+ 'curl-cffi',
9
+ 'wasmtime',
10
+ 'numpy',
11
+ ]
12
+
13
+ DESCRIPTION = (
14
+ 'The official gpt4free repository | various collection of powerful language models'
15
+ )
16
+
17
+ # Setting up
18
+ setup(
19
+ name='dsk',
20
+ version='0.0.1.0',
21
+ author='Tekky',
22
+ author_email='<[email protected]>',
23
+ description=DESCRIPTION,
24
+ long_description_content_type='text/markdown',
25
+ long_description='',
26
+ packages=find_packages(),
27
+ include_package_data=True,
28
+ install_requires=INSTALL_REQUIRE
29
+ )" >> ./deepseek4free/setup.py
30
+
31
+ pip install ./deepseek4free --break-system-packages
generated_images/1738132089_386b1b49-9cb2-48b7-90e6-9d37c0af6604.jpg ADDED
generated_images/1738132162_f4cc45c3-33d3-48ed-a8ea-1bfcf78d114f.jpg ADDED
generated_images/1738132256_4694db5d-2328-4200-9dc9-bf1246ab3761.jpg ADDED
generated_images/1738132532_6333b9d5-8fc4-4222-851e-fc65aecb3afd.jpg ADDED
generated_images/1738202142_hallo.webp ADDED
generated_images/1738235394_13e87a97-0e40-4692-90f1-390b722a7c2b.jpg ADDED
generated_images/1738235522_10d7ae46-18a0-4b25-8813-7d1838877abb.jpg ADDED
har_and_cookies/.logging/2025-01-30-.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
har_and_cookies/.usage/2025-01-29.jsonl ADDED
@@ -0,0 +1,182 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"user": "MTadder", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 11, "completion_tokens": 27, "total_tokens": 38}
2
+ {"user": "MTadder", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 11, "completion_tokens": 26, "total_tokens": 37}
3
+ {"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 8, "completion_tokens": 12, "total_tokens": 20}
4
+ {"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 8, "completion_tokens": 23, "total_tokens": 31}
5
+ {"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 45, "completion_tokens": 180, "total_tokens": 225}
6
+ {"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 45, "completion_tokens": 106, "total_tokens": 151}
7
+ {"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 45, "completion_tokens": 44, "total_tokens": 89}
8
+ {"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 45, "completion_tokens": 159, "total_tokens": 204}
9
+ {"user": "roxky", "model": "black-forest-labs/FLUX.1-schnell", "provider": "HuggingFace", "prompt_tokens": 45, "completion_tokens": 142, "total_tokens": 187}
10
+ {"user": "roxky", "model": "black-forest-labs/FLUX.1-schnell", "provider": "HuggingSpace", "prompt_tokens": 45, "completion_tokens": 192, "total_tokens": 237}
11
+ {"user": "roxky", "model": "black-forest-labs/FLUX.1-schnell", "provider": "HuggingSpace", "prompt_tokens": 45, "completion_tokens": 191, "total_tokens": 236}
12
+ {"user": "roxky", "model": "black-forest-labs/FLUX.1-schnell", "provider": "HuggingChat", "prompt_tokens": 45, "completion_tokens": 153, "total_tokens": 198}
13
+ {"user": "roxky", "model": "black-forest-labs/FLUX.1-schnell", "provider": "HuggingSpace", "prompt_tokens": 45, "completion_tokens": 137, "total_tokens": 182}
14
+ {"user": "roxky", "model": "Pollinations AI: flux", "provider": "HuggingSpace", "prompt_tokens": 16, "completion_tokens": 152, "total_tokens": 168}
15
+ {"user": "roxky", "model": "Pollinations AI: flux", "provider": "Blackbox", "prompt_tokens": 16, "completion_tokens": 139, "total_tokens": 155}
16
+ {"user": "MTadder", "model": "qwq-32b", "provider": "HuggingFace", "prompt_tokens": 30, "completion_tokens": 523, "total_tokens": 553}
17
+ {"user": "roxky", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "Feature", "prompt_tokens": 9, "completion_tokens": 20, "total_tokens": 29}
18
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 13, "completion_tokens": 242, "total_tokens": 255}
19
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 263, "completion_tokens": 137, "total_tokens": 400}
20
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 255, "completion_tokens": 57, "total_tokens": 312}
21
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 10088, "completion_tokens": 39, "total_tokens": 10127}
22
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
23
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 31, "completion_tokens": 92, "total_tokens": 123}
24
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 132, "completion_tokens": 20, "total_tokens": 152}
25
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 157, "completion_tokens": 67, "total_tokens": 224}
26
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 232, "completion_tokens": 74, "total_tokens": 306}
27
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 307, "completion_tokens": 81, "total_tokens": 388}
28
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 497, "completion_tokens": 614, "total_tokens": 1111}
29
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
30
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 40, "completion_tokens": 117, "total_tokens": 157}
31
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 179, "completion_tokens": 208, "total_tokens": 387}
32
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 382, "completion_tokens": 239, "total_tokens": 621}
33
+ {"user": "maverick792", "model": "o1", "provider": "Feature", "prompt_tokens": 19, "completion_tokens": 810, "total_tokens": 829}
34
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 610, "completion_tokens": 196, "total_tokens": 806}
35
+ {"user": "maverick792", "model": "o1", "provider": "Feature", "prompt_tokens": 1031, "completion_tokens": 1481, "total_tokens": 2512}
36
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 793, "completion_tokens": 313, "total_tokens": 1106}
37
+ {"user": "maverick792", "model": "o1", "provider": "Feature", "prompt_tokens": 2772, "completion_tokens": 216, "total_tokens": 2988}
38
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 4599, "completion_tokens": 55, "total_tokens": 4654}
39
+ {"user": "maverick792", "model": "o1-preview", "provider": "Feature", "prompt_tokens": 3066, "completion_tokens": 1905, "total_tokens": 4971}
40
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 5852, "completion_tokens": 461, "total_tokens": 6313}
41
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 6270, "completion_tokens": 382, "total_tokens": 6652}
42
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 6642, "completion_tokens": 234, "total_tokens": 6876}
43
+ {"user": "maverick792", "model": "o1-preview", "provider": "Feature", "prompt_tokens": 5280, "completion_tokens": 1858, "total_tokens": 7138}
44
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 6874, "completion_tokens": 299, "total_tokens": 7173}
45
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 7169, "completion_tokens": 349, "total_tokens": 7518}
46
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 7507, "completion_tokens": 432, "total_tokens": 7939}
47
+ {"user": "maverick792", "model": "o1-preview", "provider": "Feature", "prompt_tokens": 7449, "completion_tokens": 1157, "total_tokens": 8606}
48
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 7897, "completion_tokens": 3069, "total_tokens": 10966}
49
+ {"user": "maverick792", "model": "o1-preview", "provider": "Feature", "prompt_tokens": 16, "completion_tokens": 38, "total_tokens": 54}
50
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 9952, "completion_tokens": 398, "total_tokens": 10350}
51
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 7897, "completion_tokens": 3069, "total_tokens": 10966}
52
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 7897, "completion_tokens": 3069, "total_tokens": 10966}
53
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 14401, "completion_tokens": 29, "total_tokens": 14430}
54
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 16279, "completion_tokens": 256, "total_tokens": 16535}
55
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 18379, "completion_tokens": 2038, "total_tokens": 20417}
56
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 20020, "completion_tokens": 2230, "total_tokens": 22250}
57
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 18379, "completion_tokens": 2038, "total_tokens": 20417}
58
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 18379, "completion_tokens": 2038, "total_tokens": 20417}
59
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 26051, "completion_tokens": 1704, "total_tokens": 27755}
60
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 27458, "completion_tokens": 239, "total_tokens": 27697}
61
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 28631, "completion_tokens": 666, "total_tokens": 29297}
62
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 29348, "completion_tokens": 1353, "total_tokens": 30701}
63
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 31059, "completion_tokens": 1226, "total_tokens": 32285}
64
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 33004, "completion_tokens": 1502, "total_tokens": 34506}
65
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 35888, "completion_tokens": 1580, "total_tokens": 37468}
66
+ {"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 39076, "completion_tokens": 2652, "total_tokens": 41728}
67
+ {"user": "MTadder", "model": "qwq-32b", "provider": "HuggingFace", "prompt_tokens": 75, "completion_tokens": 685, "total_tokens": 760}
68
+ {"user": "roxky", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "HuggingChat", "prompt_tokens": 17, "completion_tokens": 23, "total_tokens": 40}
69
+ {"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 17, "completion_tokens": 55, "total_tokens": 72}
70
+ {"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 8, "completion_tokens": 14, "total_tokens": 22}
71
+ {"user": "roxky", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "HuggingChat", "prompt_tokens": 8, "completion_tokens": 12, "total_tokens": 20}
72
+ {"user": "nelivol732", "model": "sd-3.5", "provider": "HuggingSpace", "prompt_tokens": 31, "completion_tokens": 220, "total_tokens": 251}
73
+ {"user": "nelivol732", "model": "sd-3.5", "provider": "HuggingFace", "prompt_tokens": 312, "completion_tokens": 102, "total_tokens": 414}
74
+ {"user": "nelivol732", "model": "sd-3.5", "provider": "HuggingSpace", "prompt_tokens": 437, "completion_tokens": 219, "total_tokens": 656}
75
+ {"user": "nelivol732", "model": "flux-dev", "provider": "HuggingSpace", "prompt_tokens": 31, "completion_tokens": 214, "total_tokens": 245}
76
+ {"user": "nelivol732", "model": "flux-dev", "provider": "HuggingFace", "prompt_tokens": 306, "completion_tokens": 98, "total_tokens": 404}
77
+ {"user": "nelivol732", "model": "flux-schnell", "provider": "HuggingFace", "prompt_tokens": 31, "completion_tokens": 88, "total_tokens": 119}
78
+ {"user": "roxky", "model": "Qwen/Qwen2.5-72B-Instruct", "provider": "HuggingChat", "prompt_tokens": 9, "completion_tokens": 30, "total_tokens": 39}
79
+ {"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 9, "completion_tokens": 28, "total_tokens": 37}
80
+ {"user": "nelivol732", "model": "flux-schnell", "provider": "HuggingFace", "prompt_tokens": 180, "completion_tokens": 98, "total_tokens": 278}
81
+ {"user": "roxky", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 7, "total_tokens": 16}
82
+ {"user": "nelivol732", "model": "sd-3.5", "provider": "HuggingSpace", "prompt_tokens": 31, "completion_tokens": 214, "total_tokens": 245}
83
+ {"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 57, "completion_tokens": 176, "total_tokens": 233}
84
+ {"user": "nelivol732", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 37, "completion_tokens": 30, "total_tokens": 67}
85
+ {"user": "nelivol732", "completion_tokens": 74, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 179, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 253}
86
+ {"user": "nelivol732", "completion_tokens": 166, "prompt_tokens": 313, "total_tokens": 479}
87
+ {"user": "nelivol732", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 2522, "completion_tokens": 6, "total_tokens": 2528}
88
+ {"user": "nelivol732", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 2522, "completion_tokens": 6, "total_tokens": 2528}
89
+ {"user": "nelivol732", "completion_tokens": 136, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 2559, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 2695}
90
+ {"user": "fabler43", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 25, "completion_tokens": 59, "total_tokens": 84}
91
+ {"user": "fabler43", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 21, "completion_tokens": 198, "total_tokens": 219}
92
+ {"user": "fabler43", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 21, "completion_tokens": 201, "total_tokens": 222}
93
+ {"user": "MTadder", "model": "qwq-32b", "provider": "HuggingFace", "prompt_tokens": 347, "completion_tokens": 855, "total_tokens": 1202}
94
+ {"user": "MTadder", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 347, "completion_tokens": 326, "total_tokens": 673}
95
+ {"user": "MTadder", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 1219, "completion_tokens": 53, "total_tokens": 1272}
96
+ {"user": "MTadder", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 1286, "completion_tokens": 449, "total_tokens": 1735}
97
+ {"user": "MTadder", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 45, "completion_tokens": 254, "total_tokens": 299}
98
+ {"user": "MTadder", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 154, "completion_tokens": 1008, "total_tokens": 1162}
99
+ {"user": "MTadder", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 1177, "completion_tokens": 1026, "total_tokens": 2203}
100
+ {"user": "roxky", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "HuggingChat", "prompt_tokens": 8, "completion_tokens": 12, "total_tokens": 20}
101
+ {"user": "roxky", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "HuggingChat", "prompt_tokens": 8, "completion_tokens": 31, "total_tokens": 39}
102
+ {"user": "silait", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 12, "total_tokens": 20}
103
+ {"user": "silait", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 31, "completion_tokens": 47, "total_tokens": 78}
104
+ {"user": "silait", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 95, "completion_tokens": 682, "total_tokens": 777}
105
+ {"user": "roxky", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "HuggingChat", "prompt_tokens": 9, "completion_tokens": 12, "total_tokens": 21}
106
+ {"user": "roxky", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "HuggingChat", "prompt_tokens": 43, "completion_tokens": 496, "total_tokens": 539}
107
+ {"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 21, "completion_tokens": 162, "total_tokens": 183}
108
+ {"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 58, "completion_tokens": 173, "total_tokens": 231}
109
+ {"user": "fabler43", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 303, "completion_tokens": 1704, "total_tokens": 2007}
110
+ {"user": "Thanh6868", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 25, "total_tokens": 33}
111
+ {"user": "Dotmazy", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 11, "completion_tokens": 19, "total_tokens": 30}
112
+ {"user": "Dotmazy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 13, "completion_tokens": 11, "total_tokens": 24}
113
+ {"user": "Dotmazy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 21, "completion_tokens": 1012, "total_tokens": 1033}
114
+ {"user": "Dotmazy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 1054, "completion_tokens": 1181, "total_tokens": 2235}
115
+ {"user": "Dotmazy", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 2248, "completion_tokens": 660, "total_tokens": 2908}
116
+ {"user": "andrewseverniy", "model": "qwen-2-72b", "provider": "HuggingSpace", "prompt_tokens": 15, "completion_tokens": 5, "total_tokens": 20}
117
+ {"user": "andrewseverniy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 36, "completion_tokens": 395, "total_tokens": 431}
118
+ {"user": "Stremty", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 12, "total_tokens": 20}
119
+ {"user": "Stremty", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 31, "completion_tokens": 44, "total_tokens": 75}
120
+ {"user": "Stremty", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 69, "completion_tokens": 372, "total_tokens": 441}
121
+ {"user": "dfed", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 11, "completion_tokens": 24, "total_tokens": 35}
122
+ {"user": "dfed", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 59, "completion_tokens": 129, "total_tokens": 188}
123
+ {"user": "dfed", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 208, "completion_tokens": 130, "total_tokens": 338}
124
+ {"user": "dfed", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 208, "completion_tokens": 377, "total_tokens": 585}
125
+ {"user": "dfed", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 731, "completion_tokens": 59, "total_tokens": 790}
126
+ {"user": "dfed", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 803, "completion_tokens": 295, "total_tokens": 1098}
127
+ {"user": "maverick792", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 17, "completion_tokens": 686, "total_tokens": 703}
128
+ {"user": "vlasdadsda", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 8, "completion_tokens": 20, "total_tokens": 28}
129
+ {"user": "vlasdadsda", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 37, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 153, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 190}
130
+ {"user": "vlasdadsda", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 223, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 1108, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 1331}
131
+ {"user": "vlasdadsda", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 313, "prompt_tokens": 440, "total_tokens": 753}
132
+ {"user": "vlasdadsda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 10, "completion_tokens": 90, "total_tokens": 100}
133
+ {"user": "misharBSD", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 8, "completion_tokens": 9, "total_tokens": 17}
134
+ {"user": "misharBSD", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 37, "completion_tokens": 38, "total_tokens": 75}
135
+ {"user": "misharBSD", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 42, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 178, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 220}
136
+ {"user": "misharBSD", "model": "sd-3.5", "provider": "HuggingSpace", "prompt_tokens": 22, "completion_tokens": 233, "total_tokens": 255}
137
+ {"user": "misharBSD", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 1029, "prompt_tokens": 177, "total_tokens": 1206}
138
+ {"user": "misharBSD", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 45, "completion_tokens": 1075, "total_tokens": 1120}
139
+ {"user": "misharBSD", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 2332, "completion_tokens": 1026, "total_tokens": 3358}
140
+ {"user": "misharBSD", "model": "flux-dev", "provider": "HuggingSpace", "prompt_tokens": 19, "completion_tokens": 205, "total_tokens": 224}
141
+ {"user": "misharBSD", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 1414, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 159, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 1573}
142
+ {"user": "misharBSD", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 847, "prompt_tokens": 166, "total_tokens": 1013}
143
+ {"user": "misharBSD", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 10, "completion_tokens": 9, "total_tokens": 19}
144
+ {"user": "misharBSD", "model": "Qwen/Qwen2.5-72B-Instruct", "provider": "HuggingChat", "prompt_tokens": 10, "completion_tokens": 20, "total_tokens": 30}
145
+ {"user": "misharBSD", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 10, "completion_tokens": 192, "total_tokens": 202}
146
+ {"user": "misharBSD", "model": "black-forest-labs/FLUX.1-dev", "provider": "HuggingChat", "prompt_tokens": 8, "completion_tokens": 65, "total_tokens": 73}
147
+ {"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 8, "completion_tokens": 14, "total_tokens": 22}
148
+ {"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 278, "completion_tokens": 30, "total_tokens": 308}
149
+ {"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 336, "completion_tokens": 155, "total_tokens": 491}
150
+ {"user": "misharBSD", "model": "CohereForAI/c4ai-command-r-plus-08-2024", "provider": "HuggingChat", "prompt_tokens": 336, "completion_tokens": 665, "total_tokens": 1001}
151
+ {"user": "misharBSD", "model": "CohereForAI/c4ai-command-r-plus-08-2024", "provider": "HuggingChat", "prompt_tokens": 551, "completion_tokens": 288, "total_tokens": 839}
152
+ {"user": "misharBSD", "model": "CohereForAI/c4ai-command-r-plus-08-2024", "provider": "HuggingChat", "prompt_tokens": 551, "completion_tokens": 302, "total_tokens": 853}
153
+ {"user": "misharBSD", "model": "meta-llama/Llama-3.3-70B-Instruct", "provider": "HuggingChat", "prompt_tokens": 551, "completion_tokens": 20, "total_tokens": 571}
154
+ {"user": "misharBSD", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "HuggingChat", "prompt_tokens": 551, "completion_tokens": 151, "total_tokens": 702}
155
+ {"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 26, "completion_tokens": 370, "total_tokens": 396}
156
+ {"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 423, "completion_tokens": 534, "total_tokens": 957}
157
+ {"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 994, "completion_tokens": 969, "total_tokens": 1963}
158
+ {"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 1982, "completion_tokens": 953, "total_tokens": 2935}
159
+ {"user": "adi6409", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 11, "completion_tokens": 28, "total_tokens": 39}
160
+ {"user": "adi6409", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 45, "completion_tokens": 276, "total_tokens": 321}
161
+ {"user": "adi6409", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 11, "completion_tokens": 84, "total_tokens": 95}
162
+ {"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 2969, "completion_tokens": 1772, "total_tokens": 4741}
163
+ {"user": "NDeriy", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 253, "prompt_tokens": 3732, "total_tokens": 3985}
164
+ {"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 4755, "completion_tokens": 1363, "total_tokens": 6118}
165
+ {"user": "NDeriy", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 456, "completion_tokens": 14, "total_tokens": 470}
166
+ {"user": "NDeriy", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 492, "completion_tokens": 53, "total_tokens": 545}
167
+ {"user": "NDeriy", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 522, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 6932, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 7454}
168
+ {"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 9, "completion_tokens": 31, "total_tokens": 40}
169
+ {"user": "roxky", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
170
+ {"user": "rdmistra", "model": "flux-dev", "provider": "HuggingFace", "prompt_tokens": 26, "completion_tokens": 1944, "total_tokens": 1970}
171
+ {"user": "roxky", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "HuggingChat", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
172
+ {"user": "NDeriy", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 856, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 10330, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 11186}
173
+ {"user": "rdmistra", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 38, "completion_tokens": 131, "total_tokens": 169}
174
+ {"user": "rdmistra", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 824, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 2953, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 3777}
175
+ {"user": "Docfile", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 10, "completion_tokens": 16, "total_tokens": 26}
176
+ {"user": "playjymer", "model": "Qwen/QwQ-32B-Preview", "provider": "HuggingChat", "prompt_tokens": 17, "completion_tokens": 9, "total_tokens": 26}
177
+ {"user": "playjymer", "model": "dall-e-3", "provider": "OpenaiAccount", "prompt_tokens": 41, "completion_tokens": 70, "total_tokens": 111}
178
+ {"user": "NDeriy", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 2368, "completion_tokens": 475, "total_tokens": 2843}
179
+ {"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 36, "completion_tokens": 244, "total_tokens": 280}
180
+ {"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 295, "completion_tokens": 180, "total_tokens": 475}
181
+ {"user": "DarkyMan", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 11, "completion_tokens": 17, "total_tokens": 28}
182
+ {"user": "DarkyMan", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 72, "completion_tokens": 483, "total_tokens": 555}
har_and_cookies/.usage/2025-01-30-.jsonl ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"user": "NDeriy", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 35, "completion_tokens": 292, "total_tokens": 327}
2
+ {"user": "roxky", "model": "dall-e-3", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 8, "total_tokens": 16}
3
+ {"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 8, "completion_tokens": 32, "total_tokens": 40}
4
+ {"user": "maverick792", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 30, "completion_tokens": 461, "total_tokens": 491}
5
+ {"user": "roxky", "model": "qwen-2-72b", "provider": "HuggingSpace", "prompt_tokens": 8, "completion_tokens": 29, "total_tokens": 37}
6
+ {"user": "roxky", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 15, "completion_tokens": 19, "total_tokens": 34}
7
+ {"user": "roxky", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 44, "completion_tokens": 10, "total_tokens": 54}
8
+ {"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 44, "completion_tokens": 30, "total_tokens": 74}
9
+ {"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 44, "completion_tokens": 17, "total_tokens": 61}
10
+ {"user": "roxky", "model": "flux", "provider": "HuggingSpace", "prompt_tokens": 44, "completion_tokens": 141, "total_tokens": 185}
11
+ {"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 44, "completion_tokens": 8, "total_tokens": 52}
12
+ {"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 44, "completion_tokens": 8, "total_tokens": 52}
13
+ {"user": "roxky", "model": "gpt-4o-mini", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 7, "total_tokens": 16}
14
+ {"user": "roxky", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 9, "completion_tokens": 122, "total_tokens": 131}
15
+ {"user": "Yazee83", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 13, "completion_tokens": 225, "total_tokens": 238}
16
+ {"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
17
+ {"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 8, "completion_tokens": 30, "total_tokens": 38}
18
+ {"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 8, "completion_tokens": 17, "total_tokens": 25}
19
+ {"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 31, "completion_tokens": 44, "total_tokens": 75}
20
+ {"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 95, "completion_tokens": 378, "total_tokens": 473}
21
+ {"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 95, "completion_tokens": 378, "total_tokens": 473}
22
+ {"user": "Yazee83", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 352, "completion_tokens": 864, "total_tokens": 1216}
23
+ {"user": "Yazee83", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 522, "completion_tokens": 46, "total_tokens": 568}
24
+ {"user": "Yazee83", "model": "claude-3-haiku", "provider": "DDG", "prompt_tokens": 3931, "completion_tokens": 80, "total_tokens": 4011}
25
+ {"user": "Yazee83", "model": "claude-3-haiku", "provider": "DDG", "prompt_tokens": 3931, "completion_tokens": 80, "total_tokens": 4011}
26
+ {"user": "Yazee83", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 3458, "completion_tokens": 11, "total_tokens": 3469}
27
+ {"user": "roxky", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 9, "completion_tokens": 176, "total_tokens": 185}
28
+ {"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 195, "completion_tokens": 200, "total_tokens": 395}
29
+ {"user": "roxky", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 404, "completion_tokens": 12, "total_tokens": 416}
30
+ {"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 404, "completion_tokens": 12, "total_tokens": 416}
31
+ {"user": "L-X-u", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 18, "prompt_tokens": 152, "total_tokens": 170}
32
+ {"user": "L-X-u", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 34, "completion_tokens": 136, "total_tokens": 170}
33
+ {"user": "L-X-u", "model": "dall-e-3", "provider": "OpenaiAccount", "prompt_tokens": 34, "completion_tokens": 327, "total_tokens": 361}
34
+ {"user": "L-X-u", "model": "flux", "provider": "HuggingSpace", "prompt_tokens": 34, "completion_tokens": 3433, "total_tokens": 3467}
35
+ {"user": "L-X-u", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 34, "completion_tokens": 932, "total_tokens": 966}
36
+ {"user": "Yazee83", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 22, "completion_tokens": 381, "total_tokens": 403}
37
+ {"user": "L-X-u", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 1221, "total_tokens": 1230}
38
+ {"user": "L-X-u", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 1587, "completion_tokens": 2460, "total_tokens": 4047}
39
+ {"user": "Yazee83", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 27, "completion_tokens": 544, "total_tokens": 571}
40
+ {"user": "L-X-u", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 1587, "completion_tokens": 471, "total_tokens": 2058}
41
+ {"user": "L-X-u", "model": "flux", "provider": "HuggingSpace", "prompt_tokens": 9, "completion_tokens": 19696, "total_tokens": 19705}
42
+ {"user": "L-X-u", "model": "dall-e-3", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 133, "total_tokens": 142}
43
+ {"user": "L-X-u", "model": "flux", "provider": "HuggingSpace", "prompt_tokens": 9, "completion_tokens": 19716, "total_tokens": 19725}
44
+ {"user": "roxky", "model": "dall-e-3", "provider": "OpenaiAccount", "prompt_tokens": 14, "completion_tokens": 108, "total_tokens": 122}
45
+ {"user": "roxky", "model": "dall-e-3", "provider": "OpenaiAccount", "prompt_tokens": 133, "completion_tokens": 82, "total_tokens": 215}
46
+ {"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 14, "completion_tokens": 128, "total_tokens": 142}
47
+ {"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 13, "completion_tokens": 1022, "total_tokens": 1035}
48
+ {"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 10, "completion_tokens": 912, "total_tokens": 922}
49
+ {"user": "Harvinderkno", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 316, "completion_tokens": 789, "total_tokens": 1105}
50
+ {"user": "Harvinderkno", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 935, "completion_tokens": 490, "total_tokens": 1425}
51
+ {"user": "Harvinderkno", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 1640, "completion_tokens": 817, "total_tokens": 2457}
52
+ {"user": "Harvinderkno", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 2290, "completion_tokens": 816, "total_tokens": 3106}
53
+ {"user": "Harvinderkno", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 11, "completion_tokens": 281, "total_tokens": 292}
54
+ {"user": "Harvinderkno", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 265, "completion_tokens": 536, "total_tokens": 801}
55
+ {"user": "OWO111", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 16, "completion_tokens": 96, "total_tokens": 112}
56
+ {"user": "OWO111", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 260, "prompt_tokens": 8452, "total_tokens": 8712}
57
+ {"user": "OWO111", "model": "qwen-2-72b", "provider": "HuggingSpace", "prompt_tokens": 513, "completion_tokens": 326, "total_tokens": 839}
58
+ {"user": "OWO111", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 856, "completion_tokens": 539, "total_tokens": 1395}
59
+ {"user": "OWO111", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 1425, "completion_tokens": 967, "total_tokens": 2392}
60
+ {"user": "andoajpr", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 441, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 158, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 599}
61
+ {"user": "OWO111", "model": "gpt-4o-mini", "provider": "OpenaiAccount", "prompt_tokens": 2409, "completion_tokens": 965, "total_tokens": 3374}
62
+ {"user": "andoajpr", "model": "qvq-72b", "provider": "HuggingSpace", "prompt_tokens": 11, "completion_tokens": 41, "total_tokens": 52}
63
+ {"user": "OWO111", "model": "gpt-4o-mini", "provider": "OpenaiAccount", "prompt_tokens": 11, "completion_tokens": 796, "total_tokens": 807}
64
+ {"user": "OWO111", "model": "gpt-4o-mini", "provider": "OpenaiAccount", "prompt_tokens": 1005, "completion_tokens": 608, "total_tokens": 1613}
65
+ {"user": "OWO111", "model": "qwq-32b", "provider": "HuggingFace", "prompt_tokens": 55, "completion_tokens": 18, "total_tokens": 73}
66
+ {"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 17, "completion_tokens": 531, "total_tokens": 548}
67
+ {"user": "roxky", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 207, "prompt_tokens": 2115, "total_tokens": 2322}
68
+ {"user": "roxky", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 835, "completion_tokens": 573, "total_tokens": 1408}
69
+ {"user": "roxky", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 1506, "completion_tokens": 785, "total_tokens": 2291}
70
+ {"user": "KoLLchIK", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 8, "completion_tokens": 9, "total_tokens": 17}
71
+ {"user": "roxky", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 178, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 10024, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 10202}