roxky commited on
Commit
a90f20f
·
1 Parent(s): ec410ac

Commit new images and usage data

Browse files
README.md CHANGED
@@ -5,4 +5,8 @@ colorFrom: green
5
  colorTo: purple
6
  sdk: docker
7
  pinned: false
 
 
 
 
8
  ---
 
5
  colorTo: purple
6
  sdk: docker
7
  pinned: false
8
+
9
+ hf_oauth: true
10
+ hf_oauth_scopes:
11
+ - inference-api
12
  ---
app.py CHANGED
@@ -1,221 +1,23 @@
1
  import os
2
- import json
3
- from g4f.providers.response import Reasoning, JsonConversation, FinishReason
4
- from g4f.typing import AsyncResult, Messages
5
- import json
6
- import re
7
- import time
8
- import logging
9
- from urllib.parse import quote_plus
10
  from fastapi import FastAPI, Response, Request
11
  from fastapi.responses import RedirectResponse
12
 
13
- from g4f.image import images_dir, copy_images
14
-
15
  import g4f.api
16
  import g4f.Provider
17
 
18
- from g4f.Provider.base_provider import AsyncGeneratorProvider, ProviderModelMixin
19
- from g4f.typing import AsyncResult, Messages
20
- from g4f.requests import StreamSession
21
- from g4f.providers.response import ProviderInfo, JsonConversation, PreviewResponse, SynthesizeData, TitleGeneration, RequestLogin
22
- from g4f.providers.response import Parameters, FinishReason, Usage, Reasoning
23
- from g4f.errors import ModelNotSupportedError
24
- from g4f import debug
25
-
26
  import demo
27
-
28
- class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
29
- url = "https://ahe.hopto.org"
30
- working = True
31
- ssl = False
32
-
33
- models = [
34
- *g4f.Provider.OpenaiAccount.get_models(),
35
- *g4f.Provider.PerplexityLabs.get_models(),
36
- "flux",
37
- "flux-pro",
38
- "MiniMax-01",
39
- "Microsoft Copilot",
40
- ]
41
-
42
- @classmethod
43
- def get_model(cls, model):
44
- if "MiniMax" in model:
45
- model = "MiniMax"
46
- elif "Copilot" in model:
47
- model = "Copilot"
48
- elif "FLUX" in model:
49
- model = f"flux-{model.split('-')[-1]}"
50
- elif "flux" in model:
51
- model = model.split(' ')[-1]
52
- elif model in g4f.Provider.OpenaiAccount.get_models():
53
- pass
54
- elif model in g4f.Provider.PerplexityLabs.get_models():
55
- pass
56
- else:
57
- raise ModelNotSupportedError(f"Model: {model}")
58
- return model
59
-
60
- @classmethod
61
- def get_provider(cls, model):
62
- if model.startswith("MiniMax"):
63
- return "HailuoAI"
64
- elif model == "Copilot" or "dall-e" in model:
65
- return "CopilotAccount"
66
- elif model in g4f.Provider.OpenaiAccount.get_models():
67
- return "OpenaiAccount"
68
- elif model in g4f.Provider.PerplexityLabs.get_models():
69
- return "PerplexityLabs"
70
- return None
71
-
72
- @classmethod
73
- async def create_async_generator(
74
- cls,
75
- model: str,
76
- messages: Messages,
77
- api_key: str = None,
78
- proxy: str = None,
79
- timeout: int = 0,
80
- **kwargs
81
- ) -> AsyncResult:
82
- debug.log(f"{__name__}: {api_key}")
83
- if "dall-e" in model and "prompt" not in kwargs:
84
- kwargs["prompt"] = messages[-1]["content"]
85
- messages[-1]["content"] = f"Generate a image: {kwargs['prompt']}"
86
-
87
- async with StreamSession(
88
- proxy=proxy,
89
- headers={"Accept": "text/event-stream", **demo.headers},
90
- timeout=timeout
91
- ) as session:
92
- model = cls.get_model(model)
93
- provider = cls.get_provider(model)
94
- async with session.post(f"{cls.url}/backend-api/v2/conversation", json={
95
- **kwargs,
96
- "model": model,
97
- "messages": messages,
98
- "provider": provider
99
- }, ssl=cls.ssl) as response:
100
- is_thinking = 0
101
- async for line in response.iter_lines():
102
- response.raise_for_status()
103
- data = json.loads(line)
104
- data_type = data.pop("type")
105
- if data_type == "provider":
106
- yield ProviderInfo(**data[data_type])
107
- provider = data[data_type]["name"]
108
- elif data_type == "conversation":
109
- yield JsonConversation(**data[data_type][provider] if provider in data[data_type] else data[data_type][""])
110
- elif data_type == "conversation_id":
111
- pass
112
- elif data_type == "message":
113
- yield Exception(data)
114
- elif data_type == "preview":
115
- yield PreviewResponse(data[data_type])
116
- elif data_type == "content":
117
- def on_image(match):
118
- extension = match.group(3).split(".")[-1].split("?")[0]
119
- extension = "" if not extension or len(extension) > 4 else f".{extension}"
120
- filename = f"{int(time.time())}_{quote_plus(match.group(1)[:100], '')}{extension}"
121
- download_url = f"/download/{filename}?url={cls.url}{match.group(3)}"
122
- return f"[![{match.group(1)}]({download_url})](/images/{filename})"
123
- if "<think>" in data[data_type]:
124
- data[data_type] = data[data_type].split("<think>", 1)
125
- yield data[data_type][0]
126
- yield Reasoning(data[data_type][1])
127
- yield Reasoning(None, "Is thinking...")
128
- is_thinking = time.time()
129
- if "</think>" in data[data_type][1]:
130
- data[data_type][1] = data[data_type].split("</think>", 1)
131
- yield Reasoning(data[data_type][0])
132
- yield Reasoning(None, f"Finished in {round(time.time()-is_thinking, 2)} seconds")
133
- yield data[data_type][1]
134
- is_thinking = 0
135
- elif is_thinking:
136
- yield Reasoning(data[data_type])
137
- else:
138
- yield re.sub(r'\[\!\[(.+?)\]\(([^)]+?)\)\]\(([^)]+?)\)', on_image, data[data_type])
139
- elif data_type =="synthesize":
140
- yield SynthesizeData(**data[data_type])
141
- elif data_type == "parameters":
142
- yield Parameters(**data[data_type])
143
- elif data_type == "usage":
144
- yield Usage(**data[data_type])
145
- elif data_type == "reasoning":
146
- yield Reasoning(**data)
147
- elif data_type == "login":
148
- pass
149
- elif data_type == "title":
150
- yield TitleGeneration(data[data_type])
151
- elif data_type == "finish":
152
- yield FinishReason(data[data_type]["reason"])
153
- elif data_type == "log":
154
- debug.log(data[data_type])
155
- else:
156
- debug.log(f"Unknown data: ({data_type}) {data}")
157
-
158
  g4f.Provider.__map__["Feature"] = BackendApi
159
 
160
- import asyncio
161
- import uuid
162
- from aiohttp import ClientSession, ClientError
163
- from g4f.typing import Optional, Cookies
164
- from g4f.image import is_accepted_format
165
 
166
- async def copy_images(
167
- images: list[str],
168
- cookies: Optional[Cookies] = None,
169
- headers: dict = None,
170
- proxy: Optional[str] = None,
171
- add_url: bool = True,
172
- target: str = None,
173
- ssl: bool = None
174
- ) -> list[str]:
175
- if add_url:
176
- add_url = not cookies
177
- #ensure_images_dir()
178
- async with ClientSession(
179
- #connector=get_connector(proxy=proxy),
180
- cookies=cookies,
181
- headers=headers,
182
- ) as session:
183
- async def copy_image(image: str, target: str = None) -> str:
184
- if target is None or len(images) > 1:
185
- target = os.path.join(images_dir, f"{int(time.time())}_{str(uuid.uuid4())}")
186
- try:
187
- if image.startswith("data:"):
188
- pass
189
- #with open(target, "wb") as f:
190
- # f.write(extract_data_uri(image))
191
- else:
192
- try:
193
- async with session.get(image, proxy=proxy, ssl=ssl) as response:
194
- response.raise_for_status()
195
- with open(target, "wb") as f:
196
- async for chunk in response.content.iter_chunked(4096):
197
- f.write(chunk)
198
- except ClientError as e:
199
- debug.log(f"copy_images failed: {e.__class__.__name__}: {e}")
200
- return image
201
- if "." not in target:
202
- with open(target, "rb") as f:
203
- extension = is_accepted_format(f.read(12)).split("/")[-1]
204
- extension = "jpg" if extension == "jpeg" else extension
205
- new_target = f"{target}.{extension}"
206
- os.rename(target, new_target)
207
- target = new_target
208
- finally:
209
- if "." not in target and os.path.exists(target):
210
- os.unlink(target)
211
- return f"/images/{os.path.basename(target)}{'?url=' + image if add_url and not image.startswith('data:') else ''}"
212
-
213
- return await asyncio.gather(*[copy_image(image, target) for image in images])
214
 
215
  def create_app():
216
  g4f.debug.logging = True
217
  g4f.api.AppConfig.gui = True
218
- g4f.api.AppConfig.demo = False
219
 
220
  app = FastAPI()
221
 
@@ -259,14 +61,4 @@ def create_app():
259
 
260
  return app
261
 
262
-
263
- class NoHomeFilter(logging.Filter):
264
- def filter(self, record):
265
- if '"GET / HTTP/1.1" 200 OK' in record.getMessage():
266
- return False
267
- if '"GET /static/' in record.getMessage():
268
- return False
269
- return True
270
- logging.getLogger("uvicorn.access").addFilter(NoHomeFilter())
271
-
272
  app = create_app()
 
1
  import os
 
 
 
 
 
 
 
 
2
  from fastapi import FastAPI, Response, Request
3
  from fastapi.responses import RedirectResponse
4
 
5
+ from g4f.image import images_dir
 
6
  import g4f.api
7
  import g4f.Provider
8
 
 
 
 
 
 
 
 
 
9
  import demo
10
+ from demo.BackendApi import BackendApi
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  g4f.Provider.__map__["Feature"] = BackendApi
12
 
 
 
 
 
 
13
 
14
+ g4f.models.demo_models
15
+ from g4f.image import copy_images
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
 
17
  def create_app():
18
  g4f.debug.logging = True
19
  g4f.api.AppConfig.gui = True
20
+ g4f.api.AppConfig.demo = True
21
 
22
  app = FastAPI()
23
 
 
61
 
62
  return app
63
 
 
 
 
 
 
 
 
 
 
 
64
  app = create_app()
demo/BackendApi.py ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import json
2
+ import re
3
+ import time
4
+ from urllib.parse import quote_plus
5
+
6
+ import g4f.api
7
+ import g4f.Provider
8
+
9
+ from g4f.Provider.base_provider import AsyncGeneratorProvider, ProviderModelMixin
10
+ from g4f.typing import AsyncResult, Messages
11
+ from g4f.requests import StreamSession
12
+ from g4f.providers.response import *
13
+ from g4f.errors import ModelNotSupportedError
14
+ from g4f import debug
15
+ from . import headers
16
+
17
+ class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
18
+ url = "https://ahe.hopto.org"
19
+ working = True
20
+ ssl = False
21
+
22
+ models = [
23
+ *g4f.Provider.OpenaiAccount.get_models(),
24
+ *g4f.Provider.PerplexityLabs.get_models(),
25
+ "flux",
26
+ "flux-pro",
27
+ "MiniMax-01",
28
+ "Microsoft Copilot",
29
+ ]
30
+
31
+ @classmethod
32
+ def get_model(cls, model):
33
+ if "MiniMax" in model:
34
+ model = "MiniMax"
35
+ elif "Copilot" in model:
36
+ model = "Copilot"
37
+ elif "FLUX" in model:
38
+ model = f"flux-{model.split('-')[-1]}"
39
+ elif "flux" in model:
40
+ model = model.split(' ')[-1]
41
+ elif model in g4f.Provider.OpenaiAccount.get_models():
42
+ pass
43
+ elif model in g4f.Provider.PerplexityLabs.get_models():
44
+ pass
45
+ else:
46
+ raise ModelNotSupportedError(f"Model: {model}")
47
+ return model
48
+
49
+ @classmethod
50
+ def get_provider(cls, model):
51
+ if model.startswith("MiniMax"):
52
+ return "HailuoAI"
53
+ elif model == "Copilot" or "dall-e" in model:
54
+ return "CopilotAccount"
55
+ elif model in g4f.Provider.OpenaiAccount.get_models():
56
+ return "OpenaiAccount"
57
+ elif model in g4f.Provider.PerplexityLabs.get_models():
58
+ return "PerplexityLabs"
59
+ return None
60
+
61
+ @classmethod
62
+ async def create_async_generator(
63
+ cls,
64
+ model: str,
65
+ messages: Messages,
66
+ api_key: str = None,
67
+ proxy: str = None,
68
+ timeout: int = 0,
69
+ **kwargs
70
+ ) -> AsyncResult:
71
+ debug.log(f"{cls.__name__}: {api_key}")
72
+
73
+ if "dall-e" in model and "prompt" not in kwargs:
74
+ kwargs["prompt"] = messages[-1]["content"]
75
+ messages[-1]["content"] = f"Generate a image: {kwargs['prompt']}"
76
+
77
+ async with StreamSession(
78
+ proxy=proxy,
79
+ headers={"Accept": "text/event-stream", **headers},
80
+ timeout=timeout
81
+ ) as session:
82
+ model = cls.get_model(model)
83
+ provider = cls.get_provider(model)
84
+ async with session.post(f"{cls.url}/backend-api/v2/conversation", json={
85
+ **kwargs,
86
+ "model": model,
87
+ "messages": messages,
88
+ "provider": provider
89
+ }, ssl=cls.ssl) as response:
90
+ is_thinking = 0
91
+ async for line in response.iter_lines():
92
+ response.raise_for_status()
93
+ data = json.loads(line)
94
+ data_type = data.pop("type")
95
+ if data_type == "provider":
96
+ yield ProviderInfo(**data[data_type])
97
+ provider = data[data_type]["name"]
98
+ elif data_type == "conversation":
99
+ yield JsonConversation(**data[data_type][provider] if provider in data[data_type] else data[data_type][""])
100
+ elif data_type == "conversation_id":
101
+ pass
102
+ elif data_type == "message":
103
+ yield Exception(data)
104
+ elif data_type == "preview":
105
+ yield PreviewResponse(data[data_type])
106
+ elif data_type == "content":
107
+ def on_image(match):
108
+ extension = match.group(3).split(".")[-1].replace("%3F", "?").split("?")[0]
109
+ extension = ".jpg" if not extension or len(extension) > 4 else f".{extension}"
110
+ filename = f"{int(time.time())}_{quote_plus(model, '')}.{quote_plus(match.group(1)[:100], '')}{extension}"
111
+ download_url = f"/download/{filename}?url={cls.url}{match.group(3)}"
112
+ return f"[![{match.group(1)}]({download_url})](/images/{filename})"
113
+ if "<think>" in data[data_type]:
114
+ data[data_type] = data[data_type].split("<think>", 1)
115
+ yield data[data_type][0]
116
+ yield Reasoning(data[data_type][1])
117
+ yield Reasoning(None, "Is thinking...")
118
+ is_thinking = time.time()
119
+ if "</think>" in data[data_type]:
120
+ data[data_type][1] = data[data_type].split("</think>", 1)
121
+ yield Reasoning(data[data_type][0])
122
+ yield Reasoning(None, f"Finished in {round(time.time()-is_thinking, 2)} seconds")
123
+ yield data[data_type][1]
124
+ is_thinking = 0
125
+ elif is_thinking:
126
+ yield Reasoning(data[data_type])
127
+ else:
128
+ yield re.sub(r'\[\!\[(.+?)\]\(([^)]+?)\)\]\(([^)]+?)\)', on_image, data[data_type])
129
+ elif data_type =="synthesize":
130
+ yield SynthesizeData(**data[data_type])
131
+ elif data_type == "parameters":
132
+ yield Parameters(**data[data_type])
133
+ elif data_type == "usage":
134
+ yield Usage(**data[data_type])
135
+ elif data_type == "reasoning":
136
+ yield Reasoning(**data)
137
+ elif data_type == "login":
138
+ pass
139
+ elif data_type == "title":
140
+ yield TitleGeneration(data[data_type])
141
+ elif data_type == "finish":
142
+ yield FinishReason(data[data_type]["reason"])
143
+ elif data_type == "log":
144
+ debug.log(data[data_type])
145
+ else:
146
+ debug.log(f"Unknown data: ({data_type}) {data}")
generated_images/1738296597_what+is+best+ai+model+yet+so+far,+compare+and+tell+me+final+result+only.jpg ADDED
generated_images/1738332473_gpt-4o.A+modern+commercial+airplane+soaring+through+a+clear+blue+sky,+with+white+contrails+streaming+behind.jpg ADDED
generated_images/1738347359_dall-e-3.Ask+Blackbox+AI+Anything.jpg ADDED
generated_images/1738347377_flux-pro.Ask+Blackbox+AI+Anything.jpg ADDED
generated_images/1738371043_flux-pro.Create+a+image:+A+futuristic+cityscape+at+sunset,+with+towering+skyscrapers+and+flying+cars+zipping+.jpg ADDED
generated_images/1738372570_Copilot.A+futuristic+cityscape+at+sunset,+with+towering+skyscrapers+and+flying+cars+zipping+by.jpg ADDED
generated_images/1738372600_dall-e-3.A+futuristic+cityscape+at+sunset,+with+towering+skyscrapers+and+flying+cars+zipping+by.jpg ADDED
har_and_cookies/.usage/2025-01-30.jsonl ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"user": "roxky", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 8, "completion_tokens": 115, "total_tokens": 123}
2
+ {"user": "roxky", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 18, "completion_tokens": 10, "total_tokens": 28}
3
+ {"user": "roxky", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 38, "completion_tokens": 10, "total_tokens": 48}
4
+ {"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 58, "completion_tokens": 12, "total_tokens": 70}
5
+ {"user": "roxky", "model": "flux-schnell", "provider": "HuggingFace", "prompt_tokens": 8, "completion_tokens": 85, "total_tokens": 93}
6
+ {"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 10, "completion_tokens": 68, "total_tokens": 78}
7
+ {"user": "roxky", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 11, "total_tokens": 19}
8
+ {"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
9
+ {"user": "roxky", "model": "flux-schnell", "provider": "HuggingFace", "prompt_tokens": 9, "completion_tokens": 76, "total_tokens": 85}
10
+ {"user": "RubielLabarta", "model": "flux-dev", "provider": "HuggingSpace", "prompt_tokens": 17, "completion_tokens": 220, "total_tokens": 237}
11
+ {"user": "RubielLabarta", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 8, "total_tokens": 16}
12
+ {"user": "RubielLabarta", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 28, "completion_tokens": 25, "total_tokens": 53}
13
+ {"user": "RubielLabarta", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 60, "completion_tokens": 28, "total_tokens": 88}
14
+ {"user": "RubielLabarta", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 60, "completion_tokens": 28, "total_tokens": 88}
15
+ {"user": "RubielLabarta", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 60, "completion_tokens": 83, "total_tokens": 143}
16
+ {"user": "roxky", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
17
+ {"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 236, "total_tokens": 244}
18
+ {"user": "roxky", "model": "flux", "provider": "PollinationsAI", "prompt_tokens": 8, "completion_tokens": 117, "total_tokens": 125}
19
+ {"user": "roxky", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 8, "completion_tokens": 123, "total_tokens": 131}
20
+ {"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 8, "completion_tokens": 32, "total_tokens": 40}
21
+ {"user": "sheeee2222", "model": "qwen-2-72b", "provider": "HuggingSpace", "prompt_tokens": 15, "completion_tokens": 17, "total_tokens": 32}
22
+ {"user": "sheeee2222", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 10, "prompt_tokens": 174, "total_tokens": 184}
23
+ {"user": "sheeee2222", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 41, "prompt_tokens": 213, "total_tokens": 254}
24
+ {"user": "sheeee2222", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 566, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 3531, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 4097}
25
+ {"user": "roxky", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 11, "total_tokens": 20}
26
+ {"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
27
+ {"user": "roxky", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 7, "total_tokens": 16}
28
+ {"user": "roxky", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 6, "total_tokens": 15}
29
+ {"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 20, "completion_tokens": 15, "total_tokens": 35}
30
+ {"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 46, "completion_tokens": 73, "total_tokens": 119}
31
+ {"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 126, "completion_tokens": 124, "total_tokens": 250}
32
+ {"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 246, "completion_tokens": 22, "total_tokens": 268}
33
+ {"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 274, "completion_tokens": 63, "total_tokens": 337}
34
+ {"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 342, "completion_tokens": 12, "total_tokens": 354}
35
+ {"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 359, "completion_tokens": 41, "total_tokens": 400}
36
+ {"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 420, "completion_tokens": 120, "total_tokens": 540}
37
+ {"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 420, "completion_tokens": 120, "total_tokens": 540}
38
+ {"user": "cihairottoilcazzo", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 18, "completion_tokens": 261, "total_tokens": 279}
39
+ {"user": "sanch1tx", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 8, "completion_tokens": 17, "total_tokens": 25}
40
+ {"user": "nasrulam", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 19, "completion_tokens": 1225, "total_tokens": 1244}
41
+ {"user": "nasrulam", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 19, "completion_tokens": 69, "total_tokens": 88}
42
+ {"user": "nasrulam", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 35, "completion_tokens": 1442, "total_tokens": 1477}
43
+ {"user": "nasrulam", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 35, "completion_tokens": 298, "total_tokens": 333}
44
+ {"user": "nasrulam", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 1574, "completion_tokens": 136, "total_tokens": 1710}
45
+ {"user": "sfewewf232", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 87, "completion_tokens": 160, "total_tokens": 247}
46
+ {"user": "hashim1", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 13, "completion_tokens": 9, "total_tokens": 22}
47
+ {"user": "hashim1", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 36, "completion_tokens": 20, "total_tokens": 56}
48
+ {"user": "roxky", "model": "flux-dev", "provider": "HuggingSpace", "prompt_tokens": 30, "completion_tokens": 200, "total_tokens": 230}
49
+ {"user": "roxky", "model": "flux-schnell", "provider": "HuggingFace", "prompt_tokens": 30, "completion_tokens": 80, "total_tokens": 110}
50
+ {"user": "roxky", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 9, "completion_tokens": 353, "total_tokens": 362}
51
+ {"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
52
+ {"user": "roxky", "model": "gpt-4o-mini", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 7, "total_tokens": 16}
53
+ {"user": "roxky", "model": "claude-3-haiku", "provider": "DDG", "prompt_tokens": 2447, "completion_tokens": 798, "total_tokens": 3245}
har_and_cookies/.usage/2025-01-31.jsonl ADDED
@@ -0,0 +1,126 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"user": "roxky", "model": "qvq-72b", "provider": "HuggingSpace", "prompt_tokens": 15, "completion_tokens": 10, "total_tokens": 25}
2
+ {"user": "roxky", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 16, "completion_tokens": 1904, "total_tokens": 1920}
3
+ {"user": "roxky", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 1490, "completion_tokens": 1736, "total_tokens": 3226}
4
+ {"user": "tanu360", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 8, "completion_tokens": 105, "total_tokens": 113}
5
+ {"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
6
+ {"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 41, "completion_tokens": 240, "total_tokens": 281}
7
+ {"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 41, "completion_tokens": 440, "total_tokens": 481}
8
+ {"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 41, "completion_tokens": 393, "total_tokens": 434}
9
+ {"user": "bbdwbayu", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 694, "completion_tokens": 561, "total_tokens": 1255}
10
+ {"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 1273, "completion_tokens": 44, "total_tokens": 1317}
11
+ {"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 1329, "completion_tokens": 166, "total_tokens": 1495}
12
+ {"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 1508, "completion_tokens": 206, "total_tokens": 1714}
13
+ {"user": "JoniSenior", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 9, "completion_tokens": 299, "total_tokens": 308}
14
+ {"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 9, "completion_tokens": 448, "total_tokens": 457}
15
+ {"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 476, "completion_tokens": 1010, "total_tokens": 1486}
16
+ {"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 70, "completion_tokens": 1075, "total_tokens": 1145}
17
+ {"user": "BruceStudios", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 9, "completion_tokens": 9, "total_tokens": 18}
18
+ {"user": "BruceStudios", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 29, "completion_tokens": 130, "total_tokens": 159}
19
+ {"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 1223, "completion_tokens": 1606, "total_tokens": 2829}
20
+ {"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 2863, "completion_tokens": 1563, "total_tokens": 4426}
21
+ {"user": "herokulastly", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 19, "completion_tokens": 10, "total_tokens": 29}
22
+ {"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 4473, "completion_tokens": 1377, "total_tokens": 5850}
23
+ {"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 5862, "completion_tokens": 755, "total_tokens": 6617}
24
+ {"user": "zayuvegda", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 25, "total_tokens": 33}
25
+ {"user": "zayuvegda", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 43, "completion_tokens": 28, "total_tokens": 71}
26
+ {"user": "zayuvegda", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 76, "completion_tokens": 347, "total_tokens": 423}
27
+ {"user": "zayuvegda", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 434, "completion_tokens": 253, "total_tokens": 687}
28
+ {"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 6635, "completion_tokens": 1026, "total_tokens": 7661}
29
+ {"user": "zayuvegda", "model": "gpt-4o-mini", "provider": "OpenaiAccount", "prompt_tokens": 728, "completion_tokens": 8, "total_tokens": 736}
30
+ {"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 75, "completion_tokens": 945, "total_tokens": 1020}
31
+ {"user": "zayuvegda", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 65, "completion_tokens": 229, "total_tokens": 294}
32
+ {"user": "zayuvegda", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 91, "completion_tokens": 143, "total_tokens": 234}
33
+ {"user": "zayuvegda", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 939, "completion_tokens": 399, "total_tokens": 1338}
34
+ {"user": "zayuvegda", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 1682, "completion_tokens": 489, "total_tokens": 2171}
35
+ {"user": "zayuvegda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 2823, "completion_tokens": 1513, "total_tokens": 4336}
36
+ {"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 1052, "completion_tokens": 833, "total_tokens": 1885}
37
+ {"user": "zayuvegda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 188, "completion_tokens": 730, "total_tokens": 918}
38
+ {"user": "zayuvegda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 928, "completion_tokens": 1272, "total_tokens": 2200}
39
+ {"user": "zayuvegda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 2299, "completion_tokens": 2075, "total_tokens": 4374}
40
+ {"user": "zayuvegda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 4444, "completion_tokens": 2914, "total_tokens": 7358}
41
+ {"user": "cruelzade", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 287, "completion_tokens": 464, "total_tokens": 751}
42
+ {"user": "seotube", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 11, "total_tokens": 19}
43
+ {"user": "seotube", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 30, "completion_tokens": 378, "total_tokens": 408}
44
+ {"user": "seotube", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 379, "completion_tokens": 214, "total_tokens": 593}
45
+ {"user": "setsunafjava", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 30, "completion_tokens": 15, "total_tokens": 45}
46
+ {"user": "setsunafjava", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 69, "completion_tokens": 17, "total_tokens": 86}
47
+ {"user": "setsunafjava", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 115, "completion_tokens": 1204, "total_tokens": 1319}
48
+ {"user": "setsunafjava", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 1890, "completion_tokens": 1940, "total_tokens": 3830}
49
+ {"user": "roxky", "model": "flux-schnell", "provider": "HuggingFace", "prompt_tokens": 9, "completion_tokens": 76, "total_tokens": 85}
50
+ {"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 9, "completion_tokens": 12, "total_tokens": 21}
51
+ {"user": "roxky", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 10, "prompt_tokens": 153, "total_tokens": 163}
52
+ {"user": "roxky", "model": "flux-dev", "provider": "HuggingSpace", "prompt_tokens": 9, "completion_tokens": 212, "total_tokens": 221}
53
+ {"user": "roxky", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
54
+ {"user": "roxky", "model": "sonar-pro", "provider": "PerplexityLabs", "prompt_tokens": 9, "completion_tokens": 27, "total_tokens": 36}
55
+ {"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 9, "completion_tokens": 53, "total_tokens": 62}
56
+ {"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
57
+ {"user": "ballm4n", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 9, "completion_tokens": 9, "total_tokens": 18}
58
+ {"user": "NexThinkLabs", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 8, "total_tokens": 16}
59
+ {"user": "NexThinkLabs", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 25, "completion_tokens": 24, "total_tokens": 49}
60
+ {"user": "NexThinkLabs", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 8, "total_tokens": 16}
61
+ {"user": "roxky", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 10, "completion_tokens": 10, "total_tokens": 20}
62
+ {"user": "roxky", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 37, "completion_tokens": 106, "total_tokens": 143}
63
+ {"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 37, "completion_tokens": 376, "total_tokens": 413}
64
+ {"user": "roxky", "model": "sonar-pro", "provider": "PerplexityLabs", "prompt_tokens": 37, "completion_tokens": 154, "total_tokens": 191}
65
+ {"user": "MAKERBOT", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 9, "completion_tokens": 16, "total_tokens": 25}
66
+ {"user": "XannX", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 11, "completion_tokens": 608, "total_tokens": 619}
67
+ {"user": "MAKERBOT", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 56, "completion_tokens": 114, "total_tokens": 170}
68
+ {"user": "MAKERBOT", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 155, "completion_tokens": 636, "total_tokens": 791}
69
+ {"user": "XannX", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 569, "completion_tokens": 3049, "total_tokens": 3618}
70
+ {"user": "dgdgdsgs", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 24, "total_tokens": 32}
71
+ {"user": "dgdgdsgs", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 185, "completion_tokens": 368, "total_tokens": 553}
72
+ {"user": "roxky", "model": "sonar-pro", "provider": "PerplexityLabs", "prompt_tokens": 10, "completion_tokens": 21, "total_tokens": 31}
73
+ {"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 1490, "completion_tokens": 2074, "total_tokens": 3564}
74
+ {"user": "roxky", "model": "claude-3-haiku", "provider": "DDG", "prompt_tokens": 9, "completion_tokens": 9, "total_tokens": 18}
75
+ {"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 8, "completion_tokens": 14, "total_tokens": 22}
76
+ {"user": "roxky", "model": "flux", "provider": "Blackbox", "prompt_tokens": 8, "completion_tokens": 112, "total_tokens": 120}
77
+ {"user": "roxky", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
78
+ {"user": "KA4E6", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 12, "total_tokens": 20}
79
+ {"user": "Majidgdr", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 13, "completion_tokens": 13, "total_tokens": 26}
80
+ {"user": "Majidgdr", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 39, "completion_tokens": 71, "total_tokens": 110}
81
+ {"user": "Majidgdr", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 121, "completion_tokens": 70, "total_tokens": 191}
82
+ {"user": "Majidgdr", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 187, "completion_tokens": 6, "total_tokens": 193}
83
+ {"user": "Jackolord", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 10, "prompt_tokens": 152, "total_tokens": 162}
84
+ {"user": "Jackolord", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 27, "completion_tokens": 129, "total_tokens": 156}
85
+ {"user": "Jackolord", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 37, "completion_tokens": 361, "total_tokens": 398}
86
+ {"user": "Jackolord", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 342, "completion_tokens": 465, "total_tokens": 807}
87
+ {"user": "Jackolord", "model": "o1-preview", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
88
+ {"user": "Jackolord", "model": "o1-preview", "provider": "OpenaiAccount", "prompt_tokens": 36, "completion_tokens": 45, "total_tokens": 81}
89
+ {"user": "Jackolord", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 17, "completion_tokens": 90, "total_tokens": 107}
90
+ {"user": "Jackolord", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 17, "completion_tokens": 53, "total_tokens": 70}
91
+ {"user": "Jackolord", "model": "sonar-pro", "provider": "PerplexityLabs", "prompt_tokens": 17, "completion_tokens": 14, "total_tokens": 31}
92
+ {"user": "Jackolord", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 17, "completion_tokens": 54, "total_tokens": 71}
93
+ {"user": "Jackolord", "model": "dall-e-3", "provider": "CopilotAccount", "prompt_tokens": 12, "completion_tokens": 157, "total_tokens": 169}
94
+ {"user": "Jackolord", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 12, "completion_tokens": 152, "total_tokens": 164}
95
+ {"user": "Jackolord", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 12, "completion_tokens": 616, "total_tokens": 628}
96
+ {"user": "NDeriy", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 110, "completion_tokens": 84, "total_tokens": 194}
97
+ {"user": "NDeriy", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 302, "completion_tokens": 249, "total_tokens": 551}
98
+ {"user": "NDeriy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 739, "completion_tokens": 825, "total_tokens": 1564}
99
+ {"user": "NDeriy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 1610, "completion_tokens": 647, "total_tokens": 2257}
100
+ {"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 17, "completion_tokens": 57, "total_tokens": 74}
101
+ {"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 87, "completion_tokens": 66, "total_tokens": 153}
102
+ {"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 167, "completion_tokens": 121, "total_tokens": 288}
103
+ {"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 303, "completion_tokens": 230, "total_tokens": 533}
104
+ {"user": "NDeriy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 2353, "completion_tokens": 1051, "total_tokens": 3404}
105
+ {"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 25, "completion_tokens": 183, "total_tokens": 208}
106
+ {"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 322, "completion_tokens": 217, "total_tokens": 539}
107
+ {"user": "NDeriy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 3416, "completion_tokens": 687, "total_tokens": 4103}
108
+ {"user": "Intgrr", "model": "o1-preview", "provider": "OpenaiAccount", "prompt_tokens": 689, "completion_tokens": 283, "total_tokens": 972}
109
+ {"user": "Intgrr", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 1144, "completion_tokens": 260, "total_tokens": 1404}
110
+ {"user": "NDeriy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 4223, "completion_tokens": 1023, "total_tokens": 5246}
111
+ {"user": "macyn", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 15, "completion_tokens": 95, "total_tokens": 110}
112
+ {"user": "macyn", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 14, "completion_tokens": 101, "total_tokens": 115}
113
+ {"user": "NDeriy", "model": "qwq-32b", "provider": "HuggingFace", "prompt_tokens": 5262, "completion_tokens": 286, "total_tokens": 5548}
114
+ {"user": "macyn", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 15, "completion_tokens": 740, "total_tokens": 755}
115
+ {"user": "macyn", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 19, "completion_tokens": 226, "total_tokens": 245}
116
+ {"user": "macyn", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 19, "completion_tokens": 327, "total_tokens": 346}
117
+ {"user": "roxky", "model": "claude-3-haiku", "provider": "DDG", "prompt_tokens": 9, "completion_tokens": 9, "total_tokens": 18}
118
+ {"user": "roxky", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 1930, "completion_tokens": 300, "total_tokens": 2230}
119
+ {"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 1930, "completion_tokens": 1026, "total_tokens": 2956}
120
+ {"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 3265, "completion_tokens": 56, "total_tokens": 3321}
121
+ {"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 1930, "completion_tokens": 1026, "total_tokens": 2956}
122
+ {"user": "Erkaperka502", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 83, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 169, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 252}
123
+ {"user": "h4mid007", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 12, "total_tokens": 20}
124
+ {"user": "h4mid007", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 29, "completion_tokens": 10, "total_tokens": 39}
125
+ {"user": "h4mid007", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 53, "completion_tokens": 25, "total_tokens": 78}
126
+ {"user": "h4mid007", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 87, "completion_tokens": 294, "total_tokens": 381}
har_and_cookies/.usage/2025-02-01.jsonl ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {"user": "amsepehrian", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 13, "completion_tokens": 233, "total_tokens": 246}
2
+ {"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 15, "completion_tokens": 12, "total_tokens": 27}
3
+ {"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 15, "completion_tokens": 7, "total_tokens": 22}
4
+ {"user": "roxky", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 15, "completion_tokens": 177, "total_tokens": 192}
5
+ {"user": "roxky", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 178, "completion_tokens": 258, "total_tokens": 436}
6
+ {"user": "amsepehrian", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 80, "total_tokens": 88}
7
+ {"user": "amsepehrian", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 15, "completion_tokens": 371, "total_tokens": 386}
8
+ {"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 178, "completion_tokens": 209, "total_tokens": 387}
9
+ {"user": "roxky", "model": "dall-e-3", "provider": "CopilotAccount", "prompt_tokens": 178, "completion_tokens": 234, "total_tokens": 412}