Spaces:
Sleeping
Sleeping
Commit new images and usage data
Browse files- README.md +4 -0
- app.py +5 -213
- demo/BackendApi.py +146 -0
- generated_images/1738296597_what+is+best+ai+model+yet+so+far,+compare+and+tell+me+final+result+only.jpg +0 -0
- generated_images/1738332473_gpt-4o.A+modern+commercial+airplane+soaring+through+a+clear+blue+sky,+with+white+contrails+streaming+behind.jpg +0 -0
- generated_images/1738347359_dall-e-3.Ask+Blackbox+AI+Anything.jpg +0 -0
- generated_images/1738347377_flux-pro.Ask+Blackbox+AI+Anything.jpg +0 -0
- generated_images/1738371043_flux-pro.Create+a+image:+A+futuristic+cityscape+at+sunset,+with+towering+skyscrapers+and+flying+cars+zipping+.jpg +0 -0
- generated_images/1738372570_Copilot.A+futuristic+cityscape+at+sunset,+with+towering+skyscrapers+and+flying+cars+zipping+by.jpg +0 -0
- generated_images/1738372600_dall-e-3.A+futuristic+cityscape+at+sunset,+with+towering+skyscrapers+and+flying+cars+zipping+by.jpg +0 -0
- har_and_cookies/.usage/2025-01-30.jsonl +53 -0
- har_and_cookies/.usage/2025-01-31.jsonl +126 -0
- har_and_cookies/.usage/2025-02-01.jsonl +9 -0
README.md
CHANGED
@@ -5,4 +5,8 @@ colorFrom: green
|
|
5 |
colorTo: purple
|
6 |
sdk: docker
|
7 |
pinned: false
|
|
|
|
|
|
|
|
|
8 |
---
|
|
|
5 |
colorTo: purple
|
6 |
sdk: docker
|
7 |
pinned: false
|
8 |
+
|
9 |
+
hf_oauth: true
|
10 |
+
hf_oauth_scopes:
|
11 |
+
- inference-api
|
12 |
---
|
app.py
CHANGED
@@ -1,221 +1,23 @@
|
|
1 |
import os
|
2 |
-
import json
|
3 |
-
from g4f.providers.response import Reasoning, JsonConversation, FinishReason
|
4 |
-
from g4f.typing import AsyncResult, Messages
|
5 |
-
import json
|
6 |
-
import re
|
7 |
-
import time
|
8 |
-
import logging
|
9 |
-
from urllib.parse import quote_plus
|
10 |
from fastapi import FastAPI, Response, Request
|
11 |
from fastapi.responses import RedirectResponse
|
12 |
|
13 |
-
from g4f.image import images_dir
|
14 |
-
|
15 |
import g4f.api
|
16 |
import g4f.Provider
|
17 |
|
18 |
-
from g4f.Provider.base_provider import AsyncGeneratorProvider, ProviderModelMixin
|
19 |
-
from g4f.typing import AsyncResult, Messages
|
20 |
-
from g4f.requests import StreamSession
|
21 |
-
from g4f.providers.response import ProviderInfo, JsonConversation, PreviewResponse, SynthesizeData, TitleGeneration, RequestLogin
|
22 |
-
from g4f.providers.response import Parameters, FinishReason, Usage, Reasoning
|
23 |
-
from g4f.errors import ModelNotSupportedError
|
24 |
-
from g4f import debug
|
25 |
-
|
26 |
import demo
|
27 |
-
|
28 |
-
class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
|
29 |
-
url = "https://ahe.hopto.org"
|
30 |
-
working = True
|
31 |
-
ssl = False
|
32 |
-
|
33 |
-
models = [
|
34 |
-
*g4f.Provider.OpenaiAccount.get_models(),
|
35 |
-
*g4f.Provider.PerplexityLabs.get_models(),
|
36 |
-
"flux",
|
37 |
-
"flux-pro",
|
38 |
-
"MiniMax-01",
|
39 |
-
"Microsoft Copilot",
|
40 |
-
]
|
41 |
-
|
42 |
-
@classmethod
|
43 |
-
def get_model(cls, model):
|
44 |
-
if "MiniMax" in model:
|
45 |
-
model = "MiniMax"
|
46 |
-
elif "Copilot" in model:
|
47 |
-
model = "Copilot"
|
48 |
-
elif "FLUX" in model:
|
49 |
-
model = f"flux-{model.split('-')[-1]}"
|
50 |
-
elif "flux" in model:
|
51 |
-
model = model.split(' ')[-1]
|
52 |
-
elif model in g4f.Provider.OpenaiAccount.get_models():
|
53 |
-
pass
|
54 |
-
elif model in g4f.Provider.PerplexityLabs.get_models():
|
55 |
-
pass
|
56 |
-
else:
|
57 |
-
raise ModelNotSupportedError(f"Model: {model}")
|
58 |
-
return model
|
59 |
-
|
60 |
-
@classmethod
|
61 |
-
def get_provider(cls, model):
|
62 |
-
if model.startswith("MiniMax"):
|
63 |
-
return "HailuoAI"
|
64 |
-
elif model == "Copilot" or "dall-e" in model:
|
65 |
-
return "CopilotAccount"
|
66 |
-
elif model in g4f.Provider.OpenaiAccount.get_models():
|
67 |
-
return "OpenaiAccount"
|
68 |
-
elif model in g4f.Provider.PerplexityLabs.get_models():
|
69 |
-
return "PerplexityLabs"
|
70 |
-
return None
|
71 |
-
|
72 |
-
@classmethod
|
73 |
-
async def create_async_generator(
|
74 |
-
cls,
|
75 |
-
model: str,
|
76 |
-
messages: Messages,
|
77 |
-
api_key: str = None,
|
78 |
-
proxy: str = None,
|
79 |
-
timeout: int = 0,
|
80 |
-
**kwargs
|
81 |
-
) -> AsyncResult:
|
82 |
-
debug.log(f"{__name__}: {api_key}")
|
83 |
-
if "dall-e" in model and "prompt" not in kwargs:
|
84 |
-
kwargs["prompt"] = messages[-1]["content"]
|
85 |
-
messages[-1]["content"] = f"Generate a image: {kwargs['prompt']}"
|
86 |
-
|
87 |
-
async with StreamSession(
|
88 |
-
proxy=proxy,
|
89 |
-
headers={"Accept": "text/event-stream", **demo.headers},
|
90 |
-
timeout=timeout
|
91 |
-
) as session:
|
92 |
-
model = cls.get_model(model)
|
93 |
-
provider = cls.get_provider(model)
|
94 |
-
async with session.post(f"{cls.url}/backend-api/v2/conversation", json={
|
95 |
-
**kwargs,
|
96 |
-
"model": model,
|
97 |
-
"messages": messages,
|
98 |
-
"provider": provider
|
99 |
-
}, ssl=cls.ssl) as response:
|
100 |
-
is_thinking = 0
|
101 |
-
async for line in response.iter_lines():
|
102 |
-
response.raise_for_status()
|
103 |
-
data = json.loads(line)
|
104 |
-
data_type = data.pop("type")
|
105 |
-
if data_type == "provider":
|
106 |
-
yield ProviderInfo(**data[data_type])
|
107 |
-
provider = data[data_type]["name"]
|
108 |
-
elif data_type == "conversation":
|
109 |
-
yield JsonConversation(**data[data_type][provider] if provider in data[data_type] else data[data_type][""])
|
110 |
-
elif data_type == "conversation_id":
|
111 |
-
pass
|
112 |
-
elif data_type == "message":
|
113 |
-
yield Exception(data)
|
114 |
-
elif data_type == "preview":
|
115 |
-
yield PreviewResponse(data[data_type])
|
116 |
-
elif data_type == "content":
|
117 |
-
def on_image(match):
|
118 |
-
extension = match.group(3).split(".")[-1].split("?")[0]
|
119 |
-
extension = "" if not extension or len(extension) > 4 else f".{extension}"
|
120 |
-
filename = f"{int(time.time())}_{quote_plus(match.group(1)[:100], '')}{extension}"
|
121 |
-
download_url = f"/download/{filename}?url={cls.url}{match.group(3)}"
|
122 |
-
return f"[![{match.group(1)}]({download_url})](/images/{filename})"
|
123 |
-
if "<think>" in data[data_type]:
|
124 |
-
data[data_type] = data[data_type].split("<think>", 1)
|
125 |
-
yield data[data_type][0]
|
126 |
-
yield Reasoning(data[data_type][1])
|
127 |
-
yield Reasoning(None, "Is thinking...")
|
128 |
-
is_thinking = time.time()
|
129 |
-
if "</think>" in data[data_type][1]:
|
130 |
-
data[data_type][1] = data[data_type].split("</think>", 1)
|
131 |
-
yield Reasoning(data[data_type][0])
|
132 |
-
yield Reasoning(None, f"Finished in {round(time.time()-is_thinking, 2)} seconds")
|
133 |
-
yield data[data_type][1]
|
134 |
-
is_thinking = 0
|
135 |
-
elif is_thinking:
|
136 |
-
yield Reasoning(data[data_type])
|
137 |
-
else:
|
138 |
-
yield re.sub(r'\[\!\[(.+?)\]\(([^)]+?)\)\]\(([^)]+?)\)', on_image, data[data_type])
|
139 |
-
elif data_type =="synthesize":
|
140 |
-
yield SynthesizeData(**data[data_type])
|
141 |
-
elif data_type == "parameters":
|
142 |
-
yield Parameters(**data[data_type])
|
143 |
-
elif data_type == "usage":
|
144 |
-
yield Usage(**data[data_type])
|
145 |
-
elif data_type == "reasoning":
|
146 |
-
yield Reasoning(**data)
|
147 |
-
elif data_type == "login":
|
148 |
-
pass
|
149 |
-
elif data_type == "title":
|
150 |
-
yield TitleGeneration(data[data_type])
|
151 |
-
elif data_type == "finish":
|
152 |
-
yield FinishReason(data[data_type]["reason"])
|
153 |
-
elif data_type == "log":
|
154 |
-
debug.log(data[data_type])
|
155 |
-
else:
|
156 |
-
debug.log(f"Unknown data: ({data_type}) {data}")
|
157 |
-
|
158 |
g4f.Provider.__map__["Feature"] = BackendApi
|
159 |
|
160 |
-
import asyncio
|
161 |
-
import uuid
|
162 |
-
from aiohttp import ClientSession, ClientError
|
163 |
-
from g4f.typing import Optional, Cookies
|
164 |
-
from g4f.image import is_accepted_format
|
165 |
|
166 |
-
|
167 |
-
|
168 |
-
cookies: Optional[Cookies] = None,
|
169 |
-
headers: dict = None,
|
170 |
-
proxy: Optional[str] = None,
|
171 |
-
add_url: bool = True,
|
172 |
-
target: str = None,
|
173 |
-
ssl: bool = None
|
174 |
-
) -> list[str]:
|
175 |
-
if add_url:
|
176 |
-
add_url = not cookies
|
177 |
-
#ensure_images_dir()
|
178 |
-
async with ClientSession(
|
179 |
-
#connector=get_connector(proxy=proxy),
|
180 |
-
cookies=cookies,
|
181 |
-
headers=headers,
|
182 |
-
) as session:
|
183 |
-
async def copy_image(image: str, target: str = None) -> str:
|
184 |
-
if target is None or len(images) > 1:
|
185 |
-
target = os.path.join(images_dir, f"{int(time.time())}_{str(uuid.uuid4())}")
|
186 |
-
try:
|
187 |
-
if image.startswith("data:"):
|
188 |
-
pass
|
189 |
-
#with open(target, "wb") as f:
|
190 |
-
# f.write(extract_data_uri(image))
|
191 |
-
else:
|
192 |
-
try:
|
193 |
-
async with session.get(image, proxy=proxy, ssl=ssl) as response:
|
194 |
-
response.raise_for_status()
|
195 |
-
with open(target, "wb") as f:
|
196 |
-
async for chunk in response.content.iter_chunked(4096):
|
197 |
-
f.write(chunk)
|
198 |
-
except ClientError as e:
|
199 |
-
debug.log(f"copy_images failed: {e.__class__.__name__}: {e}")
|
200 |
-
return image
|
201 |
-
if "." not in target:
|
202 |
-
with open(target, "rb") as f:
|
203 |
-
extension = is_accepted_format(f.read(12)).split("/")[-1]
|
204 |
-
extension = "jpg" if extension == "jpeg" else extension
|
205 |
-
new_target = f"{target}.{extension}"
|
206 |
-
os.rename(target, new_target)
|
207 |
-
target = new_target
|
208 |
-
finally:
|
209 |
-
if "." not in target and os.path.exists(target):
|
210 |
-
os.unlink(target)
|
211 |
-
return f"/images/{os.path.basename(target)}{'?url=' + image if add_url and not image.startswith('data:') else ''}"
|
212 |
-
|
213 |
-
return await asyncio.gather(*[copy_image(image, target) for image in images])
|
214 |
|
215 |
def create_app():
|
216 |
g4f.debug.logging = True
|
217 |
g4f.api.AppConfig.gui = True
|
218 |
-
g4f.api.AppConfig.demo =
|
219 |
|
220 |
app = FastAPI()
|
221 |
|
@@ -259,14 +61,4 @@ def create_app():
|
|
259 |
|
260 |
return app
|
261 |
|
262 |
-
|
263 |
-
class NoHomeFilter(logging.Filter):
|
264 |
-
def filter(self, record):
|
265 |
-
if '"GET / HTTP/1.1" 200 OK' in record.getMessage():
|
266 |
-
return False
|
267 |
-
if '"GET /static/' in record.getMessage():
|
268 |
-
return False
|
269 |
-
return True
|
270 |
-
logging.getLogger("uvicorn.access").addFilter(NoHomeFilter())
|
271 |
-
|
272 |
app = create_app()
|
|
|
1 |
import os
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
from fastapi import FastAPI, Response, Request
|
3 |
from fastapi.responses import RedirectResponse
|
4 |
|
5 |
+
from g4f.image import images_dir
|
|
|
6 |
import g4f.api
|
7 |
import g4f.Provider
|
8 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
9 |
import demo
|
10 |
+
from demo.BackendApi import BackendApi
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
g4f.Provider.__map__["Feature"] = BackendApi
|
12 |
|
|
|
|
|
|
|
|
|
|
|
13 |
|
14 |
+
g4f.models.demo_models
|
15 |
+
from g4f.image import copy_images
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
|
17 |
def create_app():
|
18 |
g4f.debug.logging = True
|
19 |
g4f.api.AppConfig.gui = True
|
20 |
+
g4f.api.AppConfig.demo = True
|
21 |
|
22 |
app = FastAPI()
|
23 |
|
|
|
61 |
|
62 |
return app
|
63 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
64 |
app = create_app()
|
demo/BackendApi.py
ADDED
@@ -0,0 +1,146 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
import json
|
2 |
+
import re
|
3 |
+
import time
|
4 |
+
from urllib.parse import quote_plus
|
5 |
+
|
6 |
+
import g4f.api
|
7 |
+
import g4f.Provider
|
8 |
+
|
9 |
+
from g4f.Provider.base_provider import AsyncGeneratorProvider, ProviderModelMixin
|
10 |
+
from g4f.typing import AsyncResult, Messages
|
11 |
+
from g4f.requests import StreamSession
|
12 |
+
from g4f.providers.response import *
|
13 |
+
from g4f.errors import ModelNotSupportedError
|
14 |
+
from g4f import debug
|
15 |
+
from . import headers
|
16 |
+
|
17 |
+
class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
|
18 |
+
url = "https://ahe.hopto.org"
|
19 |
+
working = True
|
20 |
+
ssl = False
|
21 |
+
|
22 |
+
models = [
|
23 |
+
*g4f.Provider.OpenaiAccount.get_models(),
|
24 |
+
*g4f.Provider.PerplexityLabs.get_models(),
|
25 |
+
"flux",
|
26 |
+
"flux-pro",
|
27 |
+
"MiniMax-01",
|
28 |
+
"Microsoft Copilot",
|
29 |
+
]
|
30 |
+
|
31 |
+
@classmethod
|
32 |
+
def get_model(cls, model):
|
33 |
+
if "MiniMax" in model:
|
34 |
+
model = "MiniMax"
|
35 |
+
elif "Copilot" in model:
|
36 |
+
model = "Copilot"
|
37 |
+
elif "FLUX" in model:
|
38 |
+
model = f"flux-{model.split('-')[-1]}"
|
39 |
+
elif "flux" in model:
|
40 |
+
model = model.split(' ')[-1]
|
41 |
+
elif model in g4f.Provider.OpenaiAccount.get_models():
|
42 |
+
pass
|
43 |
+
elif model in g4f.Provider.PerplexityLabs.get_models():
|
44 |
+
pass
|
45 |
+
else:
|
46 |
+
raise ModelNotSupportedError(f"Model: {model}")
|
47 |
+
return model
|
48 |
+
|
49 |
+
@classmethod
|
50 |
+
def get_provider(cls, model):
|
51 |
+
if model.startswith("MiniMax"):
|
52 |
+
return "HailuoAI"
|
53 |
+
elif model == "Copilot" or "dall-e" in model:
|
54 |
+
return "CopilotAccount"
|
55 |
+
elif model in g4f.Provider.OpenaiAccount.get_models():
|
56 |
+
return "OpenaiAccount"
|
57 |
+
elif model in g4f.Provider.PerplexityLabs.get_models():
|
58 |
+
return "PerplexityLabs"
|
59 |
+
return None
|
60 |
+
|
61 |
+
@classmethod
|
62 |
+
async def create_async_generator(
|
63 |
+
cls,
|
64 |
+
model: str,
|
65 |
+
messages: Messages,
|
66 |
+
api_key: str = None,
|
67 |
+
proxy: str = None,
|
68 |
+
timeout: int = 0,
|
69 |
+
**kwargs
|
70 |
+
) -> AsyncResult:
|
71 |
+
debug.log(f"{cls.__name__}: {api_key}")
|
72 |
+
|
73 |
+
if "dall-e" in model and "prompt" not in kwargs:
|
74 |
+
kwargs["prompt"] = messages[-1]["content"]
|
75 |
+
messages[-1]["content"] = f"Generate a image: {kwargs['prompt']}"
|
76 |
+
|
77 |
+
async with StreamSession(
|
78 |
+
proxy=proxy,
|
79 |
+
headers={"Accept": "text/event-stream", **headers},
|
80 |
+
timeout=timeout
|
81 |
+
) as session:
|
82 |
+
model = cls.get_model(model)
|
83 |
+
provider = cls.get_provider(model)
|
84 |
+
async with session.post(f"{cls.url}/backend-api/v2/conversation", json={
|
85 |
+
**kwargs,
|
86 |
+
"model": model,
|
87 |
+
"messages": messages,
|
88 |
+
"provider": provider
|
89 |
+
}, ssl=cls.ssl) as response:
|
90 |
+
is_thinking = 0
|
91 |
+
async for line in response.iter_lines():
|
92 |
+
response.raise_for_status()
|
93 |
+
data = json.loads(line)
|
94 |
+
data_type = data.pop("type")
|
95 |
+
if data_type == "provider":
|
96 |
+
yield ProviderInfo(**data[data_type])
|
97 |
+
provider = data[data_type]["name"]
|
98 |
+
elif data_type == "conversation":
|
99 |
+
yield JsonConversation(**data[data_type][provider] if provider in data[data_type] else data[data_type][""])
|
100 |
+
elif data_type == "conversation_id":
|
101 |
+
pass
|
102 |
+
elif data_type == "message":
|
103 |
+
yield Exception(data)
|
104 |
+
elif data_type == "preview":
|
105 |
+
yield PreviewResponse(data[data_type])
|
106 |
+
elif data_type == "content":
|
107 |
+
def on_image(match):
|
108 |
+
extension = match.group(3).split(".")[-1].replace("%3F", "?").split("?")[0]
|
109 |
+
extension = ".jpg" if not extension or len(extension) > 4 else f".{extension}"
|
110 |
+
filename = f"{int(time.time())}_{quote_plus(model, '')}.{quote_plus(match.group(1)[:100], '')}{extension}"
|
111 |
+
download_url = f"/download/{filename}?url={cls.url}{match.group(3)}"
|
112 |
+
return f"[![{match.group(1)}]({download_url})](/images/{filename})"
|
113 |
+
if "<think>" in data[data_type]:
|
114 |
+
data[data_type] = data[data_type].split("<think>", 1)
|
115 |
+
yield data[data_type][0]
|
116 |
+
yield Reasoning(data[data_type][1])
|
117 |
+
yield Reasoning(None, "Is thinking...")
|
118 |
+
is_thinking = time.time()
|
119 |
+
if "</think>" in data[data_type]:
|
120 |
+
data[data_type][1] = data[data_type].split("</think>", 1)
|
121 |
+
yield Reasoning(data[data_type][0])
|
122 |
+
yield Reasoning(None, f"Finished in {round(time.time()-is_thinking, 2)} seconds")
|
123 |
+
yield data[data_type][1]
|
124 |
+
is_thinking = 0
|
125 |
+
elif is_thinking:
|
126 |
+
yield Reasoning(data[data_type])
|
127 |
+
else:
|
128 |
+
yield re.sub(r'\[\!\[(.+?)\]\(([^)]+?)\)\]\(([^)]+?)\)', on_image, data[data_type])
|
129 |
+
elif data_type =="synthesize":
|
130 |
+
yield SynthesizeData(**data[data_type])
|
131 |
+
elif data_type == "parameters":
|
132 |
+
yield Parameters(**data[data_type])
|
133 |
+
elif data_type == "usage":
|
134 |
+
yield Usage(**data[data_type])
|
135 |
+
elif data_type == "reasoning":
|
136 |
+
yield Reasoning(**data)
|
137 |
+
elif data_type == "login":
|
138 |
+
pass
|
139 |
+
elif data_type == "title":
|
140 |
+
yield TitleGeneration(data[data_type])
|
141 |
+
elif data_type == "finish":
|
142 |
+
yield FinishReason(data[data_type]["reason"])
|
143 |
+
elif data_type == "log":
|
144 |
+
debug.log(data[data_type])
|
145 |
+
else:
|
146 |
+
debug.log(f"Unknown data: ({data_type}) {data}")
|
generated_images/1738296597_what+is+best+ai+model+yet+so+far,+compare+and+tell+me+final+result+only.jpg
ADDED
![]() |
generated_images/1738332473_gpt-4o.A+modern+commercial+airplane+soaring+through+a+clear+blue+sky,+with+white+contrails+streaming+behind.jpg
ADDED
![]() |
generated_images/1738347359_dall-e-3.Ask+Blackbox+AI+Anything.jpg
ADDED
![]() |
generated_images/1738347377_flux-pro.Ask+Blackbox+AI+Anything.jpg
ADDED
![]() |
generated_images/1738371043_flux-pro.Create+a+image:+A+futuristic+cityscape+at+sunset,+with+towering+skyscrapers+and+flying+cars+zipping+.jpg
ADDED
![]() |
generated_images/1738372570_Copilot.A+futuristic+cityscape+at+sunset,+with+towering+skyscrapers+and+flying+cars+zipping+by.jpg
ADDED
![]() |
generated_images/1738372600_dall-e-3.A+futuristic+cityscape+at+sunset,+with+towering+skyscrapers+and+flying+cars+zipping+by.jpg
ADDED
![]() |
har_and_cookies/.usage/2025-01-30.jsonl
ADDED
@@ -0,0 +1,53 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{"user": "roxky", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 8, "completion_tokens": 115, "total_tokens": 123}
|
2 |
+
{"user": "roxky", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 18, "completion_tokens": 10, "total_tokens": 28}
|
3 |
+
{"user": "roxky", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 38, "completion_tokens": 10, "total_tokens": 48}
|
4 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 58, "completion_tokens": 12, "total_tokens": 70}
|
5 |
+
{"user": "roxky", "model": "flux-schnell", "provider": "HuggingFace", "prompt_tokens": 8, "completion_tokens": 85, "total_tokens": 93}
|
6 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 10, "completion_tokens": 68, "total_tokens": 78}
|
7 |
+
{"user": "roxky", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 11, "total_tokens": 19}
|
8 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
|
9 |
+
{"user": "roxky", "model": "flux-schnell", "provider": "HuggingFace", "prompt_tokens": 9, "completion_tokens": 76, "total_tokens": 85}
|
10 |
+
{"user": "RubielLabarta", "model": "flux-dev", "provider": "HuggingSpace", "prompt_tokens": 17, "completion_tokens": 220, "total_tokens": 237}
|
11 |
+
{"user": "RubielLabarta", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 8, "total_tokens": 16}
|
12 |
+
{"user": "RubielLabarta", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 28, "completion_tokens": 25, "total_tokens": 53}
|
13 |
+
{"user": "RubielLabarta", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 60, "completion_tokens": 28, "total_tokens": 88}
|
14 |
+
{"user": "RubielLabarta", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 60, "completion_tokens": 28, "total_tokens": 88}
|
15 |
+
{"user": "RubielLabarta", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 60, "completion_tokens": 83, "total_tokens": 143}
|
16 |
+
{"user": "roxky", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
|
17 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 236, "total_tokens": 244}
|
18 |
+
{"user": "roxky", "model": "flux", "provider": "PollinationsAI", "prompt_tokens": 8, "completion_tokens": 117, "total_tokens": 125}
|
19 |
+
{"user": "roxky", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 8, "completion_tokens": 123, "total_tokens": 131}
|
20 |
+
{"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 8, "completion_tokens": 32, "total_tokens": 40}
|
21 |
+
{"user": "sheeee2222", "model": "qwen-2-72b", "provider": "HuggingSpace", "prompt_tokens": 15, "completion_tokens": 17, "total_tokens": 32}
|
22 |
+
{"user": "sheeee2222", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 10, "prompt_tokens": 174, "total_tokens": 184}
|
23 |
+
{"user": "sheeee2222", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 41, "prompt_tokens": 213, "total_tokens": 254}
|
24 |
+
{"user": "sheeee2222", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 566, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 3531, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 4097}
|
25 |
+
{"user": "roxky", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 11, "total_tokens": 20}
|
26 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
|
27 |
+
{"user": "roxky", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 7, "total_tokens": 16}
|
28 |
+
{"user": "roxky", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 6, "total_tokens": 15}
|
29 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 20, "completion_tokens": 15, "total_tokens": 35}
|
30 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 46, "completion_tokens": 73, "total_tokens": 119}
|
31 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 126, "completion_tokens": 124, "total_tokens": 250}
|
32 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 246, "completion_tokens": 22, "total_tokens": 268}
|
33 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 274, "completion_tokens": 63, "total_tokens": 337}
|
34 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 342, "completion_tokens": 12, "total_tokens": 354}
|
35 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 359, "completion_tokens": 41, "total_tokens": 400}
|
36 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 420, "completion_tokens": 120, "total_tokens": 540}
|
37 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 420, "completion_tokens": 120, "total_tokens": 540}
|
38 |
+
{"user": "cihairottoilcazzo", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 18, "completion_tokens": 261, "total_tokens": 279}
|
39 |
+
{"user": "sanch1tx", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 8, "completion_tokens": 17, "total_tokens": 25}
|
40 |
+
{"user": "nasrulam", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 19, "completion_tokens": 1225, "total_tokens": 1244}
|
41 |
+
{"user": "nasrulam", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 19, "completion_tokens": 69, "total_tokens": 88}
|
42 |
+
{"user": "nasrulam", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 35, "completion_tokens": 1442, "total_tokens": 1477}
|
43 |
+
{"user": "nasrulam", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 35, "completion_tokens": 298, "total_tokens": 333}
|
44 |
+
{"user": "nasrulam", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 1574, "completion_tokens": 136, "total_tokens": 1710}
|
45 |
+
{"user": "sfewewf232", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 87, "completion_tokens": 160, "total_tokens": 247}
|
46 |
+
{"user": "hashim1", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 13, "completion_tokens": 9, "total_tokens": 22}
|
47 |
+
{"user": "hashim1", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 36, "completion_tokens": 20, "total_tokens": 56}
|
48 |
+
{"user": "roxky", "model": "flux-dev", "provider": "HuggingSpace", "prompt_tokens": 30, "completion_tokens": 200, "total_tokens": 230}
|
49 |
+
{"user": "roxky", "model": "flux-schnell", "provider": "HuggingFace", "prompt_tokens": 30, "completion_tokens": 80, "total_tokens": 110}
|
50 |
+
{"user": "roxky", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 9, "completion_tokens": 353, "total_tokens": 362}
|
51 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
|
52 |
+
{"user": "roxky", "model": "gpt-4o-mini", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 7, "total_tokens": 16}
|
53 |
+
{"user": "roxky", "model": "claude-3-haiku", "provider": "DDG", "prompt_tokens": 2447, "completion_tokens": 798, "total_tokens": 3245}
|
har_and_cookies/.usage/2025-01-31.jsonl
ADDED
@@ -0,0 +1,126 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{"user": "roxky", "model": "qvq-72b", "provider": "HuggingSpace", "prompt_tokens": 15, "completion_tokens": 10, "total_tokens": 25}
|
2 |
+
{"user": "roxky", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 16, "completion_tokens": 1904, "total_tokens": 1920}
|
3 |
+
{"user": "roxky", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 1490, "completion_tokens": 1736, "total_tokens": 3226}
|
4 |
+
{"user": "tanu360", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 8, "completion_tokens": 105, "total_tokens": 113}
|
5 |
+
{"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
|
6 |
+
{"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 41, "completion_tokens": 240, "total_tokens": 281}
|
7 |
+
{"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 41, "completion_tokens": 440, "total_tokens": 481}
|
8 |
+
{"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 41, "completion_tokens": 393, "total_tokens": 434}
|
9 |
+
{"user": "bbdwbayu", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 694, "completion_tokens": 561, "total_tokens": 1255}
|
10 |
+
{"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 1273, "completion_tokens": 44, "total_tokens": 1317}
|
11 |
+
{"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 1329, "completion_tokens": 166, "total_tokens": 1495}
|
12 |
+
{"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 1508, "completion_tokens": 206, "total_tokens": 1714}
|
13 |
+
{"user": "JoniSenior", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 9, "completion_tokens": 299, "total_tokens": 308}
|
14 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 9, "completion_tokens": 448, "total_tokens": 457}
|
15 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 476, "completion_tokens": 1010, "total_tokens": 1486}
|
16 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 70, "completion_tokens": 1075, "total_tokens": 1145}
|
17 |
+
{"user": "BruceStudios", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 9, "completion_tokens": 9, "total_tokens": 18}
|
18 |
+
{"user": "BruceStudios", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 29, "completion_tokens": 130, "total_tokens": 159}
|
19 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 1223, "completion_tokens": 1606, "total_tokens": 2829}
|
20 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 2863, "completion_tokens": 1563, "total_tokens": 4426}
|
21 |
+
{"user": "herokulastly", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 19, "completion_tokens": 10, "total_tokens": 29}
|
22 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 4473, "completion_tokens": 1377, "total_tokens": 5850}
|
23 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 5862, "completion_tokens": 755, "total_tokens": 6617}
|
24 |
+
{"user": "zayuvegda", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 25, "total_tokens": 33}
|
25 |
+
{"user": "zayuvegda", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 43, "completion_tokens": 28, "total_tokens": 71}
|
26 |
+
{"user": "zayuvegda", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 76, "completion_tokens": 347, "total_tokens": 423}
|
27 |
+
{"user": "zayuvegda", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 434, "completion_tokens": 253, "total_tokens": 687}
|
28 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 6635, "completion_tokens": 1026, "total_tokens": 7661}
|
29 |
+
{"user": "zayuvegda", "model": "gpt-4o-mini", "provider": "OpenaiAccount", "prompt_tokens": 728, "completion_tokens": 8, "total_tokens": 736}
|
30 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 75, "completion_tokens": 945, "total_tokens": 1020}
|
31 |
+
{"user": "zayuvegda", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 65, "completion_tokens": 229, "total_tokens": 294}
|
32 |
+
{"user": "zayuvegda", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 91, "completion_tokens": 143, "total_tokens": 234}
|
33 |
+
{"user": "zayuvegda", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 939, "completion_tokens": 399, "total_tokens": 1338}
|
34 |
+
{"user": "zayuvegda", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 1682, "completion_tokens": 489, "total_tokens": 2171}
|
35 |
+
{"user": "zayuvegda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 2823, "completion_tokens": 1513, "total_tokens": 4336}
|
36 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 1052, "completion_tokens": 833, "total_tokens": 1885}
|
37 |
+
{"user": "zayuvegda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 188, "completion_tokens": 730, "total_tokens": 918}
|
38 |
+
{"user": "zayuvegda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 928, "completion_tokens": 1272, "total_tokens": 2200}
|
39 |
+
{"user": "zayuvegda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 2299, "completion_tokens": 2075, "total_tokens": 4374}
|
40 |
+
{"user": "zayuvegda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 4444, "completion_tokens": 2914, "total_tokens": 7358}
|
41 |
+
{"user": "cruelzade", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 287, "completion_tokens": 464, "total_tokens": 751}
|
42 |
+
{"user": "seotube", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 11, "total_tokens": 19}
|
43 |
+
{"user": "seotube", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 30, "completion_tokens": 378, "total_tokens": 408}
|
44 |
+
{"user": "seotube", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 379, "completion_tokens": 214, "total_tokens": 593}
|
45 |
+
{"user": "setsunafjava", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 30, "completion_tokens": 15, "total_tokens": 45}
|
46 |
+
{"user": "setsunafjava", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 69, "completion_tokens": 17, "total_tokens": 86}
|
47 |
+
{"user": "setsunafjava", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 115, "completion_tokens": 1204, "total_tokens": 1319}
|
48 |
+
{"user": "setsunafjava", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 1890, "completion_tokens": 1940, "total_tokens": 3830}
|
49 |
+
{"user": "roxky", "model": "flux-schnell", "provider": "HuggingFace", "prompt_tokens": 9, "completion_tokens": 76, "total_tokens": 85}
|
50 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 9, "completion_tokens": 12, "total_tokens": 21}
|
51 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 10, "prompt_tokens": 153, "total_tokens": 163}
|
52 |
+
{"user": "roxky", "model": "flux-dev", "provider": "HuggingSpace", "prompt_tokens": 9, "completion_tokens": 212, "total_tokens": 221}
|
53 |
+
{"user": "roxky", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
|
54 |
+
{"user": "roxky", "model": "sonar-pro", "provider": "PerplexityLabs", "prompt_tokens": 9, "completion_tokens": 27, "total_tokens": 36}
|
55 |
+
{"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 9, "completion_tokens": 53, "total_tokens": 62}
|
56 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
|
57 |
+
{"user": "ballm4n", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 9, "completion_tokens": 9, "total_tokens": 18}
|
58 |
+
{"user": "NexThinkLabs", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 8, "total_tokens": 16}
|
59 |
+
{"user": "NexThinkLabs", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 25, "completion_tokens": 24, "total_tokens": 49}
|
60 |
+
{"user": "NexThinkLabs", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 8, "total_tokens": 16}
|
61 |
+
{"user": "roxky", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 10, "completion_tokens": 10, "total_tokens": 20}
|
62 |
+
{"user": "roxky", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 37, "completion_tokens": 106, "total_tokens": 143}
|
63 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 37, "completion_tokens": 376, "total_tokens": 413}
|
64 |
+
{"user": "roxky", "model": "sonar-pro", "provider": "PerplexityLabs", "prompt_tokens": 37, "completion_tokens": 154, "total_tokens": 191}
|
65 |
+
{"user": "MAKERBOT", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 9, "completion_tokens": 16, "total_tokens": 25}
|
66 |
+
{"user": "XannX", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 11, "completion_tokens": 608, "total_tokens": 619}
|
67 |
+
{"user": "MAKERBOT", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 56, "completion_tokens": 114, "total_tokens": 170}
|
68 |
+
{"user": "MAKERBOT", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 155, "completion_tokens": 636, "total_tokens": 791}
|
69 |
+
{"user": "XannX", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 569, "completion_tokens": 3049, "total_tokens": 3618}
|
70 |
+
{"user": "dgdgdsgs", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 24, "total_tokens": 32}
|
71 |
+
{"user": "dgdgdsgs", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 185, "completion_tokens": 368, "total_tokens": 553}
|
72 |
+
{"user": "roxky", "model": "sonar-pro", "provider": "PerplexityLabs", "prompt_tokens": 10, "completion_tokens": 21, "total_tokens": 31}
|
73 |
+
{"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 1490, "completion_tokens": 2074, "total_tokens": 3564}
|
74 |
+
{"user": "roxky", "model": "claude-3-haiku", "provider": "DDG", "prompt_tokens": 9, "completion_tokens": 9, "total_tokens": 18}
|
75 |
+
{"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 8, "completion_tokens": 14, "total_tokens": 22}
|
76 |
+
{"user": "roxky", "model": "flux", "provider": "Blackbox", "prompt_tokens": 8, "completion_tokens": 112, "total_tokens": 120}
|
77 |
+
{"user": "roxky", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
|
78 |
+
{"user": "KA4E6", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 12, "total_tokens": 20}
|
79 |
+
{"user": "Majidgdr", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 13, "completion_tokens": 13, "total_tokens": 26}
|
80 |
+
{"user": "Majidgdr", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 39, "completion_tokens": 71, "total_tokens": 110}
|
81 |
+
{"user": "Majidgdr", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 121, "completion_tokens": 70, "total_tokens": 191}
|
82 |
+
{"user": "Majidgdr", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 187, "completion_tokens": 6, "total_tokens": 193}
|
83 |
+
{"user": "Jackolord", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 10, "prompt_tokens": 152, "total_tokens": 162}
|
84 |
+
{"user": "Jackolord", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 27, "completion_tokens": 129, "total_tokens": 156}
|
85 |
+
{"user": "Jackolord", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 37, "completion_tokens": 361, "total_tokens": 398}
|
86 |
+
{"user": "Jackolord", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 342, "completion_tokens": 465, "total_tokens": 807}
|
87 |
+
{"user": "Jackolord", "model": "o1-preview", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
|
88 |
+
{"user": "Jackolord", "model": "o1-preview", "provider": "OpenaiAccount", "prompt_tokens": 36, "completion_tokens": 45, "total_tokens": 81}
|
89 |
+
{"user": "Jackolord", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 17, "completion_tokens": 90, "total_tokens": 107}
|
90 |
+
{"user": "Jackolord", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 17, "completion_tokens": 53, "total_tokens": 70}
|
91 |
+
{"user": "Jackolord", "model": "sonar-pro", "provider": "PerplexityLabs", "prompt_tokens": 17, "completion_tokens": 14, "total_tokens": 31}
|
92 |
+
{"user": "Jackolord", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 17, "completion_tokens": 54, "total_tokens": 71}
|
93 |
+
{"user": "Jackolord", "model": "dall-e-3", "provider": "CopilotAccount", "prompt_tokens": 12, "completion_tokens": 157, "total_tokens": 169}
|
94 |
+
{"user": "Jackolord", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 12, "completion_tokens": 152, "total_tokens": 164}
|
95 |
+
{"user": "Jackolord", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 12, "completion_tokens": 616, "total_tokens": 628}
|
96 |
+
{"user": "NDeriy", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 110, "completion_tokens": 84, "total_tokens": 194}
|
97 |
+
{"user": "NDeriy", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 302, "completion_tokens": 249, "total_tokens": 551}
|
98 |
+
{"user": "NDeriy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 739, "completion_tokens": 825, "total_tokens": 1564}
|
99 |
+
{"user": "NDeriy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 1610, "completion_tokens": 647, "total_tokens": 2257}
|
100 |
+
{"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 17, "completion_tokens": 57, "total_tokens": 74}
|
101 |
+
{"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 87, "completion_tokens": 66, "total_tokens": 153}
|
102 |
+
{"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 167, "completion_tokens": 121, "total_tokens": 288}
|
103 |
+
{"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 303, "completion_tokens": 230, "total_tokens": 533}
|
104 |
+
{"user": "NDeriy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 2353, "completion_tokens": 1051, "total_tokens": 3404}
|
105 |
+
{"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 25, "completion_tokens": 183, "total_tokens": 208}
|
106 |
+
{"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 322, "completion_tokens": 217, "total_tokens": 539}
|
107 |
+
{"user": "NDeriy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 3416, "completion_tokens": 687, "total_tokens": 4103}
|
108 |
+
{"user": "Intgrr", "model": "o1-preview", "provider": "OpenaiAccount", "prompt_tokens": 689, "completion_tokens": 283, "total_tokens": 972}
|
109 |
+
{"user": "Intgrr", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 1144, "completion_tokens": 260, "total_tokens": 1404}
|
110 |
+
{"user": "NDeriy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 4223, "completion_tokens": 1023, "total_tokens": 5246}
|
111 |
+
{"user": "macyn", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 15, "completion_tokens": 95, "total_tokens": 110}
|
112 |
+
{"user": "macyn", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 14, "completion_tokens": 101, "total_tokens": 115}
|
113 |
+
{"user": "NDeriy", "model": "qwq-32b", "provider": "HuggingFace", "prompt_tokens": 5262, "completion_tokens": 286, "total_tokens": 5548}
|
114 |
+
{"user": "macyn", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 15, "completion_tokens": 740, "total_tokens": 755}
|
115 |
+
{"user": "macyn", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 19, "completion_tokens": 226, "total_tokens": 245}
|
116 |
+
{"user": "macyn", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 19, "completion_tokens": 327, "total_tokens": 346}
|
117 |
+
{"user": "roxky", "model": "claude-3-haiku", "provider": "DDG", "prompt_tokens": 9, "completion_tokens": 9, "total_tokens": 18}
|
118 |
+
{"user": "roxky", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 1930, "completion_tokens": 300, "total_tokens": 2230}
|
119 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 1930, "completion_tokens": 1026, "total_tokens": 2956}
|
120 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 3265, "completion_tokens": 56, "total_tokens": 3321}
|
121 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 1930, "completion_tokens": 1026, "total_tokens": 2956}
|
122 |
+
{"user": "Erkaperka502", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 83, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 169, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 252}
|
123 |
+
{"user": "h4mid007", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 12, "total_tokens": 20}
|
124 |
+
{"user": "h4mid007", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 29, "completion_tokens": 10, "total_tokens": 39}
|
125 |
+
{"user": "h4mid007", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 53, "completion_tokens": 25, "total_tokens": 78}
|
126 |
+
{"user": "h4mid007", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 87, "completion_tokens": 294, "total_tokens": 381}
|
har_and_cookies/.usage/2025-02-01.jsonl
ADDED
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{"user": "amsepehrian", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 13, "completion_tokens": 233, "total_tokens": 246}
|
2 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 15, "completion_tokens": 12, "total_tokens": 27}
|
3 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 15, "completion_tokens": 7, "total_tokens": 22}
|
4 |
+
{"user": "roxky", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 15, "completion_tokens": 177, "total_tokens": 192}
|
5 |
+
{"user": "roxky", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 178, "completion_tokens": 258, "total_tokens": 436}
|
6 |
+
{"user": "amsepehrian", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 80, "total_tokens": 88}
|
7 |
+
{"user": "amsepehrian", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 15, "completion_tokens": 371, "total_tokens": 386}
|
8 |
+
{"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 178, "completion_tokens": 209, "total_tokens": 387}
|
9 |
+
{"user": "roxky", "model": "dall-e-3", "provider": "CopilotAccount", "prompt_tokens": 178, "completion_tokens": 234, "total_tokens": 412}
|