Spaces:
Sleeping
Sleeping
Commit changes in Space
Browse files- app.py +103 -11
- deepseek4free-setup.sh +31 -0
- generated_images/1738132089_386b1b49-9cb2-48b7-90e6-9d37c0af6604.jpg +0 -0
- generated_images/1738132162_f4cc45c3-33d3-48ed-a8ea-1bfcf78d114f.jpg +0 -0
- generated_images/1738132256_4694db5d-2328-4200-9dc9-bf1246ab3761.jpg +0 -0
- generated_images/1738132532_6333b9d5-8fc4-4222-851e-fc65aecb3afd.jpg +0 -0
- generated_images/1738202142_hallo.webp +0 -0
- generated_images/1738235394_13e87a97-0e40-4692-90f1-390b722a7c2b.jpg +0 -0
- generated_images/1738235522_10d7ae46-18a0-4b25-8813-7d1838877abb.jpg +0 -0
- har_and_cookies/.logging/2025-01-30-.jsonl +0 -0
- har_and_cookies/.usage/2025-01-29.jsonl +182 -0
- har_and_cookies/.usage/2025-01-30-.jsonl +71 -0
app.py
CHANGED
@@ -5,6 +5,7 @@ from g4f.typing import AsyncResult, Messages
|
|
5 |
import json
|
6 |
import re
|
7 |
import time
|
|
|
8 |
from urllib.parse import quote_plus
|
9 |
from fastapi import FastAPI, Response, Request
|
10 |
from fastapi.responses import RedirectResponse
|
@@ -29,7 +30,7 @@ class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
|
|
29 |
|
30 |
models = [
|
31 |
*g4f.Provider.OpenaiAccount.get_models(),
|
32 |
-
*g4f.Provider.
|
33 |
"flux",
|
34 |
"flux-pro",
|
35 |
"MiniMax-01",
|
@@ -48,7 +49,7 @@ class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
|
|
48 |
model = model.split(' ')[-1]
|
49 |
elif model in g4f.Provider.OpenaiAccount.get_models():
|
50 |
pass
|
51 |
-
elif model in g4f.Provider.
|
52 |
pass
|
53 |
else:
|
54 |
raise ModelNotSupportedError(f"Model: {model}")
|
@@ -58,12 +59,12 @@ class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
|
|
58 |
def get_provider(cls, model):
|
59 |
if model.startswith("MiniMax"):
|
60 |
return "HailuoAI"
|
61 |
-
elif model == "Copilot":
|
62 |
return "CopilotAccount"
|
63 |
elif model in g4f.Provider.OpenaiAccount.get_models():
|
64 |
return "OpenaiAccount"
|
65 |
-
elif model in g4f.Provider.
|
66 |
-
return "
|
67 |
return None
|
68 |
|
69 |
@classmethod
|
@@ -77,6 +78,9 @@ class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
|
|
77 |
**kwargs
|
78 |
) -> AsyncResult:
|
79 |
debug.log(f"{__name__}: {api_key}")
|
|
|
|
|
|
|
80 |
|
81 |
async with StreamSession(
|
82 |
proxy=proxy,
|
@@ -86,12 +90,14 @@ class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
|
|
86 |
model = cls.get_model(model)
|
87 |
provider = cls.get_provider(model)
|
88 |
async with session.post(f"{cls.url}/backend-api/v2/conversation", json={
|
|
|
89 |
"model": model,
|
90 |
"messages": messages,
|
91 |
-
"provider": provider
|
92 |
-
**kwargs
|
93 |
}, ssl=cls.ssl) as response:
|
|
|
94 |
async for line in response.iter_lines():
|
|
|
95 |
data = json.loads(line)
|
96 |
data_type = data.pop("type")
|
97 |
if data_type == "provider":
|
@@ -112,7 +118,22 @@ class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
|
|
112 |
filename = f"{int(time.time())}_{quote_plus(match.group(1)[:100], '')}{extension}"
|
113 |
download_url = f"/download/{filename}?url={cls.url}{match.group(3)}"
|
114 |
return f"[![{match.group(1)}]({download_url})](/images/{filename})"
|
115 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
116 |
elif data_type =="synthesize":
|
117 |
yield SynthesizeData(**data[data_type])
|
118 |
elif data_type == "parameters":
|
@@ -134,10 +155,74 @@ class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
|
|
134 |
|
135 |
g4f.Provider.__map__["Feature"] = BackendApi
|
136 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
137 |
def create_app():
|
138 |
g4f.debug.logging = True
|
139 |
g4f.api.AppConfig.gui = True
|
140 |
-
g4f.api.AppConfig.demo =
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
141 |
|
142 |
app = FastAPI()
|
143 |
|
@@ -156,6 +241,8 @@ def create_app():
|
|
156 |
api.register_authorization()
|
157 |
api.register_validation_exception_handler()
|
158 |
|
|
|
|
|
159 |
@app.get("/download/{filename}", response_class=RedirectResponse)
|
160 |
async def download(filename, request: Request):
|
161 |
filename = os.path.basename(filename)
|
@@ -163,9 +250,14 @@ def create_app():
|
|
163 |
if not os.path.exists(target):
|
164 |
url = str(request.query_params).split("url=", 1)[1]
|
165 |
if url:
|
166 |
-
source_url = url.replace("%2F", "/").replace("%3A", ":").replace("%3F", "?")
|
167 |
-
await copy_images(
|
|
|
|
|
|
|
|
|
168 |
if not os.path.exists(target):
|
|
|
169 |
return Response(status_code=404)
|
170 |
return RedirectResponse(f"/images/{filename}")
|
171 |
|
|
|
5 |
import json
|
6 |
import re
|
7 |
import time
|
8 |
+
import logging
|
9 |
from urllib.parse import quote_plus
|
10 |
from fastapi import FastAPI, Response, Request
|
11 |
from fastapi.responses import RedirectResponse
|
|
|
30 |
|
31 |
models = [
|
32 |
*g4f.Provider.OpenaiAccount.get_models(),
|
33 |
+
*g4f.Provider.PerplexityLabs.get_models(),
|
34 |
"flux",
|
35 |
"flux-pro",
|
36 |
"MiniMax-01",
|
|
|
49 |
model = model.split(' ')[-1]
|
50 |
elif model in g4f.Provider.OpenaiAccount.get_models():
|
51 |
pass
|
52 |
+
elif model in g4f.Provider.PerplexityLabs.get_models():
|
53 |
pass
|
54 |
else:
|
55 |
raise ModelNotSupportedError(f"Model: {model}")
|
|
|
59 |
def get_provider(cls, model):
|
60 |
if model.startswith("MiniMax"):
|
61 |
return "HailuoAI"
|
62 |
+
elif model == "Copilot" or "dall-e" in model:
|
63 |
return "CopilotAccount"
|
64 |
elif model in g4f.Provider.OpenaiAccount.get_models():
|
65 |
return "OpenaiAccount"
|
66 |
+
elif model in g4f.Provider.PerplexityLabs.get_models():
|
67 |
+
return "PerplexityLabs"
|
68 |
return None
|
69 |
|
70 |
@classmethod
|
|
|
78 |
**kwargs
|
79 |
) -> AsyncResult:
|
80 |
debug.log(f"{__name__}: {api_key}")
|
81 |
+
if "dall-e" in model and "prompt" not in kwargs:
|
82 |
+
kwargs["prompt"] = messages[-1]["content"]
|
83 |
+
messages[-1]["content"] = f"Generate a image: {kwargs['prompt']}"
|
84 |
|
85 |
async with StreamSession(
|
86 |
proxy=proxy,
|
|
|
90 |
model = cls.get_model(model)
|
91 |
provider = cls.get_provider(model)
|
92 |
async with session.post(f"{cls.url}/backend-api/v2/conversation", json={
|
93 |
+
**kwargs,
|
94 |
"model": model,
|
95 |
"messages": messages,
|
96 |
+
"provider": provider
|
|
|
97 |
}, ssl=cls.ssl) as response:
|
98 |
+
is_thinking = 0
|
99 |
async for line in response.iter_lines():
|
100 |
+
response.raise_for_status()
|
101 |
data = json.loads(line)
|
102 |
data_type = data.pop("type")
|
103 |
if data_type == "provider":
|
|
|
118 |
filename = f"{int(time.time())}_{quote_plus(match.group(1)[:100], '')}{extension}"
|
119 |
download_url = f"/download/{filename}?url={cls.url}{match.group(3)}"
|
120 |
return f"[![{match.group(1)}]({download_url})](/images/{filename})"
|
121 |
+
if "<think>" in data[data_type]:
|
122 |
+
data[data_type] = data[data_type].split("<think>", 1)
|
123 |
+
yield data[data_type][0]
|
124 |
+
yield Reasoning(data[data_type][1])
|
125 |
+
yield Reasoning(None, "Is thinking...")
|
126 |
+
is_thinking = time.time()
|
127 |
+
if "</think>" in data[data_type][1]:
|
128 |
+
data[data_type][1] = data[data_type].split("</think>", 1)
|
129 |
+
yield Reasoning(data[data_type][0])
|
130 |
+
yield Reasoning(None, f"Finished in {round(time.time()-is_thinking, 2)} seconds")
|
131 |
+
yield data[data_type][1]
|
132 |
+
is_thinking = 0
|
133 |
+
elif is_thinking:
|
134 |
+
yield Reasoning(data[data_type])
|
135 |
+
else:
|
136 |
+
yield re.sub(r'\[\!\[(.+?)\]\(([^)]+?)\)\]\(([^)]+?)\)', on_image, data[data_type])
|
137 |
elif data_type =="synthesize":
|
138 |
yield SynthesizeData(**data[data_type])
|
139 |
elif data_type == "parameters":
|
|
|
155 |
|
156 |
g4f.Provider.__map__["Feature"] = BackendApi
|
157 |
|
158 |
+
import asyncio
|
159 |
+
import uuid
|
160 |
+
from aiohttp import ClientSession, ClientError
|
161 |
+
from g4f.typing import Optional, Cookies
|
162 |
+
from g4f.image import is_accepted_format
|
163 |
+
|
164 |
+
async def copy_images(
|
165 |
+
images: list[str],
|
166 |
+
cookies: Optional[Cookies] = None,
|
167 |
+
headers: dict = None,
|
168 |
+
proxy: Optional[str] = None,
|
169 |
+
add_url: bool = True,
|
170 |
+
target: str = None,
|
171 |
+
ssl: bool = None
|
172 |
+
) -> list[str]:
|
173 |
+
if add_url:
|
174 |
+
add_url = not cookies
|
175 |
+
#ensure_images_dir()
|
176 |
+
async with ClientSession(
|
177 |
+
#connector=get_connector(proxy=proxy),
|
178 |
+
cookies=cookies,
|
179 |
+
headers=headers,
|
180 |
+
) as session:
|
181 |
+
async def copy_image(image: str, target: str = None) -> str:
|
182 |
+
if target is None or len(images) > 1:
|
183 |
+
target = os.path.join(images_dir, f"{int(time.time())}_{str(uuid.uuid4())}")
|
184 |
+
try:
|
185 |
+
if image.startswith("data:"):
|
186 |
+
pass
|
187 |
+
#with open(target, "wb") as f:
|
188 |
+
# f.write(extract_data_uri(image))
|
189 |
+
else:
|
190 |
+
try:
|
191 |
+
async with session.get(image, proxy=proxy, ssl=ssl) as response:
|
192 |
+
response.raise_for_status()
|
193 |
+
with open(target, "wb") as f:
|
194 |
+
async for chunk in response.content.iter_chunked(4096):
|
195 |
+
f.write(chunk)
|
196 |
+
except ClientError as e:
|
197 |
+
debug.log(f"copy_images failed: {e.__class__.__name__}: {e}")
|
198 |
+
return image
|
199 |
+
if "." not in target:
|
200 |
+
with open(target, "rb") as f:
|
201 |
+
extension = is_accepted_format(f.read(12)).split("/")[-1]
|
202 |
+
extension = "jpg" if extension == "jpeg" else extension
|
203 |
+
new_target = f"{target}.{extension}"
|
204 |
+
os.rename(target, new_target)
|
205 |
+
target = new_target
|
206 |
+
finally:
|
207 |
+
if "." not in target and os.path.exists(target):
|
208 |
+
os.unlink(target)
|
209 |
+
return f"/images/{os.path.basename(target)}{'?url=' + image if add_url and not image.startswith('data:') else ''}"
|
210 |
+
|
211 |
+
return await asyncio.gather(*[copy_image(image, target) for image in images])
|
212 |
+
|
213 |
def create_app():
|
214 |
g4f.debug.logging = True
|
215 |
g4f.api.AppConfig.gui = True
|
216 |
+
g4f.api.AppConfig.demo = False
|
217 |
+
|
218 |
+
class NoHomeFilter(logging.Filter):
|
219 |
+
def filter(self, record):
|
220 |
+
if '"GET / HTTP/1.1" 200 OK' in record.getMessage():
|
221 |
+
return False
|
222 |
+
if '"GET /static/' in record.getMessage():
|
223 |
+
return False
|
224 |
+
return True
|
225 |
+
logging.getLogger("uvicorn.access").addFilter(NoHomeFilter())
|
226 |
|
227 |
app = FastAPI()
|
228 |
|
|
|
241 |
api.register_authorization()
|
242 |
api.register_validation_exception_handler()
|
243 |
|
244 |
+
headers = {}
|
245 |
+
|
246 |
@app.get("/download/{filename}", response_class=RedirectResponse)
|
247 |
async def download(filename, request: Request):
|
248 |
filename = os.path.basename(filename)
|
|
|
250 |
if not os.path.exists(target):
|
251 |
url = str(request.query_params).split("url=", 1)[1]
|
252 |
if url:
|
253 |
+
source_url = url.replace("%2F", "/").replace("%3A", ":").replace("%3F", "?")
|
254 |
+
await copy_images(
|
255 |
+
[source_url],
|
256 |
+
target=target,
|
257 |
+
ssl=False,
|
258 |
+
headers=headers if source_url.startswith(BackendApi.url) else None)
|
259 |
if not os.path.exists(target):
|
260 |
+
|
261 |
return Response(status_code=404)
|
262 |
return RedirectResponse(f"/images/{filename}")
|
263 |
|
deepseek4free-setup.sh
ADDED
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
git clone https://github.com/xtekky/deepseek4free.git
|
2 |
+
|
3 |
+
echo "recursive-include dsk/wasm *" >> ./deepseek4free/MANIFEST.in
|
4 |
+
|
5 |
+
echo "from setuptools import find_packages, setup
|
6 |
+
|
7 |
+
INSTALL_REQUIRE = [
|
8 |
+
'curl-cffi',
|
9 |
+
'wasmtime',
|
10 |
+
'numpy',
|
11 |
+
]
|
12 |
+
|
13 |
+
DESCRIPTION = (
|
14 |
+
'The official gpt4free repository | various collection of powerful language models'
|
15 |
+
)
|
16 |
+
|
17 |
+
# Setting up
|
18 |
+
setup(
|
19 |
+
name='dsk',
|
20 |
+
version='0.0.1.0',
|
21 |
+
author='Tekky',
|
22 |
+
author_email='<[email protected]>',
|
23 |
+
description=DESCRIPTION,
|
24 |
+
long_description_content_type='text/markdown',
|
25 |
+
long_description='',
|
26 |
+
packages=find_packages(),
|
27 |
+
include_package_data=True,
|
28 |
+
install_requires=INSTALL_REQUIRE
|
29 |
+
)" >> ./deepseek4free/setup.py
|
30 |
+
|
31 |
+
pip install ./deepseek4free --break-system-packages
|
generated_images/1738132089_386b1b49-9cb2-48b7-90e6-9d37c0af6604.jpg
ADDED
![]() |
generated_images/1738132162_f4cc45c3-33d3-48ed-a8ea-1bfcf78d114f.jpg
ADDED
![]() |
generated_images/1738132256_4694db5d-2328-4200-9dc9-bf1246ab3761.jpg
ADDED
![]() |
generated_images/1738132532_6333b9d5-8fc4-4222-851e-fc65aecb3afd.jpg
ADDED
![]() |
generated_images/1738202142_hallo.webp
ADDED
![]() |
generated_images/1738235394_13e87a97-0e40-4692-90f1-390b722a7c2b.jpg
ADDED
![]() |
generated_images/1738235522_10d7ae46-18a0-4b25-8813-7d1838877abb.jpg
ADDED
![]() |
har_and_cookies/.logging/2025-01-30-.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
har_and_cookies/.usage/2025-01-29.jsonl
ADDED
@@ -0,0 +1,182 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{"user": "MTadder", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 11, "completion_tokens": 27, "total_tokens": 38}
|
2 |
+
{"user": "MTadder", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 11, "completion_tokens": 26, "total_tokens": 37}
|
3 |
+
{"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 8, "completion_tokens": 12, "total_tokens": 20}
|
4 |
+
{"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 8, "completion_tokens": 23, "total_tokens": 31}
|
5 |
+
{"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 45, "completion_tokens": 180, "total_tokens": 225}
|
6 |
+
{"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 45, "completion_tokens": 106, "total_tokens": 151}
|
7 |
+
{"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 45, "completion_tokens": 44, "total_tokens": 89}
|
8 |
+
{"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 45, "completion_tokens": 159, "total_tokens": 204}
|
9 |
+
{"user": "roxky", "model": "black-forest-labs/FLUX.1-schnell", "provider": "HuggingFace", "prompt_tokens": 45, "completion_tokens": 142, "total_tokens": 187}
|
10 |
+
{"user": "roxky", "model": "black-forest-labs/FLUX.1-schnell", "provider": "HuggingSpace", "prompt_tokens": 45, "completion_tokens": 192, "total_tokens": 237}
|
11 |
+
{"user": "roxky", "model": "black-forest-labs/FLUX.1-schnell", "provider": "HuggingSpace", "prompt_tokens": 45, "completion_tokens": 191, "total_tokens": 236}
|
12 |
+
{"user": "roxky", "model": "black-forest-labs/FLUX.1-schnell", "provider": "HuggingChat", "prompt_tokens": 45, "completion_tokens": 153, "total_tokens": 198}
|
13 |
+
{"user": "roxky", "model": "black-forest-labs/FLUX.1-schnell", "provider": "HuggingSpace", "prompt_tokens": 45, "completion_tokens": 137, "total_tokens": 182}
|
14 |
+
{"user": "roxky", "model": "Pollinations AI: flux", "provider": "HuggingSpace", "prompt_tokens": 16, "completion_tokens": 152, "total_tokens": 168}
|
15 |
+
{"user": "roxky", "model": "Pollinations AI: flux", "provider": "Blackbox", "prompt_tokens": 16, "completion_tokens": 139, "total_tokens": 155}
|
16 |
+
{"user": "MTadder", "model": "qwq-32b", "provider": "HuggingFace", "prompt_tokens": 30, "completion_tokens": 523, "total_tokens": 553}
|
17 |
+
{"user": "roxky", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "Feature", "prompt_tokens": 9, "completion_tokens": 20, "total_tokens": 29}
|
18 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 13, "completion_tokens": 242, "total_tokens": 255}
|
19 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 263, "completion_tokens": 137, "total_tokens": 400}
|
20 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 255, "completion_tokens": 57, "total_tokens": 312}
|
21 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 10088, "completion_tokens": 39, "total_tokens": 10127}
|
22 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
|
23 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 31, "completion_tokens": 92, "total_tokens": 123}
|
24 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 132, "completion_tokens": 20, "total_tokens": 152}
|
25 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 157, "completion_tokens": 67, "total_tokens": 224}
|
26 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 232, "completion_tokens": 74, "total_tokens": 306}
|
27 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 307, "completion_tokens": 81, "total_tokens": 388}
|
28 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 497, "completion_tokens": 614, "total_tokens": 1111}
|
29 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
|
30 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 40, "completion_tokens": 117, "total_tokens": 157}
|
31 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 179, "completion_tokens": 208, "total_tokens": 387}
|
32 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 382, "completion_tokens": 239, "total_tokens": 621}
|
33 |
+
{"user": "maverick792", "model": "o1", "provider": "Feature", "prompt_tokens": 19, "completion_tokens": 810, "total_tokens": 829}
|
34 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 610, "completion_tokens": 196, "total_tokens": 806}
|
35 |
+
{"user": "maverick792", "model": "o1", "provider": "Feature", "prompt_tokens": 1031, "completion_tokens": 1481, "total_tokens": 2512}
|
36 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 793, "completion_tokens": 313, "total_tokens": 1106}
|
37 |
+
{"user": "maverick792", "model": "o1", "provider": "Feature", "prompt_tokens": 2772, "completion_tokens": 216, "total_tokens": 2988}
|
38 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 4599, "completion_tokens": 55, "total_tokens": 4654}
|
39 |
+
{"user": "maverick792", "model": "o1-preview", "provider": "Feature", "prompt_tokens": 3066, "completion_tokens": 1905, "total_tokens": 4971}
|
40 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 5852, "completion_tokens": 461, "total_tokens": 6313}
|
41 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 6270, "completion_tokens": 382, "total_tokens": 6652}
|
42 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 6642, "completion_tokens": 234, "total_tokens": 6876}
|
43 |
+
{"user": "maverick792", "model": "o1-preview", "provider": "Feature", "prompt_tokens": 5280, "completion_tokens": 1858, "total_tokens": 7138}
|
44 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 6874, "completion_tokens": 299, "total_tokens": 7173}
|
45 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 7169, "completion_tokens": 349, "total_tokens": 7518}
|
46 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 7507, "completion_tokens": 432, "total_tokens": 7939}
|
47 |
+
{"user": "maverick792", "model": "o1-preview", "provider": "Feature", "prompt_tokens": 7449, "completion_tokens": 1157, "total_tokens": 8606}
|
48 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 7897, "completion_tokens": 3069, "total_tokens": 10966}
|
49 |
+
{"user": "maverick792", "model": "o1-preview", "provider": "Feature", "prompt_tokens": 16, "completion_tokens": 38, "total_tokens": 54}
|
50 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 9952, "completion_tokens": 398, "total_tokens": 10350}
|
51 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 7897, "completion_tokens": 3069, "total_tokens": 10966}
|
52 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 7897, "completion_tokens": 3069, "total_tokens": 10966}
|
53 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 14401, "completion_tokens": 29, "total_tokens": 14430}
|
54 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 16279, "completion_tokens": 256, "total_tokens": 16535}
|
55 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 18379, "completion_tokens": 2038, "total_tokens": 20417}
|
56 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 20020, "completion_tokens": 2230, "total_tokens": 22250}
|
57 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 18379, "completion_tokens": 2038, "total_tokens": 20417}
|
58 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 18379, "completion_tokens": 2038, "total_tokens": 20417}
|
59 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 26051, "completion_tokens": 1704, "total_tokens": 27755}
|
60 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 27458, "completion_tokens": 239, "total_tokens": 27697}
|
61 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 28631, "completion_tokens": 666, "total_tokens": 29297}
|
62 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 29348, "completion_tokens": 1353, "total_tokens": 30701}
|
63 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 31059, "completion_tokens": 1226, "total_tokens": 32285}
|
64 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 33004, "completion_tokens": 1502, "total_tokens": 34506}
|
65 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 35888, "completion_tokens": 1580, "total_tokens": 37468}
|
66 |
+
{"user": "evanforsale", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 39076, "completion_tokens": 2652, "total_tokens": 41728}
|
67 |
+
{"user": "MTadder", "model": "qwq-32b", "provider": "HuggingFace", "prompt_tokens": 75, "completion_tokens": 685, "total_tokens": 760}
|
68 |
+
{"user": "roxky", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "HuggingChat", "prompt_tokens": 17, "completion_tokens": 23, "total_tokens": 40}
|
69 |
+
{"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 17, "completion_tokens": 55, "total_tokens": 72}
|
70 |
+
{"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 8, "completion_tokens": 14, "total_tokens": 22}
|
71 |
+
{"user": "roxky", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "HuggingChat", "prompt_tokens": 8, "completion_tokens": 12, "total_tokens": 20}
|
72 |
+
{"user": "nelivol732", "model": "sd-3.5", "provider": "HuggingSpace", "prompt_tokens": 31, "completion_tokens": 220, "total_tokens": 251}
|
73 |
+
{"user": "nelivol732", "model": "sd-3.5", "provider": "HuggingFace", "prompt_tokens": 312, "completion_tokens": 102, "total_tokens": 414}
|
74 |
+
{"user": "nelivol732", "model": "sd-3.5", "provider": "HuggingSpace", "prompt_tokens": 437, "completion_tokens": 219, "total_tokens": 656}
|
75 |
+
{"user": "nelivol732", "model": "flux-dev", "provider": "HuggingSpace", "prompt_tokens": 31, "completion_tokens": 214, "total_tokens": 245}
|
76 |
+
{"user": "nelivol732", "model": "flux-dev", "provider": "HuggingFace", "prompt_tokens": 306, "completion_tokens": 98, "total_tokens": 404}
|
77 |
+
{"user": "nelivol732", "model": "flux-schnell", "provider": "HuggingFace", "prompt_tokens": 31, "completion_tokens": 88, "total_tokens": 119}
|
78 |
+
{"user": "roxky", "model": "Qwen/Qwen2.5-72B-Instruct", "provider": "HuggingChat", "prompt_tokens": 9, "completion_tokens": 30, "total_tokens": 39}
|
79 |
+
{"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 9, "completion_tokens": 28, "total_tokens": 37}
|
80 |
+
{"user": "nelivol732", "model": "flux-schnell", "provider": "HuggingFace", "prompt_tokens": 180, "completion_tokens": 98, "total_tokens": 278}
|
81 |
+
{"user": "roxky", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 7, "total_tokens": 16}
|
82 |
+
{"user": "nelivol732", "model": "sd-3.5", "provider": "HuggingSpace", "prompt_tokens": 31, "completion_tokens": 214, "total_tokens": 245}
|
83 |
+
{"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 57, "completion_tokens": 176, "total_tokens": 233}
|
84 |
+
{"user": "nelivol732", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 37, "completion_tokens": 30, "total_tokens": 67}
|
85 |
+
{"user": "nelivol732", "completion_tokens": 74, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 179, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 253}
|
86 |
+
{"user": "nelivol732", "completion_tokens": 166, "prompt_tokens": 313, "total_tokens": 479}
|
87 |
+
{"user": "nelivol732", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 2522, "completion_tokens": 6, "total_tokens": 2528}
|
88 |
+
{"user": "nelivol732", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 2522, "completion_tokens": 6, "total_tokens": 2528}
|
89 |
+
{"user": "nelivol732", "completion_tokens": 136, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 2559, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 2695}
|
90 |
+
{"user": "fabler43", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 25, "completion_tokens": 59, "total_tokens": 84}
|
91 |
+
{"user": "fabler43", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 21, "completion_tokens": 198, "total_tokens": 219}
|
92 |
+
{"user": "fabler43", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 21, "completion_tokens": 201, "total_tokens": 222}
|
93 |
+
{"user": "MTadder", "model": "qwq-32b", "provider": "HuggingFace", "prompt_tokens": 347, "completion_tokens": 855, "total_tokens": 1202}
|
94 |
+
{"user": "MTadder", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 347, "completion_tokens": 326, "total_tokens": 673}
|
95 |
+
{"user": "MTadder", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 1219, "completion_tokens": 53, "total_tokens": 1272}
|
96 |
+
{"user": "MTadder", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 1286, "completion_tokens": 449, "total_tokens": 1735}
|
97 |
+
{"user": "MTadder", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 45, "completion_tokens": 254, "total_tokens": 299}
|
98 |
+
{"user": "MTadder", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 154, "completion_tokens": 1008, "total_tokens": 1162}
|
99 |
+
{"user": "MTadder", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 1177, "completion_tokens": 1026, "total_tokens": 2203}
|
100 |
+
{"user": "roxky", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "HuggingChat", "prompt_tokens": 8, "completion_tokens": 12, "total_tokens": 20}
|
101 |
+
{"user": "roxky", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "HuggingChat", "prompt_tokens": 8, "completion_tokens": 31, "total_tokens": 39}
|
102 |
+
{"user": "silait", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 12, "total_tokens": 20}
|
103 |
+
{"user": "silait", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 31, "completion_tokens": 47, "total_tokens": 78}
|
104 |
+
{"user": "silait", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 95, "completion_tokens": 682, "total_tokens": 777}
|
105 |
+
{"user": "roxky", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "HuggingChat", "prompt_tokens": 9, "completion_tokens": 12, "total_tokens": 21}
|
106 |
+
{"user": "roxky", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "HuggingChat", "prompt_tokens": 43, "completion_tokens": 496, "total_tokens": 539}
|
107 |
+
{"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 21, "completion_tokens": 162, "total_tokens": 183}
|
108 |
+
{"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 58, "completion_tokens": 173, "total_tokens": 231}
|
109 |
+
{"user": "fabler43", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 303, "completion_tokens": 1704, "total_tokens": 2007}
|
110 |
+
{"user": "Thanh6868", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 25, "total_tokens": 33}
|
111 |
+
{"user": "Dotmazy", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 11, "completion_tokens": 19, "total_tokens": 30}
|
112 |
+
{"user": "Dotmazy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 13, "completion_tokens": 11, "total_tokens": 24}
|
113 |
+
{"user": "Dotmazy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 21, "completion_tokens": 1012, "total_tokens": 1033}
|
114 |
+
{"user": "Dotmazy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 1054, "completion_tokens": 1181, "total_tokens": 2235}
|
115 |
+
{"user": "Dotmazy", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 2248, "completion_tokens": 660, "total_tokens": 2908}
|
116 |
+
{"user": "andrewseverniy", "model": "qwen-2-72b", "provider": "HuggingSpace", "prompt_tokens": 15, "completion_tokens": 5, "total_tokens": 20}
|
117 |
+
{"user": "andrewseverniy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 36, "completion_tokens": 395, "total_tokens": 431}
|
118 |
+
{"user": "Stremty", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 12, "total_tokens": 20}
|
119 |
+
{"user": "Stremty", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 31, "completion_tokens": 44, "total_tokens": 75}
|
120 |
+
{"user": "Stremty", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 69, "completion_tokens": 372, "total_tokens": 441}
|
121 |
+
{"user": "dfed", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 11, "completion_tokens": 24, "total_tokens": 35}
|
122 |
+
{"user": "dfed", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 59, "completion_tokens": 129, "total_tokens": 188}
|
123 |
+
{"user": "dfed", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 208, "completion_tokens": 130, "total_tokens": 338}
|
124 |
+
{"user": "dfed", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 208, "completion_tokens": 377, "total_tokens": 585}
|
125 |
+
{"user": "dfed", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 731, "completion_tokens": 59, "total_tokens": 790}
|
126 |
+
{"user": "dfed", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 803, "completion_tokens": 295, "total_tokens": 1098}
|
127 |
+
{"user": "maverick792", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 17, "completion_tokens": 686, "total_tokens": 703}
|
128 |
+
{"user": "vlasdadsda", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 8, "completion_tokens": 20, "total_tokens": 28}
|
129 |
+
{"user": "vlasdadsda", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 37, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 153, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 190}
|
130 |
+
{"user": "vlasdadsda", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 223, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 1108, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 1331}
|
131 |
+
{"user": "vlasdadsda", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 313, "prompt_tokens": 440, "total_tokens": 753}
|
132 |
+
{"user": "vlasdadsda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 10, "completion_tokens": 90, "total_tokens": 100}
|
133 |
+
{"user": "misharBSD", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 8, "completion_tokens": 9, "total_tokens": 17}
|
134 |
+
{"user": "misharBSD", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 37, "completion_tokens": 38, "total_tokens": 75}
|
135 |
+
{"user": "misharBSD", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 42, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 178, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 220}
|
136 |
+
{"user": "misharBSD", "model": "sd-3.5", "provider": "HuggingSpace", "prompt_tokens": 22, "completion_tokens": 233, "total_tokens": 255}
|
137 |
+
{"user": "misharBSD", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 1029, "prompt_tokens": 177, "total_tokens": 1206}
|
138 |
+
{"user": "misharBSD", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 45, "completion_tokens": 1075, "total_tokens": 1120}
|
139 |
+
{"user": "misharBSD", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 2332, "completion_tokens": 1026, "total_tokens": 3358}
|
140 |
+
{"user": "misharBSD", "model": "flux-dev", "provider": "HuggingSpace", "prompt_tokens": 19, "completion_tokens": 205, "total_tokens": 224}
|
141 |
+
{"user": "misharBSD", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 1414, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 159, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 1573}
|
142 |
+
{"user": "misharBSD", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 847, "prompt_tokens": 166, "total_tokens": 1013}
|
143 |
+
{"user": "misharBSD", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 10, "completion_tokens": 9, "total_tokens": 19}
|
144 |
+
{"user": "misharBSD", "model": "Qwen/Qwen2.5-72B-Instruct", "provider": "HuggingChat", "prompt_tokens": 10, "completion_tokens": 20, "total_tokens": 30}
|
145 |
+
{"user": "misharBSD", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 10, "completion_tokens": 192, "total_tokens": 202}
|
146 |
+
{"user": "misharBSD", "model": "black-forest-labs/FLUX.1-dev", "provider": "HuggingChat", "prompt_tokens": 8, "completion_tokens": 65, "total_tokens": 73}
|
147 |
+
{"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 8, "completion_tokens": 14, "total_tokens": 22}
|
148 |
+
{"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 278, "completion_tokens": 30, "total_tokens": 308}
|
149 |
+
{"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 336, "completion_tokens": 155, "total_tokens": 491}
|
150 |
+
{"user": "misharBSD", "model": "CohereForAI/c4ai-command-r-plus-08-2024", "provider": "HuggingChat", "prompt_tokens": 336, "completion_tokens": 665, "total_tokens": 1001}
|
151 |
+
{"user": "misharBSD", "model": "CohereForAI/c4ai-command-r-plus-08-2024", "provider": "HuggingChat", "prompt_tokens": 551, "completion_tokens": 288, "total_tokens": 839}
|
152 |
+
{"user": "misharBSD", "model": "CohereForAI/c4ai-command-r-plus-08-2024", "provider": "HuggingChat", "prompt_tokens": 551, "completion_tokens": 302, "total_tokens": 853}
|
153 |
+
{"user": "misharBSD", "model": "meta-llama/Llama-3.3-70B-Instruct", "provider": "HuggingChat", "prompt_tokens": 551, "completion_tokens": 20, "total_tokens": 571}
|
154 |
+
{"user": "misharBSD", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "HuggingChat", "prompt_tokens": 551, "completion_tokens": 151, "total_tokens": 702}
|
155 |
+
{"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 26, "completion_tokens": 370, "total_tokens": 396}
|
156 |
+
{"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 423, "completion_tokens": 534, "total_tokens": 957}
|
157 |
+
{"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 994, "completion_tokens": 969, "total_tokens": 1963}
|
158 |
+
{"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 1982, "completion_tokens": 953, "total_tokens": 2935}
|
159 |
+
{"user": "adi6409", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 11, "completion_tokens": 28, "total_tokens": 39}
|
160 |
+
{"user": "adi6409", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 45, "completion_tokens": 276, "total_tokens": 321}
|
161 |
+
{"user": "adi6409", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 11, "completion_tokens": 84, "total_tokens": 95}
|
162 |
+
{"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 2969, "completion_tokens": 1772, "total_tokens": 4741}
|
163 |
+
{"user": "NDeriy", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 253, "prompt_tokens": 3732, "total_tokens": 3985}
|
164 |
+
{"user": "misharBSD", "model": "Qwen/Qwen2.5-Coder-32B-Instruct", "provider": "HuggingChat", "prompt_tokens": 4755, "completion_tokens": 1363, "total_tokens": 6118}
|
165 |
+
{"user": "NDeriy", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 456, "completion_tokens": 14, "total_tokens": 470}
|
166 |
+
{"user": "NDeriy", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 492, "completion_tokens": 53, "total_tokens": 545}
|
167 |
+
{"user": "NDeriy", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 522, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 6932, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 7454}
|
168 |
+
{"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 9, "completion_tokens": 31, "total_tokens": 40}
|
169 |
+
{"user": "roxky", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
|
170 |
+
{"user": "rdmistra", "model": "flux-dev", "provider": "HuggingFace", "prompt_tokens": 26, "completion_tokens": 1944, "total_tokens": 1970}
|
171 |
+
{"user": "roxky", "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "provider": "HuggingChat", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
|
172 |
+
{"user": "NDeriy", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 856, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 10330, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 11186}
|
173 |
+
{"user": "rdmistra", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 38, "completion_tokens": 131, "total_tokens": 169}
|
174 |
+
{"user": "rdmistra", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 824, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 2953, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 3777}
|
175 |
+
{"user": "Docfile", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 10, "completion_tokens": 16, "total_tokens": 26}
|
176 |
+
{"user": "playjymer", "model": "Qwen/QwQ-32B-Preview", "provider": "HuggingChat", "prompt_tokens": 17, "completion_tokens": 9, "total_tokens": 26}
|
177 |
+
{"user": "playjymer", "model": "dall-e-3", "provider": "OpenaiAccount", "prompt_tokens": 41, "completion_tokens": 70, "total_tokens": 111}
|
178 |
+
{"user": "NDeriy", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 2368, "completion_tokens": 475, "total_tokens": 2843}
|
179 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 36, "completion_tokens": 244, "total_tokens": 280}
|
180 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 295, "completion_tokens": 180, "total_tokens": 475}
|
181 |
+
{"user": "DarkyMan", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 11, "completion_tokens": 17, "total_tokens": 28}
|
182 |
+
{"user": "DarkyMan", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 72, "completion_tokens": 483, "total_tokens": 555}
|
har_and_cookies/.usage/2025-01-30-.jsonl
ADDED
@@ -0,0 +1,71 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{"user": "NDeriy", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 35, "completion_tokens": 292, "total_tokens": 327}
|
2 |
+
{"user": "roxky", "model": "dall-e-3", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 8, "total_tokens": 16}
|
3 |
+
{"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 8, "completion_tokens": 32, "total_tokens": 40}
|
4 |
+
{"user": "maverick792", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 30, "completion_tokens": 461, "total_tokens": 491}
|
5 |
+
{"user": "roxky", "model": "qwen-2-72b", "provider": "HuggingSpace", "prompt_tokens": 8, "completion_tokens": 29, "total_tokens": 37}
|
6 |
+
{"user": "roxky", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 15, "completion_tokens": 19, "total_tokens": 34}
|
7 |
+
{"user": "roxky", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 44, "completion_tokens": 10, "total_tokens": 54}
|
8 |
+
{"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 44, "completion_tokens": 30, "total_tokens": 74}
|
9 |
+
{"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 44, "completion_tokens": 17, "total_tokens": 61}
|
10 |
+
{"user": "roxky", "model": "flux", "provider": "HuggingSpace", "prompt_tokens": 44, "completion_tokens": 141, "total_tokens": 185}
|
11 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 44, "completion_tokens": 8, "total_tokens": 52}
|
12 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 44, "completion_tokens": 8, "total_tokens": 52}
|
13 |
+
{"user": "roxky", "model": "gpt-4o-mini", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 7, "total_tokens": 16}
|
14 |
+
{"user": "roxky", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 9, "completion_tokens": 122, "total_tokens": 131}
|
15 |
+
{"user": "Yazee83", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 13, "completion_tokens": 225, "total_tokens": 238}
|
16 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
|
17 |
+
{"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 8, "completion_tokens": 30, "total_tokens": 38}
|
18 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 8, "completion_tokens": 17, "total_tokens": 25}
|
19 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 31, "completion_tokens": 44, "total_tokens": 75}
|
20 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 95, "completion_tokens": 378, "total_tokens": 473}
|
21 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 95, "completion_tokens": 378, "total_tokens": 473}
|
22 |
+
{"user": "Yazee83", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 352, "completion_tokens": 864, "total_tokens": 1216}
|
23 |
+
{"user": "Yazee83", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 522, "completion_tokens": 46, "total_tokens": 568}
|
24 |
+
{"user": "Yazee83", "model": "claude-3-haiku", "provider": "DDG", "prompt_tokens": 3931, "completion_tokens": 80, "total_tokens": 4011}
|
25 |
+
{"user": "Yazee83", "model": "claude-3-haiku", "provider": "DDG", "prompt_tokens": 3931, "completion_tokens": 80, "total_tokens": 4011}
|
26 |
+
{"user": "Yazee83", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 3458, "completion_tokens": 11, "total_tokens": 3469}
|
27 |
+
{"user": "roxky", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 9, "completion_tokens": 176, "total_tokens": 185}
|
28 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 195, "completion_tokens": 200, "total_tokens": 395}
|
29 |
+
{"user": "roxky", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 404, "completion_tokens": 12, "total_tokens": 416}
|
30 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 404, "completion_tokens": 12, "total_tokens": 416}
|
31 |
+
{"user": "L-X-u", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 18, "prompt_tokens": 152, "total_tokens": 170}
|
32 |
+
{"user": "L-X-u", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 34, "completion_tokens": 136, "total_tokens": 170}
|
33 |
+
{"user": "L-X-u", "model": "dall-e-3", "provider": "OpenaiAccount", "prompt_tokens": 34, "completion_tokens": 327, "total_tokens": 361}
|
34 |
+
{"user": "L-X-u", "model": "flux", "provider": "HuggingSpace", "prompt_tokens": 34, "completion_tokens": 3433, "total_tokens": 3467}
|
35 |
+
{"user": "L-X-u", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 34, "completion_tokens": 932, "total_tokens": 966}
|
36 |
+
{"user": "Yazee83", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 22, "completion_tokens": 381, "total_tokens": 403}
|
37 |
+
{"user": "L-X-u", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 1221, "total_tokens": 1230}
|
38 |
+
{"user": "L-X-u", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 1587, "completion_tokens": 2460, "total_tokens": 4047}
|
39 |
+
{"user": "Yazee83", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 27, "completion_tokens": 544, "total_tokens": 571}
|
40 |
+
{"user": "L-X-u", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 1587, "completion_tokens": 471, "total_tokens": 2058}
|
41 |
+
{"user": "L-X-u", "model": "flux", "provider": "HuggingSpace", "prompt_tokens": 9, "completion_tokens": 19696, "total_tokens": 19705}
|
42 |
+
{"user": "L-X-u", "model": "dall-e-3", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 133, "total_tokens": 142}
|
43 |
+
{"user": "L-X-u", "model": "flux", "provider": "HuggingSpace", "prompt_tokens": 9, "completion_tokens": 19716, "total_tokens": 19725}
|
44 |
+
{"user": "roxky", "model": "dall-e-3", "provider": "OpenaiAccount", "prompt_tokens": 14, "completion_tokens": 108, "total_tokens": 122}
|
45 |
+
{"user": "roxky", "model": "dall-e-3", "provider": "OpenaiAccount", "prompt_tokens": 133, "completion_tokens": 82, "total_tokens": 215}
|
46 |
+
{"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 14, "completion_tokens": 128, "total_tokens": 142}
|
47 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 13, "completion_tokens": 1022, "total_tokens": 1035}
|
48 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 10, "completion_tokens": 912, "total_tokens": 922}
|
49 |
+
{"user": "Harvinderkno", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 316, "completion_tokens": 789, "total_tokens": 1105}
|
50 |
+
{"user": "Harvinderkno", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 935, "completion_tokens": 490, "total_tokens": 1425}
|
51 |
+
{"user": "Harvinderkno", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 1640, "completion_tokens": 817, "total_tokens": 2457}
|
52 |
+
{"user": "Harvinderkno", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 2290, "completion_tokens": 816, "total_tokens": 3106}
|
53 |
+
{"user": "Harvinderkno", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 11, "completion_tokens": 281, "total_tokens": 292}
|
54 |
+
{"user": "Harvinderkno", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 265, "completion_tokens": 536, "total_tokens": 801}
|
55 |
+
{"user": "OWO111", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 16, "completion_tokens": 96, "total_tokens": 112}
|
56 |
+
{"user": "OWO111", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 260, "prompt_tokens": 8452, "total_tokens": 8712}
|
57 |
+
{"user": "OWO111", "model": "qwen-2-72b", "provider": "HuggingSpace", "prompt_tokens": 513, "completion_tokens": 326, "total_tokens": 839}
|
58 |
+
{"user": "OWO111", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 856, "completion_tokens": 539, "total_tokens": 1395}
|
59 |
+
{"user": "OWO111", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 1425, "completion_tokens": 967, "total_tokens": 2392}
|
60 |
+
{"user": "andoajpr", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 441, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 158, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 599}
|
61 |
+
{"user": "OWO111", "model": "gpt-4o-mini", "provider": "OpenaiAccount", "prompt_tokens": 2409, "completion_tokens": 965, "total_tokens": 3374}
|
62 |
+
{"user": "andoajpr", "model": "qvq-72b", "provider": "HuggingSpace", "prompt_tokens": 11, "completion_tokens": 41, "total_tokens": 52}
|
63 |
+
{"user": "OWO111", "model": "gpt-4o-mini", "provider": "OpenaiAccount", "prompt_tokens": 11, "completion_tokens": 796, "total_tokens": 807}
|
64 |
+
{"user": "OWO111", "model": "gpt-4o-mini", "provider": "OpenaiAccount", "prompt_tokens": 1005, "completion_tokens": 608, "total_tokens": 1613}
|
65 |
+
{"user": "OWO111", "model": "qwq-32b", "provider": "HuggingFace", "prompt_tokens": 55, "completion_tokens": 18, "total_tokens": 73}
|
66 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 17, "completion_tokens": 531, "total_tokens": 548}
|
67 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 207, "prompt_tokens": 2115, "total_tokens": 2322}
|
68 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 835, "completion_tokens": 573, "total_tokens": 1408}
|
69 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 1506, "completion_tokens": 785, "total_tokens": 2291}
|
70 |
+
{"user": "KoLLchIK", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 8, "completion_tokens": 9, "total_tokens": 17}
|
71 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 178, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 10024, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 10202}
|