Spaces:
Running
Running
Commit new images and usage data
Browse files- README.md +4 -0
- app.py +5 -213
- demo/BackendApi.py +146 -0
- generated_images/1738296597_what+is+best+ai+model+yet+so+far,+compare+and+tell+me+final+result+only.jpg +0 -0
- generated_images/1738332473_gpt-4o.A+modern+commercial+airplane+soaring+through+a+clear+blue+sky,+with+white+contrails+streaming+behind.jpg +0 -0
- generated_images/1738347359_dall-e-3.Ask+Blackbox+AI+Anything.jpg +0 -0
- generated_images/1738347377_flux-pro.Ask+Blackbox+AI+Anything.jpg +0 -0
- generated_images/1738371043_flux-pro.Create+a+image:+A+futuristic+cityscape+at+sunset,+with+towering+skyscrapers+and+flying+cars+zipping+.jpg +0 -0
- generated_images/1738372570_Copilot.A+futuristic+cityscape+at+sunset,+with+towering+skyscrapers+and+flying+cars+zipping+by.jpg +0 -0
- generated_images/1738372600_dall-e-3.A+futuristic+cityscape+at+sunset,+with+towering+skyscrapers+and+flying+cars+zipping+by.jpg +0 -0
- har_and_cookies/.usage/2025-01-30.jsonl +53 -0
- har_and_cookies/.usage/2025-01-31.jsonl +126 -0
- har_and_cookies/.usage/2025-02-01.jsonl +9 -0
README.md
CHANGED
|
@@ -5,4 +5,8 @@ colorFrom: green
|
|
| 5 |
colorTo: purple
|
| 6 |
sdk: docker
|
| 7 |
pinned: false
|
|
|
|
|
|
|
|
|
|
|
|
|
| 8 |
---
|
|
|
|
| 5 |
colorTo: purple
|
| 6 |
sdk: docker
|
| 7 |
pinned: false
|
| 8 |
+
|
| 9 |
+
hf_oauth: true
|
| 10 |
+
hf_oauth_scopes:
|
| 11 |
+
- inference-api
|
| 12 |
---
|
app.py
CHANGED
|
@@ -1,221 +1,23 @@
|
|
| 1 |
import os
|
| 2 |
-
import json
|
| 3 |
-
from g4f.providers.response import Reasoning, JsonConversation, FinishReason
|
| 4 |
-
from g4f.typing import AsyncResult, Messages
|
| 5 |
-
import json
|
| 6 |
-
import re
|
| 7 |
-
import time
|
| 8 |
-
import logging
|
| 9 |
-
from urllib.parse import quote_plus
|
| 10 |
from fastapi import FastAPI, Response, Request
|
| 11 |
from fastapi.responses import RedirectResponse
|
| 12 |
|
| 13 |
-
from g4f.image import images_dir
|
| 14 |
-
|
| 15 |
import g4f.api
|
| 16 |
import g4f.Provider
|
| 17 |
|
| 18 |
-
from g4f.Provider.base_provider import AsyncGeneratorProvider, ProviderModelMixin
|
| 19 |
-
from g4f.typing import AsyncResult, Messages
|
| 20 |
-
from g4f.requests import StreamSession
|
| 21 |
-
from g4f.providers.response import ProviderInfo, JsonConversation, PreviewResponse, SynthesizeData, TitleGeneration, RequestLogin
|
| 22 |
-
from g4f.providers.response import Parameters, FinishReason, Usage, Reasoning
|
| 23 |
-
from g4f.errors import ModelNotSupportedError
|
| 24 |
-
from g4f import debug
|
| 25 |
-
|
| 26 |
import demo
|
| 27 |
-
|
| 28 |
-
class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
|
| 29 |
-
url = "https://ahe.hopto.org"
|
| 30 |
-
working = True
|
| 31 |
-
ssl = False
|
| 32 |
-
|
| 33 |
-
models = [
|
| 34 |
-
*g4f.Provider.OpenaiAccount.get_models(),
|
| 35 |
-
*g4f.Provider.PerplexityLabs.get_models(),
|
| 36 |
-
"flux",
|
| 37 |
-
"flux-pro",
|
| 38 |
-
"MiniMax-01",
|
| 39 |
-
"Microsoft Copilot",
|
| 40 |
-
]
|
| 41 |
-
|
| 42 |
-
@classmethod
|
| 43 |
-
def get_model(cls, model):
|
| 44 |
-
if "MiniMax" in model:
|
| 45 |
-
model = "MiniMax"
|
| 46 |
-
elif "Copilot" in model:
|
| 47 |
-
model = "Copilot"
|
| 48 |
-
elif "FLUX" in model:
|
| 49 |
-
model = f"flux-{model.split('-')[-1]}"
|
| 50 |
-
elif "flux" in model:
|
| 51 |
-
model = model.split(' ')[-1]
|
| 52 |
-
elif model in g4f.Provider.OpenaiAccount.get_models():
|
| 53 |
-
pass
|
| 54 |
-
elif model in g4f.Provider.PerplexityLabs.get_models():
|
| 55 |
-
pass
|
| 56 |
-
else:
|
| 57 |
-
raise ModelNotSupportedError(f"Model: {model}")
|
| 58 |
-
return model
|
| 59 |
-
|
| 60 |
-
@classmethod
|
| 61 |
-
def get_provider(cls, model):
|
| 62 |
-
if model.startswith("MiniMax"):
|
| 63 |
-
return "HailuoAI"
|
| 64 |
-
elif model == "Copilot" or "dall-e" in model:
|
| 65 |
-
return "CopilotAccount"
|
| 66 |
-
elif model in g4f.Provider.OpenaiAccount.get_models():
|
| 67 |
-
return "OpenaiAccount"
|
| 68 |
-
elif model in g4f.Provider.PerplexityLabs.get_models():
|
| 69 |
-
return "PerplexityLabs"
|
| 70 |
-
return None
|
| 71 |
-
|
| 72 |
-
@classmethod
|
| 73 |
-
async def create_async_generator(
|
| 74 |
-
cls,
|
| 75 |
-
model: str,
|
| 76 |
-
messages: Messages,
|
| 77 |
-
api_key: str = None,
|
| 78 |
-
proxy: str = None,
|
| 79 |
-
timeout: int = 0,
|
| 80 |
-
**kwargs
|
| 81 |
-
) -> AsyncResult:
|
| 82 |
-
debug.log(f"{__name__}: {api_key}")
|
| 83 |
-
if "dall-e" in model and "prompt" not in kwargs:
|
| 84 |
-
kwargs["prompt"] = messages[-1]["content"]
|
| 85 |
-
messages[-1]["content"] = f"Generate a image: {kwargs['prompt']}"
|
| 86 |
-
|
| 87 |
-
async with StreamSession(
|
| 88 |
-
proxy=proxy,
|
| 89 |
-
headers={"Accept": "text/event-stream", **demo.headers},
|
| 90 |
-
timeout=timeout
|
| 91 |
-
) as session:
|
| 92 |
-
model = cls.get_model(model)
|
| 93 |
-
provider = cls.get_provider(model)
|
| 94 |
-
async with session.post(f"{cls.url}/backend-api/v2/conversation", json={
|
| 95 |
-
**kwargs,
|
| 96 |
-
"model": model,
|
| 97 |
-
"messages": messages,
|
| 98 |
-
"provider": provider
|
| 99 |
-
}, ssl=cls.ssl) as response:
|
| 100 |
-
is_thinking = 0
|
| 101 |
-
async for line in response.iter_lines():
|
| 102 |
-
response.raise_for_status()
|
| 103 |
-
data = json.loads(line)
|
| 104 |
-
data_type = data.pop("type")
|
| 105 |
-
if data_type == "provider":
|
| 106 |
-
yield ProviderInfo(**data[data_type])
|
| 107 |
-
provider = data[data_type]["name"]
|
| 108 |
-
elif data_type == "conversation":
|
| 109 |
-
yield JsonConversation(**data[data_type][provider] if provider in data[data_type] else data[data_type][""])
|
| 110 |
-
elif data_type == "conversation_id":
|
| 111 |
-
pass
|
| 112 |
-
elif data_type == "message":
|
| 113 |
-
yield Exception(data)
|
| 114 |
-
elif data_type == "preview":
|
| 115 |
-
yield PreviewResponse(data[data_type])
|
| 116 |
-
elif data_type == "content":
|
| 117 |
-
def on_image(match):
|
| 118 |
-
extension = match.group(3).split(".")[-1].split("?")[0]
|
| 119 |
-
extension = "" if not extension or len(extension) > 4 else f".{extension}"
|
| 120 |
-
filename = f"{int(time.time())}_{quote_plus(match.group(1)[:100], '')}{extension}"
|
| 121 |
-
download_url = f"/download/{filename}?url={cls.url}{match.group(3)}"
|
| 122 |
-
return f"[](/images/{filename})"
|
| 123 |
-
if "<think>" in data[data_type]:
|
| 124 |
-
data[data_type] = data[data_type].split("<think>", 1)
|
| 125 |
-
yield data[data_type][0]
|
| 126 |
-
yield Reasoning(data[data_type][1])
|
| 127 |
-
yield Reasoning(None, "Is thinking...")
|
| 128 |
-
is_thinking = time.time()
|
| 129 |
-
if "</think>" in data[data_type][1]:
|
| 130 |
-
data[data_type][1] = data[data_type].split("</think>", 1)
|
| 131 |
-
yield Reasoning(data[data_type][0])
|
| 132 |
-
yield Reasoning(None, f"Finished in {round(time.time()-is_thinking, 2)} seconds")
|
| 133 |
-
yield data[data_type][1]
|
| 134 |
-
is_thinking = 0
|
| 135 |
-
elif is_thinking:
|
| 136 |
-
yield Reasoning(data[data_type])
|
| 137 |
-
else:
|
| 138 |
-
yield re.sub(r'\[\!\[(.+?)\]\(([^)]+?)\)\]\(([^)]+?)\)', on_image, data[data_type])
|
| 139 |
-
elif data_type =="synthesize":
|
| 140 |
-
yield SynthesizeData(**data[data_type])
|
| 141 |
-
elif data_type == "parameters":
|
| 142 |
-
yield Parameters(**data[data_type])
|
| 143 |
-
elif data_type == "usage":
|
| 144 |
-
yield Usage(**data[data_type])
|
| 145 |
-
elif data_type == "reasoning":
|
| 146 |
-
yield Reasoning(**data)
|
| 147 |
-
elif data_type == "login":
|
| 148 |
-
pass
|
| 149 |
-
elif data_type == "title":
|
| 150 |
-
yield TitleGeneration(data[data_type])
|
| 151 |
-
elif data_type == "finish":
|
| 152 |
-
yield FinishReason(data[data_type]["reason"])
|
| 153 |
-
elif data_type == "log":
|
| 154 |
-
debug.log(data[data_type])
|
| 155 |
-
else:
|
| 156 |
-
debug.log(f"Unknown data: ({data_type}) {data}")
|
| 157 |
-
|
| 158 |
g4f.Provider.__map__["Feature"] = BackendApi
|
| 159 |
|
| 160 |
-
import asyncio
|
| 161 |
-
import uuid
|
| 162 |
-
from aiohttp import ClientSession, ClientError
|
| 163 |
-
from g4f.typing import Optional, Cookies
|
| 164 |
-
from g4f.image import is_accepted_format
|
| 165 |
|
| 166 |
-
|
| 167 |
-
|
| 168 |
-
cookies: Optional[Cookies] = None,
|
| 169 |
-
headers: dict = None,
|
| 170 |
-
proxy: Optional[str] = None,
|
| 171 |
-
add_url: bool = True,
|
| 172 |
-
target: str = None,
|
| 173 |
-
ssl: bool = None
|
| 174 |
-
) -> list[str]:
|
| 175 |
-
if add_url:
|
| 176 |
-
add_url = not cookies
|
| 177 |
-
#ensure_images_dir()
|
| 178 |
-
async with ClientSession(
|
| 179 |
-
#connector=get_connector(proxy=proxy),
|
| 180 |
-
cookies=cookies,
|
| 181 |
-
headers=headers,
|
| 182 |
-
) as session:
|
| 183 |
-
async def copy_image(image: str, target: str = None) -> str:
|
| 184 |
-
if target is None or len(images) > 1:
|
| 185 |
-
target = os.path.join(images_dir, f"{int(time.time())}_{str(uuid.uuid4())}")
|
| 186 |
-
try:
|
| 187 |
-
if image.startswith("data:"):
|
| 188 |
-
pass
|
| 189 |
-
#with open(target, "wb") as f:
|
| 190 |
-
# f.write(extract_data_uri(image))
|
| 191 |
-
else:
|
| 192 |
-
try:
|
| 193 |
-
async with session.get(image, proxy=proxy, ssl=ssl) as response:
|
| 194 |
-
response.raise_for_status()
|
| 195 |
-
with open(target, "wb") as f:
|
| 196 |
-
async for chunk in response.content.iter_chunked(4096):
|
| 197 |
-
f.write(chunk)
|
| 198 |
-
except ClientError as e:
|
| 199 |
-
debug.log(f"copy_images failed: {e.__class__.__name__}: {e}")
|
| 200 |
-
return image
|
| 201 |
-
if "." not in target:
|
| 202 |
-
with open(target, "rb") as f:
|
| 203 |
-
extension = is_accepted_format(f.read(12)).split("/")[-1]
|
| 204 |
-
extension = "jpg" if extension == "jpeg" else extension
|
| 205 |
-
new_target = f"{target}.{extension}"
|
| 206 |
-
os.rename(target, new_target)
|
| 207 |
-
target = new_target
|
| 208 |
-
finally:
|
| 209 |
-
if "." not in target and os.path.exists(target):
|
| 210 |
-
os.unlink(target)
|
| 211 |
-
return f"/images/{os.path.basename(target)}{'?url=' + image if add_url and not image.startswith('data:') else ''}"
|
| 212 |
-
|
| 213 |
-
return await asyncio.gather(*[copy_image(image, target) for image in images])
|
| 214 |
|
| 215 |
def create_app():
|
| 216 |
g4f.debug.logging = True
|
| 217 |
g4f.api.AppConfig.gui = True
|
| 218 |
-
g4f.api.AppConfig.demo =
|
| 219 |
|
| 220 |
app = FastAPI()
|
| 221 |
|
|
@@ -259,14 +61,4 @@ def create_app():
|
|
| 259 |
|
| 260 |
return app
|
| 261 |
|
| 262 |
-
|
| 263 |
-
class NoHomeFilter(logging.Filter):
|
| 264 |
-
def filter(self, record):
|
| 265 |
-
if '"GET / HTTP/1.1" 200 OK' in record.getMessage():
|
| 266 |
-
return False
|
| 267 |
-
if '"GET /static/' in record.getMessage():
|
| 268 |
-
return False
|
| 269 |
-
return True
|
| 270 |
-
logging.getLogger("uvicorn.access").addFilter(NoHomeFilter())
|
| 271 |
-
|
| 272 |
app = create_app()
|
|
|
|
| 1 |
import os
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
from fastapi import FastAPI, Response, Request
|
| 3 |
from fastapi.responses import RedirectResponse
|
| 4 |
|
| 5 |
+
from g4f.image import images_dir
|
|
|
|
| 6 |
import g4f.api
|
| 7 |
import g4f.Provider
|
| 8 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
import demo
|
| 10 |
+
from demo.BackendApi import BackendApi
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
g4f.Provider.__map__["Feature"] = BackendApi
|
| 12 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 13 |
|
| 14 |
+
g4f.models.demo_models
|
| 15 |
+
from g4f.image import copy_images
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 16 |
|
| 17 |
def create_app():
|
| 18 |
g4f.debug.logging = True
|
| 19 |
g4f.api.AppConfig.gui = True
|
| 20 |
+
g4f.api.AppConfig.demo = True
|
| 21 |
|
| 22 |
app = FastAPI()
|
| 23 |
|
|
|
|
| 61 |
|
| 62 |
return app
|
| 63 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 64 |
app = create_app()
|
demo/BackendApi.py
ADDED
|
@@ -0,0 +1,146 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import json
|
| 2 |
+
import re
|
| 3 |
+
import time
|
| 4 |
+
from urllib.parse import quote_plus
|
| 5 |
+
|
| 6 |
+
import g4f.api
|
| 7 |
+
import g4f.Provider
|
| 8 |
+
|
| 9 |
+
from g4f.Provider.base_provider import AsyncGeneratorProvider, ProviderModelMixin
|
| 10 |
+
from g4f.typing import AsyncResult, Messages
|
| 11 |
+
from g4f.requests import StreamSession
|
| 12 |
+
from g4f.providers.response import *
|
| 13 |
+
from g4f.errors import ModelNotSupportedError
|
| 14 |
+
from g4f import debug
|
| 15 |
+
from . import headers
|
| 16 |
+
|
| 17 |
+
class BackendApi(AsyncGeneratorProvider, ProviderModelMixin):
|
| 18 |
+
url = "https://ahe.hopto.org"
|
| 19 |
+
working = True
|
| 20 |
+
ssl = False
|
| 21 |
+
|
| 22 |
+
models = [
|
| 23 |
+
*g4f.Provider.OpenaiAccount.get_models(),
|
| 24 |
+
*g4f.Provider.PerplexityLabs.get_models(),
|
| 25 |
+
"flux",
|
| 26 |
+
"flux-pro",
|
| 27 |
+
"MiniMax-01",
|
| 28 |
+
"Microsoft Copilot",
|
| 29 |
+
]
|
| 30 |
+
|
| 31 |
+
@classmethod
|
| 32 |
+
def get_model(cls, model):
|
| 33 |
+
if "MiniMax" in model:
|
| 34 |
+
model = "MiniMax"
|
| 35 |
+
elif "Copilot" in model:
|
| 36 |
+
model = "Copilot"
|
| 37 |
+
elif "FLUX" in model:
|
| 38 |
+
model = f"flux-{model.split('-')[-1]}"
|
| 39 |
+
elif "flux" in model:
|
| 40 |
+
model = model.split(' ')[-1]
|
| 41 |
+
elif model in g4f.Provider.OpenaiAccount.get_models():
|
| 42 |
+
pass
|
| 43 |
+
elif model in g4f.Provider.PerplexityLabs.get_models():
|
| 44 |
+
pass
|
| 45 |
+
else:
|
| 46 |
+
raise ModelNotSupportedError(f"Model: {model}")
|
| 47 |
+
return model
|
| 48 |
+
|
| 49 |
+
@classmethod
|
| 50 |
+
def get_provider(cls, model):
|
| 51 |
+
if model.startswith("MiniMax"):
|
| 52 |
+
return "HailuoAI"
|
| 53 |
+
elif model == "Copilot" or "dall-e" in model:
|
| 54 |
+
return "CopilotAccount"
|
| 55 |
+
elif model in g4f.Provider.OpenaiAccount.get_models():
|
| 56 |
+
return "OpenaiAccount"
|
| 57 |
+
elif model in g4f.Provider.PerplexityLabs.get_models():
|
| 58 |
+
return "PerplexityLabs"
|
| 59 |
+
return None
|
| 60 |
+
|
| 61 |
+
@classmethod
|
| 62 |
+
async def create_async_generator(
|
| 63 |
+
cls,
|
| 64 |
+
model: str,
|
| 65 |
+
messages: Messages,
|
| 66 |
+
api_key: str = None,
|
| 67 |
+
proxy: str = None,
|
| 68 |
+
timeout: int = 0,
|
| 69 |
+
**kwargs
|
| 70 |
+
) -> AsyncResult:
|
| 71 |
+
debug.log(f"{cls.__name__}: {api_key}")
|
| 72 |
+
|
| 73 |
+
if "dall-e" in model and "prompt" not in kwargs:
|
| 74 |
+
kwargs["prompt"] = messages[-1]["content"]
|
| 75 |
+
messages[-1]["content"] = f"Generate a image: {kwargs['prompt']}"
|
| 76 |
+
|
| 77 |
+
async with StreamSession(
|
| 78 |
+
proxy=proxy,
|
| 79 |
+
headers={"Accept": "text/event-stream", **headers},
|
| 80 |
+
timeout=timeout
|
| 81 |
+
) as session:
|
| 82 |
+
model = cls.get_model(model)
|
| 83 |
+
provider = cls.get_provider(model)
|
| 84 |
+
async with session.post(f"{cls.url}/backend-api/v2/conversation", json={
|
| 85 |
+
**kwargs,
|
| 86 |
+
"model": model,
|
| 87 |
+
"messages": messages,
|
| 88 |
+
"provider": provider
|
| 89 |
+
}, ssl=cls.ssl) as response:
|
| 90 |
+
is_thinking = 0
|
| 91 |
+
async for line in response.iter_lines():
|
| 92 |
+
response.raise_for_status()
|
| 93 |
+
data = json.loads(line)
|
| 94 |
+
data_type = data.pop("type")
|
| 95 |
+
if data_type == "provider":
|
| 96 |
+
yield ProviderInfo(**data[data_type])
|
| 97 |
+
provider = data[data_type]["name"]
|
| 98 |
+
elif data_type == "conversation":
|
| 99 |
+
yield JsonConversation(**data[data_type][provider] if provider in data[data_type] else data[data_type][""])
|
| 100 |
+
elif data_type == "conversation_id":
|
| 101 |
+
pass
|
| 102 |
+
elif data_type == "message":
|
| 103 |
+
yield Exception(data)
|
| 104 |
+
elif data_type == "preview":
|
| 105 |
+
yield PreviewResponse(data[data_type])
|
| 106 |
+
elif data_type == "content":
|
| 107 |
+
def on_image(match):
|
| 108 |
+
extension = match.group(3).split(".")[-1].replace("%3F", "?").split("?")[0]
|
| 109 |
+
extension = ".jpg" if not extension or len(extension) > 4 else f".{extension}"
|
| 110 |
+
filename = f"{int(time.time())}_{quote_plus(model, '')}.{quote_plus(match.group(1)[:100], '')}{extension}"
|
| 111 |
+
download_url = f"/download/{filename}?url={cls.url}{match.group(3)}"
|
| 112 |
+
return f"[](/images/{filename})"
|
| 113 |
+
if "<think>" in data[data_type]:
|
| 114 |
+
data[data_type] = data[data_type].split("<think>", 1)
|
| 115 |
+
yield data[data_type][0]
|
| 116 |
+
yield Reasoning(data[data_type][1])
|
| 117 |
+
yield Reasoning(None, "Is thinking...")
|
| 118 |
+
is_thinking = time.time()
|
| 119 |
+
if "</think>" in data[data_type]:
|
| 120 |
+
data[data_type][1] = data[data_type].split("</think>", 1)
|
| 121 |
+
yield Reasoning(data[data_type][0])
|
| 122 |
+
yield Reasoning(None, f"Finished in {round(time.time()-is_thinking, 2)} seconds")
|
| 123 |
+
yield data[data_type][1]
|
| 124 |
+
is_thinking = 0
|
| 125 |
+
elif is_thinking:
|
| 126 |
+
yield Reasoning(data[data_type])
|
| 127 |
+
else:
|
| 128 |
+
yield re.sub(r'\[\!\[(.+?)\]\(([^)]+?)\)\]\(([^)]+?)\)', on_image, data[data_type])
|
| 129 |
+
elif data_type =="synthesize":
|
| 130 |
+
yield SynthesizeData(**data[data_type])
|
| 131 |
+
elif data_type == "parameters":
|
| 132 |
+
yield Parameters(**data[data_type])
|
| 133 |
+
elif data_type == "usage":
|
| 134 |
+
yield Usage(**data[data_type])
|
| 135 |
+
elif data_type == "reasoning":
|
| 136 |
+
yield Reasoning(**data)
|
| 137 |
+
elif data_type == "login":
|
| 138 |
+
pass
|
| 139 |
+
elif data_type == "title":
|
| 140 |
+
yield TitleGeneration(data[data_type])
|
| 141 |
+
elif data_type == "finish":
|
| 142 |
+
yield FinishReason(data[data_type]["reason"])
|
| 143 |
+
elif data_type == "log":
|
| 144 |
+
debug.log(data[data_type])
|
| 145 |
+
else:
|
| 146 |
+
debug.log(f"Unknown data: ({data_type}) {data}")
|
generated_images/1738296597_what+is+best+ai+model+yet+so+far,+compare+and+tell+me+final+result+only.jpg
ADDED
|
generated_images/1738332473_gpt-4o.A+modern+commercial+airplane+soaring+through+a+clear+blue+sky,+with+white+contrails+streaming+behind.jpg
ADDED
|
generated_images/1738347359_dall-e-3.Ask+Blackbox+AI+Anything.jpg
ADDED
|
generated_images/1738347377_flux-pro.Ask+Blackbox+AI+Anything.jpg
ADDED
|
generated_images/1738371043_flux-pro.Create+a+image:+A+futuristic+cityscape+at+sunset,+with+towering+skyscrapers+and+flying+cars+zipping+.jpg
ADDED
|
generated_images/1738372570_Copilot.A+futuristic+cityscape+at+sunset,+with+towering+skyscrapers+and+flying+cars+zipping+by.jpg
ADDED
|
generated_images/1738372600_dall-e-3.A+futuristic+cityscape+at+sunset,+with+towering+skyscrapers+and+flying+cars+zipping+by.jpg
ADDED
|
har_and_cookies/.usage/2025-01-30.jsonl
ADDED
|
@@ -0,0 +1,53 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"user": "roxky", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 8, "completion_tokens": 115, "total_tokens": 123}
|
| 2 |
+
{"user": "roxky", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 18, "completion_tokens": 10, "total_tokens": 28}
|
| 3 |
+
{"user": "roxky", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 38, "completion_tokens": 10, "total_tokens": 48}
|
| 4 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 58, "completion_tokens": 12, "total_tokens": 70}
|
| 5 |
+
{"user": "roxky", "model": "flux-schnell", "provider": "HuggingFace", "prompt_tokens": 8, "completion_tokens": 85, "total_tokens": 93}
|
| 6 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 10, "completion_tokens": 68, "total_tokens": 78}
|
| 7 |
+
{"user": "roxky", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 11, "total_tokens": 19}
|
| 8 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
|
| 9 |
+
{"user": "roxky", "model": "flux-schnell", "provider": "HuggingFace", "prompt_tokens": 9, "completion_tokens": 76, "total_tokens": 85}
|
| 10 |
+
{"user": "RubielLabarta", "model": "flux-dev", "provider": "HuggingSpace", "prompt_tokens": 17, "completion_tokens": 220, "total_tokens": 237}
|
| 11 |
+
{"user": "RubielLabarta", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 8, "total_tokens": 16}
|
| 12 |
+
{"user": "RubielLabarta", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 28, "completion_tokens": 25, "total_tokens": 53}
|
| 13 |
+
{"user": "RubielLabarta", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 60, "completion_tokens": 28, "total_tokens": 88}
|
| 14 |
+
{"user": "RubielLabarta", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 60, "completion_tokens": 28, "total_tokens": 88}
|
| 15 |
+
{"user": "RubielLabarta", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 60, "completion_tokens": 83, "total_tokens": 143}
|
| 16 |
+
{"user": "roxky", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
|
| 17 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 236, "total_tokens": 244}
|
| 18 |
+
{"user": "roxky", "model": "flux", "provider": "PollinationsAI", "prompt_tokens": 8, "completion_tokens": 117, "total_tokens": 125}
|
| 19 |
+
{"user": "roxky", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 8, "completion_tokens": 123, "total_tokens": 131}
|
| 20 |
+
{"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 8, "completion_tokens": 32, "total_tokens": 40}
|
| 21 |
+
{"user": "sheeee2222", "model": "qwen-2-72b", "provider": "HuggingSpace", "prompt_tokens": 15, "completion_tokens": 17, "total_tokens": 32}
|
| 22 |
+
{"user": "sheeee2222", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 10, "prompt_tokens": 174, "total_tokens": 184}
|
| 23 |
+
{"user": "sheeee2222", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 41, "prompt_tokens": 213, "total_tokens": 254}
|
| 24 |
+
{"user": "sheeee2222", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 566, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 3531, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 4097}
|
| 25 |
+
{"user": "roxky", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 11, "total_tokens": 20}
|
| 26 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
|
| 27 |
+
{"user": "roxky", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 7, "total_tokens": 16}
|
| 28 |
+
{"user": "roxky", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 6, "total_tokens": 15}
|
| 29 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 20, "completion_tokens": 15, "total_tokens": 35}
|
| 30 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 46, "completion_tokens": 73, "total_tokens": 119}
|
| 31 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 126, "completion_tokens": 124, "total_tokens": 250}
|
| 32 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 246, "completion_tokens": 22, "total_tokens": 268}
|
| 33 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 274, "completion_tokens": 63, "total_tokens": 337}
|
| 34 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 342, "completion_tokens": 12, "total_tokens": 354}
|
| 35 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 359, "completion_tokens": 41, "total_tokens": 400}
|
| 36 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 420, "completion_tokens": 120, "total_tokens": 540}
|
| 37 |
+
{"user": "Cocfff", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 420, "completion_tokens": 120, "total_tokens": 540}
|
| 38 |
+
{"user": "cihairottoilcazzo", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 18, "completion_tokens": 261, "total_tokens": 279}
|
| 39 |
+
{"user": "sanch1tx", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 8, "completion_tokens": 17, "total_tokens": 25}
|
| 40 |
+
{"user": "nasrulam", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 19, "completion_tokens": 1225, "total_tokens": 1244}
|
| 41 |
+
{"user": "nasrulam", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 19, "completion_tokens": 69, "total_tokens": 88}
|
| 42 |
+
{"user": "nasrulam", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 35, "completion_tokens": 1442, "total_tokens": 1477}
|
| 43 |
+
{"user": "nasrulam", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 35, "completion_tokens": 298, "total_tokens": 333}
|
| 44 |
+
{"user": "nasrulam", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 1574, "completion_tokens": 136, "total_tokens": 1710}
|
| 45 |
+
{"user": "sfewewf232", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 87, "completion_tokens": 160, "total_tokens": 247}
|
| 46 |
+
{"user": "hashim1", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 13, "completion_tokens": 9, "total_tokens": 22}
|
| 47 |
+
{"user": "hashim1", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 36, "completion_tokens": 20, "total_tokens": 56}
|
| 48 |
+
{"user": "roxky", "model": "flux-dev", "provider": "HuggingSpace", "prompt_tokens": 30, "completion_tokens": 200, "total_tokens": 230}
|
| 49 |
+
{"user": "roxky", "model": "flux-schnell", "provider": "HuggingFace", "prompt_tokens": 30, "completion_tokens": 80, "total_tokens": 110}
|
| 50 |
+
{"user": "roxky", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 9, "completion_tokens": 353, "total_tokens": 362}
|
| 51 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
|
| 52 |
+
{"user": "roxky", "model": "gpt-4o-mini", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 7, "total_tokens": 16}
|
| 53 |
+
{"user": "roxky", "model": "claude-3-haiku", "provider": "DDG", "prompt_tokens": 2447, "completion_tokens": 798, "total_tokens": 3245}
|
har_and_cookies/.usage/2025-01-31.jsonl
ADDED
|
@@ -0,0 +1,126 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"user": "roxky", "model": "qvq-72b", "provider": "HuggingSpace", "prompt_tokens": 15, "completion_tokens": 10, "total_tokens": 25}
|
| 2 |
+
{"user": "roxky", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 16, "completion_tokens": 1904, "total_tokens": 1920}
|
| 3 |
+
{"user": "roxky", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 1490, "completion_tokens": 1736, "total_tokens": 3226}
|
| 4 |
+
{"user": "tanu360", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 8, "completion_tokens": 105, "total_tokens": 113}
|
| 5 |
+
{"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
|
| 6 |
+
{"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 41, "completion_tokens": 240, "total_tokens": 281}
|
| 7 |
+
{"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 41, "completion_tokens": 440, "total_tokens": 481}
|
| 8 |
+
{"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 41, "completion_tokens": 393, "total_tokens": 434}
|
| 9 |
+
{"user": "bbdwbayu", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 694, "completion_tokens": 561, "total_tokens": 1255}
|
| 10 |
+
{"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 1273, "completion_tokens": 44, "total_tokens": 1317}
|
| 11 |
+
{"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 1329, "completion_tokens": 166, "total_tokens": 1495}
|
| 12 |
+
{"user": "bbdwbayu", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 1508, "completion_tokens": 206, "total_tokens": 1714}
|
| 13 |
+
{"user": "JoniSenior", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 9, "completion_tokens": 299, "total_tokens": 308}
|
| 14 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 9, "completion_tokens": 448, "total_tokens": 457}
|
| 15 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 476, "completion_tokens": 1010, "total_tokens": 1486}
|
| 16 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 70, "completion_tokens": 1075, "total_tokens": 1145}
|
| 17 |
+
{"user": "BruceStudios", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 9, "completion_tokens": 9, "total_tokens": 18}
|
| 18 |
+
{"user": "BruceStudios", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 29, "completion_tokens": 130, "total_tokens": 159}
|
| 19 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 1223, "completion_tokens": 1606, "total_tokens": 2829}
|
| 20 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 2863, "completion_tokens": 1563, "total_tokens": 4426}
|
| 21 |
+
{"user": "herokulastly", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 19, "completion_tokens": 10, "total_tokens": 29}
|
| 22 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 4473, "completion_tokens": 1377, "total_tokens": 5850}
|
| 23 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 5862, "completion_tokens": 755, "total_tokens": 6617}
|
| 24 |
+
{"user": "zayuvegda", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 25, "total_tokens": 33}
|
| 25 |
+
{"user": "zayuvegda", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 43, "completion_tokens": 28, "total_tokens": 71}
|
| 26 |
+
{"user": "zayuvegda", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 76, "completion_tokens": 347, "total_tokens": 423}
|
| 27 |
+
{"user": "zayuvegda", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 434, "completion_tokens": 253, "total_tokens": 687}
|
| 28 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 6635, "completion_tokens": 1026, "total_tokens": 7661}
|
| 29 |
+
{"user": "zayuvegda", "model": "gpt-4o-mini", "provider": "OpenaiAccount", "prompt_tokens": 728, "completion_tokens": 8, "total_tokens": 736}
|
| 30 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 75, "completion_tokens": 945, "total_tokens": 1020}
|
| 31 |
+
{"user": "zayuvegda", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 65, "completion_tokens": 229, "total_tokens": 294}
|
| 32 |
+
{"user": "zayuvegda", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 91, "completion_tokens": 143, "total_tokens": 234}
|
| 33 |
+
{"user": "zayuvegda", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 939, "completion_tokens": 399, "total_tokens": 1338}
|
| 34 |
+
{"user": "zayuvegda", "model": "gpt-4", "provider": "OpenaiAccount", "prompt_tokens": 1682, "completion_tokens": 489, "total_tokens": 2171}
|
| 35 |
+
{"user": "zayuvegda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 2823, "completion_tokens": 1513, "total_tokens": 4336}
|
| 36 |
+
{"user": "JoniSenior", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 1052, "completion_tokens": 833, "total_tokens": 1885}
|
| 37 |
+
{"user": "zayuvegda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 188, "completion_tokens": 730, "total_tokens": 918}
|
| 38 |
+
{"user": "zayuvegda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 928, "completion_tokens": 1272, "total_tokens": 2200}
|
| 39 |
+
{"user": "zayuvegda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 2299, "completion_tokens": 2075, "total_tokens": 4374}
|
| 40 |
+
{"user": "zayuvegda", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 4444, "completion_tokens": 2914, "total_tokens": 7358}
|
| 41 |
+
{"user": "cruelzade", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 287, "completion_tokens": 464, "total_tokens": 751}
|
| 42 |
+
{"user": "seotube", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 11, "total_tokens": 19}
|
| 43 |
+
{"user": "seotube", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 30, "completion_tokens": 378, "total_tokens": 408}
|
| 44 |
+
{"user": "seotube", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 379, "completion_tokens": 214, "total_tokens": 593}
|
| 45 |
+
{"user": "setsunafjava", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 30, "completion_tokens": 15, "total_tokens": 45}
|
| 46 |
+
{"user": "setsunafjava", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 69, "completion_tokens": 17, "total_tokens": 86}
|
| 47 |
+
{"user": "setsunafjava", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 115, "completion_tokens": 1204, "total_tokens": 1319}
|
| 48 |
+
{"user": "setsunafjava", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 1890, "completion_tokens": 1940, "total_tokens": 3830}
|
| 49 |
+
{"user": "roxky", "model": "flux-schnell", "provider": "HuggingFace", "prompt_tokens": 9, "completion_tokens": 76, "total_tokens": 85}
|
| 50 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 9, "completion_tokens": 12, "total_tokens": 21}
|
| 51 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 10, "prompt_tokens": 153, "total_tokens": 163}
|
| 52 |
+
{"user": "roxky", "model": "flux-dev", "provider": "HuggingSpace", "prompt_tokens": 9, "completion_tokens": 212, "total_tokens": 221}
|
| 53 |
+
{"user": "roxky", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
|
| 54 |
+
{"user": "roxky", "model": "sonar-pro", "provider": "PerplexityLabs", "prompt_tokens": 9, "completion_tokens": 27, "total_tokens": 36}
|
| 55 |
+
{"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 9, "completion_tokens": 53, "total_tokens": 62}
|
| 56 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 9, "completion_tokens": 10, "total_tokens": 19}
|
| 57 |
+
{"user": "ballm4n", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 9, "completion_tokens": 9, "total_tokens": 18}
|
| 58 |
+
{"user": "NexThinkLabs", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 8, "total_tokens": 16}
|
| 59 |
+
{"user": "NexThinkLabs", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 25, "completion_tokens": 24, "total_tokens": 49}
|
| 60 |
+
{"user": "NexThinkLabs", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 8, "total_tokens": 16}
|
| 61 |
+
{"user": "roxky", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 10, "completion_tokens": 10, "total_tokens": 20}
|
| 62 |
+
{"user": "roxky", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 37, "completion_tokens": 106, "total_tokens": 143}
|
| 63 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 37, "completion_tokens": 376, "total_tokens": 413}
|
| 64 |
+
{"user": "roxky", "model": "sonar-pro", "provider": "PerplexityLabs", "prompt_tokens": 37, "completion_tokens": 154, "total_tokens": 191}
|
| 65 |
+
{"user": "MAKERBOT", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 9, "completion_tokens": 16, "total_tokens": 25}
|
| 66 |
+
{"user": "XannX", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 11, "completion_tokens": 608, "total_tokens": 619}
|
| 67 |
+
{"user": "MAKERBOT", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 56, "completion_tokens": 114, "total_tokens": 170}
|
| 68 |
+
{"user": "MAKERBOT", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 155, "completion_tokens": 636, "total_tokens": 791}
|
| 69 |
+
{"user": "XannX", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 569, "completion_tokens": 3049, "total_tokens": 3618}
|
| 70 |
+
{"user": "dgdgdsgs", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 24, "total_tokens": 32}
|
| 71 |
+
{"user": "dgdgdsgs", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 185, "completion_tokens": 368, "total_tokens": 553}
|
| 72 |
+
{"user": "roxky", "model": "sonar-pro", "provider": "PerplexityLabs", "prompt_tokens": 10, "completion_tokens": 21, "total_tokens": 31}
|
| 73 |
+
{"user": "roxky", "model": "MiniMax-01", "provider": "HailuoAI", "prompt_tokens": 1490, "completion_tokens": 2074, "total_tokens": 3564}
|
| 74 |
+
{"user": "roxky", "model": "claude-3-haiku", "provider": "DDG", "prompt_tokens": 9, "completion_tokens": 9, "total_tokens": 18}
|
| 75 |
+
{"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 8, "completion_tokens": 14, "total_tokens": 22}
|
| 76 |
+
{"user": "roxky", "model": "flux", "provider": "Blackbox", "prompt_tokens": 8, "completion_tokens": 112, "total_tokens": 120}
|
| 77 |
+
{"user": "roxky", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
|
| 78 |
+
{"user": "KA4E6", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 12, "total_tokens": 20}
|
| 79 |
+
{"user": "Majidgdr", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 13, "completion_tokens": 13, "total_tokens": 26}
|
| 80 |
+
{"user": "Majidgdr", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 39, "completion_tokens": 71, "total_tokens": 110}
|
| 81 |
+
{"user": "Majidgdr", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 121, "completion_tokens": 70, "total_tokens": 191}
|
| 82 |
+
{"user": "Majidgdr", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 187, "completion_tokens": 6, "total_tokens": 193}
|
| 83 |
+
{"user": "Jackolord", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 10, "prompt_tokens": 152, "total_tokens": 162}
|
| 84 |
+
{"user": "Jackolord", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 27, "completion_tokens": 129, "total_tokens": 156}
|
| 85 |
+
{"user": "Jackolord", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 37, "completion_tokens": 361, "total_tokens": 398}
|
| 86 |
+
{"user": "Jackolord", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 342, "completion_tokens": 465, "total_tokens": 807}
|
| 87 |
+
{"user": "Jackolord", "model": "o1-preview", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 10, "total_tokens": 18}
|
| 88 |
+
{"user": "Jackolord", "model": "o1-preview", "provider": "OpenaiAccount", "prompt_tokens": 36, "completion_tokens": 45, "total_tokens": 81}
|
| 89 |
+
{"user": "Jackolord", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 17, "completion_tokens": 90, "total_tokens": 107}
|
| 90 |
+
{"user": "Jackolord", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 17, "completion_tokens": 53, "total_tokens": 70}
|
| 91 |
+
{"user": "Jackolord", "model": "sonar-pro", "provider": "PerplexityLabs", "prompt_tokens": 17, "completion_tokens": 14, "total_tokens": 31}
|
| 92 |
+
{"user": "Jackolord", "model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 17, "completion_tokens": 54, "total_tokens": 71}
|
| 93 |
+
{"user": "Jackolord", "model": "dall-e-3", "provider": "CopilotAccount", "prompt_tokens": 12, "completion_tokens": 157, "total_tokens": 169}
|
| 94 |
+
{"user": "Jackolord", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 12, "completion_tokens": 152, "total_tokens": 164}
|
| 95 |
+
{"user": "Jackolord", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 12, "completion_tokens": 616, "total_tokens": 628}
|
| 96 |
+
{"user": "NDeriy", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 110, "completion_tokens": 84, "total_tokens": 194}
|
| 97 |
+
{"user": "NDeriy", "model": "gpt-4o", "provider": "Blackbox", "prompt_tokens": 302, "completion_tokens": 249, "total_tokens": 551}
|
| 98 |
+
{"user": "NDeriy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 739, "completion_tokens": 825, "total_tokens": 1564}
|
| 99 |
+
{"user": "NDeriy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 1610, "completion_tokens": 647, "total_tokens": 2257}
|
| 100 |
+
{"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 17, "completion_tokens": 57, "total_tokens": 74}
|
| 101 |
+
{"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 87, "completion_tokens": 66, "total_tokens": 153}
|
| 102 |
+
{"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 167, "completion_tokens": 121, "total_tokens": 288}
|
| 103 |
+
{"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 303, "completion_tokens": 230, "total_tokens": 533}
|
| 104 |
+
{"user": "NDeriy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 2353, "completion_tokens": 1051, "total_tokens": 3404}
|
| 105 |
+
{"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 25, "completion_tokens": 183, "total_tokens": 208}
|
| 106 |
+
{"user": "Intgrr", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 322, "completion_tokens": 217, "total_tokens": 539}
|
| 107 |
+
{"user": "NDeriy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 3416, "completion_tokens": 687, "total_tokens": 4103}
|
| 108 |
+
{"user": "Intgrr", "model": "o1-preview", "provider": "OpenaiAccount", "prompt_tokens": 689, "completion_tokens": 283, "total_tokens": 972}
|
| 109 |
+
{"user": "Intgrr", "model": "o1-mini", "provider": "OpenaiAccount", "prompt_tokens": 1144, "completion_tokens": 260, "total_tokens": 1404}
|
| 110 |
+
{"user": "NDeriy", "model": "auto", "provider": "OpenaiAccount", "prompt_tokens": 4223, "completion_tokens": 1023, "total_tokens": 5246}
|
| 111 |
+
{"user": "macyn", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 15, "completion_tokens": 95, "total_tokens": 110}
|
| 112 |
+
{"user": "macyn", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 14, "completion_tokens": 101, "total_tokens": 115}
|
| 113 |
+
{"user": "NDeriy", "model": "qwq-32b", "provider": "HuggingFace", "prompt_tokens": 5262, "completion_tokens": 286, "total_tokens": 5548}
|
| 114 |
+
{"user": "macyn", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 15, "completion_tokens": 740, "total_tokens": 755}
|
| 115 |
+
{"user": "macyn", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 19, "completion_tokens": 226, "total_tokens": 245}
|
| 116 |
+
{"user": "macyn", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 19, "completion_tokens": 327, "total_tokens": 346}
|
| 117 |
+
{"user": "roxky", "model": "claude-3-haiku", "provider": "DDG", "prompt_tokens": 9, "completion_tokens": 9, "total_tokens": 18}
|
| 118 |
+
{"user": "roxky", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 1930, "completion_tokens": 300, "total_tokens": 2230}
|
| 119 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 1930, "completion_tokens": 1026, "total_tokens": 2956}
|
| 120 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 3265, "completion_tokens": 56, "total_tokens": 3321}
|
| 121 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 1930, "completion_tokens": 1026, "total_tokens": 2956}
|
| 122 |
+
{"user": "Erkaperka502", "model": "gpt-4o", "provider": "PollinationsAI", "completion_tokens": 83, "completion_tokens_details": {"accepted_prediction_tokens": 0, "audio_tokens": 0, "reasoning_tokens": 0, "rejected_prediction_tokens": 0}, "prompt_tokens": 169, "prompt_tokens_details": {"audio_tokens": 0, "cached_tokens": 0}, "total_tokens": 252}
|
| 123 |
+
{"user": "h4mid007", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 8, "completion_tokens": 12, "total_tokens": 20}
|
| 124 |
+
{"user": "h4mid007", "model": "qwen-2.5-coder-32b", "provider": "HuggingFace", "prompt_tokens": 29, "completion_tokens": 10, "total_tokens": 39}
|
| 125 |
+
{"user": "h4mid007", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 53, "completion_tokens": 25, "total_tokens": 78}
|
| 126 |
+
{"user": "h4mid007", "model": "deepseek-r1", "provider": "HuggingFace", "prompt_tokens": 87, "completion_tokens": 294, "total_tokens": 381}
|
har_and_cookies/.usage/2025-02-01.jsonl
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"user": "amsepehrian", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 13, "completion_tokens": 233, "total_tokens": 246}
|
| 2 |
+
{"user": "roxky", "model": "deepseek-r1", "provider": "HuggingFaceAPI", "prompt_tokens": 15, "completion_tokens": 12, "total_tokens": 27}
|
| 3 |
+
{"user": "roxky", "model": "gpt-4o", "provider": "OpenaiAccount", "prompt_tokens": 15, "completion_tokens": 7, "total_tokens": 22}
|
| 4 |
+
{"user": "roxky", "model": "llama-3.2-11b", "provider": "HuggingFaceAPI", "prompt_tokens": 15, "completion_tokens": 177, "total_tokens": 192}
|
| 5 |
+
{"user": "roxky", "model": "flux-pro", "provider": "PollinationsAI", "prompt_tokens": 178, "completion_tokens": 258, "total_tokens": 436}
|
| 6 |
+
{"user": "amsepehrian", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 8, "completion_tokens": 80, "total_tokens": 88}
|
| 7 |
+
{"user": "amsepehrian", "model": "o1", "provider": "OpenaiAccount", "prompt_tokens": 15, "completion_tokens": 371, "total_tokens": 386}
|
| 8 |
+
{"user": "roxky", "model": "Microsoft Copilot", "provider": "CopilotAccount", "prompt_tokens": 178, "completion_tokens": 209, "total_tokens": 387}
|
| 9 |
+
{"user": "roxky", "model": "dall-e-3", "provider": "CopilotAccount", "prompt_tokens": 178, "completion_tokens": 234, "total_tokens": 412}
|