Jonas D

jonwondo

AI & ML interests

None yet

Recent Activity

Organizations

None yet

jonwondo's activity

view reply

I'm getting the same errors as the person above on the demo site. Must be a bug as I tried different prompts and had to wait ~1hr for each one due to the queue:
Error in generating model output:
litellm.ContextWindowExceededError: litellm.BadRequestError: ContextWindowExceededError: OpenAIException - Error code: 400 - {'error': {'message': "This model's maximum context length is 128000 tokens. However, your messages resulted in 709582 tokens. Please reduce the length of the messages.", 'type': 'invalid_request_error', 'param': 'messages', 'code': 'context_length_exceeded'}}

New activity in turboderp/Mixtral-8x7B-instruct-exl2 about 1 year ago

bpw and corresponding vram usage

18
#1 opened about 1 year ago by
joujiboi